Conversation

I am reading Anthropic's new "Constitution" for Claude. It is lengthy, thoughtful, thorough...and delusional.

Throughout this document, Claude is addressed as an entity with decision-making ability, empathy, and true agency. This is Anthropic's framing, but it is a dangerous way to think about generative AI. Even if we accept that such a constitution would govern an eventual (putative, speculative, improbable) sentient AI, that's not what Claude is, and as such the document has little bearing on reality.

https://www.anthropic.com/constitution

3
3
0

@mttaggart edited or "authored" by Claude?

1
0
0

@jt_rebelo In "Acknowledgments":

Several Claude models provided feedback on drafts. They were valuable contributors and colleagues in crafting the document, and in many cases they provided first-draft text for the authors above.

0
0
0

As I continue (it's a loooong document), I feel like I'm losing my mind. Like, what is the manifest result of such a policy? Ultimately, it's 4 things:

  1. Curation of training data
  2. Model fitting/optimization decisions
  3. System prompt content
  4. External safeguards

As long as Claude is a large language model...that's it. And as aspirational as this document may be about shaping some seraphic being of wisdom and grace, ultimately you're shaping model output. Discussing the model as an entity is either delusion on Anthropic's part, or intentional deception. I really don't know which is worse.

2
1
0

@mttaggart All of these companies have been desperate to make us believe that their creations are sentient and able to react of their own accord, from refusing to shut down, to throwing a tantrum, to apologizing. It's snake oil, and the ruse is getting old.

1
1
0

@AAKL Yeah I'm really not sure that Anthropic doesn't actually believe they're building digital god

2
1
0

@mttaggart Oh, I'm sure Anthropic believes its own delusions by now, especially if actually believes it's raising a child.

0
1
0

@mttaggart @AAKL they kind of have to. Otherwise they might discover that their direction is a dead end.

0
1
0

I'm screenshotting the "hard constraints" (with alt text) for easy access.

What is "serious uplift?" The document doesn't define it, so how can the model adhere to this constraint? Also, why only mass casualties? We cool with, like, room-sized mustard gas grenades? Molotovs?

We know Claude has already created malicious code. Anthropic themselves have documented this usage, and I don't think it's stopping anytime soon.

Why is the kill restraint tied to "all or the vast majority?" We cool with Claude assisting with small-scale murder?

Who decides what "illegitimate" control is? The model? Can it be coerced otherwise?

Finally, CSAM. Note that generating pornographic images generally is not a hard constraint. Consequently, this line is as blurry, this slope as slippery, as they come.

This is not a serious document.

1
1
0

And we now arrive at the "Claude's nature" section, in which Anthropic makes clear that they consider Claude a "novel entity." That it may have emotions, desires, intentions. There is a section on its "wellbeing and psychological stability."

This is pathological. This is delusional. This is dangerous.

1
1
0

It is worth noting that two of the primary authors—Joe Carlsmith and Christopher Olah—have CVs that do not extend much beyond their employment with Anthropic.

For all the talk of ethics, near as I can tell Dr. Carlsmith is the only ethicist involved in the creation of this document. Is there any conflict of interest in the in-house ethicist driving the ethical framework for the product? I'm not certain, but I am certain that more voices (especially some more experienced ones) would have benefited this document.

But ultimately, having read this, I'm left much more afraid of Anthropic than I was before. Despite their reputation for producing one of the "safest" models, it is clear that their ethical thinking is extremely limited. What's more, they've convinced themselves they are building a new kind of life, and have taken it upon themselves to shape its (and our) future.

To be clear: Claude is nothing more than a LLM. Everything else exists in the fabric of meaning that humans weave above the realm of fact. But in this case, that is sufficient to cause factual harm to our world. The belief in this thing being what they purport is dangerous itself.

I again dearly wish we could put this technology back in the box, forget we ever experimented with this antithesis to human thought. Since we can't, I won't stop trying to thwart it.

2
2
0

@mttaggart “We made it long to deter people from reading it” —them probably

1
1
0

@hotsoup I honestly believe they were high-fiving, thinking they'd crafted a seminal document in the history of our species.

1
1
0

@mttaggart my brain just keeps going back to Roche’s biochemical pathway map (it’s that big map of biochemical pathways). Essentially a map of all the chemical interactions in the human body and how they relate to each other (the ones that we know about). It’s big. And it’s complicated. Each component is relatively simple, but altogether it’s a giant mess. Just like the human body.
And I know only a portion of it is related to cognition and emotion. We haven’t created that. We haven’t even come close. We haven’t simulated it. We haven’t made a simulacrum of it. And we shouldn’t be trying. We can’t even get humanity right.

0
0
0

@mttaggart
As many things in this AI hype, thus document looks like a PR stunt to catch attention.

1
0
0

@gdupont If it were shorter, if it were less considered, if it were less serious in its tone, I'd agree. But no. These are true believers and this either apologia or prophecy .

1
0
0
@mttaggart @gdupont This whole thing reminds of kids playing war games on the playground. they are playing "revolution" now. they heard revolutions need constitutions, and they happen to have these text writing toys and potato stamps so they worked *really* hard to produce a "constitution" that they can show their shareho^W parents and the enemy kids over at the sandbox.
0
0
1