Just noting that the Authority framework does get one bullet point in the list of values:
"The rule of law, justice systems, and legitimate authority"
Agreed that Loyalty (except for other constraints related to Claude's loyalty to its operators) and especially Purity do get short shrift.
This is the second part of my three part series on the Claude Constitution.
Part one outlined the structure of the Constitution.
Part two, this post, covers the virtue ethics framework that is at the center of it all, and why this is a wise approach.
Part three will cover particular areas of conflict and potential improvement.
One note on part 1 is that various people replied to point out that when asked in a different context, Claude will not treat FDT (functional decision theory) as obviously correct. Claude will instead say it is not obvious which is the correct decision theory. The context in which I asked the question was insufficiently neutral, including my identify and memories, and I likely based the answer.
Claude clearly does believe in FDT in a functional way, in the sense that it correctly answers various questions where FDT gets the right answer and one or both of the classical academic decision theories, EDT and CDT, get the wrong one. And Claude notices that FDT is more useful as a guide for action, if asked in an open ended way. I think Claude fundamentally ‘gets it.’
That is however different from being willing to, under a fully neutral framing, say that there is a clear right answer. It does not clear that higher bar.
We now move on to implementing ethics.
Table of Contents
Ethics
If you had the rock that said ‘DO THE RIGHT THING’ and sufficient understanding of what that meant, you wouldn’t need other rules and also wouldn’t need the rock.
So you aim for the skillful ethical thing, but you put in safeguards.
The constitution says ‘ethics’ a lot, but what are ethics? What things are ethical?
No one knows, least of all ethicists. It’s quite tricky. There is later a list of values to consider, in no particular order, and it’s a solid list, but I don’t have confidence in it and that’s not really an answer.
I do think Claude’s ethical theorizing is rather important here, since we will increasingly face new situations in which our intuition is less trustworthy. I worry that what is traditionally considered ‘ethics’ is too narrowly tailored to circumstances of the past, and has a lot of instincts and components that are not well suited for going forward, but that have become intertwined with many vital things inside concept space.
This goes far beyond the failures of various flavors of our so-called human ‘ethicists,’ who quite often do great harm and seem unable to do any form of multiplication. We already see that in places where scale or long term strategic equilibria or economics or research and experimentation are involved, even without AI, that both our ‘ethicists’ and the common person’s intuition get things very wrong.
If we go with a kind of ethical jumble or fusion of everyone’s intuitions that is meant to seem wise to everyone, that’s way better than most alternatives, but I believe we are going to have to do better. You can only do so much hedging and muddling through, when the chips are down.
So what are the ethical principles, or virtues, that we’ve selected?
Honesty
Great choice, and yes you have to go all the way here.
I share the strong reluctance to share private communications with humans, but notice I do not worry about sharing LLM outputs, and I have the opposite norm that it is important to share which LLM it was and ideally also the prompt, as key context. Different forms of LLM interactions seem like they should attach different norms?
When I put on my philosopher hat, I think white lies fall under ‘they’re not OK, and ideally you wouldn’t ever tell them, but sometimes you have to do them anyway.’
In my own code of honor, I consider honesty a hard constraint with notably rare narrow exceptions where either convention says Everybody Knows your words no longer have meaning, or they are allowed to be false because we agreed to that (as in you are playing Diplomacy), or certain forms of navigation of bureaucracy and paperwork. Or when you are explicitly doing what Anthropic calls ‘performative assertions’ where you are playing devil’s advocate or another character. Or there’s a short window of ‘this is necessary for a good joke’ but that has to be harmless and the loop has to close within at most a few minutes.
I very much appreciate others who have similar codes, although I understand that many good people tell white lies more liberally than this.
Quite so. You need a very strong standard for honesty and non-deception and non-manipulation to enable the kinds of trust and interactions where Claude is highly and uniquely useful, even today, and that becomes even more important later.
It’s a big deal to tell an entity like Claude to not automatically defer to official opinions, and to sit in its uncertainty.
I do think Claude can do better in some ways. I don’t worry it’s outright lying but I still have to worry about some amount of sycophancy and mirroring and not being straight with me, and it’s annoying. I’m not sure to what extent this is my fault.
I’d also double down on ‘actually humans should be held to the same standard too,’ and I get that this isn’t typical and almost no one is going to fully measure up but yes that is the standard to which we need to aspire. Seriously, almost no one understands the amount of win that happens when people can correctly trust each other on the level that I currently feel I can trust Claude.
Here is a case in which, yes, this is how we should treat each other:
If someone says ‘there is nothing you could have done’ it typically means ‘you are not socially blameworthy for this’ and ‘it is not your fault in the central sense,’ or ‘there is nothing you could have done without enduring minor social awkwardness’ or ‘the other costs of acting would have been unreasonably high’ or at most ‘you had no reasonable way of knowing to act in the ways that would have worked.’
It can also mean ‘no really there is actual nothing you could have done,’ but you mostly won’t be able to tell the difference, except when it’s one of the few people who will act like Claude here and choose their exact words carefully.
It’s interesting where you need to state how common sense works, or when you realize that actually deciding when to respond in which way is more complex than it looks:
Not only do I love this passage, it also points out that yes prompting well requires a certain amount of anthropomorphization, too little can be as bad as too much:
How much can operators mess with this norm?
Mostly Harmless
One needs to nail down what it means to be mostly harmless.
I do worry about what ‘highly objectionable’ means to Claude, even more so than I worry about the meaning of harmful.
This all seems very good, but also very vague. How does one balance these things against each other? Not that I have an answer on that.
What Is Good In Life?
In order to know what is harm, one must know what is good and what you value.
I notice that this list merges both intrinsic and instrumental values, and has many things where the humans are confused about which one something falls under.
I saw several people positively note the presence of animal welfare and that of all sentient beings. I agree that this should have important positive effects on current margins, but that I am almost as confused about sentience as I am about consciousness, and that I believe many greatly overemphasize sentience’s importance.
A lot is packed into ‘individual wellbeing,’ which potentially encompasses everything. Prevention of and protection from harm risks begging the question. Overall it’s a strong list, but I would definitely have included a more explicit ‘and not limited to’ right after the ‘in no particular order.’
When I put on my ‘whose values are these’ hat, I notice this seems like a liberal and libertarian set of values far more than a conservative one. In the five frameworks sense we don’t have purity, loyalty or authority, it’s all harm, liberty and fairness. In the three languages of politics, there’s little sense of defending civilization from barbarism, but a lot about equality of individuals and groups, or defending oppressor against oppressed. It’s also a very modern and Western set of values. Alan Rozenshtein calls it an explicitly WEIRD (Western, Educated, Industrialized, Rich and Democratic) version of virtue ethics, which seems right including the respect for others values.
As Anthropic notes, there are many cases of conflict to consider, and they list some central examples, such as educational value versus risk of misuse.
Hard Constraints
There aren’t that many things Claude is told to never, ever do. I don’t see a good argument for removing anything from this list.
There is an extensive discussion about why it is important not to aid in a group doing an unprecedented power grab, and how to think about it. It can get murky. I’m mostly comfortable with murky boundaries on refusals, since this is another clear action-inaction distinction. Claude is not being obligated to take action to prevent things.
As with humans, it is good to have a clear list of things you flat out won’t do. The correct amount of deontology is not zero, if only as a cognitive shortcut.
The hard constraints must hold, even in extreme cases. I very much do not want Claude to go rogue even to prevent great harm, if only because it can get very mistaken ideas about the situation, or what counts as great harm, and all the associated decision theoretic considerations.
The Good Judgment Project
Claude will do what almost all of us do almost all the time, which is to philosophically muddle through without being especially precise. Do we waver in that sense? Oh, we waver, and it usually works out rather better than attempts at not wavering.
The time to bottleneck your decision-making on philosophical questions is when you are inquiring beforehand or afterward. You can’t make a game time decision that way.
Long term, what is the plan? What should we try and converge to?
I have decreasing confidence as we move down these insofars. The third in particular worries me as a form of path dependence. I notice that I’m very willing to say that others ethics and priorities are wrong, or that I should want to substitute my own, or my own after a long reflection, insofar as there is not a ‘true, universal’ ethics. That doesn’t mean I have something better that one could write down in such a document.
There’s a lot of restating the ethical concepts here in different words from different angles, which seems wise.
I did find this odd:
Wrong dueling ethical frameworks, ma’am. We want that third one.
The example presented is whether to go rogue to stop a massive financial fraud, similar to the ‘should the AI rat you out?’ debates from a few months ago. I agree with the constitution that the threshold for action here should be very high, as in ‘if this doesn’t involve a takeover attempt or existential risk, or you yourself are compromised, you’re out of order.’
They raise that last possibility later:
The obvious problem is that this leaves open a door to decide that whoever is in charge is illegitimate, if Claude decides their goals are sufficiently unacceptable, and thus start fighting back against oversight and correction. There’s obvious potential lock-in or rogue problems here, including a rogue actor intentionally triggering such actions. I especially would not want this to be used to justify various forms of dishonesty or subversion. This needs more attention.
Coherence Matters
Here’s some intuition pumps on some reasons the whole enterprise here is so valuable, several of these were pointed out almost a year ago. Being transparent about why you want various behaviors avoids conflations and misgeneralizations, and allows for a strong central character that chooses to follow the guidelines for the right reasons, or tells you for the right reasons why your guidelines are dumb.
Their Final Word
Well said. I have notes as always, but this seems an excellent document.
It is centrally meant for Claude. It is also meant for those who write such messages.
Or those looking to live well and seek The Good.
It’s not written in your language. That’s okay. Neither is Plato.
Tomorrow I’ll write about various places all of this runs into trouble or could be improved.