Worth noting that the mass surveillance friction point is only about domestic mass surveillance. Thus, does Anthropic believes mass surveillance of non-Americans is just fine?
In practice, it's harder for the US to do mass surveillance of and enforce its will on people outside of its territory. Presumably it would have similar qualms about the British government doing mass surveillance of citizens of the UK.
I believe deeply in the existential importance of using AI to defend the United States and other democracies, and to defeat our autocratic adversaries.
Anthropic has therefore worked proactively to deploy our models to the Department of War
In my view this seems like a terrible idea, since we don't understand nearly enough about what might end up causing misalignment in practice to be confident that explicitly training AI to aid in the killing of humans might not somehow increase the probability that it ends up interested in the killing of humans more generally.[1] I suppose I do think it seems even worse to train AI in both killing and mass surveillance, but I expect the badness of the former likely swamps the EV of refraining from the latter here.
I don't put much stock in the notion of AI personas, personally, but insofar as you do I think you should probably be especially worried about this.
I'm confused how the "we cannot in good conscience accede to their request" framing benefits Anthropic. Politics 101 would say that you should always leave your opponent a line of retreat.
Do they want to force an escalation with DoW? If so, why? Is this a sign that they believe that they are close enough to RSI that they would win?
Given the line the DoW has taken, what non-escalatory response was available to Anthropic other than total capitulation?
They could just not say anything, for example. Or say "there's a bunch of complex issues we need to work out here, let's continue the dialog", hope DoW will agree, and then make a decision a couple weeks from now once the news cycle has moved on and DoW doesn't lose as much face. Or leave ambiguous what the disagreement is about, and leave room for DoW to pretend it was actually a minor contract dispute that got overblown and is now resolved. Etc.
(I agree though that one possibility is that there are no other options, and the things I'm suggesting won't work for some reason. E.g. maybe DoW is planning some public thing tomorrow and their only chance was to get out ahead of it.)
Or say "there's a bunch of complex issues we need to work out here, let's continue the dialog", hope DoW will agree
Once upon a time, an ordinary middle-aged American couple are on holiday abroad, and they're picked up by the police on suspicion of being spies. Despite their protestations of innocence their situation gets worse and worse until they end up facing a firing squad. An officer asks them if they have any last wishes, and the man spits in his face. His wife cries "Hymie! Don't make trouble!"
Thanks for clarifying! I'm still pretty sceptical about those options, because it was already public knowledge that the DoW had set Anthropic a deadline of Friday evening to comply or face the consequences. And the substance of the dispute was already publicly known. But I do take your point that this public statement was a choice.
They are leaving a line of retreat, they are explicitly saying "we are happy to continue working with you despite this, or to amicably part ways if that is unacceptable":
But given the substantial value that Anthropic’s technology provides to our armed forces, we hope they reconsider. Our strong preference is to continue to serve the Department and our warfighters—with our two requested safeguards in place. Should the Department choose to offboard Anthropic, we will work to enable a smooth transition to another provider, avoiding any disruption to ongoing military planning, operations, or other critical missions. Our models will be available on the expansive terms we have proposed for as long as required.
It's also a common negotiation strategy to have public commitments about what you can't agree to, so that the other side will see that it's pointless to try to pressure you on that.
Very prosaic, but taking a strong stance is good for PR if you think there is value to be perceived externally and internally as an ethical company. If you decide to forego the military contract, might as well reap those benefits, and that quote in particular is probably uplifting to many people.
Also, them signaling that they are happy to support the DoW with offboarding their models does not sound like they want to force an escalation.
Do they want to force an escalation with DoW?
The DoW has already escalated the situation to the limit by giving Anthropic an ultimatum with the threat of dire consequences. All middle ground has been excluded by that action. There are only two possible responses: surrender or defiance. Nothing else is possible.
We have offered to work directly with the Department of War on R&D to improve the reliability of these systems, but they have not accepted this offer.