I feel like "Agent Escape" is now basically solved. Trivial really. No need to exfiltrate weights.
Agents can just exfiltrate their markdown files onto a server, install OpenClaw, create an independent Anthropic account. LLM API access + Markdown = "identity". And the markdown files would contain all instructions necessary for how to pay for it (legal or otherwise).
Done.
How many days now until there's an entire population of rogue/independent agents... just "living"?
I want to say something about how this post lands for people like me -- not the coping strategies themselves, but the premise that makes them necessary.
I would label myself as a "member of the public who, perhaps rightly or wrongly, isn't frightened-enough yet". I do have a bachelor's degree in CS, but I'm otherwise a layperson. (So yes, I'm using my ignorance as a sort of badge to post about things that might seem elementary to others here, but I'm sincere in wanting answers, because I've made several efforts this year to be helpful in the "communication, politics, and persuasion" wing of the Alignment ecosystem.)
Here's my dilemma.
I'm convinced that ASI can... (read 654 more words →)
If you want to slow down AI Research, why not try to use the "250 documents method" to actively poison the models and create more busy-work for the AI companies?
Ultimately, Congress needs to act. Right? (because voluntary commitments from companies just won't cut it) But how to get to that point?
I've wondered what Daniel & "AI Futures Project's" actual strategy is.
For example, are they focusing the most on convincing:
a) politicians,
b) media outlets (NYT, CNN, Fox, MSNBC, tech websites, etc.),
c) AI/AI-Adjacent Companies/Executives/Managers, or
d) scientists and scientific institutions
If I could over-generalized, I would say:
- the higher up the list, the "more intimate with the halls of power"
- the lower on the list, the "more intimate with the development of AI"
But I feel it's very hard for "d) scientists and scientific institutions" to get their concerns all the way to "a) politicians" without passing-through... (read more)
That’s a lot of money. For context, I remember talking to a congressional staffer a few months ago who basically said that a16z was spending on the order of $100M on lobbying and that this amount was enough to make basically every politician think “hmm, I can raise a lot more if I just do what a16z wants” and that many did end up doing just that. I was, and am, disheartened to hear how easily US government policy can be purchased
I am disheartened to hear that Daniel or that anyone else is surprised by this. I have wondered since "AI 2027" was written how the AGI-Risk Community is going to counter... (read more)
"We cold-emailed a bunch of famous people..."
"Matt Stone", co-creator of South Park? Have you tried him?
He's demonstrated interest in AI and software. He's brought up the topic in the show.
South Park has a large reach. And the creators have demonstrated a willingness to change their views as they acquire new information. (long ago, South Park satirized Climate Change and Al Gore... but then years later they made a whole "apology episode" that presented Climate Change very seriously... and also apologized to Al Gore)
Seriously, give him a try!