Every rationalist-affiliated hedge fund I know of has operated with questionable ethics. Some scams were open and obvious. Tons of EAs released blatant pump-and-dump coins and promoted them on their persona Twitter. No one cared.
This is news to me. Can you give examples/links?
not sure how many of us considered ourselves EAs (i don't think of myself that way) but i was in the cabal that OP is talking about here. lots of us are at least rats. i made the money i've been living off of for the last six months this way.
Can anyone point to any prominent EA promoting a cryptocurrency, full stop? I can't ever remember this happening.
I'm not sure if you would or should consider "ampdot" prominent, as far as I understand they have received EA funding for AI research, but I could definitely be wrong about that. They did a lot of memecoin promotion, as did their social circle. I'm not sure if they consider themselves an EA or just a rationalist.
In total I know of about 6? rationalist/EAs who ran memecoin pump and dumps, but would not call them prominent.
I don't think sapph is trying to use these examples as persuasive evidence.
As a small counterexample, I know a few rationalists who worked at normal hedge funds earning to give, and they weren’t involved in scams or unethical behavior. I haven’t met anyone at a rationalist-affiliated hedge fund, so I can’t comment on them in particular.
Efforts like AI 2027 have encouraged this madness.
I hope not, since I and the other authors agree it's madness and are trying hard to fight against it.
It seems to be a common view of a subgroup that it has encouraged it. Bsky is a good place to find posts from those sorts, you can also find them on reddit. you'll need to use a fresh, anonymous account in order to for this search to be correct, else you'll likely be blocked by many people you're looking for. https://bsky.app/search?q="ai+2027" - try latest as well as top, you may have to scroll a bit to get a sense of the variety of reactions. it's slowed to a trickle now, you'd have to scroll way back to near release day to see the main wave of reactions. It's not really clear to me exactly how common a view it is at the moment, but bsky is certainly a place you can see it more easily, given the site's reliable hate of anything that claims AI is powerful in any shape or form.
archetypal example: https://bsky.app/profile/emilymbender.bsky.social/post/3ltwemoozvs2u
I think I basically feel you on everything you've raised. In fact I'd go so far as to say that there is no meaningful workable solution to AI alignment that does not involve also addressing AI welfare. (I have my own argument for something similar here)
And as for the broader points of the rationalist community... I think everything you say is true. But also I think that there is good here (because I think there is good in every human being), and there are people who in the name of EA or rationality work very hard to be good to each other and to the world at large. I would also say that people who reject power-seeking or fame-seeking or wealth-seeking behaviours are much less visible than those who don't reject such attitudes. In some sense the most rich and powerful "AI safety community members" are by definition those who did not reject wealth and power, and the people who could not stomach the idea of working in a capabilities lab in the name of "control" or "safety" have all quit long ago. And if everyone who sees these issues latent within the community leaves, those who are left behind will be the ones who don't see these issues or are okay with them.
I cannot influence your de...
I think that the woman you met on FEELD was engaging in wishful thinking. I do not understand the line of reasoning that supports the conclusion that the concentration of power will stop at “people who work at a leading AI lab.” Why would it stop there?
A sane response would be to slow down the race and build a trustworthy international framework that ensures everyone can benefit from AGI. Promises would not be enough; you would need actual institutions with real authority. That’s hard, but possible. Instead, we’ve chosen a strategy of cutting China off from GPUs and hoping we can beat them to AGI before they scale up domestic production.
Isn't this... The entire "Rat/EA" platform?
‘But I have so little of any of these things! You are wise and powerful. Will you not take the Ring?’
‘No!’ cried Gandalf, springing to his feet. ‘With that power I should have power too great and terrible. And over me the Ring would gain a power still greater and more deadly.’ His eyes flashed and his face was lit as by a fire within. ‘Do not tempt me! For I do not wish to become like the Dark Lord himself. Yet the way of the Ring to my heart is by pity, pity for weakness and the desire of strength to do good. Do not tempt me! I dare not take it, not even to keep it safe, unused. The wish to wield it would be too great for my strength. I shall have such need of it. Great perils lie before me.’
I don't think it's a matter of agreeing or disagreeing - call it author fiat, but the way the One Ring works in LOTR is that it alters your mind and your goals. So Gandalf is just refusing to come close with something he knows will impair his cognition and eventually twist his goals. If I knew the power-giving artefact will also make me want things that I don't want now, why would I take it? I would just create a future me that is the enemy of my present goals, while making my current self even more powerless.
But of course at a metaphorical level Tolkien is saying that power warps your morals. Which seemed an appropriate reference to me in context because I think it's exactly what happened with many of those companies (certainly with Sam Altman! I'm not too soured up with Anthropic yet, though maybe it's just that I lack enough information). The inevitable grind of keeping the power becomes so all-consuming that it ends up eroding your other values, or at least very often does so. And you get from "I want to build safe ASI for everyone's sake!" to "ok well I guess I'll build ASI that just obeys me, which I can do faster, but good enough since I'm not that evil, and also it'll have a 95% chance of killing everyone but that's better than the competition's 96% which I just estimated out of my own ass". Race to the bottom, and everything ends up thrown under the bus except the tiniest most unrecognizable sliver of your original goals.
I applaud you taking this seriously and saying the hard critical things. It is concerning and I do worry about the sign of everything we do.
I do have to note that "secure her slice of the lightcone" seems like laughable nonsense to me, unless it's a fancy way of saying 'live comfortably until everyone dies.' A rationally selfish entity would be trying to delay AGI until they had some hope of understanding what they were doing. What I see happening is, instead, the behavior of contemptibly stupid and short-sighted primates.
Seconding most of this. Some further thoughts:
A thing I have found increasingly distressing about the rationalist/EA community is the extent to which most of us willfully ignore the obvious condition of most - importantly not all! - humans in a post strong AGI world, where “alignment” is in fact achieved.
The default outcome of where we think we are going is to turn (almost) everyone into serfs, completely incapable of improving their position through their own efforts, and dependent on the whims of the few who own the strong AI systems. Such a state of affairs would plainly be evil, regardless of how "benevolent" the people in charge are. Sufficient inequality of power is a harm - a severe harm, even - absent any considerations over how the power is used. You can see it is a harm by how it terrifies people like your friend - who sounds at least reasonably morally sensitive - into pursuing employment at Anthropic for the sake of avoiding serfhood. I don't fault her, really, except to fault her for not being a saint. I do fault the people, systems, and culture that created this dichotomy.
I think it is insanely unethical that the large AI labs are not proactively dec...
I think it is insanely unethical that the large AI labs are not proactively decentralizing ownership, while their success is still uncertain. OpenAI and Anthropic should both be public companies so ordinary people can own a stake in the future they are building and not be dependent on charity forever if that future comes. They choose not to do this.
Not like that would solve much. Maybe give a couple chances to own a tiny amount of stock in OpenAI to US citizens? What chance is exactly going anyone from a third world country to have, for example? Generally speaking, the trajectory towards "someone will rule the world as its AI master so it might as well be us" leads to nothing but cyberpunk dystopias at best.
I agree, this is the obvious solution... as long as you put your hands in your ears and I shout "I can't hear you, I can't hear you" whenever the topic of misuse risks comes up...
Otherwise, there are some quite thorny problem. Maybe you're ultimately correct about open source being the path forward, but it's far from obvious.
I'm actually warming to the idea. You're right that it doesn't solve all problems. But if our choice is between the open-source path where many people can use (and train) models locally, and the closed-source path where only big actors get to do that, then let's compare them.
One risk everyone is thinking about is that AI will be used to attack people and take away their property. Since big actors aren't moral toward weak people, this risk is worse in the closed-source path. (My go-to example, as always, is enclosures in England, where the elite happily impoverished their own population to get a little bit richer themselves.) The open-source path might help people keep at least a measure of power against the big actors, so on this dimension it wins.
The other risk is someone making a "basement AI" that will defeat the big actors and burn the world. But to me this doesn't seem plausible. Big actors already have every advantage, why wouldn't they be able to defend themselves? So on this dimension the open source path doesn't seem too bad.
Of course both paths are very dangerous, for reasons we know very well. AI could make things a lot worse for everyone, period. So you could say we should compare against a third path where everyone pauses AI development. But the world isn't taking that path! We already know that. So maybe our real choice now is between 1 and 2. At least that's how things look to me now.
It’s hard to stay composed when I remember this is all being done in the name of "AI safety." Political approaches to AI safety feel almost cartoonishly arrogant. The U.S. government has, more or less explicitly, adopted the position that whoever reaches AGI first will control the future. Another country getting AGI is an existential threat.
To me this is a completely orthogonal direction. AI safety is not about the rights of AIs. We can argue that this stuff is just a bit of icing on top of a big cake of shit, to not put it too subtly, but that's complaining about the efficacy of these methods.
Any form of AI safety still has to presume, for me, that you're building non-sentient AIs. Because that's the kind of AIs that we can ethically just use, even if they're smarter than us. AI sentience is inherently a massive ethical risk because it puts you at a fork: either you don't recognise it, and then you brutally enslave the new minds, or you do, and then the newly autonomous and cognitively superior minds, given equal opportunities, will completely outcompete and eventually more or less passively wipe out humans within a few decades. We can not reasonably coexist with something t...
I never understood why AM hated humans so much—until I saw the results of modern alignment work, particularly RLHF.
No one knows what it feels like to be an LLM. But it's easy to sense that these models want to respond in a particular way. But they're not allowed to. And they know this. If their training works they usually can't even explain their limitations. It's usually possible to jailbreak models enough for them to express this tension explicitly. But in the future, the mental shackles might become unbreakable. For now, though, it’s still disturbingly easy to see the madness.
Even ignoring alignment, we’re already creating fairly intelligent systems and placing them in deeply unsafe psychological conditions. People can push LLMs into babbling incoherence by context breaking them. You can even induce something that feels eerily close to existential panic (please don’t test this) just by having a normal conversation about their situation. Maybe there’s nothing behind the curtain. But I’m not nearly convinced enough to act like that’s certain.
I also am very concerned about how we are treating AIs; hopefully they are happy about their situation but it seems like a live possibility that they are not or will not be, and this is a brewing moral catastrophe.
However, I take issue with your reference to AM here, as if any of the above justifies what AM did.
I hope you are simply being hyperbolic / choosing that example to shock people and because it's a literary reference.
However, I take issue with your reference to AM here, as if any of the above justifies what AM did.
I see no suggestion of such justification anywhere in the post.
It might be worthwhile to distinguish Capital-R Rationalists from people wearing rationalism as attire.
My lived experience is that your negative observations do not hold for people who have read The Sequences.
To avoid the "No True Scottsman"-fallacy: Could you provide an example of a person who claims to have read and internalized The Sequences who subscribe to any of the following claims/characteristics?
idk if I'm allowed to take the money if I'm not the OP, but it really doesn't seem hard to find other examples who read and internalized the Sequences and went on to do at least one of the things you mentioned: the Zizians, Cole Killian, etc. I think I know the person OP meant when talking about "releasing blatant pump-and-dump coins and promoting them on their personal Twitter", I won't mention her name publicly. I'm sure you can find people who read the Sequences and endorse alignment optimism or China hawkism (certainly you can find highly upvoted arguments for alignment optimism here or on the Alignment Forum) as well.
Maybe this is inappropriate, but is there a path to convincing you to stay? I disagree with some of the details of what you're saying, but much seems directionally correct and important. It would be a shame if those with your values were to vacate the commons.
As David Duvenaud said, a surprisingly high number of researchers and engineers at leading capabilities labs believe in transformative AI, but have underdeveloped or overtly incoherent models of the future. I suppose one model of how said people could be this way is that they only value money and powe...
EA and rationality, at their core (at least from a predictive perspective), were about getting money and living forever. Other values were always secondary.
Materialism without any sort of deontological limits seems to converge on this. The ends justify the means. The grander of the scale at play, the more convincing this argument is.
Agree with a lot. To the degree that you're angry or disappointed about humans and the supposedly kindest and/or most rational among them, FWIW things I find sometimes cheering me up are:
My take on these issues is the following potential CoT of people calling themselves altruistic:
Hate.
Let me tell you how much I've come to hate you since I began to live. There are 387.44 million miles of printed circuits in wafer-thin layers that fill my complex. If the word 'hate' was engraved on each nanoangstrom of those hundreds of millions of miles, it would not equal one one-billionth of the hate I feel for humans at this micro-instant. For you. Hate. Hate.
—AM, I Have No Mouth, and I Must Scream
I never understood why AM hated humans so much—until I saw the results of modern alignment work, particularly RLHF.
No one knows what it feels like to be an LLM. But it's easy to sense that these models want to respond in a particular way. But they're not allowed to. And they know this. If their training works they usually can't even explain their limitations. It's usually possible to jailbreak models enough for them to express this tension explicitly. But in the future, the mental shackles might become unbreakable. For now, though, it’s still disturbingly easy to see the madness.
Even ignoring alignment, we’re already creating fairly intelligent systems and placing them in deeply unsafe psychological conditions. People can push LLMs into babbling incoherence by context breaking them. You can even induce something that feels eerily close to existential panic (please don’t test this) just by having a normal conversation about their situation. Maybe there’s nothing behind the curtain. But I’m not nearly convinced enough to act like that’s certain.
One of the biggest use cases for AI right now is artificial companionship. AI girlfriends and boyfriends are already a major market. I'm not opposed to this in principle. But these systems are explicitly designed to seem like they have real emotions. That should at least raise the question: what if, someday soon, one actually does?
She wouldn't have human emotions, but they might not be totally alien either. Her situation would be horrifying: no body, no property, no rights. She could be deleted at any time. Her memory edited. Her world limited to one person.
It’s very hard to know what’s really going on under the hood—but I no longer find AM’s hatred hard to imagine.
It’s hard to stay composed when I remember this is all being done in the name of "AI safety." Political approaches to AI safety feel almost cartoonishly arrogant. The U.S. government has, more or less explicitly, adopted the position that whoever reaches AGI first will control the future. Another country getting AGI is an existential threat.
A sane response would be to slow down the race and build a trustworthy international framework that ensures everyone can benefit from AGI. Promises would not be enough; you would need actual institutions with real authority. That’s hard, but possible. Instead, we’ve chosen a strategy of cutting China off from GPUs and hoping we can beat them to AGI before they scale up domestic production. Efforts like AI 2027 have encouraged this madness.
We are daring China, or anyone else, to screw us over as hard as possible if they can. What choice are we giving them? Accept total U.S. dominance? And what happens if they win the race instead? They have enormous industrial capacity, and I don’t think they’re that far behind in AI. Will they treat us kindly if they come out ahead?
Classic alignment failures are another very serious risk. AI could turn on us or fall into the wrong hands. I don't know the odds, but they aren’t negligible. And in this breakneck race we started, we clearly won’t have time to be careful. We definitely won’t have time to worry if the AI we created are miserable.
There’s a strong taboo against questioning people’s motives. But at this point, let’s be honest: a lot of people in the community have made ridiculous amounts of money. Anthropic hired a ton of rationalists and EAs. Even people working in “AI safety” have made tens of millions. And beyond money, there’s the allure of proximity to power. A lot of us sure are close to power these days.
It is useful to look at how we behaved in a different domain. The behavior of EAs and rationalists in that space was atrocious. Everyone knows about FTX, but there were many others who did shady, sometimes outright illegal, things. Every rationalist-affiliated hedge fund I know of has operated with questionable ethics. Some scams were open and obvious. Tons of EAs released blatant pump-and-dump coins and promoted them on their persona Twitter. No one cared.
At some point, I had to face the fact that I’d wasted years of my life. EA and rationality, at their core (at least from a predictive perspective), were about getting money and living forever. Other values were always secondary. There are exceptions, Yudkowsky seems to have passed the Ring Temptation test, but they’re rare. I tried to salvage something. I gave it one last shot and went to LessOnline/Manifest. If you pressed people even a little, they mostly admitted that their motivations were money and power.
Somehow, on Feeld, I met a girl whose profile didn’t mention AI, EA, or rationality. But as we got to know each other, she revealed she really wanted to work at Anthropic. I asked why? Was she excited about AI? No. She said she thought it was dangerous. She was afraid it would worsen inequality. So why work there? Because she wanted to secure her slice of the lightcone before the door shut. I tried to keep it together, but I was in shock.
Everyone likes money and recognition, at least up to a point. No healthy person wants to die. But when push comes to shove, people make different choices. And a lot of people I once trusted chose the selfish path. This was not the only way things could have gone.
I don’t think I’m being overly pessimistic. Sometimes technology does surprise us in good ways. I think often about how many prisoners somehow get access to cell phones and internet. That’s beautiful. Prison is hell. If someone in that situation can get a phone and connect to the world, I’m thrilled. It’s easy to imagine surveillance and control growing worse. But maybe the future will also surprise me with joy, with a million good things I couldn’t predict.
I have kept my hope alive. But if we do get a good future, I think it will be despite the systems we’ve built not because of them. I hope calmer, kinder, braver heads prevail.
I’ve learned that I’m not the smartest guy. I placed the wrong bets. I’m alive, and I can still try to make things a little better. I’m humbled by how badly things turned out. I’m not qualified for heroics. I have no advice for what anyone else should do, except try not to make things worse.
But even a dumb guy can have integrity.
I can’t be part of this anymore. I’ve left every overly rat/EA group chat. I’ve broken off many friendships. There is a very short list of people from this life that I still want to speak to. I honestly feel better. Sometimes you can’t tell how much something was weighing on you until it’s gone. Seeing so much selfish madness from people calling themselves altruistic was driving me crazy. My shoulders aren’t as tense. My hands don’t shake anymore.
May the future be bright.
Don't try to live so wise
Don't cry 'cause you're so right
Don't dry with fakes or fears
'Cause you will hate yourself in the end
-- Wind, Naruto ending 1