Agent-foundations researcher. Working on Synthesizing Standalone World-Models, aiming at a timely technical solution to the AGI risk fit for worlds where alignment is punishingly hard and we only get one try.
Currently looking for additional funders ($1k+, details). Consider reaching out if you're interested, or donating directly.
Or get me to pay you money ($5-$100) by spotting holes in my agenda or providing other useful information.
Eh, I don't think this really disproves the OP (who mentions bejeweled iPhones as well). This isn't really a "$100k phone", it's a normal $1k phone with $99k of jewelry tastelessly bolted on. You're not getting a 50,000 mAh battery, a gaming GPU, and Wi-Fi 12 plus 6G support in a 300 gram package here. Which, indeed: why not?
I've previously left a comment describing my being fairly unimpressed with Claude Code Opus 4.5, and leaving a few guesses regarding what causes the difference in people's opinion regarding its usefulness. Eight more days into it, I have new comments and new guesses.
tl;dr: Very useful (or maybe I'm also deluding myself), very hard to use (or maybe I have skill issues). If you want to build something genuinely complicated with it, it's probably worth it, but it will be an uphill battle against superhuman-speed codebase rot, and you will need significant technical expertise and/or the ability to learn that expertise quickly.
First, what I'm attempting is to use it to implement an app that's not really all that complicated, but which is still pretty involved, still runs fairly nontrivial logic at the backend (along the lines of this comment), and about whose functionality I have precise desiderata and very few vibes.
Is Opus 4.5 helpful and significantly speeding me up? Yes, of that there is no doubt. Asking it questions about the packages to use, what tools they offer, and how I could architecture solutions to various problems I run into, is incredibly helpful. Its answers are so much more precise than Google's, it distill information so much better than raw code documentation does, and it's both orders of magnitude faster than StackOverflow and is smarter than the median answer you'd get there.
Is Claude Code helpful and speeding me up? That is more of an open question. Some loose thoughts:
To sum my current view up: Seems useful, but hard to use. You'll have to fight it/the decay it spreads in its wake every step of the way, and making a misstep will give your codebase lethal cancer.
We'll see how I feel about it in one more week, I suppose.
Seconded. Went from a skeptical "big if true" at the post title to rolling my eyes once I saw "iruletheworldmo".
For reference, check out this leak by that guy from February 2025:
ok. i’m tired of holding back. some of labs are holding things back from you.
the acceleration curve is fucking vertical now. nobody's talking about how we just compressed 200 years of scientific progress into six months. every lab hitting capability jumps that would've been sci-fi last quarter. we're beyond mere benchmarks and into territory where intelligence is creating entirely new forms of intelligence.
watched a demo yesterday that casually solved protein folding while simultaneously developing metamaterials that shouldn't be physically possible. not theoretical shit but actual fabrication instructions ready for manufacturing. the researchers presenting it looked shell shocked. some were laughing uncontrollably while others sat in stunned silence. there's no roadmap for this level of cognitive explosion.
we've crossed into recursive intelligence territory and it's no longer possible to predict second order effects. forget mars terraforming or fusion. those are already solved problems just waiting for implementation. the real story is the complete collapse of every barrier between conceivable and achievable. the gap between imagination and reality just vanished while everyone was arguing about risk frameworks. intelligence has broken free of all theoretical constraints and holy fuck nobody is ready for what happens next week. reality itself is now negotiable.
I guess it's kind of entertainingly written.
It seems like, yes, he is saying that wealth levels get locked in by early investment choices, and then that it is ‘hard to justify’ high levels of ‘inequality’ and that even if you can make 10 million a year in real income in the post-abundance future Larry Page’s heirs owning galaxies is not okay.
I say, actually, yes that’s perfectly okay, provided there is stable political economy and we’ve solved the other concerns so you can enjoy that 10 million a year in peace.
I dunno about that. I think it is not okay for directionally the same reasons it wouldn't be okay if we got an "infinitesimally aligned" paperclip maximizer who leaves the Solar System alone but paperclips the rest of the universe: astronomical waste.
Like, suppose 99% of the universe ends up split between twenty people, with them using it as they please, in ways that don't generate much happiness for others. Arguably it's not going to be that bad even in the "tech-oligarch capture" future (because Dario Amodei has made a pledge to donate 10% of his earnings or whatever[1]), but let's use that to examine our intuitions.
One way to look at it is: this means the rest of civilization will only end up 1% as big as it could be. This argument may or may not feel motivating to you; I know "more people is better" is not a very visceral-feeling intuition.
Another way to look at it is: this means all the other people will only end up with 1% of the lifespan they could have had. Like, in the very long term, post-scarcity isn't real, the universe's resources are finite (as far as we currently know), and physical entities need to continuously consume those to keep living. If 99% of resources are captured by people who don't care to share them, everyone else will end up succumbing to the heat death much faster than in the counterfactual.
This is isomorphic to "the rich have their soldiers take all timber and leave the poor to freeze to death in the winter". The only reason it doesn't feel the same way is because it's hard to wrap your head around large numbers: surely you'd be okay with only living for 100 billion years, instead of 10 trillion years, right? In the here and now, both numbers just round up to "effectively forever". But no, once you actually get to the point of you and all your loved ones dying of negentropic starvation 100 billion years in, it would feel just as infuriatingly unfair.
I understand the tactical moves of "we have to pretend it's okay if the currently-powerful capture most of the value of the universe, so that they're more amicable to listening to our arguments for AI safety and don't get so scared of taxes they accelerate AI even further" and "we have to shut down the discussion of 'but which monkey gets the banana?' at every turn because it competes with the 'the banana is poisoned' messaging". But if we're keeping to Simulacrum Level 1, no, I do not in fact believe it's okay.
I also don't necessarily agree that those moves are pragmatically good. It's mostly pointless to keep talking to AI-industry insiders; if we're doing any rhetoric, it should focus on "outsiders". And if it is in fact true that the current default trajectory of worlds in which the AGI labs' plans succeed may lead to something like the above astronomical-waste scenarios, making those arguments to the general public is potentially a high-impact move. "Ban the AGI research because otherwise the rich will take all the stuff" is a much more memetically viral message than "ban the AGI because Terminator".
(To be clear, I'm not arguing we should join various coalitions making false arguments to that end, e. g. the datacenter water thing. But if there are true arguments of that form, as I believe there are...)
I do not trust that guy to keep such non-binding promises, by the way. His track record isn't good, what with "Anthropic won't advance the AI frontier".
I note that the wording in the more direct sources (rather than paraphrases) is "preventive war" and "bomb them", which doesn't actually strictly imply preventive nuclear bombings. It's plausible that "bomb them" and "war with the USSR" could only mean "nuclear war" in-context... But it'd also be really funny if this is another "Eliezer advocates nuking foreign datacenters" situation.
I don't think we necessarily disagree, here? I can imagine it being useful in this context, and I nod along at your "no loose leash" and "doesn't feel like an employee, but a powerful tool" caveats.
What sort of codebase are you working on?
This specifically was an attempt to vibe-code from scratch a note-management app that's to my tastes; think Obsidian/Logseq, broken down into small modules/steps.
As a simple example, I've had a "real project" benchmark for awhile to convert ~2000 lines of test cases from an old framework to a new one
My guess is that "convert this already-written code from this representation/framework/language/factorization to this other one" may be one of the things LLMs are decent at, yep! This is a direction I'm exploring right now: hack the codebase together using some language + factorization in which it's easy (but perhaps unwise) to write, then try to use LLMs to "compile" it into a better format.
But this isn't really "vibe-coding"/"describe the spec in natural language and watch the LLM implement it!"/"programming as a job is gone/dramatically transformed!", the way it's being advertised. LLMs are not, it seems, actually good at mapping natural-language descriptions into non-hack-y, robust background logic. You need a "code-level" prompt to specify the task precisely enough. And there's only one way to bring that code into existence if LLMs can't do it for you.
Language Models Offer Mundane Utility
Very preliminary opinion here, I've not yet spent enough time messing with it to be confident, but all these "Opus 4.5 in Claude Code can do anything!!!" experiences seem completely alien to mine. I can make Opus 4.5 sort of kind of implement not-entirely-trivial features if I do enough chewing-up and hand-holding and manual bug-reporting (its self-written tests are not sufficient). But it can't autonomously code its way out of a wet paper bag.
And yes, I've been to Twitter, I've tried everything people's been suggesting. We designed a detailed tech specification, a solid architecture, and a step-by-step implementation plan with it beforehand, and I asked it to do test-driven development and to liberally use AskUserQuestionTool at me. I've also tried to do the opposite, starting with a minimal "user-facing features" spec and letting it take the wheel. The frustrated tone of this comment wasn't a factor either, I've been aiming to convey myself in a clear and polite manner.[1] None of that worked, I detect basically no change since August.
My current guess is that we have a massive case of this happening. All the people raving about CCO4.5 being an AGI with no limits happen to be using it on some narrow suite of tasks,[2] and everyone else just thinks they have skill issues, so they sit quiet.
Or maybe I indeed have skill issues. We'll see, I suppose. I'll keep trying to figure out how to use it/collaborate with it.
I expect there's indeed some way to wring utility out of LLMs for serious coding projects. But I'm also guessing that most of this frippery:
agents, subagents, their prompts, contexts, memory, modes, permissions, tools, plugins, skills, hooks, MCP, LSP, slash commands, workflows, IDE integrations
– will not end up very useful for that task.
(People often say that AI progress feels slow/stalled-out only if you're not interacting with frontier LLMs in a technical capacity, if you're operating off of months-outdated beliefs and chatbot conversations. It's been the opposite for me: every time I take some break from heavier coding and only update based on other people's experiences with newer models, I get psyop'd into believing that there's indeed been a massive leap in AI capabilities, I get concerned, and then I go back and I find my timelines growing to unprecedented lengths.)
I suppose I haven't tried the opposite on that matter. Maybe you do need to yell at LLMs for them to start working?
E. g., maybe it's:
Have you tried double-blinding yourself, Gwern-style? (I. e., prepare a box with glucose capsules and a box with some placebo (empty capsules + close your eyes when taking them out of the box?); at the start of the day, pick the box to take capsules from at random; at the end of the day, try to guess whether it was the placebo box; then check; repeat for several days.)
One obvious problem is that this turns getting funded into a popularity contest, which makes Goodhart kick in. It might work fine as a one-off thing, but in the long run, it will predictably get gamed, and will likely have negative effects on the whole LW discussion ecosystem by setting up perverse incentives for engaging with it (and, unless the list of eligible people is frozen forever, attracting new people who are only interested in promoting themselves to get money).
What should be the amount? Thiel gave 200k. Is it too much for 2 years? Too little?
You should almost certainly have some mechanism for deciding the amount to pay on a case-by-case basis, rather than having it be flat.
Could there be an entirely different approach to finding fellows? How would you do it?
What I would want to experiment with is using prediction markets to "amplify" the judgement of well-known people with unusually good AGI Ruin models who are otherwise too busy to review thousands of mostly-terrible-by-their-lights proposals (e. g., Eliezer or John Wentworth). Fund the top N proposals the market expects the "amplified individual" to consider most promising, subject to their veto.
This would be notably harder to game than a straightforward popularity contest, especially if the amplifee is high-percentile disagreeable (as my suggested picks are).
Tried it out a couple times just now, it appears specialized for low-level, syntax-level rephrasings. It will inline functions and intermediate-variable computations that are only used once and try to distill if-else blocks into something more elegant, but it won't even attempt doing things at a higher level. Was very eager to remove Claude's own overly verbose/obvious comments, though. Very relatable.
Overall, it would be mildly useful in isolation, but I'm pretty sure you can get the same job done ten times faster using Haiku 4.5 or Composer-1 (Cursor's own blazing-fast LLM).
Curious if you get a different experience.