A lot of the rationalist discourse around birthrates don't seem to square away with AGI predictions.
Like the most negative predictions of AGI destroying humanity in the next century or two leaves birthrates completely negligent as an issue. The positive predictions with AGI leave a high possibility of robotic child rearing and artificial wombs (when considering the amount of progress even us puny humans have already made) in the next century or two which also makes natural birthrates irrelevant because we could just make and raise more humans without the n...
Though noteworthy that he has taken no steps towards technical solutions for the birth rate issue
I mean, as individual men in the West go, having fourteen children is pretty above-average, and he seems to have gotten the process down to a science. He's not a Saudi oil baron with three digits of offspring, but he's certainly taken a Silicon Valley approach to it.
Pro-Natalists, in general, seem to take a 'lead-by-example' tack, which isn't horrible considering that it demonstrates an understanding of the consequences of materially encouraging people who woul...
This shortpost is just a reference post for the following point:
It's very easy for conversations about LLM beliefs or goals or values to get derailed by questions about whether an LLM can genuinely said to believe something, or to have a goal, or to hold a value. These are valid questions! But there are other important questions about LLMs that touch on these subjects, which don't turn on whether an LLM belief is a "real" belief. It's not productive for those discussions to be so frequently derailed.
I've taken various approaches to this proble...
Assuming people accept the model that LLM behavior is primarily determined by modeling the behavior of some subset (such that fine-tuning works primarily by shaping the subset of humans that the model emulates) of the human writers it was trained on, it might be simplest to ask whether the model "behaves like a person who believes X".
This framing carries practical benefits (again, so long as you agree with the assumption above), in that the fine-tuning paradigm can be examined in the context of identifying what causes the model to upweight, say, the "a bla...
A response to someone asking about my criticisms of EA (crossposted from twitter):
EA started off with global health and ended up pivoting hard to AI safety, AI governance, etc. You can think of this as “we started with one cause area and we found another using the same core assumptions” but another way to think about it is “the worldview which generated ‘work on global health’ was wrong about some crucial things”, and the ideology hasn’t been adequately refactored to take those things out.
Some of them include:
Some of them include:
imo a larger one is something like not rooting the foundations in "build your own models of the world so that you contain within you a stack trace of why you're doing what you're doing" + "be willing to be challenges and update major beliefs based on deep-in-the-weeds technical arguments, and do so from a highly truth-seeking stance which knows what it feels like to actually understand something not just have an opinion".
Lack of this is fineish in global health, but in AI Safety generates a crop of people with only surface deferral flavor understanding of the issues, which is insufficient to orient in a much less straightforward technical domain.
I'm working on a piece for Asterisk magazine on how biohackers (broadly defined) think about risk. Are you interested in being interviewed? Do you know someone who is? I'm taking a very broad definition of biohacker (self-medicating to save money counts), but some people I'm especially interested in talking to are those who:
Depends. Are you strictly following standard of care, or personalizing for yourself?
"Most people make the mistake of generalizing from a single data point. Or at least, I do." - SA
When can you learn a lot from one data point? People, especially stats- or science- brained people, are often confused about this, and frequently give answers that (imo) are the opposite of useful. Eg they say that usually you can’t know much but if you know a lot about the meta-structure of your distribution (eg you’re interested in the mean of a distribution with low variance), sometimes a single data point can be a significant update.
This type of limited conc...
wow thanks! It's the same point but he puts it better.
I'm interested in soliciting takes on pretty much anything people think Anthropic should be doing differently. One of Alignment Stress-Testing's core responsibilities is identifying any places where Anthropic might be making a mistake from a safety perspective—or even any places where Anthropic might have an opportunity to do something really good that we aren't taking—so I'm interested in hearing pretty much any idea there that I haven't heard before.[1] I'll read all the responses here, but I probably won't reply to any of them to avoid revealing anythin...
Glad to see prefill was disabled for Opus 4.6!
One thing that I find somewhat confusing is that the "time horizon"-equivalent for AIs reading blog posts seems so short. Like this is very vibes-y, but if I were to think of a question operationalized as "I read a blog post for X period of time, at what X would I think Claude has a >50% chance of identifying more central errors than I could?" intuitively it feels like X is very short. Well under an hour, and likely under 10 minutes.
This is in some sense surprising, since reading feels like a task they're extremely natively suited for, and on other tasks like programming their time horizons tend to be in multiple hours.
I don't have a good resolution to this.
The task you're talking about isn't reading, it's logical analysis or something like that.
I think it is noteworthy that current llms are bad at this.
Of course, what counts as an error within the local context of the essay and what counts as an error? Given all the shared context, the writer and reader rely on to interpret. It, is highly subjective and debatable. So you need some sort of committee of expert humans to compare to.
links 2/11/26: https://roamresearch.com/#/app/srcpublic/page/02-11-2026
Elon Musk is what Scott Adams was trying to be.
In Scott Alexander's "Dilbert Afterlife" he pointed out that Adams had the idea that if one was so smart, the highest ROI use of that strength was to try to control lesser minds by taking advantage of their vulnerabilities. His hypnosis and "linguistic killshots" were attempts at this.
Listening to Elon on the Dwarkesh pod and I feel like they are finally exposing him. Dwarkesh and John Collison are engaging in more or less rationalist discourse and asking all the obvious follow-up questions, and Elon just has ...
we try to operate on simulacra level 1 (genuinely saying what we mean) and it turns out that that is by far the weakest of the four.
Level 1 allows you to work on something alone. So if you can't find other people who would cooperate with you (e.g. because you are a little autistic), level 1 is the only one you can work at. Your options are to work at level 1 alone, or serve other people, or fail.
Which I suspect might have also been Scott Adams' problem. He could see that the other levels work better, but he couldn't work at them efficiently. So he just kep...
Are models like Opus 4.6 doing a similar thing to o1/o3 when reasoning?
There was a lot of talk about reasoning models like o1/o3 devolving into uninterpretable gibberish in their chains-of-thought, and that these were fundamentally a different kind of thing than previous LLMs. This was (to my understanding) one of the reasons only a summary of the thinking was available.
But when I use models like Opus 4.5/4.6 with extended thinking, the chains-of-thought (appear to be?) fully reported, and completely legible.
I've just realised that I'm not sure what's goin...
Hmm, but when you use these models in the chat interface, you can literally open up the reasoning tab and watch it be generated in real time? It feels like there isn’t enough time here for that reasoning to have been generated by a summarizer
The benefits of lazy parenting.
My wife and I often look at ourselves as lazy parents. From as early as the kids could manage, we would expect them to get up and get their own water or snack or dress themselves or even shower themselves. On some occasions, the kids, at 4 and 6, would do the entire bedtime routine, saying goodnight downstairs, and then go up, shower, dress, teeth, and bed on their own.
Sometimes we feel guilty, but the kids love the independence we give them. We give them endless love, we show it, but we avoid doing things for the...
I agree. The question is when to start being lazy about what.
I have a concept that I am considering making a post about: How culture war and political signaling is all about boundary maintenance rather than truth seeking. And that if you have a label, which you decide to believe in as your identity, then the person who controls the definition of that label by proxy controls you. If you are a son of god, then who are under the control of whoever decides what god is and what he wants. Similarly, if you believe that being a liberal is the only way to be a person who “believes in human rights” than you are under the con...
The idea seems correct. If you identify as X, and an authority that you respect says "all true X believe Y", you are more likely to accept Y uncritically. Especially if other Xs around you accept it; that creates social pressure.
Politics is considered a minefield here, so you would have to write the article carefully, to avoid tribalism. (Basically, do not mix "this is an analysis of how political tribes work" with "this is my favorite political tribe" in the same article.) It would probably be better to use multiple examples from various sides, rather tha...
For machine learning, it is desirable for the trained model to have absolutely no random information left over from the initialization; in this short post, I will mathematically prove an interesting (to me) but simple consequence of this desirable behavior.
This post is a result of some research that I am doing for machine learning algorithms related to my investigation of cryptographic functions for the cryptocurrency that I launched (to discuss crypto, leave me a personal message so we can discuss this off this site).
This post shall be about linear ...
Yes. When we take convex combinations of finitely many point mass measures, the integral is just a sum. I use the sum of finitely many elements for ease of calculations, but to prove theorems, I should use measures for full generality.
The idea of finding an object along with distinct local optima with maximized looks like an interesting problem to work on. I have not worked on this kind of objective before, but I can certainly try this, as I have a few ideas of how to do this. This might work better for dis...
In case I don't write a full post about this:
The question whether reversible computation produces minds with relevant moral status is extremely important. Claude estimates me that it'd be a difference between having and mind-seconds instantiable in the reachable universe. (Because reversible minds could stretch our entropy budget long into the black hole era.)
Question is whether the reversing of the computation that makes up the mind and the lack of output (that'd imply bit-erasure) entail that the mind "didn't really exist".
There ar...
Yeah… seems right. I could cop out with "ah, the number of bit erasures by a program matters, not just whether it did", but I don't have any good reason for believing this.
Update: Brushing after eating acidic food is likely fine.
Context: 7 months ago, me in Adam Zerner's shortform:
...I remember something about not brushing immediately after eating though. Here is a random article I googled. This says don't brush after eating acidic food, not sure about the general case.
https://www.cuimc.columbia.edu/news/brushing-immediately-after-meals-you-may-want-wait
“The reason for that is that when acids are in the mouth, they weaken the enamel of the tooth, which is the outer layer of the tooth,” Rolle says. Brushing immediately afte
Thank you for double checking.
Overall I am still very uncertain, but lean towards it being fine. Even dentists are giving mixed signals.
Unfortunately Chinese sources but these are dentists saying you can brush immediately
Definitely there are dentists saying you should wait too
Just thinking out loud: If AI (or something else) causes the S&P500 to go up by 50%...
By January 21st, 2028 then you can achieve a 1324% ROI (https://finance.yahoo.com/quote/SPY280121C01000000/)
By June 16, 2028 then you can achieve a 659% ROI (https://finance.yahoo.com/quote/SPY280616C01000000/)
By December 15, 2028 then you can achieve a 315% ROI (https://finance.yahoo.com/quote/SPY281215C01000000/)
Curious to hear others thoughts, especially those who have short AGI timelines. *Of course this is not financial advice
The music industry has already passed the threshold for digital AGI, but seems to be experiencing disruption dramatically less than AI song generation capabilities might suggest. Anyone can create a song of expert level quality in under 10s with an automatically generated prompt (or write one themselves if they want). The Chainsmokers took about 8 hours in total to make the single Roses, which has over 1.3bn streams as of 02/10/2026 and I believe is on the quicker end of writing time for pop songs, although I am not sure reliable data exists for this anywh...
I remember a hygienist at the dentist once telling me that toothpaste isn't a huge deal and that it's the mechanical friction of the toothbrush that provides most of the value. Since being told that, after a meal, I often wet my toothbrush with water and brush for 10 seconds or so.
I just researched it some more and from what I understand, after eating, food debris that remains on your teeth forms a sort of biofilm. Once the biofilm is formed you need those traditional 2 minute long tooth brushing sessions to break it down and remove it. But it takes 30+ mi...
Update: Brushing after eating acidic food is likely fine.
Should We Wait to Brush Our Teeth? A Scoping Review Regarding Dental Caries and Erosive Tooth Wear
Key messages: Although the available evidence lacked robust clinical studies, tooth brushing using fluoridated products immediately after an erosive challenge does not increase the risk of ETW [Erosive Tooth Wear] and can be recommended, which is in line with recommendations for dental caries prevention. Furthermore, we suggest updating the international guidelines to promote individualized recommendatio...
Even now and then I meet someone who tries to argue that if I don't agree with them this is because I'm not open mided enough. Is there a term for this?
Epistemically I'm not convinced buy this type of arugment, but socialy it feels like I'm beeing shamed, and I hate it.
I also find it hard to call out this type of behaviur when it happens, even when I can tell exactly what is going on. I think it I had a name for this behaviour it would be easier? Not sure though?
Edit to add:
I've now got some more time to figure out what I want and don't want out of this th...
A more transparent term would be psychologizing:
psychologize: to speculate in psychological terms or on psychological motivations
See also Ayn Rand on this topic:
(...) Just as reasoning, to an irrational person, becomes rationalizing, and moral judgment becomes moralizing, so psychological theories become psychologizing. The common denominator is the corruption of a cognitive process to serve an ulterior motive.
...Psychologizing consists in condemning or excusing specific individuals on the grounds of their psychological problems, real or invented, in
Model to track: You get 80% of the current max value LLMs could provide you from standard-issue chat models and any decent out-of-the-box coding agent, both prompted the obvious way. Trying to get the remaining 20% that are locked behind figuring out agent swarms, optimizing your prompts, setting up ad-hoc continuous-memory setups, doing comparative analyses of different frontier models' performance on your tasks, inventing new galaxy-brained workflows, writing custom software, et cetera, would not be worth it: it would take too long for too little payoff....
I lean towards this, despite being a guy currently heavily invested in AI tools.
Cluster of things that all seem true to me:
But: