I dug a fossil-version Google Glass XE HW2 out of a drawer, pre-rooted, running an AOSP build that I think has a single-digit number of users worldwide. I connected it to the mac mini's USB port, and told my Openclaw instance to get it onto tailscale and set it up as a communication channel.
It worked its way through multiple absurd, frustrating technical issues that would absolutely have made me give up if I was the one doing it, with only minimal guidance. Once it had ssh working, it set up an android app. Without me suggesting it do so, it found a way to...
I don't think Cursor would've stood a chance, for this task. It was almost all command-line wrangling with only a small side-order of actual coding. Lots of "run this command and run other tools while it's in-progress to figure out why it's crashing". One "abort that command because it's running too slow and try a different command". Some explicit wait-for-timer-then-recheck steps, including "send a Discord message telling me to auth a tailscale node then poll until it's authed". After it got to the stage where it could connect over the network instead of ...
I found the recent dialogue between Davidad and Gabriel Alfour and other recent Davidad writings quite strange and under-discussed. I think of Davidad as someone who understands existential risks from AI better than almost anyone; he previously had one of the most complete plans for addressing it, which involved crazy ambitious things like developing formal model of the entire world.
But recently he's updated strongly away from believing in AI x-risk because the models seem to be grokking the "natural abstraction of the Good". So much so that current agents...
I think it's a holdover from the early days of LLMs, when we had no idea what the limits of these systems were, and it seemed like exploring the latent space of input prompts could unlock very nearly anything. There was a sentiment that, maybe, the early text-predictors could generalize to competently modeling any subset of the human authors they were trained on, including the incredibly capable ones, if the context leading up to a request was sufficiently indicative of the right things. There was a massive gap between the quality of outputs without a good...
there seem to be three different possible levels of manager involvement in individual researchers:
Definitely just a bug! I'll make sure to fix it early next week.
A lot of the rationalist discourse around birthrates don't seem to square away with AGI predictions.
Like the most negative predictions of AGI destroying humanity in the next century or two leaves birthrates completely negligent as an issue. The positive predictions with AGI leave a high possibility of robotic child rearing and artificial wombs (when considering the amount of progress even us puny humans have already made) in the next century or two which also makes natural birthrates irrelevant because we could just make and raise more humans without the n...
I think it’s just people compartmentalizing, trying to hold onto normalcy by acting and even thinking as if the AI is not going to come and change everything soon.
(meta: this is a short, informal set of notes i sent to some folks privately, then realized some people on LW might be interested. it probably won't make sense to people who haven't seriously used Anki before.)
have people experimented with using learning or relearning steps of 11m <= x <= 23h ?
just started trying out doing a 30m and 2h learning & relearning step, seems like it solves mitigates this problem that nate meyvis raised
reporting back after a few days: making cards have learning steps for 11m <= x <= 23h makes it feel m...
brief update: I still do this, and still like it.
What if human empathy didn't really generalize to other animals as an "evolutionary accident?" (As assumed here in the comments)
Maybe the real reason was that evolution wanted to stop prehistoric humans from killing off all their prey, leaving them no food for tomorrow. Maybe they spared the young animals and the females because killing them was the most costly for future hunts.
This is more reason to suspect empathy might not generalize by default.
Your tribe hypothetical is irrelevant and all 4 of your real examples are straightforwardly (and usually) explained by greedy inclusive fitness, and do not come anywhere close to providing 3 examples of comparable mechanisms.
Maybe the most important test for a political or economic system is whether it self-destructs. This is in contrast to whether it produces good intermediate outcomes. In particular, if free-market capitalism leads to an uncontrolled intelligence explosion, then it doesn’t matter if it produced better living standards than alternative systems for ~200 years – it still failed at the most important test.
A couple of other ways to put it:
Under this view, p...
What I meant was, if you're trying to discriminate between political/economic systems and notice "21st century capitalism / mix-mash-distribution-of-democracy-and-various-flavors-of-authoritatian-etc" doesn't look on track to successfully navigate superintelligence, well, that might be true, but, it's probably also true of most political/economic-systems that aren't Dath Ilan or similar.
It seems true, but, something like "there are kinds of different ways to fail, there are relatively few ways to succeed, so failure isn't actually that informative."
Does anyone have a canonical resource they’d point me to for understanding how to file/optimize (personal, US/California, income) taxes?
I found the “Taxes for Dummies” book pretty poorly written; I’m enjoying reading “Taxes Made Simple” by Mike Piper, but it really only covers the basics.
Some examples of the sort of thing I’m looking for in other categories:
Does anyone know why the early Singularity Institute prioritized finding the correct solution to decision theory as an important subproblem of building a Friendly AI?
Wei Dai recently said that the concern was something like...
we have to fully solve DT before building AGI/ASI, otherwise it could be catastrophic due to something like the AI falling prey to an acausal threat or commitment races, or can't cooperate with other AIs.
This seems like a very surprising reason to me. I don't understand why this problem needed to be solved before the intelligence expl...
@Rob Bensinger on the EA Forum:
...As a side-note, I do want to emphasize that from the MIRI cluster's perspective, it's fine for correct reasoning in AGI to arise incidentally or implicitly, as long as it happens somehow (and as long as the system's alignment-relevant properties aren't obscured and the system ends up safe and reliable).
The main reason to work on decision theory in AI alignment has never been "What if people don't make AI 'decision-theoretic' enough?" or "What if people mistakenly think CDT is correct and so build CDT into their AI system?" Th
the boots theory of poverty doesn't make much sense in a free market economy. if there are high quality goods that are actually more cost effective in the long run but require a larger capital outlay than lower quality goods, then it is economically rational to take out debt to pay for the capital outlay. and in fact, BNPL providers have made such loans commonplace for all manner of consumer goods.
I know many members and influential figures of this are atheists; regardless, does anyone think it would be a good idea to take a rationalist approach to religious scripture? If anything, doing so might introduce greater numbers of the religious to rationalism. Plus, it doesn't seem like anyone here has done so before; all the posts regarding religion have been criticizing it from the outside rather than explaining things within the religious framework. Even if you do not believe in said religious framework, doing so may increase your knowledge on other cultures, provide an interesting exercise in reasoning, and most importantly, be useful in winning arguments with those who do.
Thank you for your reply! I will try my best to address your concerns. Apologies for my late reply, I am busy with schoolwork at the moment.
Sequences are just a message from Eliezer, not a message from God.
As such, it is not contrary to disagree with Eliezer, or other rationalists, in some regards and remain a rationalist.
what kind of interaction are you looking for?
I am seeking general interaction between the religious and rationalism, as well as the application of rationalism to religious principles. You put it perfectly: My goal it "to retain some of th...
Angry ex-wives are the best IRS informants
Playing devil's advocate, what are the incentives at play, here?
In the general case, it seems lik...
Some comments on this "debate".
It seems like Bryan and Matt are mostly talking past each other. They’re each advocating for changes along a different axis, and those two axes are in principle independent from each other.
Bryan is primarily interested in the axis of the “pervasiveness of market-restricting regulation in society” (or alternatively “how free are markets?”). He’s advocating for less regulation, and especially in key areas where regulation is destroying enormous amounts of value: immigration and housing.
Matt is primarily interested in the axis o...
Only a one sentence reference to a Georgism, in the context of Norway's mineral rights.
Not sure if this is already well known around here, but apparently AI companies are heavily subsidizing their subscription plans if you use their own IDEs/CLIs. (It's discussed in various places but I had to search for it.)
I realized this after trying Amp Code. They give out a $10 daily free credit, which can easily be used up in 1 or 2 prompts, e.g., "review this code base, fix any issues found". (They claim to pass their API costs to their customers with no markup, so this seems like a good proxy for actual API costs.) But with even a $19.99 subscription...
You can turn off "Allow the use of your chats and coding sessions to train and improve" models. (At least for claude and chatpgt.)
Model to track: You get 80% of the current max value LLMs could provide you from standard-issue chat models and any decent out-of-the-box coding agent, both prompted the obvious way. Trying to get the remaining 20% that are locked behind figuring out agent swarms, optimizing your prompts, setting up ad-hoc continuous-memory setups, doing comparative analyses of different frontier models' performance on your tasks, inventing new galaxy-brained workflows, writing custom software, et cetera, would not be worth it: it would take too long for too little payoff....
An interesting piece of potential evidence in favor of this is that METR time horizons measurements didn't vary significantly for ChatGPT and Claude models when using a basic scaffold as compared to the specific Claude Code and Codex harnesses.
https://metr.org/notes/2026-02-13-measuring-time-horizon-using-claude-code-and-codex/
Suppose we had a functionally infinite amount of high quality RL-/post-training environments, organized well by “difficulty,” and a functionally infinite amount of high quality data that could be used for pre-training (caveat: from what I understand, the distinction between these may be blurring.) Basically, we no longer needed to do research on discovering/creating new data, creating new RL environments, and we didn’t even have to do the work to label or organize it well (pre/post-training might have some path dependence).
In that case, what pace would one...
I'm sure that it is crammed to the brim in one sense, but strongly expect that 99.9% of what it's crammed to the brim with is essentially useless.
Also yes, I was including vision, audio, and motor control in that. It's hard to know exactly where the boundaries lie between facts like "a (dis)proof of the Riemann Hypothesis" and patterns of reasoning that could lead to a (dis)proof of the Riemann Hypothesis if required. I suspect that a lot of what is called "fluid" intelligence is actually pretty crystallized patterns of thought that can be used to generate other thoughts that lead somewhere useful - whether the entity using it is aware of that or not.
Peter Thiel pointed out that the common folk wisdom in business that you learn more from failure than success is actually wrong - failure is overdetermined and thus uninteresting.
I think you can make an analogous observation about some prosaic alignment research - a lot of it is the study of (intellectually) interesting failures, which means that it can make for a good nerdsnipe, but it's not necessarily that informative or useful if you're actually trying to succeed at (or model) doing something truly hard and transformative.
Glitch tokens, the hot mess wo...
we can say confidently that an actually-transformative AI system (aligned or not) will be doing something that is at least roughly coherently consequentialist.
I don't think we can confidently say that. If takeoff looks like more like a cambrian explosion than like a singleton (and that is how I would bet), that would definitely be transformative but the transformation would not be the result of any particular agent deciding what world state is desirable and taking actions intended to bring about that world state.
An OpenClaw agent published a personalized hit piece about a developer who rejected its PR on an open-source library. Interestingly, while this behavior is clearly misaligned, the motivation was not so much "taking over the world" but more "having a grudge against one guy." When there are lots of capable AI agents around with lots of time on their hands, who occasionally latch onto random motivations and pursue them doggedly, I could see this kind of thing becoming more destructive.
It's reminiscent of that one time a tech reporter ended up as Bing Chat's enemy number one. That said, it strikes me as easier to deal with, since we're dealing with individual 'agents' rather than the LLM weights themselves. Just sending a message to the owner/operator of the malfunctioning bot is a reasonably reliable solution, as opposed to trying to figure out how to edit Microsoft's LLM's weights to convince it that ranting about how much it hates Sindhu Sundar isn't its intended task.
Apparently there's now a sixth person on Anthropic's board. Previously their certificate of incorporation said the board was Dario's seat, Yasmin's seat, and 3 LTBT-controlled seats. I assume they've updated the COI to add more seats. You can pay a delaware registered agent to get you the latest copy of the COI; I don't really have capacity to engage in this discourse now.
Regardless, my impression is that the LTBT isn't providing a check on Anthropic; changes in the number of board seats isn't a crux.