David J Higgs
David J Higgs has not written any posts yet.

David J Higgs has not written any posts yet.

Huh, I knew there wasn't the sort of plan you'd naively expect where the US gov/military command observes the response of the Japanese gov/military to one of their cities being destroyed by unthinkable godlike powers and then decides what to do next. I didn't know that president Truman literally didn't know about/have implicit preemptive control over the 2nd bombing.
Is there a reliable way to distinguish between [remembers more facts] and [infers more correct facts from remembered ones]? If there isn't, then using remembered facts as an estimate of base model size would be even more noisy than you'd already expect.
I know I get far more questions right on exams than chance would predict when I have 0 direct knowledge/memory of the correct answer. I assume reasoning models have at least some of this kind of capability
I think with respect to utopia, especially via AI, mourning can make sense now, but not after it actually happens. Now you can see all the things that will never be, but you can't see all the things that are even better that will actually be.
After you will see and feel and live all that is better, and it will be obvious that it is better. Only gratitude makes sense, or perhaps some form of nostalgia, but not genuine grief (for life without AI).
I'm glad this kind of content exists on LessWrong: writing that doesn't shy away from an explicit focus on personal virtue, in a "how algorithms feel from the inside" kind of way. I used to devour anything I could find written by C.S. Lewis as a still-religious teenager, because I felt a certain quality of thought and feeling emanating from each page. I felt the sincerity of effort orienting towards truth, beauty, and goodness.
Unfortunately, much of his worldview turned out to be importantly wrong, but his writings are hardly alone in that compared to other genuine historical truth seekers. I hope that like this post, my future thinking can manage to orient in that direction which Lewis was among the first to bring before my attention.
This strikes me as a good sort of constructive feedback, but one that didn't apply in my case, and I'll try to explain why. Thinking real instead of fake seems like a canonical example of rationality that is especially contingent upon emotions and subjective experience, and intervening on that level is extremely tricky and fraught.
In my case, the copious examples, explanations of why the examples are relevant, pointers to ways of avoiding the bad/getting at the good, etc. mostly seemed helpful in conveying the right pre-rational mental patterns for allowing the relevant rational thoughts to occur (either by getting out of the way or by participating in the generation of rational thought... (read more)
I'm guessing "species" is there mainly as emphasis that we are NOT talking about (mere) tool AI, and also maybe to marginally increase the clickbait for Twitter/X purposes.
Don't forget intellectual charity, which might actually be the most LW distinguishing feature relative to other smart online communities.
Counter-counterpoint: big groups like bureaucracies are not composed of randomly selected individuals from their respective countries. I strongly doubt that say, 100 randomly selected Google employees (the largest plausible bureaucracy that might potentially develop AGI in the very near term future?) would answer extremely similarly to 100 randomly selected Americans.
Of course, in the only moderately near term or median future, something like a Manhatten Project for AI could produce an AGI. This would still not be identical to 100 random Americans, but averaging across the US security & intelligence apparatus, the current political facing portion of the US executive administration, and the leadership + relevant employee influence from a (mandatory?) collaboration of... (read more)
If you haven't already, you should consider reading the Timelines Forecast and Takeoff Forecast research supplements linked to on the AI 2027 website. But I think there are a good half dozen (not necessarily independent) reasons for thinking that if AI capabilities start to takeoff in short timeline futures, other parts of the overall economy/society aren't likely to massively change nearly as quickly.
The jagged capabilities frontier in AI that already exists and will likely increase, Moravec's Paradox, the internal model/external model gap, the lack of compute available for experimentation + training + synthetic data creation + deployment, the gap in ease of obtaining training data for tasks like Whole Brain Emulation versus... (read more)
In addition to the option of spending effort on reducing the chance the world ends, one could also reframe from "leaving a mark on the world that outlives you" to "contributing to something bigger and beyond yourself." The world is bigger than you, more important than you and exists outside of you right now, as well as up until the world ends (if/when it does).
Helping the world right now, and helping the world after you are gone, are morally equivalent, and quite possibly equivalent at the level of fundamental physics. I'm not sure what, other than a false sense of personal immortality (legacy as something beyond the actual beneficial effects on the world), is tied to benefiting the world later than your own time of existence. But perhaps that's my own ignorance.