In this post, I proclaim/endorse forum participation (aka commenting) as a productive research strategy that I've managed to stumble upon, and recommend it to others (at least to try). Note that this is different from saying that forum/blog posts are a good way for a research community to communicate. It's about individually doing better as researchers.
Hi Ali, sorry for my slow response, too! Needed to think on it for a bit.
This is my personal opinion, and in particular, does not represent anything like a MIRI consensus; I've gotten push-back from almost everyone I've spoken with about this, although in most cases I believe I eventually convinced them of the narrow terminological point I'm making.
In the AI x-risk community, I think there is a tendency to ask people to estimate "time to AGI" when what is meant is really something more like "time to doom" (or, better, point-of-no-return). For about a year, I've been answering this question "zero" when asked.
This strikes some people as absurd or at best misleading. I disagree.
The term "Artificial General Intelligence" (AGI) was coined in the early 00s, to contrast with the prevalent paradigm of Narrow AI. I was getting my undergraduate computer science...
On 16 March 2024, I sat down to chat with New York Times technology reporter Cade Metz! In part of our conversation, transcribed below, we discussed his February 2021 article "Silicon Valley's Safe Space", covering Scott Alexander's Slate Star Codex blog and the surrounding community.
The transcript has been significantly edited for clarity. (It turns out that real-time conversation transcribed completely verbatim is full of filler words, false starts, crosstalk, "uh huh"s, "yeah"s, pauses while one party picks up their coffee order, &c. that do not seem particularly substantive.)
ZMD: I actually have some questions for you.
CM: Great, let's start with that.
ZMD: They're critical questions, but one of the secret-lore-of-rationality things is that a lot of people think criticism is bad, because if someone criticizes you, it hurts your...
Hm. I think we like Slate Star Codex in this thread, so let's enjoy a throwback:
...It was wrong of me to say I hate poor minorities. I meant I hate Poor Minorities! Poor Minorities is a category I made up that includes only poor minorities who complain about poverty or racism.
No, wait! I can be even more charitable! A poor minority is only a Poor Minority if their compaints about poverty and racism come from a sense of entitlement. Which I get to decide after listening to them for two seconds. And If they don’t realize that they’re doing something wrong, th
This is the eighth post in my series on Anthropics. The previous one is Lessons from Failed Attempts to Model Sleeping Beauty Problem. The next one is Beauty and the Bets.
Suppose we take the insights from the previous post, and directly try to construct a model for the Sleeping Beauty problem based on them.
We expect a halfer model, so
On the other hand, in order not repeat Lewis' Model's mistakes:
But both of these statements can only be true if
And, therefore, apparently, has to be zero, which sounds obviously wrong. Surely the Beauty can be awaken on Tuesday!
At this point, I think, you wouldn't be surprised, if I tell you that there are philosophers who are eager to bite this bullet and claim that the Beauty should, indeed, reason as...
Let it be not two different days but two different half-hour intervals. Or even two milliseconds - this doesn't change the core of the issue that sequential events are not mutually exclusive.
OUTCOME: A measurable result of a random experiment.
SAMPLE SPACE: a set of exhaustive, mutually exclusive outcomes of a random experiment.
EVENT: Any subset of the sample space of a random experiment.
INDEPENDENT EVENTS: If A and B are events from the same sample space, and the occurrence of event A does not affect the chances of the occurrence of event B, then A a...
This in the (bi-)annual ACX/SCC Schelling Meetup, where you can meet like-minded curious folks. This time I reserved an indoor space! I'm pleased to announce that we meet on Saturday 27nd of April at 15:00 at Leih-Lokal Freiräume, Gerwigstraße 41, Karlsruhe.
This is a foremost social event and there is no structure or schedule. Just come and enjoy the discourse about any topic you are interested in.
I'll try to provide some snacks so please RSVP for a better estimate of the expected number of mouths to feed.
The Karlsruhe Rationality group (currently in hiatus) aims to connect Rationalists from Karlsruhe (Germany) and surrounding areas. Everyone worries they're not serious enough about ACX to join, so you should banish that thought and come anyway. "Please feel free to come even if you feel awkward about it, even if you’re not 'the typical ACX reader', even if you’re worried people won’t like you", even if you didn't come to the previous meetings, even if you don't speak German, etc., etc.
Lots of people already know about ACX/SSC, but I think that crossposting to LW is unusually valuable in this particular case, since lots of people were waiting for a big schelling-point overview of the 15-hour Rootclaim Lab Leak debate, and unlike LW, ACX's comment section is a massive vote-less swamp that lags the entire page and gives everyone equal status.
It remains unclear whether commenting there is worth your time if you think you have something worth saying, since there's no sorting, only sifting, implying that it attracts small numbers of sifters instead of large numbers of people who expect sorting.
Here are the first 11 paragraphs:
...Saar Wilf is an ex-Israeli entrepreneur. Since 2016, he’s been developing a new form of reasoning, meant to transcend normal human bias.
His method
One thing that occurs to me is that each analysis, such as the Putin one, can be thought of as a function hypothesis.
It takes as inputs the variables:
Russian demographics
healthy lifestyle
family history
facial swelling
hair present
And is outputting the probability 86%, where the function is
P = F(demographics, lifestyle, history, swelling, hair) and then each term is being looked up in some source, which has a data quality, and the actual equation seems to be a mix of Bayes and simple probability calculations.
There are other variables not considered, and other...
I want to thank Jan Kulveit, Tomáš Gavenčiak, and Jonathan Shock for their extensive feedback and ideas they contributed to this work and for Josh Burgener and Yusuf Heylen for their proofreading and comments. I would also like to acknowledge the Epistea Residency and its organisers where much of the thinking behind this work was done.
This post aims to build towards a theory of how meditation alters the mind based on the ideas of active inference (ActInf). ActInf has been growing in its promise as a theory of how brains process information and interact with the world and has become increasingly validated with a growing body of work in the scientific literature.
Why bring the idea of ActInf and meditation together? Meditation seems to have a profound effect on...
In his method, I think the happiness of the first few Jhanas is not caused by prediction error directly, but rather indirectly through the activation of the reward circuitry. So while the method involves creating some amount of prediction error, the ultimate result is less overall prediction error, because the reward neurotransmitters bring the experiential world closer to the ideal.
After the first three Jhanas, the reward circuitry is less relevant and you start to reduce overall prediction error through other means, by allowing attention to let go of asp...
Summary: The post describes a method that allows us to use an untrustworthy optimizer to find satisficing outputs.
Acknowledgements: Thanks to Benjamin Kolb (@benjaminko), Jobst Heitzig (@Jobst Heitzig) and Thomas Kehrenberg (@Thomas Kehrenberg) for many helpful comments.
Imagine you have black-box access to a powerful but untrustworthy optimizing system, the Oracle. What do I mean by "powerful but untrustworthy"? I mean that, when you give an objective function as input to the Oracle, it will output an element that has an impressively low[1] value of . But sadly, you don't have any guarantee that it will output the optimal element and e.g. not one that's also chosen for a different purpose (which might be dangerous for many reasons, e.g. instrumental convergence).
What questions can you safely ask the Oracle? Can you use it to...
There's a particular kind of widespread human behavior that is kind on the surface, but upon closer inspection reveals quite the opposite. This post is about four such patterns.
One of the most useful ideas I got out of Algorithms to Live By is that of computational kindness. I was quite surprised to only find a single mention of the term on lesswrong. So now there's two.
Computational kindness is the antidote to a common situation: imagine a friend from a different country is visiting and will stay with you for a while. You're exchanging some text messages beforehand in order to figure out how to spend your time together. You want to show your friend the city, and you want to be very accommodating and make sure...
Forget where I read it, but this Idea seems similar. When responding to a request, being upfront about your boundaries or constraints feels intense but can be helpful for both parties. If Bob asks Alice to help him move, and Alice responds "sure thing" that leaves the interaction open to miscommunication. But if instead Alice says, " yeah! I am available 1pm to 5pm and my neck has been bothering me so no heavy lifting for me!" Although that's seems like less of a kind response Bob now doesn't have to guess at Alice's constraints and can comfortably move forward without feeling the need to tiptoe around how long and to what degree Alice can help.