People who helped Jews during WWII are intriguing. They appear to be some kind of moral supermen. They had almost nothing to gain and everything to lose. How did they differ from the general population? Can we do anything to get more of such people today?
This post was written during the agent foundations fellowship with Alex Altair funded by the LTFF. Thanks to Alex, Jose, Daniel, Cole, and Einar for reading and commenting on a draft.
The Good Regulator Theorem, as published by Conant and Ashby in their 1970 paper (cited over 1700 times!) claims to show that 'every good regulator of a system must be a model of that system', though it is a subject of debate as to whether this is actually what the paper shows. It is a fairly simple mathematical result which is worth knowing about for people who care about agent foundations and selection theorems. You might have heard about the Good Regulator Theorem in the context of John Wentworth's 'Gooder Regulator' theorem and his other improvements on...
The archetypal example for this is something like a thermostat. The variable S represents random external temperature fluctuations. The regulator R is the thermostat, which measures these fluctuations and takes an action (such as putting on heating or air conditioning) based on the information it takes in. The outcome Z is the resulting temperature of the room, which depends both on the action taken by the regulator, and the external temperature.
The ordinary room thermostat does not measure S. It measures Z. Its actions are determined by Z and the referenc...
Thanks a lot! It's a good comment by Scott on Sailor Vulcan's post. I have added it and your other links to the page's "see also" on my site.
I like this paragraph in particular. It captures the tension between the pursuit of epistemic and instrumental rationality:
...I think my complaint is: once you become a self-help community, you start developing the sorts of epistemic norms that help you be a self-help community, and you start attracting the sort of people who are attracted to self-help communities. And then, if ten years later, someone says “Hey, are w
[ Context: The Debate on Animal Consciousness, 2014 ]
There's a story in Growing Up Yanomamo where the author, Mike Dawson, a white boy from America growing up among Yanomamö hunter-gatherer Yanomamö hunter-gatherer kids in the Amazon, is woken up in the early morning by two of his friends.
One of the friends says, "We're going to go fishing".
So he goes with them.
At some point on the walk to the river he realizes that his friends haven't said whose boat they'll use [ they're too young to have their own boat ].
He considers asking, then realizes that if he asks, and they're planning to borrow an older tribesmember's boat without permission [ which is almost certainly the case, given that they didn't specify up front ], his friends will...
I like this treatment of consciousness and morality so much better than the typical EA (and elsewhere) naive idea that anything that "has consciousness" suddenly "has moral value" (even worse, and dangerous, is to combine that with symmetric population ethics). We should treat these things carefully (and imo democratically) to avoid making giant mistakes once AI allows us to put ethics into practice.
Current “unlearning” methods only suppress capabilities instead of truly unlearning the capabilities. But if you distill an unlearned model into a randomly initialized model, the resulting network is actually robust to relearning. We show why this works, how well it works, and how to trade off compute for robustness.
Produced as part of the ML Alignment & Theory Scholars Program in the winter 2024–25 cohort of the shard theory stream.
Read our paper on ArXiv and enjoy an interactive demo.
Maybe some future AI has long-term goals and humanity is in its...
"How was your first day of high school?"
"Well, in algebra, the teacher just stood in front of the class for 45 minutes, scratching his head and saying things like 'what the heck is an inequality?' and 'I've never factored an expression in my life!' Maybe he's trying to get fired?"
When I was first learning about hypnosis, one of the things that was very confusing to me is how "expectations" relate to "intent". Some hypnotists would say "All suggestion is about expectation; if they expect to have an experience they will", and frame their inductions in terms of expectation (e.g. "Your eyelids will become heavy"). The problem with this is that "I don't think it's gonna work". Other hypnotists would avoid this issue entirely by saying "I don't care if you think it will work. Follow my instructions, and you will get the results regardless of what you believe" and then say things like "Make your eyelids heavy". The problem with this is that "I don't know to do that!", which would be avoided by saying "You...
Sorry for the delayed reply... I don't get notifications of replies, and the LW RSS has been broken for me for years now, so I only poke my head here occasionally.
Well that sounds... scary, at best. I hope you've come out of it okay.
50/100. But that rather exciting story is best not told in a public forum.
Though these distinctions are kinda confusing for me.
Well, lack of appearance of something otherwise expected would be negative, and appearance of something otherwise unexpected would be positive?
For example, a false pregnancy is a "positive somatization"...
The ACX/EA/LW Sofia Meetup for June will be on the 29th (Sunday) at 16:00 in the Gradinka na Yogite (in Borisova Gradina Park).
Sofia ACX started with the 2021 Meetups Everywhere round. Attendance hovers around 4-8 people. Everyone worries they're not serious enough about ACX to join, so you should banish that thought and come anyway. "Please feel free to come even if you feel awkward about it, even if you’re not 'the typical ACX reader', even if you’re worried people won’t like you", even if you didn't come to the previous meetings, even if you don't speak Bulgarian, etc., etc.
Each month we pick something new to read and discuss. In August, we're discussing Against Empathy by Paul Bloom (Chapter 1).
We'll be in the gazebo in "Градинка на Юогите" (picture here https://maps.app.goo.gl/kYhRv6aT4WQPJKBz9). This little garden is part of Borisova Gradina near the tennis courts, roughly between the Television Tower and Levski Stadium. If you think you'll have trouble finding it, email me and I'll arrange for someone to meet you.
Coordinates: https://plus.codes/8GJ5M8GW+P6
See you there.
At Less Online, I ran a well-attended session titled "Religion for Rationalists" to help me work out how I could write a post (this one!) about one of my more controversial beliefs without getting downvoted to hell. Let's see how I do!
My thesis is that most people, including the overwhelmingly atheist and non-religious rationalist crowd, would be better off if they actively participated in an organized religion.
My argument is roughly that religions uniquely provide a source of meaning, community, and life guidance not available elsewhere, and to the extent anything that doesn't consider itself a religion provides these, it's because it's imitating the package of things that makes something a religion. Not participating in a religion is obviously fine, but I think it leaves people missing out...
Is Judaism not also based around disputation of texts?
I am not banning you because you are a critic.
Thank heaven for that! But notice that you’re responding to a strawman: I never claimed that you banned me because I am a critic, period. Obviously not; since, as you say, you haven’t banned plenty of other people.
(Although, as I pointed out upthread, you have, in at least one case, threatened to ban another person for their critical comments, after deleting several of their comments. As far as I’m aware, that person—quite unsurprisingly!—hasn’t commented on your posts since. So, no, you don’t get to claim t...
If they weren't ready to deploy these safeguards and thought that proceeding outweighed the (expected) cost in human lives, they should have publicly acknowledged the level of fatalities and explained why they thought weakening their safety policies and incurring these expected fatalities was net good.[1]
Public acknowledgements of the capabilities could be net negative in itself, especially if they resulted in media attention. I expect bringing awareness to the (possible) fact that the AI can assist with CBRN tasks likely increases the chance that pe...