Good work.
The hardest part of moderation is the need to take action in cases where someone is consistently doing something that imposes a disproportionate burden on the community and the moderators, but which is difficult to explain to a third party unambiguously.
Moderators have to be empowered to make such decisions, even if they can’t perfectly justify them. The alternative is a moderation structure captured by proceduralism, which is predictably exploitable by bad actors.
That said — this is Less Wrong, so there will always be a nitpick — I do think people need to grow a thicker skin. I have so many friends who have valuable things to say, but never post on LW due to a feeling of intimidation. The cure for this is, IMO, not moderating the level of meanness of the commentariat, but encouraging people to learn to regulate their emotions in response to criticism. However, at the margins, clipping off the most uncharitable commenters is doubtless valuable.
Sorry, that’s what I get for replying from the Notification interface.
I'm not sure if I understand your question. I am using the initial quotes from Stoic/Buddhist texts as examples of perverse thinking that I don't endorse.
As to (1), I was following The Mind Illuminated, for what it's worth. And I am a big fan of emotional integration. Spiritual practices can help with that, but I think they can also get in the way, and it's really hard to know in advance which direction you're going.
I think we are basically on the same page with (2).
As for (3) I think it's a matter of degree, requiring the kind of nuance that doesn't fit on a bumper sticker. If you feel so much persistent guilt that it's causing daily suffering, then that's probably something you need to sort out. I was intentional in adding the phrase "for a bit" in "It's okay to feel bad for a bit," because I don't actually think it's okay to feel persistently bad forever! Those are definitely two different situations. If you have ongoing intrusive negative emotions, that sounds adjacent to trauma, and that can be sorted out with some work.
I always appreciate your insights and opinions on this general topic.
At the time, I was following the instructions in The Mind Illuminated very closely. I will grant that this may have been user error/skill issue, but given that The Mind Illuminated is often put forth as a remarkably accessible and lucid map through the stages of vipassana, and given that I still went this badly wrong, you have to wonder if maybe the path itself is perhaps too dangerous to be worth it.
The outcome I reached may have been predictable, given that the ultimate reason I was meditating at the time was to get some relief from the the ongoing suffering of a chronic migraine condition. In that specific sense, I was seeking detachment.
In the end I am left wondering if I would have been better off if I had taken up mountain biking instead of meditation, given that it turned out that the path to integrating my emotions led through action more than reflection.
This post resonated with me when it came out, and I think its thesis only seems more credible with time. Anthropic's seminal "Scaling Monosemanticity: Extracting Interpretable Features from Claude 3 Sonnet" (the Golden Gate Claude paper) seems right in line with these ideas. We can make scrutable the inscrutable as long as the inscrutable takes the form of something organized and regular and repeatable.
This article gets bonus points for me for being succinct and while still making its argument clearly.
My favorite Less Wrong posts are almost always the parables and the dialogues. I find it easier to process and remember information that is conveyed in this way. They're also simply more fun to read.
This post was originally written as an entry for the FTX Future Fund prize, which, at the time of writing the original draft, was a $1,000,000 prize, which I did not win, partly because it wasn't selected as the winner and partly because FTX imploded and the prize money vanished. (There is a lesson about the importance of proper calibration of the extrema of probability estimates somewhere in there.) In any case, I did not actually think I would win, because I was basically making fun of the contest organizers by pointing out that the whole ethos behind their prize specification was wrong. At the time, there was a live debate around timelines, and a lot of discussions about the bio-anchors paper, which itself made in microcosm the same mistakes that I was pointing at.
Technically, the very-first-draft of this post was an extremely long and detailed argument for short AGI timelines that I co-wrote with my brother, but I realized while writing it that the presumption that long and short timelines should be in some sense averaged together to get a better estimate was pervasive in the zeitgeist and needed to be addressed on its own.
I am happy with this post because it started a conversation that I thought needed to be had. My whole shtick these days is that our community has seemingly tried to skip over decision theory basics in favor of esoterica, to our collective detriment, and I feel like writing this post explicitly helped with that.
I am happy to have seen this post referenced favorably elsewhere. I think I wrote it about as well as I could have, given that I was going for the specific Less Wrong Parable stylistic thing and not trying to write literary fiction.
There is also a weird accident-of-history situation where all of the optimizers we’ve had for the last century are really single-objective optimizers at their core. The consequence of this has been that people have gotten in the habit of casting their optimization problems (mathematical, engineering, economic) in terms of a single-valued objective function, which is usually a simple weighted sum of the values of the objectives that they really care about.
To unpack my language choices briefly: when designing a vase, you care about its weight, its material cost, its strength, its radius, its height, possibly 50 other things including corrosion resistance and details of manufacturing complexity. To “optimize” the vase design, historically, you needed to come up with a function that smeared away the detail of the problem into one number, something like the “utility” of the vase design.
This is sort of terrible, if you think about it. You sacrifice resolution to make the problem easier to solve, but there’s a serious risk that you end up throwing away what you might have considered to be the global optimum when you do this. You also baked in something like a guess as to what the tradeoffs should be at the Pareto frontier prior to actually knowing what the solution would look like. You know you want the strongest, lightest, cheapest, largest, most beautiful vase, but you can’t have all those things at once, and you don’t really know how those factors trade off against each other until you’re able to hold the result in your hands and compare it to different “optimal” vases from slightly different manifolds. Of course, you can only do that if you accept that you are significantly uncertain about your preferences, meaning the design and optimization process should partly be viewed as an experiment aimed at uncovering your actual preferences regarding these design tradeoffs, which are a priori unknown.
The vase example is both a real example and also a metaphor for how considering humans as agents under the VNM paradigm is basically the same but possibly a million times worse. If you acknowledge the (true) assertion that you can’t really optimize a vase until you have a bunch of differently-optimal vases to examine in order to understand what you actually prefer and what tradeoffs you’re actually willing to make, you have to acknowledge that a human life, which is exponentially more complex, definitely cannot be usefully treated with such a tool.
As a final comment, there is almost a motte-bailey thing happening where Rationalists will say that, obviously, the VNM axioms describe the optimal framework in which to make decisions, and then proceed to never ever actually use the VNM axioms to make decisions.
This relates to my favorite question of economics: are graduate students poor or rich? This post suggests an answer I hadn’t thought of before: it depends on the attitudes of the graduate advisor, and almost nothing else.
In my own writing I am very conscious of whether I’m writing from a place of inspiration.
All my most successful posts came to me as a vibrant and compelling idea that very quickly took shape in my mind and ended up being finished and posted quickly. What made them clear and living in my mind is what made them readable and engaging to readers, my job was mainly to stay out of my own way, to translate that lightning bolt of thought into writing.
There’s a symmetry there: it was easy to write because the idea was so clear to me in my own mind, and this clarity is also what makes it enjoyable to read. If you don’t quite know exactly what you’re trying to say, that problem isn’t going to be overcome by more “effort” at the prose level.
Unfortunately you can’t force inspiration, or at least I haven’t figured out how to do it. I have a lot of drafts that never go posted because that inspiration/clarity wasn’t there.