people including alignment researchers just seem more confident about their own preferred solution to metaethics, and comfortable assuming their own preferred solution is correct as part of solving other problems, like AI alignment or strategy. (E.g., moral anti-realism is true, therefore empowering humans in straightforward ways is fine as the alignment target can't be wrong about their own values.)
Obviously committed anti-realists would be right not to worry -- if they're correct! But I agree with you, we shouldn't be overconfident in our metaethics...which makes me wonder, do you really think metaethics can be "solved?"
Secondly, even if it were solved (and to avoid the anti-realist apathy, let's assume moral realism is true), how do you think that would help with alignment? Couldn't the alignment-target simply say, "this is true, but I don't care, as it doesn't help me achieve my goals?" Saying "1+1=2, but I'm going to act as if it equals 3" might keep you from achieving your goal. Saying, "stealing is wrong, but I would really like to have X" might actually help you achieve your goal.
A worthwhile cosmopolitanism should probably fight Moloch in this regard and extend its own agency to things which do not wield the tools of autopoesis themselves.
[warning: anthropomorphisation of autopoetic processes]
*autopoiesis, and autopoietic (brought to you by the pedantic spelling police)
Hi! Like others on this thread, I'm a long time reader who's finally created an account to try to join the discussion. I'm curious, if I comment on a 15 year old article or something, is anyone likely to see that? I love browsing around the Concepts pages, but are comments there (.e.g.) likely to be seen?
My intuition is that comments on recent trending articles are more likely to get engagement, but can anyone confirm or deny or give suggestions on the best ways/places to engage?
Thanks!
Imagine you're immortal. Would you rather get a dust speck in your eye for 3^^^3 straight days, or endure 50 years of torture, never to be bothered by a dust speck again?
Going with dust specks is the obvious choice to me. The way I see it, life is still very much worth living when I'm dealing with a dust speck. Torture might make me wish for death in the first day...especially with the knowledge that 18,261 days were forthcoming.
I can’t tell if:
a) I dislike dust specks less than you
b) I dislike torture more than you
c) Aggregating pain/discomfort/suffering has weird results. (e.g. perhaps 5 minutes of torture is immeasurably more than 60x as bad as 5 seconds of torture.)
or d) there's something else I'm missing.
I think this is very relevant to the complexity of values and how difficult it is for humans to grasp their own utility functions. Cards on the table, the aggregation/math issues make me unable to embrace consequentialism...but if I'm overlooking something I'd love to be shown the light.
I had bookmarked this post as fascinating (but I claim no first-hand knowledge here): https://www.lesswrong.com/posts/BgBJqPv5ogsX4fLka/the-mind-body-vicious-cycle-model-of-rsi-and-back-pain
Looks potentially pretty relevant. Is that the kind of thing you're looking for?