LESSWRONG
LW

2275
Mikhail Samin
2353263145
Message
Dialogue
Subscribe

My name is Mikhail Samin (diminutive Misha, @Mihonarium on Twitter, @misha on Telegram). 

Humanity's future can be enormous and awesome; losing it would mean our lightcone (and maybe the universe) losing most of its potential value.

My research is currently focused on AI governance and improving the understanding of AI and AI risks among stakeholders. I also have takes on what seems to me to be the very obvious shallow stuff about the technical AI notkilleveryoneism; but many AI Safety researchers told me our conversations improved their understanding of the alignment problem.

I believe a capacity for global regulation is necessary to mitigate the risks posed by future general AI systems. I'm happy to talk to policymakers and researchers about ensuring AI benefits society.

I took the Giving What We Can pledge to donate at least 10% of my income for the rest of my life or until the day I retire (why?).

In the past, I've launched the most funded crowdfunding campaign in the history of Russia (it was to print HPMOR! we printed 21 000 copies =63k books) and founded audd.io, which allowed me to donate >$100k to EA causes, including >$60k to MIRI.

[Less important: I've also started a project to translate 80,000 Hours, a career guide that helps to find a fulfilling career that does good, into Russian. The impact and the effectiveness aside, for a year, I was the head of the Russian Pastafarian Church: a movement claiming to be a parody religion, with 200 000 members in Russia at the time, trying to increase separation between religious organisations and the state. I was a political activist and a human rights advocate. I studied relevant Russian and international law and wrote appeals that won cases against the Russian government in courts; I was able to protect people from unlawful police action. I co-founded the Moscow branch of the "Vesna" democratic movement, coordinated election observers in a Moscow district, wrote dissenting opinions for members of electoral commissions, helped Navalny's Anti-Corruption Foundation, helped Telegram with internet censorship circumvention, and participated in and organized protests and campaigns. The large-scale goal was to build a civil society and turn Russia into a democracy through nonviolent resistance. This goal wasn't achieved, but some of the more local campaigns were successful. That felt important and was also mostly fun- except for being detained by the police. I think it's likely the Russian authorities would imprison me if I ever visit Russia.]

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
6Mikhail Samin's Shortform
3y
272
Mikhail Samin's Shortform
Mikhail Samin14h20

Yep. Good that he stopped. Likely bad that he started.

Reply
Mikhail Samin's Shortform
Mikhail Samin2d20

The chatbot I made that answers questions and counterarguments about existential threat from AI is public! Please ask it hard questions (especially if you're not convinced of AI x-risk yourself or can repeat the kind of questions others ask you). And feel free to share it with more people!

Suggestions for design improvements are also welcome!

https://whycare.aisgf.us 

Reply
Mikhail Samin's Shortform
Mikhail Samin2d6424

"There is no justice in the laws of Nature, no term for fairness in the equations of motion. The universe is neither evil, nor good, it simply does not care. The stars don't care, or the Sun, or the sky. But they don't have to! We care! There is light in the world, and it is us!"

Reply
Mikhail Samin's Shortform
Mikhail Samin3d2-1

the demand is that a specific company agrees to halt if everyone halts; this does not help in reality, because in fact it won't be the case that everyone halts (abscent gov intervention).

Reply
Mikhail Samin's Shortform
Mikhail Samin3d20

You are of course right that there’s no difference between reality-fluid and normal probabilities in a small world: it’s just how much you care about various branches relative to each other, regardless of whether all of them will exist or only some.

I claim that the negative utility due to stopping to exist is just not there, because you don’t actually stop to exist in a way you reflectively care about, when you have fewer instances. For normal things (e.g., how much do you care about paperclips), the expected utility is the same; but here, it’s the kind of terminal value that i expect for most people would be different; guaranteed continuation in 5% of instances is much better than 5% chance of continuing in all instances; in the first case, you don’t die!

Reply
Mikhail Samin's Shortform
Mikhail Samin3d10

Imagine they you’re an agent in the game of life. Your world, your laws of physics are computed on a very large independent computers; all performing the same computation.

You exist within the laws of causality of your world, computed as long as at least one server computes your world. If some of them stop performing the computation, it won’t be a death of a copy; you’ll just have one fewer instance of yourself.

Reply
Mikhail Samin's Shortform
Mikhail Samin3d10

In a large universe, you do not end. Like, not in expectation see some branch versus other; you just continue, the computation that is you continues. When you open your eyes, you're not likely to find yourself as a person in a branch computed only relatively rarely; still, that person continues, and does not die.

Attemted suicide reduces your reality-fluid- how much you're computed and how likely you are to find yourself there- but you will continue to experience the world. If you die in a nuclear explosion, the continuation of you will be somewhere else, sort-of isekaied; and mostly you will find yourself not in a strange world that recovers the dead but in a world where the nuclear explosion did not appear; still, in a large world, even after a nuclear explosion, you continue.

You might care about having a lot of reality-fluid, because this makes your actions more impactful, because you can spend your lightcone better, and improve the average experience in the large universe. You might also assign negative utility to others seeing you die; they'll have a lot of reality-fluid in worlds where you're dead and they can't talk to you, even as you continue. But I don't think it works out to assigning the same negative utility to dying as in branches of small worlds.

Reply
Mikhail Samin's Shortform
Mikhail Samin3d42

"People are on hunger strikes" is not really a lot of evidence for "AI CEOs are optimizing for their short-term status/profits and are not doing the due diligence" in the eyes of the public.

I don't think there's any problem with painting people and institutions as evil, I'm just not sure why you would want to do this here, as compared to other things, and would want people to have answers to how they imagine a hunger strike would paint AI companies/CEOs and what would be the impact of that, because I expect little that could move the needle.

Reply
Mikhail Samin's Shortform
Mikhail Samin3d20

"Attention on AI risk" is a somewhat very bad proxy to optimize for, where available tactics include attention that would be paid to luddites, lunatics, and crackpots caring about some issue.

The actions that we can take can:

  • Use what separates us from people everyone considers crazy: that our arguments check out and our predictions hold; communicate those;
  • Spark and mobilize existing public support;
  • Be designed to optimize for positive attention, not for any attention.

I don't think DeepMind employees really changed their minds? Like, there are people at DeepMind with p(doom) higher than Eliezer's; they would be sympathetic; would they change anything they're doing? (I can imagine it prompting them to talk to others at DeepMind, talking about the hunger strike to validate the reasons for it.)

I don't think Demis responding to the strike would make Dario look particularly villainous, happy to make conditional bets. How villainous someone looks here should be pretty independent, outside of eg Demis responding, prompting a journalist to ask Dario, which takes plausible deniability away from him.

I'm also not sure how effective it would be to use this to paint the companies (or the CEOs-- are they even the explicit targets of the hunger strikes?) as villainous.

Reply1
Mikhail Samin's Shortform
Mikhail Samin4d-1-5

Something like: Hunger strikes are optimized hard specifically for painting someone as a villain because they decide to make someone suffer or die (or be inhumanely fed), this is different from other forms of protests that are more focused on, e.g., that specific decisions are bad and should be revoked, but don't necessarily try to make people perceive the other side as evil.

Reply1
Load More
The Tree of AI Alignment on Arbital
a month ago
The Tree of AI Alignment on Arbital
a month ago
(+37676)
Decision theory
6 months ago
(+142)
Functional Decision Theory
6 months ago
(+242)
Translations Into Other Languages
3 years ago
(+84/-60)
76OpenAI Claims IMO Gold Medal
2mo
74
33No, Futarchy Doesn’t Have This EDT Flaw
3mo
28
6Superintelligence's goals are likely to be random
6mo
6
80No one has the ball on 1500 Russian olympiad winners who've received HPMOR
8mo
21
71How to Give in to Threats (without incentivizing them)
1y
34
11Can agents coordinate on randomness without outside sources?
Q
1y
Q
16
76Claude 3 claims it's conscious, doesn't want to die or be modified
2y
118
33FTX expects to return all customer money; clawbacks may go away
2y
1
25An EA used deceptive messaging to advance her project; we need mechanisms to avoid deontologically dubious plans
2y
1