Ruby

Team Lead for LessWrong

Sequences

LW Team Updates & Announcements
Novum Organum

Comments

This is basically what we do, capped by our team capacity. For most of the last ~2 years, we had ~4 people working full-time on LessWrong plus shared stuff we get from EA Forum team. Since the last few months, we reallocated people from elsewhere in the org and are at ~6 people, though several are newer to working on code. So pretty small startup. Dialogues has been the big focus of late (plus behind the scenes performance optimizations and code infrastructure).

All that to say, we could do more with more money and people. If you know skilled developers willing to live in the Berkeley area, please let us know!

My intuition (not rigorous) is there a multiple levels in the consequentialist/deontoligical/consequentialist dealio.

I believe that unconditional friendship is approximately something one can enter into, but one enters into it for contingent reasons (perhaps in a Newcomb-like way – I'll unconditionally be your friend because I'm betting that you'll unconditionally be my friend). Your ability to credibly enter such relationships (at least in my conception of them) is dependent on you not starting to be more "conditional" because you doubt that the other person is also being there. This I think is related to not being a "fair weather" friend. I continue to be your friend even when it's not fun (you're sick, need taking care of whatever) even if I wouldn't have become your friend to do that. And vice versa. Kind of a mutual insurance policy.

Same thing could be with contracts, agreements, and other collaborations. In a Newcomb-like way, I commit to being honest, being cooperative, etc to a very high degree even in the face of doubts about you. (Maybe you stop by the time someone is threatening your family, not sure what Ben, et al, think about that.) But the fact I entered into this commitment was based on the probabilities I assigned to your behavior at the start.

I see interesting points on both sides here. Something about how this comment(s) is expressed makes me feel uncomfortable, like this isn't the right tone for exploring disagreements about correct moral/cooperative behavior, it at least it makes it a lot harder for me to participate. I think it's something like it feels like performing moral outrage/indignation in a way that feels more persuadey than explainy, and more in the direction of social pressure, norms-enforcery. The phrase "shame on you" is a particularly clear thing I'll point at that makes me perceive this.

I was going to write stuff about integrity, and there's stuff to that, but the thing that is striking me most right now is that the whole effort seemed very incompetent and naive. And that's upsetting.

I am now feeling uncertain about the incompetence and naivety of it. Whether this was the best move possible that failed to work out, or best move possible that actually did get a good outcome, or a total blunder is determined by info I don't have.

I have some feeling of they were playing against a higher-level political player which both makes it hard but also means they needed to account for that? Their own level might be 80+th percentile in reference class of executive/board type-people, but still lower than Sam.

The piece that does seem most like they really made a mistake was trying to appoint an interim CEO (Mira) who didn't want the role. It seems like before doing that, you should be confident the person wants it.

I've seen it raised that the board might find the outcome to be positive (board stays independent even if current members leave?). If that's true, does change the evaluation of the competence. Feels hard for me to confidently judge that, though my gut sense is Sam got more of what he wanted/common knowledge of his sway than others.

Styling of the headers in this post is off and makes it harder to read. Maybe the result of a bad copy/paste?

These recent events have me thinking the opposite: policy and cooperation approaches to making AI go well are doomed – while many people are starting to take AI risk seriously, not enough are, and those who are worried will fail to restrain those who aren't (where not being risked in a consequence of humans often being quite insane when incentives are at play). The hope lies in somehow developing enough useful AI theory that leading labs adopt and resultantly build an aligned AI even though they never believed they were going to cause AGI ruin.

And so maybe let's just get everyone to focus on the technical stuff. Actually more doable than wrangling other people to not build unsafe stuff.

If he can lead an exodus from OpenAI to Microsoft, he can lead one from Microsoft to somewhere else.

People associated with EA are likely to decide at some point that the normal rules for the organization do not apply to them, if they expect that they can generate a large enough positive impact in the world by disregarding those rules.

I am myself consequentialist at my core, but invoking consequentialism to justify breaking commitments, non-cooperation, theft, or whatever else is just a stupid bad policy (the notion of people doing this generates some strong emotions for me) that as a policy/algorithm, won't result in accomplishing one's consequentialist goals.

I fear what you say is not wholly inaccurate and is true of at least some in EA, I hope though it's not that true of many.

Where it does get tricky is potentially unilateral pivotal acts about which I think go in this direction but also feel different from what you describe.

If I were to do it again, I might include such an option, though I'm still not terribly sad I didn't.

If we really wanted that info, I could  sample and survey people who received the message, looked at it (we have data to know this) and ask them why they didn't vote. My guess is between 1-10% of people who didn't vote because of frame commented about it,, so that's 40 to 400. 

372 people have responded by now out of 2500, so 15%. Let's guess that 50% of people saw it by now, so ~1250 (though could get better data on it).  A third responded if so, which seems great for a poll. Of those the 800 who saw but didn't, I could see 100-400 doing so because the frame didn't really seem right (lining up with the above estimate). Which seems fine. I bet if I'd spent 10x developing the poll, I wouldn't get that number down much, and knowing it with more precision doesn't really help. It's LW, people are very picky and precise (a virtue...but also makes having some nice things hard).

Load More