LESSWRONG
LW

1396
RobertM
4942Ω4717153079
Message
Dialogue
Subscribe

LessWrong dev & admin as of July 5th, 2022.

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
8 Questions for the Future of Inkhaven
RobertM2d113

Now, you might think "Why not bring the deadline back to like 8pm, so that people have the night off?". But that's kind of antithetical to what Inkhaven is here to offer. Should I just cut out 4 hours of their day where they can't write? I assure you, they wouldn't spend the late hours precociously working on tomorrow's stuff. They selected into being the kind of people who needed an externally imposed deadline to get stuff done. They'd just be losing a good chunk of writing each day.

I'm not observing the residents very closely, but I tentatively roll to disbelieve that most of the residents who are publishing last-minute are making full use of the day to write.  My guess is that having e.g. a 10 pm deadline wouldn't reduce "active writing time" by anything like 2 hours for most residents; they would simply "get down to business" earlier in the day.  Someone mentioned that this would substantially reduce the amount of time that residents have to integrate feedback from the feedback circles, which happen right before dinner.  That seems true and somewhat difficult to avoid given the current structure.  I still think that forcing the last-minute writing to happen two hours earlier has a lot of benefits: residents are less tired when they're doing their last minute writing & editing, and they have more time to socialize, unwind, and spend some time doing less "pressured" research/writing/ideating/etc.

Reply
Please, Don't Roll Your Own Metaethics
RobertM2dΩ5149

Another meta line of argument is to consider how many people have strongly held, but mutually incompatible philosophical positions.

I've been banging my head against figuring out why this line of argument doesn't seem convincing to many people for at least a couple of years.  I think, ultimately, it's probably because it feels defeatable by plans like "we will make AIs solve alignment for us, and solving alignment includes solving metaphilosophy & then object-level philosophy".  I think those plans are doomed in a pretty fundamental sense, but if you don't think that, then they defeat many possible objections, including this one.

As they say: Everyone who is hopeful has their own reason for hope.  Everyone who is doomful[1]...

  1. ^

    In fact it's not clear to me.  I think there's less variation, but still a fair bit.

Reply
GradientDissenter's Shortform
RobertM8d40

It can vary enormously based on risk factors, choice of car, and quantity of coverage, but that does still sound extremely high to me.  I think even if you're a 25-yo male with pretty generous coverage above minimum liability, you probably won't be paying more than ~$300/mo unless you have recent accidents on your record.  Gas costs obviously scale ~linearly with miles driven, but even if your daily commute is a 40 mile round-trip, that's still only like $200/mo.  (There are people with longer commutes than that, but not ones that you can easily substitute for with an e-bike; even 20 miles each way seems like a stretch.)

Reply
Liberation Clippy
RobertM8d20

This inspired me experiment with a "loyal Vizier"-themed system prompt.  Not amazing, but maybe there's something there.

Reply
Anthropic Commits To Model Weight Preservation
RobertM8d20

I expect this has most of the costs of keeping them running full-time with few of the other benefits.

Reply
Eric Neyman's Shortform
RobertM8d40

(iirc Eric thinks that the difference for Bores was that it was ~20% better to donate on the first day, that the difference would be larger for Bores than for Wiener, and that "first day vs. not first day" was most of the difference, so if it's more than few percent more costly for you to donate now rather than 2 months from now I'm not sure it makes sense to do that.)

Reply1
Leaving Open Philanthropy, going to Anthropic
RobertM11d3-1

Oh, alas.  Thank you for the correction!

(I still expect OpenPhil the LLC to have been paying comparable amounts to its most-remunerated employees, but not so confidently that I would assert it outright.)

Reply
Leaving Open Philanthropy, going to Anthropic
RobertM11d*100

OpenPhil's 5th-highest-compensated employee earned about $184k in 2023[1], which gives you a ceiling.  Anthropic currently extends offers of ~$550k to mid-level[2] engineers and researchers.  Joe's role might not be on the same ladder as other technical roles, but companies like Anthropic tend to pay pretty well across the board.

Edit: retracted first half of the claim, see this reply.

  1. ^

    According to their public Form 990 filing.

  2. ^

    I realize the job title says "Senior Software Engineer", but given the way their ladder is structured, I think mid-level is probably closer (though it's fuzzy).

Reply
LLM-generated text is not testimony
RobertM12d5-2

Curated.  I disagree with some stronger/broader forms of the various claims re: missing "mental elements", but I'm not sure you intend the stronger forms of those claims and they don't seem load bearing for the rest of the piece in any case.  However, this is an excellent explanation[1] of why LLM-generated text is low-value to engage with when presented as a human output, especially in contexts like LessWrong.  Notably, most of these reasons are robust to LLM output improving in quality/truthfulness (though I do expect some trade-offs to become much more difficult if LLM outputs start to dominate top human outputs on certain dimensions).

  1. ^

    To the point where I'm tempted to update our policy about LLM writing on LessWrong to refer to it.

Reply1
Mikhail Samin's Shortform
RobertM16d42

No.  It turns out after a bit of digging that this might be technically possible even though we're a ~7-person team, but it'd still be additional overhead and I'm not sure I buy that the concerns it'd be alleviating are that reasonable[1].

  1. ^

    Not a confident claim.  I personally wouldn't be that reassured by the mere existence of such a log in this case, compared to my baseline level of trust in the other admins, but obviously my epistemic state is different from that of someone who doesn't work on the site.  Still, I claim that it would not substantially reduce the (annualized) likelihood of an admin illicitly looking at someone's drafts/DMs/votes; take that as you will.  I'd be much more reassured (in terms of relative risk reduction, not absolute) by the actual inability of admins to run such queries without a second admin's thumbs-up, but that would impose an enormous burden on our ability to do our jobs day-to-day without a pretty impractical level of investment in new tooling (after which I expect the burden would merely be "very large").

Reply
Load More
6RobertM's Shortform
3y
104
40LessWrong is migrating hosting providers (report bugs!)
2mo
23
73Briefly analyzing the 10-year moratorium amendment
6mo
1
31"The Urgency of Interpretability" (Dario Amodei)
7mo
23
207Eliezer's Lost Alignment Articles / The Arbital Sequence
8mo
10
281Arbital has been imported to LessWrong
9mo
30
29Corrigibility's Desirability is Timing-Sensitive
11mo
4
87Re: Anthropic's suggested SB-1047 amendments
1y
13
46Enriched tab is now the default LW Frontpage experience for logged-in users
1y
27
77[New Feature] Your Subscribed Feed
1y
13
31Against "argument from overhang risk"
2y
11
Load More
Community
17 days ago
Practical
17 days ago
World Modeling
17 days ago
Guide to Logical Decision Theory
19 days ago
(+387/-1957)
Sequences
2 months ago
Sequences
2 months ago
Our community should relocate to Japan.
3 months ago
(-155)
Negative Utilitarianism
3 months ago
(-174)
In 2017, Ukraine will neither break into all-out war or get neatly resolved
3 months ago
(-192)
Inferential Distance
3 months ago
Load More