LESSWRONG
LW

1541
RobertM
4917Ω4317152879
Message
Dialogue
Subscribe

LessWrong dev & admin as of July 5th, 2022.

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
6RobertM's Shortform
3y
104
GradientDissenter's Shortform
RobertM9h30

It can vary enormously based on risk factors, choice of car, and quantity of coverage, but that does still sound extremely high to me.  I think even if you're a 25-yo male with pretty generous coverage above minimum liability, you probably won't be paying more than ~$300/mo unless you have recent accidents on your record.  Gas costs obviously scale ~linearly with miles driven, but even if your daily commute is a 40 mile round-trip, that's still only like $200/mo.  (There are people with longer commutes than that, but not ones that you can easily substitute for with an e-bike; even 20 miles each way seems like a stretch.)

Reply
Liberation Clippy
RobertM1d20

This inspired me experiment with a "loyal Vizier"-themed system prompt.  Not amazing, but maybe there's something there.

Reply
Anthropic Commits To Model Weight Preservation
RobertM1d20

I expect this has most of the costs of keeping them running full-time with few of the other benefits.

Reply
Eric Neyman's Shortform
RobertM1d40

(iirc Eric thinks that the difference for Bores was that it was ~20% better to donate on the first day, that the difference would be larger for Bores than for Wiener, and that "first day vs. not first day" was most of the difference, so if it's more than few percent more costly for you to donate now rather than 2 months from now I'm not sure it makes sense to do that.)

Reply1
Leaving Open Philanthropy, going to Anthropic
RobertM4d3-1

Oh, alas.  Thank you for the correction!

(I still expect OpenPhil the LLC to have been paying comparable amounts to its most-remunerated employees, but not so confidently that I would assert it outright.)

Reply
Leaving Open Philanthropy, going to Anthropic
RobertM4d*100

OpenPhil's 5th-highest-compensated employee earned about $184k in 2023[1], which gives you a ceiling.  Anthropic currently extends offers of ~$550k to mid-level[2] engineers and researchers.  Joe's role might not be on the same ladder as other technical roles, but companies like Anthropic tend to pay pretty well across the board.

Edit: retracted first half of the claim, see this reply.

  1. ^

    According to their public Form 990 filing.

  2. ^

    I realize the job title says "Senior Software Engineer", but given the way their ladder is structured, I think mid-level is probably closer (though it's fuzzy).

Reply
LLM-generated text is not testimony
RobertM5d5-2

Curated.  I disagree with some stronger/broader forms of the various claims re: missing "mental elements", but I'm not sure you intend the stronger forms of those claims and they don't seem load bearing for the rest of the piece in any case.  However, this is an excellent explanation[1] of why LLM-generated text is low-value to engage with when presented as a human output, especially in contexts like LessWrong.  Notably, most of these reasons are robust to LLM output improving in quality/truthfulness (though I do expect some trade-offs to become much more difficult if LLM outputs start to dominate top human outputs on certain dimensions).

  1. ^

    To the point where I'm tempted to update our policy about LLM writing on LessWrong to refer to it.

Reply1
Mikhail Samin's Shortform
RobertM9d42

No.  It turns out after a bit of digging that this might be technically possible even though we're a ~7-person team, but it'd still be additional overhead and I'm not sure I buy that the concerns it'd be alleviating are that reasonable[1].

  1. ^

    Not a confident claim.  I personally wouldn't be that reassured by the mere existence of such a log in this case, compared to my baseline level of trust in the other admins, but obviously my epistemic state is different from that of someone who doesn't work on the site.  Still, I claim that it would not substantially reduce the (annualized) likelihood of an admin illicitly looking at someone's drafts/DMs/votes; take that as you will.  I'd be much more reassured (in terms of relative risk reduction, not absolute) by the actual inability of admins to run such queries without a second admin's thumbs-up, but that would impose an enormous burden on our ability to do our jobs day-to-day without a pretty impractical level of investment in new tooling (after which I expect the burden would merely be "very large").

Reply
Mikhail Samin's Shortform
RobertM9d50

I don't think we currently have one.  As far as I know, LessWrong hasn't had any requests made of it by law enforcement that would trip a warrant canary while I've been working here (since July 5th, 2022).  I have no information about before then.  I'm not sure this is at the top of our priority list; we'd need to stand up some new infrastructure for it to be more helpful than harmful (i.e. because we forgot to update it, or something).

Reply
Mikhail Samin's Shortform
RobertM9d160

Expanding on Ruby's comment with some more detail, after talking to some other Lightcone team members:

Those of us with access to database credentials (which is all the core team members, in theory) would be physically able to run those queries without getting sign-off from another Lightcone team member.  We don't look at the contents of user's DMs without their permission unless we get complaints about spam or harassment, and in those cases also try to take care to only look at the minimum information necessary to determine whether the complaint is valid, and this has happened extremely rarely[1].  Similarly, we don't read the contents or titles of users' never-published[2] drafts.  We also don't look at users' votes except when conducting investigations into suspected voting misbehavior like targeted downvoting or brigading, and when we do we're careful to only look at the minimum amount of information necessary to render a judgment, and we try to minimize the number of moderators who conduct any given investigation.

  1. ^

    I don't recall ever having done it, Habryka remembers having done it once.

  2. ^

    We do see drafts that were previously published and then redrafted in certain moderation views.  Some users will post something that gets downvoted and then redraft it; we consider this reasonable because other users will have seen the post and it could easily have been archived by e.g. archive.org in the meantime.

Reply
Load More
40LessWrong is migrating hosting providers (report bugs!)
2mo
23
73Briefly analyzing the 10-year moratorium amendment
5mo
1
31"The Urgency of Interpretability" (Dario Amodei)
6mo
23
207Eliezer's Lost Alignment Articles / The Arbital Sequence
8mo
10
281Arbital has been imported to LessWrong
9mo
30
29Corrigibility's Desirability is Timing-Sensitive
10mo
4
87Re: Anthropic's suggested SB-1047 amendments
1y
13
46Enriched tab is now the default LW Frontpage experience for logged-in users
1y
27
77[New Feature] Your Subscribed Feed
1y
13
31Against "argument from overhang risk"
1y
11
Load More
Community
9 days ago
Practical
9 days ago
World Modeling
9 days ago
Guide to Logical Decision Theory
11 days ago
(+387/-1957)
Sequences
2 months ago
Sequences
2 months ago
Our community should relocate to Japan.
3 months ago
(-155)
Negative Utilitarianism
3 months ago
(-174)
In 2017, Ukraine will neither break into all-out war or get neatly resolved
3 months ago
(-192)
Inferential Distance
3 months ago
Load More