Hello! I work at Lightcone and like LessWrong :-)

Wiki Contributions



(No, "you need huge profits to solve alignment" isn't a good excuse — we had nowhere near exhausted the alignment research that can be done without huge profits.)

This seems insufficiently argued; the existence of any alignment research that can be done without huge profits is not enough to establish that you don't need huge profits to solve alignment (particularly when considering things like how long timelines are even absent your intervention).

To be clear, I agree that OpenAI are doing evil by creating AI hype.


Is there anything particularly quantum about this effect?

Using the simulator frame, one might think there's space to tweak:

  1. The basic physical laws
  2. The fundamental constants
  3. The "PRNG" (in an Everettian picture this looks kind of weird because its more like throwing out parts of the wavefunction to save on computation; reminds me a little of mangled worlds)

Perhaps the idea is that tweaking 1 & 2 results in worlds less interesting to the simulator?


I'm not seeing any active rate limits. Do you know when you observed it? It's certainly the case that an automatic rate limit could have kicked in and then, as voting changed, been removed.


Good question! From the Wiki-Tag FAQ:

A good heuristic is that tag ought to have three high-quality posts, preferably written by two or more authors. 

I believe all tags have to be approved. If I were going through the morning moderation queue, I wouldn't approve an empty tag.


I was trying to figure out why you believed something that seemed silly to me! I think it barely occurred to me that it's a joke.


The main subcultures that I can think of where this applies are communities based around solving some problem:

  • Weight loss, especially if based around a particular diet
  • Dealing with a particular mental health problem
  • Trying to solve a particular problem in the world (e.g. explaining some mystery or finding the identity of some criminal)

Any favourite examples?


I think my big problem with complexity science (having bounced off it a couple of times, never having engaged with it productively) is that though some of the questions seem quite interesting, none of the answers or methods seem to have much to say.

Which is exacerbated by a tendency to imply they have answers (or at least something that is clearly going to lead to an answer)


I would like to read it! Satire is sometimes helpful for me to get a perspective shift

Answer by kaveApr 04, 20243810

To answer, for now, just one piece of this post:

We're currently experimenting with a rule that flags users who've received several downvotes from "senior" users (I believe 5 downvotes from users with above 1,000 karma) on comments that are already net-negative (I believe that were posted in the last year).

We're currently in the manual review phase, so users are being flagged and then users are having the rate limit applied if it seems reasonable. For what it's worth, I don't think this rule has an amazing track record so far, but all the cases in the "rate limit wave" were reviewed by me and Habryka and he decided to apply a limit in those cases.

(We applied some rate limit in 60% of the cases of users who got flagged by the rule).

People who get manually rate-limited don't have an explanation visible when trying to comment (unlike users who are limited by an automatic rule, I think).

We have explained this to users that reached out (in fact this answer is adapted from one such conversation), but I do think we plausibly should have set up infrastructure to explain these new rate limits.

Load More