LESSWRONG
LW

Leon Lang
1863Ω132141760
Message
Dialogue
Subscribe

I'm a last-year PhD student at the University of Amsterdam working on AI Safety and Alignment, and specifically safety risks of Reinforcement Learning from Human Feedback (RLHF). Previously, I also worked on abstract multivariate information theory and equivariant deep learning. https://langleon.github.io/

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
2Leon Lang's Shortform
3y
86
No wikitag contributions to display.
Leon Lang's Shortform
Leon Lang6d150

A NeurIPS paper on scaling laws from 1993, shared by someone on twitter.

Reply
Leon Lang's Shortform
Leon Lang8d120

Is there a way to filter on Lesswrong for all posts from the alignment forum?

I often like to just see what's on the alignment forum, but I dislike that I don't see most Lesswrong comments when viewing those posts on the alignment forum. 

Reply
Futility Illusions
Leon Lang9d63

I suspect there’s a basic reason why futility claims are often successful in therapy/coaching: by claiming (and succeeding in convincing the client) that something can’t be changed, you reduce the client’s shame in not changing the thing. Now the client is without shame, and that’s a state of mind that makes it a priory easier to change, and focusing the change on aspects the client didn’t fail on in the past additionally increases the chance of succeeding since there’s no evidence of not succeeding on those aspects.

However, I also really care about truth, and so I really dislike such futility claims.

Reply
Leon Lang's Shortform
Leon Lang11d40

I feel like Cunningham's law got confirmed here. I'm really glad about all the things I learned from people who disagreed with me. 

Reply
Leon Lang's Shortform
Leon Lang12d30

Thanks a lot for this very insightful comment!

Reply
Leon Lang's Shortform
Leon Lang13d30

I think we may not disagree about any truth-claims about the world. I'm just satisfied that the north star of Solomonoff induction exists at all, and that it is as computable (albeit only semicomputable), well-predicting, science-compatible and precise as it is. I expected less from a theory that seems so unpopular. 

> It predicts well: It's provenly a really good predictor

So can you point to any example of anyone ever predicting anything using it?

No, but crucially, I've also never seen anyone predict as well as someone using Solomonoff induction with any other method :) 

Reply
Leon Lang's Shortform
Leon Lang13d161

I'm only now really learning about Solomonoff induction. I think I didn't look into it earlier since I often heard things along the lines of "It's not computable, so it's not relevant".

But...

  • It's lower semicomputable: You can actually approximate it arbitrarily well, you just don't know how good your approximations are.
  • It predicts well: It's provably a really good predictor under the reasonable assumption of a computable world.
  • It's how science works: You focus on simple hypotheses and discard/reweight them according to Bayesian reasoning.
  • It's mathematically precise. 

What more do you want?

The fact that my master's degree in AI at the UvA didn't teach this to us seems like a huge failure. 

Reply32
Enlightenment AMA
Leon Lang19d215

Thanks for doing this AMA!

1. In what sense is enlightenment permanent? E.g., will a truly enlightened person never suffer again? Or is it a weaker claim of the form "once one learns the motion of enlightenment, it can be repeated to eliminate suffering at will"? Or something else entirely?

2. I've read Shinzen Young's "The Science of Enlightenment". He describes how in some intermediate stage, he started hallucinating giant insects in his daily life. Have you made similar experiences? How do you interpret them? Do you consider such experiences dangerous? Are such experiences distinct from schizophrenia in a relevant way?

3. As far as I understand, enlightened people don't cling to a specific reality anymore, but they may still have strong desires. Is this your typical experience? How do you relate to your desires? Is it compatible to be (a) enlightened, (b) have a desire that leads to amoral actions when acted upon them, and (c) acting on those desires?

4. Is any amount of pain/anxiety/sadness/anger/etc. compatible with being in a state of zero suffering? Is zero suffering in practice harder to maintain for higher levels of these displeasures, or will a once-enlightened person not find any of these experiences difficult to combine with a state of non-suffering?

5. Are there degrees of enlightenment, or is it more of a discrete change?

6. How much control can an enlightened person have over their experience? Is it possible to decide to not hear/see/smell/feel/taste something? Is it possible to stop thinking at will? Is it possible to feel pleasure or bliss at will? Is it possible to change ones experience at will / to hallucinate arbitrary experiences at will?

7. What is it like to be enlightened?

Reply
The Coding Theorem — A Link between Complexity and Probability
Leon Lang20d20

Thanks for adding!

Not sure if you were aware, but in the glossary at the top-right of the post, there is also an explanation (albeit shorter) of "prefix-free code". I'm just mentioning this in case you weren't aware of the glossary functionality.

Reply
K-complexity is silly; use cross-entropy instead
Leon Lang22d40

Several commenters have remarked that the alt-complexity and K-complexity differ only up to a constant. I have now written a post where I write down a detailed proof of that classical result, which is known as the "coding theorem".

Reply
Load More
32The Coding Theorem — A Link between Complexity and Probability
22d
4
159X explains Z% of the variance in Y
2mo
34
36How to work through the ARENA program on your own
3mo
3
51[Paper Blogpost] When Your AIs Deceive You: Challenges with Partial Observability in RLHF
Ω
10mo
Ω
2
90We Should Prepare for a Larger Representation of Academia in AI Safety
2y
14
32Andrew Ng wants to have a conversation about extinction risk from AI
2y
2
26Evaluating Language Model Behaviours for Shutdown Avoidance in Textual Scenarios
2y
0
48[Appendix] Natural Abstractions: Key Claims, Theorems, and Critiques
Ω
2y
Ω
0
246Natural Abstractions: Key Claims, Theorems, and Critiques
Ω
2y
Ω
26
38Andrew Huberman on How to Optimize Sleep
3y
6
Load More