I'm a last-year PhD student at the University of Amsterdam working on AI Safety and Alignment, and specifically safety risks of Reinforcement Learning from Human Feedback (RLHF). Previously, I also worked on abstract multivariate information theory and equivariant deep learning. https://langleon.github.io/
Fwiw., my hair grew longer and people often point that out, but never has anyone followed with “looks good”.
I think the compute they spend on inference will also just get scaled up over time.
I think people don’t usually even try to figure something like that out, or are even aware of the option. So if you publicly announce that a user has deactivated their account X times, then this is information that almost no one would otherwise ever receive.
I also have the sense that it’s better to not do that, even though I have a hard time explaining in words why that is.
I suspect there’s a basic reason why futility claims are often successful in therapy/coaching: by claiming (and succeeding in convincing the client) that something can’t be changed, you reduce the client’s shame in not changing the thing. Now the client is without shame, and that’s a state of mind that makes it a priory easier to change, and focusing the change on aspects the client didn’t fail on in the past additionally increases the chance of succeeding since there’s no evidence of not succeeding on those aspects.
However, I also really care about truth, and so I really dislike such futility claims.
I feel like Cunningham's law got confirmed here. I'm really glad about all the things I learned from people who disagreed with me.
Thanks a lot for this very insightful comment!
I think we may not disagree about any truth-claims about the world. I'm just satisfied that the north star of Solomonoff induction exists at all, and that it is as computable (albeit only semicomputable), well-predicting, science-compatible and precise as it is. I expected less from a theory that seems so unpopular.
> It predicts well: It's provenly a really good predictor
So can you point to any example of anyone ever predicting anything using it?
No, but crucially, I've also never seen anyone predict as well as someone using Solomonoff induction with any other method :)
I don’t know what “all-too-plausibly” means. Depending on the probabilities that this implies I may agree or disagree.