LESSWRONG
LW

212
bargo
12080
Message
Dialogue
Subscribe

Posts

Sorted by New

Wikitag Contributions

Comments

Sorted by
Newest
No wikitag contributions to display.
Infra-Bayesian physicalism: a formal theory of naturalized induction
bargo2y40

A theory of physics is mathematically quite similar to a cellular automaton. This theory will usually be incomplete, something that we can represent in infra-Bayesianism by Knightian uncertainty. So, the "cellular automaton" has underspecified time evolution.

What evidence is there that incomplete models with Knightian uncertainty are a way to turn rough models of physics into loss functions? Can the ideas behind it be applied to regular Bayesianism?

Reply
Unifying Bargaining Notions (2/2)
bargo2y10

bolution

solution

Reply
Logical Probability of Goldbach’s Conjecture: Provable Rule or Coincidence?
bargo2y30

https://arxiv.org/abs/2211.06738 is related

Reply
Self-Reference Breaks the Orthogonality Thesis
bargo3y32

Sure, I mean that it is an implementation of what you mentioned in the third-to-last paragraph.

Reply
Self-Reference Breaks the Orthogonality Thesis
bargo3y50

Congratulations, you discovered [Active Inference]!

Reply
Curiosity as a Solution to AGI Alignment
bargo3y21

Solomonoff Induction and Machine Learning. How would you formulate this in terms of a machine that can only predict future observations?

Reply
Curiosity as a Solution to AGI Alignment
bargo3y10

I want to provide feedback, but can't see the actual definition of the objective function in either of the cases. Can you write down a sketch of how this would be implemented using existing primitives (SI, ML) so I can argue against what you're really intending?

Some preliminary thoughts:

  1. Curiosity (obtaining information) is an instrumental goal, so I'm not sure if making it more important will produce more or less aligned systems. How will you trade off curiosity and satisfaction of human values?
  2. It's difficult to specify correctly - depending on what you mean by curiosity, the AI can start showing itself random unpredictable data (the Noisy TV problem for RL curiosity) or manipulating humans to give it self-affirming instructions (if the goal is to minimize prediction error). So far I haven't seen a solution that isn't a hack that won't generalize to superintelligent agents.
Reply
No posts to display.