LESSWRONG
LW

Agent FoundationsAI
Frontpage

30

New Paper: Ambiguous Online Learning

by Vanessa Kosoy
25th Jun 2025
AI Alignment Forum
2 min read
2

30

Ω 15

This is a linkpost for https://arxiv.org/abs/2506.19810
Agent FoundationsAI
Frontpage

30

Ω 15

New Paper: Ambiguous Online Learning
4the gears to ascension
10Vanessa Kosoy
New Comment
2 comments, sorted by
top scoring
Click to highlight new comments since: Today at 6:28 PM
[-]the gears to ascension19d40

Could you pitch the value of this, perhaps by linking the relevant section of your agenda post, or etc? Why should someone who encounters this paper be interested as more than a a recreational math curiosity - eg, where does it fit in your agenda? why would someone who is trying to solve ASI-grade alignment need to invent this paper, or if that's not obviously the case, what lead you to be interested? If there was another potential vanessa out there who was feeling pressure to solve ASI alignment and is getting interested in the current "field consensus", why should they get interested in this line of work, and specifically in this subthread, such that it's a good idea for them to read this particular work?

this is a comment I've been meaning to make on most of your technical posts, because you seem to be one of very few people doing work like yours and I think part of why is not explaining its necessity. I'm planning to make a post pitching, among other things, your stuff, in the hope of laying out the reasons the field should consider whether there's a good reason to try paths like yours harder, but I'm still relatively a noob to it and you arguing for it yourself is probably just better for getting the actual reasons down in a coherent and justified way.

(I'm interested in chatting on discord before a response here.)

Reply
[-]Vanessa Kosoy18d100

I did link the relevant section of my agenda post:

This work is my first rigorous foray into compositional learning theory.

A brief and simplified summary:

  • In order to have powerful learning algorithms with safety guarantees, we first need learning algorithms with powerful generalization guarantees that we know how to rigorously formulate (otherwise how do you know the algorithm will correctly infer the intended goal/behavior from the training data?).
  • Additionally, in order to formally specify "aligned to human values", we need to formally specify "human values", and it seems likely that the specification of "X's values" should be something akin to "the utility function w.r.t. which X has [specific type of powerful performance guarantees]". These powerful performance guarantees are probably a form/extension of powerful generalization guarantees.
  • Both reasons require us to understand the kind of natural powerful generalization guarantees that efficient learning algorithms can satisfy. Moreover, such understanding would likely be applicable to deep learning as well, as it seems likely deep learning algorithms satisfy such guarantees, but we currently don't know how to formulate them.
  • I conjecture that a key missing ingredient in deriving efficient learning algorithms with powerful guarantees (more powerful than anything we already understand in computational learning theory), is understanding the role of compositionality in learning. This is because compositionality is a ubiquitous feature of our thinking about the world and, intuitively, particular forms of compositionality are strong candidates for properties that are both very general and strong enough to enable efficient learning. This line of thinking led me to some success in the context of control theory, which is a necessary ingredient of the kind of guarantees we will ultimately need.
  • I identified sequence prediction / online learning in the deterministic realizable case as a relatively easy (but already highly non-trivial) starting point for investigating compositional learning.
  • For the reasons stated in the OP, this led me to ambiguous online learning.

I'm open to chatting on Discord.

Reply1
Moderation Log
Curated and popular this week
2Comments

Abstract: We propose a new variant of online learning that we call "ambiguous online learning". In this setting, the learner is allowed to produce multiple predicted labels. Such an "ambiguous prediction" is considered correct when at least one of the labels is correct, and none of the labels are "predictably wrong". The definition of "predictably wrong" comes from a hypothesis class in which hypotheses are also multi-valued. Thus, a prediction is "predictably wrong" if it's not allowed by the (unknown) true hypothesis. In particular, this setting is natural in the context of multivalued dynamical systems, recommendation algorithms and lossless compression. It is also strongly related to so-called "apple tasting". We show that in this setting, there is a trichotomy of mistake bounds: up to logarithmic factors, any hypothesis class has an optimal mistake bound of either Θ(1), Θ(√(N)) or N.

This work is my first rigorous foray into compositional learning theory. Intuitively, one approach to learning compositionally is, constructing a model incrementally, adding one part each time. This raises the question of what does it mean to have a "partial" model. One notion of partial model is provided by infra-Bayesianism / robust RL. However, there the treatment of ambiguities in the model via the maximin decision rule is strongly tied to the choice of reward function. Intuitively, we might expect the process of incremental learning to not depend on the reward function that strongly.

In this work, I propose a notion of "partial" model in the online learning setting. I develop a theory of learning for this notion analogous (but more involved than) classical realizable Littlestone theory.

Does it truly lead to compositional learning? In results that don't appear in the paper, I showed that the "ambiguous Littlestone dimension" (an invariant I discovered that characterizes the complexity of ambiguous online learning) indeed behaves nicely w.r.t. some ways to decompose hypothesis classes. However, my attempts to use this framework to get computationally efficient compositional learning algorithms haven't produced much so far.

Moreover, our results on robust RL (in particular the notion of "imprecise estimator" we have there) suggests it's more natural to study a different (easier) version of online learning with ambiguity, where the learner is only allowed to predict one label instead of a set. So, the ambiguous online learning framework is a potentially useful tool, but I'm currently more optimistic about other approaches to compositional learning theory (stay tuned!).