Could you pitch the value of this, perhaps by linking the relevant section of your agenda post, or etc? Why should someone who encounters this paper be interested as more than a a recreational math curiosity - eg, where does it fit in your agenda? why would someone who is trying to solve ASI-grade alignment need to invent this paper, or if that's not obviously the case, what lead you to be interested? If there was another potential vanessa out there who was feeling pressure to solve ASI alignment and is getting interested in the current "field consensus", why should they get interested in this line of work, and specifically in this subthread, such that it's a good idea for them to read this particular work?
this is a comment I've been meaning to make on most of your technical posts, because you seem to be one of very few people doing work like yours and I think part of why is not explaining its necessity. I'm planning to make a post pitching, among other things, your stuff, in the hope of laying out the reasons the field should consider whether there's a good reason to try paths like yours harder, but I'm still relatively a noob to it and you arguing for it yourself is probably just better for getting the actual reasons down in a coherent and justified way.
(I'm interested in chatting on discord before a response here.)
I did link the relevant section of my agenda post:
This work is my first rigorous foray into compositional learning theory.
A brief and simplified summary:
I'm open to chatting on Discord.
Abstract: We propose a new variant of online learning that we call "ambiguous online learning". In this setting, the learner is allowed to produce multiple predicted labels. Such an "ambiguous prediction" is considered correct when at least one of the labels is correct, and none of the labels are "predictably wrong". The definition of "predictably wrong" comes from a hypothesis class in which hypotheses are also multi-valued. Thus, a prediction is "predictably wrong" if it's not allowed by the (unknown) true hypothesis. In particular, this setting is natural in the context of multivalued dynamical systems, recommendation algorithms and lossless compression. It is also strongly related to so-called "apple tasting". We show that in this setting, there is a trichotomy of mistake bounds: up to logarithmic factors, any hypothesis class has an optimal mistake bound of either , or .
This work is my first rigorous foray into compositional learning theory. Intuitively, one approach to learning compositionally is, constructing a model incrementally, adding one part each time. This raises the question of what does it mean to have a "partial" model. One notion of partial model is provided by infra-Bayesianism / robust RL. However, there the treatment of ambiguities in the model via the maximin decision rule is strongly tied to the choice of reward function. Intuitively, we might expect the process of incremental learning to not depend on the reward function that strongly.
In this work, I propose a notion of "partial" model in the online learning setting. I develop a theory of learning for this notion analogous (but more involved than) classical realizable Littlestone theory.
Does it truly lead to compositional learning? In results that don't appear in the paper, I showed that the "ambiguous Littlestone dimension" (an invariant I discovered that characterizes the complexity of ambiguous online learning) indeed behaves nicely w.r.t. some ways to decompose hypothesis classes. However, my attempts to use this framework to get computationally efficient compositional learning algorithms haven't produced much so far.
Moreover, our results on robust RL (in particular the notion of "imprecise estimator" we have there) suggests it's more natural to study a different (easier) version of online learning with ambiguity, where the learner is only allowed to predict one label instead of a set. So, the ambiguous online learning framework is a potentially useful tool, but I'm currently more optimistic about other approaches to compositional learning theory (stay tuned!).