No public comments will be hosted on our website as we don't have the resources for moderation of public discussion. Authors can choose to link-post their work on the Alignment Forum or LessWrong to engage with a broader audience.
I think it'd be pretty important/useful if the UI shows links to publicly commentable link-posts where those exist.
I think this is something we may do, with the caveat that we would make it an option that the author can choose (and that that would be clear to the readers.) We don’t want to get in the business of deciding which online discussions are good or bad enough to be worth endorsing as worth the reader’s time.
Thanks for raising this.
Yes, I'm imagining if the author link-posts they can add a cross-link so viewers can participate.
I was encouraged to ask these questions under this post instead of over email so I'll do so:
I wanted to ask some questions around what papers are within scope.
Firstly, what about things overlapping between Cooperative AI and Agent Foundations? I've got a paper on tying game theory together with percolation theory in order to better predict fixed points in complex systems, is this within scope or not?
Secondly, I believe there are various problems in scaling technical single-agent safety to systemic prolems of Gradual Disempowerment and similar. We need to be able to detect agency in larger systems as well and Michael Levin has a bunch of work on trying to establish "diverse intelligence". Would something that is agent foundations on diverse intelligence be something compatible? A sort of scaling of agent foundations to systemic disciplines like computational social science.
Finally, what about taxonomy papers? Something I've noticed being frustrated about is not having a taxonomy of agent definitions and therefore running into issues with specification of language in discussions. An "agent" can mean a lot of things and I was thinking that putting together a taxonomy of what existing fields consider agents might be useful.
I love the idea of ILIAD, I think it's needed and awesome.
Thanks, we really appreciate the questions.
Our general approach to scope is to ask (1) if the topic is worth studying, and (2) if there are no other venues that can offer a substantially better review. If so, we’ll probably say yes. (We generally want to avoid reviewing manuscripts where there are already good existing journals who accept submissions on the topic, e.g., almost all interpretability.) We are willing to go outside our comfort zone to get worthwhile manuscripts reviewed imperfectly if the alternative is they get reviewed nowhere. One advantage of the reviewer abstract idea is that it allows the reviewers to communicate their uncertainty to the potential reader.
Both of the interdisciplinary papers you mention sound fine. In these sorts of cases we may ask the authors to put in special effort in helping us locate qualified (and reasonably unbiased) reviewers.
Review and taxonomy papers are fine, and indeed we’d love to see something that collects and compares various definitions of “agent” in both the conventional lit and the Alignment Forum. For us the question isn’t “Is this novel enough to ‘deserve’ publication?”, it’s “Is this worth writing? Are there at least a few researchers who will find this significantly more useful than what’s already been written?”.
This post is an update on the Proceedings of ILIAD, a conference journal for AI alignment research intended to bridge the gap between the Alignment Forum and academia. Following our successful first issue with 9 workshop papers from last year's ILIAD conference, we're launching a second issue in association with ILIAD 2: ODYSSEY. The conference is August 25-29, 2025 at Lighthaven in Berkeley, CA. Submissions to the Proceedings are open now (more info) and due June 25. Our goal is to support impactful, rapid, and readable research, carefully rationing scarce researcher time, using features like public submissions, partial anonymity, partial confidentiality, reviewer-written abstracts, reviewer compensation, and open licensing. We are soliciting community feedback and suggestions for reviewers and editorial board members.
Prior to the deep learning explosion, much early work on AI alignment occurred at MIRI, the Alignment Forum, and LessWrong (and their predecessors). Although there is now vastly more alignment and safety work happening at ML conferences and inside industry labs, it's heavily slanted toward near-term concerns and ideas that are tractable with empirical techniques. This is partly for good reasons: we now have much more capable models which guide theory and allow extremely useful empirical testing.
However, conceptual, mathematically abstract, and long-term research on alignment still doesn't have a good home in traditional academic journals and conferences. Much of it is still done on the AI Alignment Forum and here on LessWrong, or is done informally (private discussion, Twitter, blogs, etc) by academic researchers without a good venue for attracting the best constructive criticism.
As a result, there remains a gulf between more traditional academic work and much of the most important alignment work:
There is substantial truth in many of these criticisms, but often they can be misaimed. There is value in the techniques and philosophy of both communities, and we would like get best of both worlds by building a central venue to bridge them.
There is currently nothing like an academic journal on alignment. Several scientific fields (e.g., game theory, cybernetics) were meaningfully accelerated or incubated by an initial conference followed up with a journal, so we decided to take this route. We were warned several times that a journal/proceedings is an enormous amount of work, but we're stubborn and decided to try out a trial version at the first ILIAD conference (~120 attendees; LW announcement here) that took place August 28 - September 3, 2024 at Lighthaven.
We have just released 9 workshop papers in the first issue of Proceedings of ILIAD. First, the bad:
And the good:
Overall we were satisfied with this experiment and have decided to push forward. We have just opened up submissions for the second issue of the Proceedings in association with the second annual conference, ILIAD 2: ODYSSEY, taking place August 25-29, 2025 at Lighthaven. The soft deadline for submission is June 25th. We are continuing to experiment on mechanisms for running these proceedings, especially with the review process. If we succeed in getting good community engagement and adding value, we may start an archival journal dedicated to AI alignment.
In the rest of this post we first describe our general philosophy and then how we expect that to cash out in terms of the design of the second issue of the Proceedings and a possible alignment journal.
We want to accelerate impactful research and make it more readable to other researchers. Our hope to combine the best features of academic journals and internet forums.
An idealized traditional journal...
On the other hand, an idealized internet forum...
We want to combine all the above advantages as much as possible. We are inspired by the Distill Journal while keeping in mind its reasons for shutting down.
A central lens through which we think about all decisions is that, when it comes to the economics of academic discussion and review, the scarce resource is researcher time. Researcher time is best used when:
Some simple and relatively uncontroversial features:
Here are some less-trivial ideas we are tentatively planning to use:
Towards our ultimate goal of combining as many advantages of academic journals and internet forums into one venue, here are some ideas we're strongly considering for an alignment journal (but not for the next issue of Proceedings):
The last bullet point is potentially the most powerful mechanism for obtaining most of the benefits of internet forums while retaining the essentials of peer review. In essence, the review process would be a conventional internet forum discussion, except that (1) participants are filtered for expertise, (2) the discussion is confidential and lightly moderated by the editor, and (3) the detail results of the discussion are released as a reviewer abstract. Importantly, the moderation workload is greatly reduced compared to an open forum because the editor doesn’t have to monitor the comments in real-time; new commenters are hidden by default until approved by the editor.
We're seeking constructive criticism of the above ideas. Please also let us know:
We thank Oliver Habryka for discussion.