I recommend experimenting with collapsible sections.
People have a lot of different objections and these vary widely. If you try to answer all of them, then the article becomes too long and people won't finish it. If you don't answer someone's pet objection, they won't be persuaded.
Collapsable sections make this trade-off much less severe.
Thanks for the feedback! I agree that it's hard to balance between "being succinct" and "answering every pet objection", and collapsing sections could help. A few questions:
Sorry, I don't have time to review the articles at the moment.
On collapsable sections vs footnotes:
• Collapsible sections work well for longer content as you've identified. A short collapsable section might seem weird.
• Collapsible sections are more visible than footnotes so people are more likely to click them. They also have a title attached so you know that there is something in the document addressing question/objection X even if you don't read it. In contrast, footnotes are better for signalling that something isn't necessary to read unless you really care about the details
• The reading flow is often nicer for collapsible sections
The BlueDot Future of AI course uses collapsible sections very well.
Holden Karnofsky often uses collapsable sections well (see example). He often recaps previous articles to fill in context without assuming someone has read the article (or to remind folks of the details).
I can also share an example of a summary I wrote. I don't think I'm very good at this yet as I'm still learning how use collapsible sections, but I found this really helpful since it's good for summaries to be short, but the collapsable sections allowed me to give readers a sense of all the main ideas included in the paper if that's what they want.
Linking can help, but the reading flow isn't as natural as with collapsable sections. On the other hand, I imagine many folk uncollapse sections by default, so it make sense to link instead if most readers wouldn't want to follow that rabbit hole.
I was recently experimenting in extreme amounts of folding (LW linkpost): I'd be interested to hear from Chris whether he thinks this is too much folding?
I think it depends on the audience. That level of collapsible sections is too much for a more "normy" audience, but there will be some folks who love it.
TL;DR: The AISafety.info team wrote two intros to AI safety for busy laypeople: a short version and a longer version. We expect them to get a few thousand to tens of thousands of views. We'd really appreciate your critiques.
AISafety.info has two new intros: a short version and a longer version. They're intended for a reasonably smart member of the general public.
Our intros are meant to be:
We want these intros to be a resource that you would be eager to share with friends, family, co-workers, etc. So, please give us some feedback! Even a 10 word comment would help. E.g. "Seems OK, but talking about AI doom to family is weird." Or "The section on instrumental convergence doesn't add much; consider shortening it." Even if you only read a tiny part of our intro, we still want your feedback! Even if it is to say why you think this project is even worth doing.
We're keen to learn what improvements to these articles would make them most useful to you as a shareable resource. (Of course, we'd also appreciate you catching any errors.) We list some of the feedback we're most interested in later on. Only then can we make the most of our upcoming opportunity to spread good models of AI safety to as many people as possible. Rob Miles is making an accompanying video which will link to our articles. We expect that to result in a big influx of viewers, who will share our articles if they find them useful.
The articles are listed below, with summaries. The first four from the long version are imported to LW as part of a sequence. (We'll eventually import the rest, but are avoiding spamming the front page.) You can give feedback 1) on this post (best for us), 2) by clicking on the links below and then clicking on the little pencil icon below the title, or 3) by commenting under the LW posts when they're published.
A self-contained introduction that aims to draw people in with more of a narrative. It's a 16-minute read. We want this to be the one article someone can read about AI x-risks if they don't have time for anything else, and still get a decent picture of what it's about.
Summary: Companies are racing to build smarter-than-human AI. Experts think they may succeed in the next decade. But rather than building AI, they’re growing it — and nobody knows how the resulting systems work. Experts argue over whether we’ll lose control of them, and whether this will lead to everyone's demise. And although some decision-makers are talking about extinction risk, humanity does not have a plan.
Feedback we'd find most helpful: Object level feedback. Is the first paragraph engaging? Is the article engaging? Is it the right length? Does it place a high cognitive load on readers? Have we conveyed the risks you're worried about? Do we foreshadow content properly? Did we convey that there are deep models implying AI x-risks? Should we have more/fewer links? More/fewer footnotes? If you share the article with friends or family, what did they say about it?
Feedback on any part of this article would be helpful.
A more detailed introduction providing a high-level overview of the arguments, which is meant to lead into the rest of the site. It has 4 subsections, with 14 articles in total. Each article should take 3–5 minutes to read.
We want this to be a set of articles that someone who's interested in the ideas around AI safety could read as their first introduction. We also think each article can serve as a stand-alone primer on its subtopic, which can be linked to without the context of the whole sequence.
Summary:
Feedback we'd find most helpful: Object level feedback. What's worth including or cutting? What's worth emphasizing or deemphasizing? Where should we change the tone? Where is it confusing or boring or illogical or factually wrong? How can it be made more engaging without compromising on concision and correctness? Where could the organization be (clearly) improved?
Feedback on any these articles, or any part, would be helpful.