I sometimes (since 2022[1]) say that I'd prefer if the widespread LM products 'spoke in wikipedia voice' rather than being specifically trained to put forward a facade of personhood (first person pronouns etc), which is especially geared to bypassing rational engagement. I don't think people mostly thought of LMs as 'the AIs' until chat posttraining hit it big.
Sadly I think that's by default a losing battle as the chat interface is both especially navigable (as you mention) and intuitively appealing to a mass market.
It's another draft post sitting in my queue for embarrassingly long ↩︎
At FLF one of our initiatives we're recruiting for is an 'epistemic stack', which I think fits the bill as a backend/foundation for many of the desiderata you're describing. LLM chat interface would be one UX form factor on top.
Epistemic stack would be a (probably distributed) cache of annotations and metadata connecting claims to supporting sources, constructable and expandable dynamically. The cheapest and widest-coverage construction would use LM-based agents over webtext, inferring support from existing links, citations, and sometimes proactive web search for supporting (or contradictory) sources. Human participants (authors and readers) would be able to provide various annotations, including endorsement of (or alterations to) inferred epistemic support links. Something like git, versioned, signed, annotated DAGs would then be available to downstream epistemic applications (including via RAG for LM consumption, but also many other imaginable formats).
MVP on-demand tree construction for a given claim is already practical, though unreliable and more expensive than a system with caching would be.
Down the line, if more verifiable sources of ground data (signed cameras, etc.) get more widespread, such data would readily integrate as leaves.
Compare also Society Library which has some similar prior work (mostly manual), and may be moving in a similar direction.
There has also been related discussion of 'wikipedia for LLMs', and though I haven't heard more technical refinement from such proponents, that term might be intended to expand to a similar overall concept.
Note that 'wikipedia by LLM', like grokipedia, does not currently have any meaningful claim to epistemic grounding, correctability, or transparency/legibility, though its form factor would at least inherit the navigability of wikipedia.
Note that we're aware of the cautionary tales of Cyc, Xanadu, and Arbital!
We're hoping a combination of:
means that 'this time is different'.
A thought that has been bouncing around in my head for the last couple years is wikipedia as a motivating metaphor for what LLMs should become. This is particularly in contrast to social media. I owe this idea to Tom Everitt, who conveyed it to me at an Agent Foundations conference at Wytham Abbey.
LLMs are """the new Wikipedia""" in some ways: it is a new tech-enabled knowledge source, which we can consult for anything from quick questions to long information-binge sessions. It has questions of reliability, similar to those Wikipedia faced at first. The advice teachers give about it is similar: "It can be a useful resource, but you should use it to find primary sources, rather than citing it directly."
However, there is a reasonable fear that LLMs will become (or are becoming) """the new social media""" instead: maximizing engagement, spreading misinformation, addicting users, etc. X and Meta are certainly trying to figure out how to do this. However, this looks like an extremely-not-good overall trajectory for machine intelligence.[1]
How might LLMs be steered in a desirable direction? I'm going to take a somewhat cargo-cult strategy, imagining LLMs being closer to wikis in as many dimensions as possible.
Co-Agentic
One aspect is co-agency, as discussed in What, if not agency? This is closest to what Tom Everitt had in mind: Wikipedia empowers people, while social media addicts people. This intuition, formalized, may lead to better alignment/corrigibility targets. If you've got to aim superintelligence at a goal, then human empowerment seems safer than trying to directly specify human values; it keeps humans in control.[2] More optimistically, the co-agency line of thinking could articulate a better alternative (to agentic systems oriented towards goals).
Navigable
A related dimension that might be relevant is paternal vs navigable formats. "The Algorithm" running social media chooses what you get to see in an opaque way, while Wikipedia focuses on easy navigation to the content you want. I'm not entirely clear on how to analogize this to LLMs, however. On the one hand, LLMs are very much an opaque algorithm deciding what to show you. On the other hand, in some sense they're intensely navigable; they show you whatever you ask.
One idea is moving more in the direction of Microscope AI. What if, when you ask a question, instead of concocting a summary for you, AI showed you statistics of what percentages and types of people might answer in specific ways? Base models are extremely good statistical models of human-written text -- they have that sort of information. A human-navigable view of the raw statistics like this feels a lot more informative and trustworthy than something fine-tuned for """truth""" by an opaque corporation.
Correctable
One thing LLMs lack, when contrasted with Wikipedia, is a good story for why it might become increasingly trustworthy over time. Wikipedia's problems can be corrected by anyone who notices them. Malicious """corrections""" can be filtered out.
I've fantasized for some time about an LLM interface with a heavy-duty correction tool:
This certainly isn't a perfect vision, but it gestures in a direction of what Wiki AI might look like. This sort of interface might be used for personal Wiki AIs (maybe without the expensive fine-tuning component), and also for a collective project (similar to Wikipedia).
Collective
Maybe a Wikipedia-like AI has to be managed in a Wikipedia-like way, down to the detailed governance mechanisms Wikipedia uses for moderators (appropriately adapted).
Wikis came before Wikipedia, however. Maybe similarly, we have to first invent Wiki AI, and the -pedia comes later?
Academic
Here, I mean the ordinary academic virtues, chief among them being clear citation trails. Modern LLMs can search the web and provide citations that way, but it would be even nicer if relevant parts of the training data could be referenced in order to give some idea of how an LLM generated a particular response. (This might be too hard and not important enough, but I'm fantasizing here. The idea does seem to fit somewhat with my earlier comment about microscope AI.)
Is there any possibility of LLMs becoming a respectable academic publishing model? (At least in the way a well-written wiki could be?)
To put it a different way: can you envision someone taking responsibility for an LLM as an intellectual product (in a field other than machine learning)? A "live theory"?
One mental image I had: suppose I have my personal LLM running on my personal academic webpage. I've fine-tuned this LLM to represent my personal ideas (a supplement to my published writing).
Editable UI?
Maybe the best UI for what I'm envisioning is not a chat interface. Wikis were innovative UIs: websites with the editor built-in. I'm just spitballing here, but perhaps a Wiki AI is more directly like that: a website with vibe-coding built in? A "live interface"?
Neither the "social" nor the "media" parts of social media are bad in themselves. The bad part is how these algorithms are out to get you, justifying paranoia.
If, as many people expect, we never solve the problem of how to ensure that superintelligent AI systems don't kill everyone, then this is a rearranging-chairs-on-Titanic type concern. However, if humans do indeed solve a version of the alignment problem adequate to deliberately point superintelligence in a desired direction, then things could go terribly anyway if superintelligence is pointed at grabbing human attention or something similar.
There are things that could go wrong. A clever human-empowerment-maximizer would first modify human values to want power more, then empower the humans. So, closer to a good idea would be to maximize human-empowerment-while-not-changing-human-values... but this could also go wrong...
One thing that would really help here would be micropayments for compute. Imagine if someone has an OpenRouter account, for example, and when they visit my webpage, there's somehow an easy way to connect their OpenRouter account, and use the credits they have there to pay for talking to my personal LLM. This does not make sense with the way OpenRouter currently works; I'm just saying, something like this would help support the sort of ecosystem I'm imagining here. (This is an example of a concrete thing one could try and build in order to help support this vision, although there might not be enough demand for it to work as a business!)