I am reminded of the posts by @Aidan Rocke (also see his papers), specifically where he argues that the Erdős–Kac theorem could not be discovered by empirical generalization. As a theorem, it can be deduced, but I suppose the question is how you'd get the idea for the theorem in the first place.
Thank you for very relevant references - diving into posts and papers.
the question is how you'd get the idea for the theorem in the first place.
This is the question indeed. I was also thinking about examples where it would be hard to make the deduction within the framework itself. The classical one is probably the solution to cubics, where it was proven that you cannot reach the closed form solution using operations over reals. But it is trivial in complex numbers, and the formula will return a triplet of reals you can easily substitute and verify.
I suspect this could be actually quite a common case, but haven't found the general formalization yet - incompleteness is somewhat analogous, but it says the statement is unverifiable. With cubics we have a way to check the correctness of the solution, which arrived through an "inexpressible" channel detour.
Hi Mitchell,
I've actually solved the puzzle of cubics being "non-findable." There was no formal definition, so I needed to make one myself. In case you'd be interested, here is a link: https://www.lesswrong.com/posts/TbMedZDimwc5TRnu7/the-two-board-problem-training-environment-for-research
It also discusses why, for some math problems, it's hard to arrive at the idea in the first place.
Quite interesting collection of quotes in the Post Scriptum, and the post itself is a good shake-up to assumptions about computability of stuff (minds, the universe). It feels directionally opposed to the reigning philosophy here on LW, at least to me, who lurks occasionally, but that's why I liked it.
P.S. The writing style is nice.
Thank you for the kind words.
It's directionally opposed indeed. In my opinion, this website, which once started with the analogy of Map and Territory, forgot that the analogy is not a Territory, but a Map. And thus it should be challenged by evidence. It was a very interesting finding that the people who made a genuine discovery don't report their process as incrementally becoming less wrong and instead talk about taking a wild leap which turned out to be correct.
There are plenty of narratives about AI self-improvement, AGI, and superintelligence - and substantial uncertainty surrounding them all. In my opinion, these are completely reasonable things to worry about, considering three years ago your PC suddenly started talking to you. Naturally, I think about these questions too. Can I trust the outputs of this thing? How will it continue to progress? Should I expect the Skynet scenario? Do I need to keep a crowbar or a bucket of water near my laptop? In this essay, I want to share three years of my research into this topic and present the observations I've collected.
Intro
This essay aims to establish a sense of what generalization is, why is it important for general intelligence, and why it would be much harder to achieve than incremental improvements. I am pretty materialistic, so no "qualia", quantum coherence at room temperature or divine blessing would be summoned throughout the paper. It won't require any special knowledge - this work will be mostly common sense, historical observations, and careful reading. One example will have a function which I will draw for the reader's convenience. I will also explain what "making generalization" feels like, and what it looks like on the brain chemistry level, since I happen to have insider access to this research area.
For reader's convenience, I attach TLDR. I don't recommend to read it right away - no explanation of the bullets will be given. Essay structure echoes GEB in a lot of ways, and much like with GEB, the essay self-answers about why it has this structure. Similarly, there is absolutely no way I could convey this property in a summary. Every part is resolved by the essay itself. That said, I understand both curiosity and impatience and absolutely respect the choice to experience confusion[1].
I also know what every sane person currently does with the long text, so:
Verse
Let's start with building the recognition of what generalization is, and even more importantly, what it is not. As I have promised, I will illustrate the process, so meet sin(x):
Sine is a simple periodic function. It's not the simplest one, like a line, but it is pretty straightforward. It is an example of the most boring repetitive behavior in mathematics one may find, and it appears in numerous processes: anything waves, including sound, pendulums, etc. It's also a computable function - meaning, given a computer (or a person with a pen and paper), we can write the value at any point with an arbitrary precision. It makes it a perfect candidate for a small-scale generalization thought experiment: sine is a dead simple abstraction which plays a foundational role in describing a lot of stuff happening around us.
As a human, you would notice that it's pretty symmetrical - given the interval π you get the opposite value, and given the interval 2π you have the same value. So if you'd like to predict what the value at x is, you just drop all unnecessary intervals of π and focus only on the meaningful part.
To compute the meaningful part the computer can't use the sine itself - it needs simpler operations. There are few ways to make sine computable, but the most straightforward is Taylor series. It's nothing more than summarizing a long sequence of steps. You can look it up, or trust me that it's a really simple thing, maybe a few lines of code.
Given how easy it is to notice the periodicity, it would be fair to expect a sufficiently smart program to learn sine and compress it to something like "filter unnecessary repetition of 2π → compute series". We could give such a program an interval to learn the values of sine from and an unlimited number of attempts to modify itself to match the samples. You could assume something like: at each step it will be closer and closer to the program we described, and that's where precision will come from, right?
Wrong. It won't. It just won't, experimentally. If we don't encode this function (or something very similar and oscillatory in nature) - we get wrong values as soon as we leave the training region. Things we tried just don't find it. To fix it we need to include sine in the definition of our search program (e.g. use in the neural net, provide as symbol etc).
This failure to infer the program globally and finding it only for approximating the trained sample is the difference between interpolation and generalization.
This specific example is not a big deal practically - we know basic functions, so we are able to encode them, and certainly is not an argument against LLMs. Though it raises some questions about how humans were able to find this function given clay and sticks. But for now let's focus on interpolation vs generalization:
If we can take sine and compute it for something like x=2π×1032+0.1235 and any other arbitrary value, after training on the range [−1000,1000] - we generalized. If we can't be precise outside of this interval - we interpolated. If the range has good extension outside of the interval - it's an extrapolation, but not necessarily generalization - the latter guarantees limitless extrapolation within the pattern.
Lift
After looking at what generalization looks like, it's equally important to develop an intuition for what it feels like. While my later examples will cite people who reshaped entire fields, I want to show you that the feeling itself is pretty familiar. Typically it just manifests a bit less dramatic than establishing foundations of computation or relativity.
Try to remember the last time you couldn't figure something out. Not for five minutes, no. For hours. Maybe for days. How you tried ideas, how you made your mind work on it, and how your thoughts were stumbling in the wall of unmoving task which just didn't crack. How there were too many pieces, and it didn't look like you understand how to fit them and connect. How your thoughts cycled, rotated the problem through different angles, disassembled and assembled with that strange feeling that something is near, but is missing, like there is something… which you just can't describe what it is. That's a feeling of serotonin keeping you in the loop and TrkB promoting plasticity on the background.
One moment you do something completely unrelated: wash your hands, or walk around the home or take a shower, and… here it is. Eureka. The missing piece of the puzzle which makes all others self-assemble. You feel satisfied and happy - that's dopamine kicks in to establish "good vibes". But it's not that kind of happiness which makes you relax. You feel sharp. The thoughts are cascading, and the resolution is following neuronal chain-reaction. You feel motivated to write it down right where you are and check it immediately, because that's what insight is, and it's the most important thing right now. That's norepinephrine kicks in to make you remember the details of this moment, and associate it with the vibe of having done something… special. Magical almost.
You can describe steps before, you can describe steps after, but that core moment feels almost divine. Like you sent an important request and a response took you completely off guard. The world shifted a bit, the castle of the problem opened the gate without exhausting siege. This part of the map is now clear and visible, and it's strange to imagine not to have it.
This is what it feels like to generalize. I hope it's familiar to anyone who reads this. I had a lot of those moments, and I find them the most wonderful feeling in the world. Like everything pleasant, it's addictive, but it would probably be among the wisest addiction choices.
And speaking about addictive things, let's dive a little deeper into the process of neurochemistry.
If we tune down the minor tones in the fugue of neuromediators and leave the main voices of the learning theme, the progression would look like:
Problem →
Serotonin → Feeling of agitation, anxiety, complexity → needs solution →
TrkB → plasticity mediation → brain is more "volatile" → search phase →
Match →
Dopamine → good vibes →
Norepinephrine → motivate → check →
Memorise
(My wife who studied TrkB looks disdainfully at this simplification)
Even compared to our crude schema, the algorithm "didn't match, compute diff, back-propagate" looks a little bit like comparing a drum machine to Bach. This fact by itself doesn't undermine our efforts in digitalization - maybe this design isn't a minimal viable insight setup. Maybe back-prop is sufficient. And of course we can always take a bit more inspiration from biology! Nevertheless, I would note it as a second minor dissonant voice joining our sine observation. But let's continue.
Bridge
Sometimes it makes sense to step aside from the track, look at the whole picture and notice forest behind the trees. We know what the tree of generalization looks like - it's one that's simple and makes insanely good predictions. We know what it feels like to touch it, and suddenly jump like Archimedes with "Eureka!". But what would "not generalization" look like and feel like? Working backwards we can observe the moments when the components somehow… don't click. Those can be really good components, they can be really important and make real improvements. But for some reason they match their peers in a patchwork style of a drunk impressionist. Simplification chain-reaction fails to bootstrap and the picture stays incoherent. Tens of thousands of small lines don't form a beautiful, smooth and symmetric sin wave. Our drunk impressionist fails to reproduce Van Gogh.
With that distinction in mind, we can start our observations. And there is no better place to observe the play of human progress than the royal box of History's theater. We look for repetitive patterns, aggregate over the Brownian motion of millions of individual lives, and statistics emerge.
I've made some effort to apply the generalization/non-generalization framework on the biggest scale. From this height, only the biggest tectonic shifts are what counts - new axioms reshaping sciences or new fields being founded. I don't aim to undermine anyone who contributed their best to science and didn't reach the Gödel/Turing/Einstein/Bohr level of change in their field. But I think we can agree that given those examples, they are the ones that count from this height. And given the criteria "foundational theoretical contributions", you probably can already feel the distribution shape.
Ironically, we found another exhibit for the wtfhappenedin1971.com collection:
You can check the data and methodology by the link.
If we account for the fact drops in the central cluster are literally WW1 and WW2, the picture is… quite remarkable. If you are confused about the 1970+ decline I would suggest one interesting angle to think about it systemically[2]. But that's not the main point.
If you happened to have a good history or literature teacher, you probably heard that history and literature go hand in hand. The people who represent an era are the ones who created it and simultaneously they were the ones shaped by this era's problems. This website (LW) is the corner of the Internet where people representing the lumps on our chart are quoted and referred to more than anywhere else. And since we have a very particular question in mind, the question of machine intelligence and capabilities, we won't find a better person than A. M. Turing to lead our choir.
Solo
Turing's writings have a wonderful property of being concrete. Unlike the context of Kafka's prose, where blue curtains would suddenly have enormous degrees of meaning, with Turing one can reliably assume that the curtains were indeed blue. If he wanted to add anything specific and elaborate on their properties - there will be a dense paragraph meticulously explaining what Turing meant, and what he didn't. That said, I am not arguing against the attempts to understand the theory of mind and the context behind the writings. If anything, the last 3 years have shown us how important the context is for attending to the right details. And attention is what we need.
So imagine for a moment, that you're in the beginning of the 1930s. Physics and mathematics have their renaissance - the fields are volatile, they change. The seeds of the ideas of the latest decades grow like enchanted beans towards the sky. They entangle in hot debates between the greatest minds in the history, but the direction is clear and only the sky is the limit (if there are any limits). Simple, intuitive models replaced by continuous ones, which look alien, but the measurements confirm the theories to absurd precision. The global integration takes off, the speed of communication is nothing like letters - telegraphs are cross-Atlantic, phone lines are getting established, the radio is not a miracle but something mundane. Digital revolution does not yet exist as a concept, the word "compute" means writing the steps on big paper sheets, and "computer" is a job title of the person who does it.
You're studying in Princeton, and one of the students there is quite a character. He somehow manages to embody a lot of stereotypes and fit none. He mixes a fairly typical chaotic genius look and a strange methodical and precise manner of speech: like the words would be written down for stenography. If he needs to think in the middle of the phrase, he takes his time without giving you a chance to speak with a pretty annoying repetitive Am… He's not that social, and doesn't need a small talk - but ready to discuss the abstract topics and great ideas, if, of course, he finds both the ideas and your opinion on them interesting enough. He is athletic and can be seen running outside alone on a regular basis. One day you hear that this incredible persona submitted a paper, and you decide that it might be quite an interesting read. And, oh my god, it starts as you might expect from this guy:
… Computable? You mean those numbers written down by the computer? It's human's work, what definition is this? Who is this guy to introduce the numbers which are "computable", as ones which are written down by a machine? And he says it's because memory is limited? What the…
Adjusting the expectations on the fly to account for a sudden cognitive dissonance, you continue, and well… It doesn't get better from here:
What the hell is even that? Who tries to "suppose" that the number of states of mind of the person is finite!? But ok, let's think through that: this person is working with Professor Church, who is highly respectable, and while his analogies are remarkably radical, they somewhat help me as a metaphor to understand the process.
You go through the math and it checks out. The paper is decent, the results are supported, and let's write off the wild attitude towards the human mind to an awkward way of doing an analogy. Machine proposal is ambitious, but the formalism is decent.
Let's step out for a moment from the perspective we just took, and return to the reality where AI routinely performs the operation of writing a program for a machine computer while being a machine which writes it. From this perspective we of course can see that things Turing has written in this foundational paper can be taken completely literally. This will help us a lot in our journey. Let's dive back in 1930s.
Years pass, and you're still studying. You became interested in the work of Gödel and Tarski who show the limits of what formal systems, including mathematics, are capable of. One day in year 1939 you hear that this strange guy released a highly relevant work. Remembering the first experience you mentally prepare, and honestly - what surprises you, you can relax and follow the logic of the work pretty clearly - he defines which problems we'd be trying to solve, goes through the math, until:
Ok, you relaxed too early. We could already learn a pattern that almost anything which starts with "suppose" has a certain chance to blow up, leaving behind only the ruins of cognitive dissonance. What does it mean!? What definition is this? And how is it structured! "Let us suppose we're supplied". "A kind of an oracle as it were". Ok, if you introduce the thought experiment, do it properly, but what is that final statement? "We shall not go any further" "apart from saying that it cannot be a machine". If you introduce the thought experiment why are you defining the thought experiment like that? Maybe he means his own formalism of the machine… Yes, there was something like that in the previous paper. Maybe he was just inaccurate with his wording, and what it means is "the machine formalism we previously discussed".
You continue reading, the math checks out and the introduced oracles are indeed useful to support it.
In the end you face a chapter, which, while written in prose, surprisingly makes perfect sense for you:
Aside from the math which always checks out, this is the sanest text you saw from this person till this point.
Let's again take a bird's view on the topic. I would remind you that we're reading Turing. The person who would write exactly what he was thinking even if you point the gun to his head - like there was a physical law making him find the most precise wording. Let's notice a dissonance with the tone of the first paper. And also note, that the paragraphs about oracle and intuition are very far away and not connected anyhow. I remind you, that if the curtains had meaning - the author would say it out loud.
Now let's proceed to the paper which probably had the most cultural effect. The one which defined the rules of the Imitation Game.
Those were very hard ten years. WW2 left Europe, USSR and Britain in ruins. Dozens of millions dead. But we're recovering. Soviets are near, and the peace is fragile, but things are coming to norm, whatever it means after the civilized world had burned for 5 years straight. You know that this weird student whose paper you followed suddenly disappeared for years and the only thing you heard from your colleagues is "military". Interestingly, the universities started to build so-called computing machines shortly after the war. Those indeed are able to perform computations and assist in them, though they are notoriously unreliable and sometimes bugs or even rats disturb the process. It's 1950, and suddenly you hear that this strange guy released a new work, and unlike the previous ones it's nothing like the things he worked on before. It is a philosophy paper. With the mixed feeling of the curiosity of a toddler glancing at the fork and the nearby socket you decide to read the work named Computing Machinery and Intelligence…
Am… Okay, this is even not as bad as I thought it would be! So it's just "can machines talk". Wild assumption, of course that they will ever be able to understand the speech, but the philosophical question is an interesting one. And the guy was working on the machines' formalisms even before they were engineered, so it's interesting what he would say.
Hm… So he proposes to have instructions for each scenario? This is impossible to write, but as a thought experiment it makes sense.
Ok, that's wildly specific. The numbers are astronomical, but what does it mean "too meaningless for the discussion", and suddenly jump to the conclusion that everyone would consider it as normal…
Ok, that's somewhat clear, that's the reference to the incompleteness and halting problem…
… Halts …
Do you think I was joking about the fork?
I understand that this is a lot to unpack, but what helps us, is the learned rule of reading Turing. Let's take everything written completely literally, and write it point by point:
I must admit that Turing was of better opinion about his successors - we missed the deadline roughly by 25 years, and heavily bloated the storage requirements. Maybe this delay is justified given the sharp decline on our chart after 1970.
Chorus
This was quite an intense crescendo, so I think we deserve a break. Sometimes it makes sense to enjoy the silence and take a metaphorical walk around your problem landscape, without engaging with any part in particular. Not force the thinking but just allow the mind to flow. I would start with a reminder, that we were concerned about current AI takeoff, and the problem of recursive self-improvement. This lead us to the question of what is the difference between interpolating and truly generalizing. The common sense intuition of "Eureka" followed, and with those lens we took a look at one of the greatest minds of the history, but from the perspective of a rather confused observer, who would see the Turing making rather impressionistic moves. But as we see from 70 years later none of those was really confusing, and the predictions which Turing made are quite coherent. From his perspective the puzzling parts were completely different to the ones which are dissonant for us.
Sometimes it makes sense not to just read an author, but to try to model his perspective. You know that numbers are not the problem, but there are some problems which don't reduce to number manipulation. You are accurate enough to not pattern match and avoid giving a convenient label for the mind to shut the door. It's clear to you that machines will be able to talk, as well as it's clear to you that people around confuse the ability to talk with the ability to think. You're careful to not dissolve the mystery and keep your observations organized. An oracle. An intuition role. The undefined thinking. The halting problem. The ability to extend to new formalisms. Your colleagues including Gödel, attributing their insight to something Platonic or Divine. Pre-symbolic.
Of course, I can't give you a map of the Turing mind, and say how one of the brightest people of the explosive period of the scientific unification would solve the problem from within. But there is one instrument, which we were mastering throughout, and which just makes most sense to use.
Sometimes, to find the thing we're interested in, we can define what this thing is not. And Turing had given us an answer multiple times. The thing he knew for sure: it couldn't be a machine he described.
The human mind has a certain bias towards symmetry and smoothness. The works of Van Gogh are magical in their wave-like statically captured motion. The music naturally sounds pleasant if it's harmonic. We have a unique ability to feel coherence. To follow melody. And to find it. As soon as we glance at the landscape in front of us, or hear the sound, our brain just does it automatically, without our conscious involvement. We are trying to find the pattern to generalize.
We have started this essay with one of the simplest symmetric and progressive mathematical forms - sine. And it resisted digitalization. We tried to formalize the insight but it was not giving up to the symbols. We went through neuroscience of learning, about which we now know much more than Turing - but what he knew for sure, it wasn't digital substrate. We went through the critical phase of brain in which it looks for the insight - and it was volatile, it was allowing more connections, more plasticity. We went through the period of greatest insights in history and from the bird's view we observed the same - the seeds of ideas forming in a boiling cauldron of 20th century science.
Sometimes to observe what the things are we need to understand what they are not. In Turing's case the answer was obvious: they are neither discrete nor organized.
The most reasonable action from within would look like a confusing impressionist move to the ones who didn't see your insight. Especially if they already labeled you in this category. But from the internal perspective, the most reasonable move if you have stuck and cannot halt is to take a step out of your typical frame of reference. And leave observers in a mild confusion.
Breakdown
Turing was prosecuted for homosexuality, or rather for a combination of several factors: being homosexual, seeing nothing wrong with it, and being unable to deceive the people or read the social clues. He reported the burglary of his house to the police, and ultimately wasn't successful in hiding his relationships with another man. The British government rewarded the war hero who helped to win it and saved countless lives with chemical castration. It disturbed his mind, broke his body, and ultimately led to suicide (officially, the theories vary), two years after.
350 years of progress after Giordano Bruno is not a sufficient time for collective humanity to stop torturing and murdering the best of its people for minor disagreements with social norms which don't make any particular sense. The reports say that Turing was deeply confused about what he is even prosecuted for - and rightfully so. Unfortunately, as history shows, the social norms can stay irrational much longer than ones who don't fit in them can survive.
But right before this tragic… Sorry, I don't have words for it, better than "yet another manifestation of how stupid collective humanity is", Alan Turing released one paper. And it was something unlike anything he wrote before.
This paper was taking the chemical processes. The chaos of diffusion in biological substrate. The core model was continuous wave dynamics. None of the basic equations in this paper are discrete. The evolution was continuous process. It was physical. And physics has a wonderful property of performing the things which are hard to model digitally with ease.
The core question of this paper was - how, despite physical preferences for minimizing action and keeping symmetry, such systems are developing the discrete patterns?
And the answer was quite precise, it's instability and the chaotic process which reaches criticality and amplifies it until the new equilibrium:
What's also interesting about this paper, is that it mentions the main results and main area we attribute to Turing's genius almost in passing, utilitarian manner:
The morphogenesis paper is an outlier. A leap of intuition. It's only now that we know about the TrkB, the molecular mechanism of neuronal plasticity, the brain waves, the critical periods and memory consolidation. If he only knew that… He wouldn't be surprised and would take it as inevitable. Because the beautiful explanation is also somehow the most probable one. The visible part which we can introspect was always just a part of something bigger.
We don't know what this bigger looks like fully yet. But one thing we could observe and which we know for sure is continuity and oscillation. And those are sufficient to construct the discrete patterns.
Waves⊃SymbolsOutro
I hope I was able to present some compelling evidence on what exactly AI will struggle to do and how important those things are for self-improvement. This is a pretty substantial amount of knowledge, and like with a lot of things, the resulting symbols are just a bleak representation of the coherent physics which happens underneath. I wouldn't be able to summarize it better than this text, but now I hope that we have a common reference for what the symbols mean.
I named this essay Exponential Takeoff of Mediocrity. Initially this name was given due to the hypothesis of AI interpolation. But during the work on the chart I have presented, I realized that there is more to it. AI is trained on our data, most of which is produced after 1970 so unsurprisingly it's also a good mirror of collective us.
Despite the comfort, unlimited supply of amplified narcissism is not something which will make us reach AGI or achieve anything meaningful - mirrors are quite a closed system. They give you a lot to observe but the most impressive leaps we're making by stepping outside of our reflection.
There are things which can be achieved architecturally, but nothing will help us to escape the limitations of the substrate.
So get your brain ready - because it's still the best machine available to us. Notice the confusion. Step outside. Get insights.
The map IS NOT the territory.
It's time to ride the wave.
Post Scriptum. Fugue.
You may have noticed that this essay is anything but an argument. It's more like a lens or a filter to show you my perspective.
As a person who grew up reading Rationality A-Z, HPMoR, Feynman, etc., I fully understand that sample size matters enormously, and an N of 1 is not sufficient. Luckily, we don't have one sample. Our chart has more than 100 data points.
So, I would like to invite you, for the last time in this essay, to take a look at the phenomenon we study and observe what it is not.
I. Founders of Computation and Formal Systems
Kurt Gödel (1906–1978)
Key contribution: Incompleteness theorems (1931), recursive functions, foundational contributions to set theory.
Emil Post (1897–1954)
Key contributions: Post production systems, Post correspondence problem, independent anticipation of Gödel's incompleteness and Church-Turing computability.
Alonzo Church (1903–1995)
Church was careful to frame the Church-Turing thesis as being about effective calculability - a precise mathematical concept - rather than about minds or cognition. He did not publicly extend his results to philosophy of mind.
Key contributions: Lambda calculus, Church's theorem, Church-Turing thesis.
John von Neumann (1903–1957)
Key contributions: Von Neumann architecture, game theory, mathematical foundations of quantum mechanics, self-reproducing automata.
II. Founders of Modern Physics
Max Planck (1858–1947)
Key contribution: Originated quantum theory (1900). Nobel Prize 1918.
Albert Einstein (1879–1955)
Key contributions: Special and general relativity, photoelectric effect, Brownian motion. Nobel Prize 1921.
Niels Bohr (1885–1962)
Key contributions: Atomic model, complementarity principle, Copenhagen interpretation. Nobel Prize 1922.
Erwin Schrödinger (1887–1961)
Key contributions: Wave equation of quantum mechanics, What is Life? Nobel Prize 1933.
Werner Heisenberg (1901–1976)
Key contributions: Uncertainty principle, matrix mechanics. Nobel Prize 1932.
Wolfgang Pauli (1900–1958)
Key contributions: Exclusion principle, spin-statistics theorem. Nobel Prize 1945.
Eugene Wigner (1902–1995)
Key contributions: Symmetries in quantum mechanics, Wigner's friend thought experiment. Nobel Prize 1963.
III. Founders of Modern Biology
Charles Darwin (1809–1882)
Key contributions: Theory of evolution by natural selection, The Expression of the Emotions in Man and Animals (1872).
Charles Scott Sherrington (1857–1952)
Key contributions: Function of neurons, concept of the synapse, integrative action of the nervous system. Nobel Prize 1932.
Santiago Ramón y Cajal (1852–1934)
Key contributions: Neuron doctrine, neuroanatomy, neuroplasticity. Nobel Prize 1906.
Barbara McClintock (1902–1992)
Key contributions: Discovery of genetic transposition (mobile genetic elements). Nobel Prize 1983.
Francis Crick (1916–2004)
Key contributions: Co-discovery of the structure of DNA, neural correlates of consciousness. Nobel Prize 1962.
IV. Information Theory
Claude Shannon (1916–2001)
Key contributions: Information theory, Boolean logic applied to circuits, foundations of digital communication.
PPS due to attention mechanisms.
Ok, here is the compact summary of the main thesis, which would work with certain kind of reader who already have made similar observations:
The "Country of geniuses in data centers", would look more of a country of clerks and bureaucrats.
Capital deployment and markets are the distributed coordination mechanism across any area of economics including science. High inflation makes it rational to operate on short cycles, because the old printed money would become cheaper and would need to be spent now. But future printed money would be much cheaper to get. The reward function of the grant would favor fast convergence of interpolation over step-like derivation of the general result.