See Paul Krugman
There are model-oriented economists, like Alan Blinder, who also write for a broader audience, and they don't put their equations in their books and articles; but the skeleton of the models that structure their thought is visible under the surface to those who know how to look. By contrast, in the writings of Reich or Galbraith what you read is what you get -- there is no hidden mathematical structure to the argument, no diagram one might draw on a blackboard or simulation one might run on a computer to clarify the point. Maynard Smith not only has a name that should have made him an economist; he writes and thinks like an economist, representing evolutionary issues with stylized mathematical models that are sometimes confronted with data, sometimes simulated on the computer, but always serve as the true structure informing the verbal argument. what makes Gould so popular with intellectuals is not merely the quality of his writing but the fact that, unlike Dawkins or Ridley, he is not trying to explain the essentially mathematical logic of modern evolutionary theory. It's not just that there are no equations or simulations in his books; he doesn't even think in terms of the mathematical models that inform the work of writers like Dawkins. That is what makes his work so appealing. The problem, of course, is that evolutionary theory -- the real thing -- is based on mathematical models;
As a physics enjoyer, I am reminded of a quote from Feynmann (or someone else?) that agreed with my personal experience: the strength and weakness of intuition is how malleable it is. The strength being that you can make the math structure become intuitive. It's not just that you assume A and see whether you actually can get B, but rather you assume A, get some weird result, and then stare at it until it makes intuitive sense. Maybe you then land upon a new concept. Alternatively you might crystallize some intuition you have to see what the true nature of the thing really is - this can be best seen in mathematics itself, where the definitions encode express many of the "True Names" of concepts. This is a lot more than just falsification, you do a lot of locating and understanding of ideas this way.
I don't know enough economics to give helpful examples, as all the ones I have are ones where I already knew the math (and so can't compare clarity after learning the math) and that could plausibly be qualitatively understood without it.
I can however think of the VNM theorem as an example where the basic theorem itself is the concept, that is that you can get utilities via preferences on lotteries by having "strength of preference" refer to what you'd choose under uncertainty.
Regardless, quantitative statements are still pretty important though. To be able to predict the size of the effect, or to have your theories make claims about the sizes, matters a whole lot (including qualitatively - it's important to know which consideration are negligible and which ones dominate).
Off the top of my head, the Lerner index is the marginal profit as a percent of the price, and the theory is that a monopolist maximizes profits when this is equal to -1/elasticity of demand. Without following the calculus you won't arrive at this direct relationship between elasticity of demand and the profit percent.
Also if you count voting theory as economics, there are some rather counterintuitive results like Condorcet paradoxes (which I once came up with an example of independently in the context of intransitive dice, but wouldn't have been able to do so if I didn't know about directed graphs - the point here is the help mathematics provided) and Arrow's impossibility theorem and other no go theorems.
Edit: Recently, I used a linear approximation and published elasticity values to estimate how much reducing consumption of a good (in the context of vegetarianism) should be expected to reduce supply. Apparently, it's the same formula as tax incidence! The important part of the derivation was just "the result is on the order of the ratio of elasticities" (to first order:
Nice post. One thing I'd add is Sahil's description here of mathematics (and other "clean" concepts) as "achiev[ing] the scaling and transport of insight (which is the business of generalization) by isolation and exclusion". But, he argues, there are other ways to scale and transport insight. I think of emotional work and meditative practices as showcasing these other ways: they don't rely on theories or theorems, and often there aren't even canonical statements of their core insights. Instead, insights too complex to formalize (yet) are transmitted person-to-person.
Where I disagree with Sahil is that I suspect other ways of scaling and transporting insight are much more vulnerable to adversarial attacks (because e.g. there's no central statement which can be criticized). So in some sense the "point of the math" is that it means you need to rely less on the honesty and integrity of the people you're learning from.
For me a key benefit of maths is to answer the question "how much?", to turn qualitative intuitions into quantitative models.
For example if someone tells you "drug X binds to receptor Y which triggers therapeutic effect Z", the first question that comes to mind is "how much X do I need to take to get that much Z?".
If you don't answer that the info is not actionable. That's where the math models (pharmacocinetics and pharmacodynamics) come in, they tell you how much, which allows you to turn info into action.
This post was written while at MATS 9.0 under the mentorship of Richard Ngo. It's only meta-related to my research.
I would like to start by quoting a point Jan Kulveit made about economics culture in a recent post.
Jan notes that the math is "less relevant than it seems" in the process. I resonate with this. The mathematical results are predetermined by the assumptions in the model, which in turn follow the insight born from intuitive reasoning. This begs the question: if the important part of the insight is the intuition, and the math is its deterministic (if laborious) consequence, then what exactly is the point of the math? What insight does it bring? These questions of course apply not only to economics, but to any mathematisation of worldly[1] phenomena.
In this post, I describe some roles math plays in delivering and communicating insight to researchers and the social structures they are embedded in. I'm interested in characterising the legitimate uses of mathematical formalism, but also in commenting on the incentives in science and society that instantiate its misuse.
Calibrating intuitions
The most innocent and consistently solid use of math is in verifying qualitative intuitions. Suppose a researcher has an interesting insight about economics concept A; they suspect that their insight has a non-trivial implication for some other concept B. The process of mathematising A, B, and the effect of A on B is in fact one of calibration. If the intuitive model is well-calibrated, then the researcher's instincts about these ideas and their consequences will check out in the mathematical formalisation. Note that this is a necessary but not a sufficient condition; your math could check out even if your intuitions are bad. However, the math not checking out probably means it could be improved.
In the possible absence of empirical feedback loops, math serves as a surrogate tool for scientific falsification. I have found math useful for this in my own research, though it's worth noting that calibrating with concrete experiments or simulations are alternative/complementary approaches.
Communicating your work
Even if a researcher has conducted clear, intuitive reasoning that they believe is well-calibrated, they might need to communicate their useful insight to other researchers and society at large. For this purpose, math has some clear benefits due to its relative unambiguity. Firstly, it allows for transfer of verifiable claims from a researcher to their field. Secondly, sufficiently concrete mathematical formalisations may be amenable to communication beyond the research bubble. The most memetically successful pieces of math are those that can be implemented by arbitrary workers or even calculators.
Even though math has these advantages, the notion that it is less lossy than the alternatives as a communication tool is rather non-trivial. What it gains in verifiability of its claims, it may lose in the amount of insight communicated. Mathematical formalism can be painstaking to generate from intuitions, and the finished product generally does not lend itself to the recovery of those intuitions. The choice of the extent to which mathematisation dominates scientific communication illustrates a trade-off between different types of loss.
Communication channels can degrade due to perverse social dynamics and incentives, losing their legitimate purpose along the way. This results in what I'll call communication traps for researchers. I'll describe a couple in this post.
The lossiness of memetically fit mathematics
Null Hypothesis Significance Testing (NHST) is (still) the dominant paradigm for statistical inference, especially in the social and medical sciences. This was not a foregone conclusion, nor is it an adequate state of affairs. Ronald Fisher, one of the early advocates for null hypothesis testing[2], was originally a bayesian[3]. When he settled on frequentist inference, he disagreed on matters of philosophy with his colleagues, Neyman and Pearson, who instead proposed a decision rule to choose between two competing hypotheses.
Fisher, Neyman and Pearson all attempted to develop paradigms for statistical inference with the aim of adoption by scientific researchers. Unfortunately, their wishes were granted in the worst of ways. NHST conflated[4] their approaches into a version of null hypothesis testing which uses a decision-rule based on the p-value, something that none of them seems to have ever advocated for. The sneerily titled "The Null Ritual" documents the emergence of this testing paradigm in psychology over the 1940s and 1950s. The authors additionally present some of their anthropological work assessing the (blood-curdling) illiteracy among psychology students and educators about the philosophy and insight that statisticians originally meant to communicate.
"The Null Ritual" also discusses various possible reasons for the emergence and maintenance of this paradigm. Without having delved into the extensive sociological literature on the topic, my instinctive guess is that NHST "won" because it is simple, easy to execute (including computationally), and easy to misinterpret as indicating a decisive conclusion. I find it particularly suggestive that Fisher recommended the p-value as a useful metric that could inform further experiments, but NHST turned it into a goal for researchers to p-hack. The main appeal of NHST was quite possibly that it provides the illusion of a computable binary decision rule authorised by a sense of mathematical legitimacy. By now, NHST probably wins by default due to its widespread use in education.
This example illustrates that our mathematisations are not nearly as unambiguous as we would hope. The communication of a mathematical formalism and any decision procedures it suggests should aim to be robust to perturbations caused by miscommunication and the incentive structures of practitioners. Science is probably better off with NHST than if we had just "vibed things out" or used whatever statistical inference we did before, but a more successful dissemination of statisticians' insight might have softened a replication crisis or two along the way.
"Proof" of work
I already discussed that mathematical formalism trades off between different types of information loss. However, the verifiability of mathematical (or statistical) claims makes math a justifiably popular means for communication between researchers. Unfortunately, research communities tend to converge to turning metrics of published research output into goals that determine status among its members. Consequently, researchers are subject to incentives to prove the value of their work by exporting it in mathematised terms.
I think this systematically results in premature mathematisation. Consider a researcher, Alice, seeking to formalise an insight she has, possibly for entirely noble reasons to start with. She judges that a formalisation that respects the spirit of the original insight may take years. Her research will additionally involve making appropriately simplified or related formalisms that may take a year or two, but will help slowly build up the full theory. She has also thought of some simplified potential formalisms that are likely tractable in the scope of a few months, but aren't ultimately promising for faithfully communicating the insight.
However, Alice's project's funding is up for renewal this year. This presents her with two options. On the one hand, she could focus on the formalisms she thinks will ultimately be fruitful for the mature theory, and risk having no legible output to show her progress for a project proposal. On the other hand, she could work on the tractable approaches that her research taste repelled her from, but will give her something to show to reviewers.
Sadly, I suspect researchers are regularly faced with such dilemmas. They'll often choose against their research tastes, and can justify their choice (with varying degrees of accuracy) as enabling them to continue their research at all.
In my day-to-day research, I have become increasingly aware of how I intuitively gravitate towards lines of inquiry I expect to be mathematically tractable within the scope of the program I'm currently in. I don't think this is directly due to a specific incentive imposed on me by anyone or anything in particular; I'm lucky to have amply sufficient levels of intellectual freedom. I rather perceive my affect as an artefact of the culture I grew up in, which taught me to value mathematical formalism for its own sake. I aspire to one day be a person free from the compulsion to show their work.
I can't use "real" since the mathematicians reserved the term
This is not exactly the same as NHST, both in terms of the philosophical motivation and the practical implementation
See this commentary of Fisher's early work for details. Fisher had disavowed some of his early work's reliance on "inverse probability", bayesian probability's old name, by 1922.
Wikipedia politely says the approaches were "combined"