There is something it feels like to make a choice. As I decide how to open this essay, I have the familiar sense that I could express these ideas in many ways. I weigh different options, imagine how each might land, and select one. This process of deliberation is what most people call "free will", and it feels undeniably real.
Yet some argue it’s an illusion. One prominent opponent of the concept of free will is the author, podcaster, and philosopher Sam Harris. He has written a book on free will, spoken about it in countlesspublicappearances, and devoted many podcast episodestoit. He has also engaged with defenders of free will, such as a lengthy back-and-forth and podcast interview with the philosopher Dan Dennett.
This essay is my attempt to convince Sam[1]of free will in the compatibilist sense, the view that free will and determinism are compatible. Compatibilists like me hold that we can live in a deterministic universe, fully governed by the laws of physics, and still have a meaningful notion of free will.
In what follows, I'll argue that this kind of free will is real: that deliberation is part of the causal pathway that produces action, not a post-hoc story we tell ourselves. Consciousness isn't merely witnessing decisions made elsewhere, but is instead an active participant in the process. And while none of us chose the raw materials we started with, we can still become genuine agents: selves that reflect on their own values, reshape them over time, and act from reasons they endorse. My aim is to explore where Sam and I disagree and to offer an account of free will that is both scientifically grounded and faithful to what people ordinarily mean by the term.
A Pledge Before We Start
Before we get too deep, I want to take a pledge that I think everyone debating free will should take:
I acknowledge that I am entering a discussion of “free will” and I solemnly swear to do my best to ensure we do not talk past each other. In pursuit of that, I will not implicitly change the definition of “free will”. If I dispute a definition, I will own it and explicitly say, “I hereby dispute the definition”.
I say this, in part, to acknowledge that some of the difference is down to semantics, but also that there’s much more than that to explore. I’ll aim to be clear about when we are and are not arguing over definitions. In defining “free will”, I’ll start with the intuitive sense in which most people use the term and I’ll sharpen it later.
While we’re on definitions, we should also distinguish between two senses of “could”. Here are the two definitions:
We’ll use Could₁ to mean “could have done otherwise if my reasons or circumstances were different”.
We’ll use Could₂ to mean “could have done otherwise even if we rewound the universe to the exact same state—same atoms, same brain state, same everything—and replayed it”.
Here's an example of each case:
Sam often uses the example of choosing between coffee and tea. Let’s say you chose coffee this morning. If your doctor had told you that you need to cut out coffee, you could have chosen tea instead. That’s Could₁. If we rewound the universe back to how it was when you made your decision and replayed the tape, you could not have chosen otherwise, no matter how many times you tried. That would be Could₂. So in this case, you Could₁ but not Could₂ have chosen tea.
Imagine instead that your choices are at least partially determined by quantum noise, and that this is a fundamentally random process. If you rewound the universe and replayed it, you really might have made a different choice. That's Could₂. But notice: if quantum noise determined your choice yet no amount of reasoning could have changed it, you'd have Could₂ without Could₁—you could have done otherwise, but only by luck, not by thinking. That would be a strange notion of freedom.
Compatibilist “free will”, which is what I’m arguing for, is about Could₁, not Could₂.
Sam’s Position
Areas of Agreement
Let me start with a list of things Sam and I agree on. I know not everyone will agree on these points, but Sam and I do, so, fair warning, some of these I’m not going to discuss in detail. I’ve used direct quotes from Sam when possible. In other cases I’ve used my wording but I believe Sam would agree with it:
No libertarian free will: Neither of us believes in libertarian[2]free will, which is the idea that a person could (Could₂) have acted differently with all physical facts held constant.
Randomness doesn’t help: The presence of randomness doesn’t create free will. If there is also some fundamental randomness in the universe (e.g. from quantum physics), that doesn’t rescue free will because you didn’t choose which random path to go down. That might give you Could₂, but it doesn’t give you Could₁, which, I believe, is what matters for free will.
Souls don’t help: Even if people have souls, this probably doesn’t change anything because you likely didn’t choose your own soul.[3]
Determinism does not mean or imply fatalism: We are both determinists, but not fatalists. It does not follow from “everything is determined” to “nothing you do matters”.
No ultimate authorship: Ultimately, you did not choose to be you. You did not choose your genes, parents, childhood environment, and so on.
Incarceration and contract enforcement still make sense: Nothing Sam or I believe suggests that, as Dan Dennett says about Sam’s position, “not only should the prisons be emptied, but no contract is valid, mortgages should be abolished, and we can never hold anybody to account for anything they do.”
We must decouple two distinct questions of free will: The metaphysical question (does free will exist?) is separate from the sociological question (what happens if people believe it does or doesn't?). Some argue for free will by saying belief in it leads to good outcomes (personal responsibility, motivation), or that disbelief leads to nihilism or fatalism. Sam and I agree these arguments are irrelevant to whether free will actually exists. The truth of a claim is independent of the consequences of believing it.
Sam’s Thought Experiment
Sam argues that we do not have free will.[4]In the podcast, “Final Thoughts on Free Will” (transcript here), he provides an excellent thought experiment explaining his position. I quote sections of it below, but if you’re interested, I recommend listening to it in his voice. (I find it quite soothing.) Click here to jump right to the thought experiment and listen for the next nine and a half minutes. But in case you don’t want to do that, here’s what he says (truncated for brevity):
Think of a movie. It can be one you’ve seen or just one you know the name of; it doesn’t have to be good, it can be bad; whatever comes to mind, doesn’t matter. Pay attention to what this experience is like.
A few films have probably come to mind. Just pick one, and pay attention to what the experience of choosing is like. Now, the first thing to notice is that this is as free a choice as you are ever going to make in your life. You are completely free. You have all the films in the world to choose from, and you can pick any one you want.
[...]
What is it like to choose? What is it like to make this completely free choice?
[...]
Did you see any evidence for free will here? Because if it’s not here, it’s not anywhere. So we better be able to find it here. So, let’s look for it.
[...]
There are many other films whose names are well known to you—many of which you’ve seen but which didn’t occur to you to pick. For instance, you absolutely know that The Wizard of Oz is a film, but you just didn’t think of it.
[…]
Consider the few films that came to mind—in light of all the films that might have come to mind but didn’t—and ask yourself, ‘Were you free to choose that which did not occur to you to choose?’ As a matter of neurophysiology, your The Wizard of Oz circuits were not in play a few moments ago for reasons that you can’t possibly know and could not control. Based on the state of your brain, The Wizard of Oz was not an option even though you absolutely know about this film. If we could return your brain to the state it was in a moment ago and account for all the noise in the system—adding back any contributions of randomness, whatever they were—you would fail to think of The Wizard of Oz again, and again, and again until the end of time. Where is the freedom in that?
[…]
The thing to notice is that you as the conscious witness of your inner life are not making decisions. All you can do is witness decisions once they’re made.
[...]
I say, ‘Pick a film’, and there’s this moment before anything has changed for you. And, then the names of films begin percolating at the margins of consciousness, and you have no control over which appear. None. Really, none. Can you feel that? You can’t pick them before they pick themselves.
[…]
If you pay attention to how your thoughts arise and how decisions actually get made, you’ll see that there’s no evidence for free will.
Free Will As a Deliberative Algorithm
I wanted to see if I could write down my process of making a decision to see if I could “find the free will” in it. I wrote down the following algorithm. Note that it is not in any way The General Algorithm for Free WillTM, but merely the process I noticed myself following for this specific task. Here’s what it felt like to me:[5]
Set a goal
In this case, the goal is just “name a movie”.
Decide on a course of actions to reach the goal
I realize I’ll need to remember some movies and select one. The selection criteria don’t matter that much.
Generate options
To generate options, I simply instruct my memory to recall movies. I can also add extra instructions in my internal dialog to see if that triggers anything: “What about Halloween movies, aren’t there more of those? Oh, yeah, that reminds me, what about more Winona Ryder movies? I must know some more of those.”
Receive response
The names of movies just pop into my head. More precisely, I should say they “become available to my consciousness” or “my consciousness becomes aware of them”.
Simulate and evaluate each option
I hold candidates in working memory and simulate saying them. I reason about each option (will this achieve my goal? What are the pros/cons?) Then I evaluate each option and each returns a response like “yes, I can say this” / “no, this doesn’t achieve the goal” (maybe it’s a book and not actually a movie). It also returns some sense of how much I “like” the answer based on my utility function[6]. This is the thing that makes "Edward Scissorhands" feel like a better answer than "Transformers 4," even though both are valid movies. Maybe I want to seem interesting, or I genuinely loved that film, or I have a thing for Winona Ryder. Whatever the reason, I get an additional response of "yes, that's a good answer" or "eh, I can do better."
I can reflect further on my choice. I hear Regis’ voice asking, “Is that your final answer?” Eventually, I tell myself that I am satisfied with my answer, and commit to it.
Say my answer
I say it out loud (if I’m with others) or just say it to myself. Either way, I feel like I have made the decision.
Reflect on my choice
I reflect on my decision. I feel ownership of my actions. I feel proud or embarrassed by my answer (“Did I really say that movie? In front of these people? Was that the best I could do?”).
So, where does this algorithm leave me? It leaves me with a vivid sense that “I chose X, but I could have chosen Y”. I can recall simulating the possibilities, and feel like I could have selected any of them (assuming they were all valid movies). In this case, when I say “could”, I’m using Could₁: I could (Could₁) have selected differently, had my reasons or preferences been different. It’s this sense of having the ability to act otherwise that makes me feel like I have free will, and it falls directly out of this algorithm.
This was simply the algorithm for selecting a movie, but this general structure can be expanded for more complex situations. The goal doesn’t have to be a response or some immediate need, but can include higher-order goals like maintaining a diet, self-improvement, or keeping promises. The evaluation phase would be significantly more elaborate for more complex tasks, such as thinking about constraints, effects on other people, whether there’s missing information, and so on. Even committing to a decision might require more steps. I might ask myself, “Was this just an impulse? Do I really want to do this?” And, importantly, I can evaluate the algorithm itself: “Do I need to change a step, or add a new step somewhere?”
In short, I’m saying free will is this control process, implemented in a physical brain, that integrates goals, reasons, desires, and so on. Some steps are conscious, some aren't. What matters is that the system is actively working through reasons for action, not passively witnessing a foregone conclusion. (Perhaps there is already a difference in definition from Sam’s, but I want to put that aside for another moment to fully explain how I think about it, then we’ll get to semantics.)
So when someone asks, "Did you have free will in situation X?" translate it to: "Did your algorithm run?"
Constraints and Influences
Let me be clear about what I'm not claiming. My compatibilist free will doesn't require:
Freedom from constraint. Sam points out that saying “Wizard of Oz” was not an option if I didn’t think of it at the time, even if I know about the film. This is true. But free will doesn’t mean you can select any movie, or any movie you’ve seen, or even any movie you’ve seen that you could remember if you thought longer. It just means that the algorithm ran. You had the free will to decide how much thought to put into this task, you had the free will to decide you had thought of enough options, and you had the free will to select one.
Consider a more extreme case: someone puts a gun to your head and demands your wallet. Do you have any free will in this situation? Your options are severely constrained—you could fight back, but I wouldn’t recommend it. However, you can still run the algorithm, so you have some diminished, yet non-zero amount of free will in this case. For legal and moral reasons, it would likely not be enough to be considered responsible for your actions (depending on the specific details, as this is a question of degree).
In these scenarios, you have constrained choices. Constraints come in many forms: physical laws (you can’t choose to fly), your subconscious (Wizard of Oz just didn’t come to mind), other people (the gunman), time, resources, and so on. None of these eliminates free will, because free will isn't about having unlimited options; it's about running the deliberative algorithm with whatever options you do have.
Freedom from influence. Sam gives many examples of how our decisions are shaped by things we're unaware of, such as priming effects, childhood memories, and neurotransmitter levels. That's fine. Free will is running the algorithm, not being immune to influence. Your algorithm incorporates these influences. It isn’t supposed to ignore them.
Perfect introspection. You don't need complete understanding as to why certain movies popped into your head or why you weighed one option over another.
We have some level of introspection into what goes on inside our brains, though it’s certainly not perfect, or maybe even very good. We confabulate more than we'd like to admit and spend a lot of time rationalizing after the fact. But the question isn't whether you can accurately report your reasoning; it's whether reasoning occurred. The algorithm works even when you can't fully explain your own preferences.
Complete unpredictability. Free will doesn’t require unpredictability. If I offer you a choice between chocolate ice cream and a poke in the eye with a sharp stick, you'll pick the ice cream every time. That predictability doesn’t mean you lack free will; it just means the algorithm reached an obvious conclusion. The question isn’t about whether the results were predictable, but whether the deliberative control process served as a guide versus being bypassed.
I think these distinctions resolve many of the issues Sam brings up. To hear them, you can listen to the thought experiment 42 minutes into the podcast episode Making Sense of Free Will. If you have these clarifications in mind, you'll find that his objections don't threaten compatibilist free will after all. See “Responding to Another Sam Harris Thought Experiment” in the appendix for my walkthrough of that thought experiment.
Objections, Your Honor
Let's address some likely objections to this algorithmic account of free will.
Exhibit A: Who Is This “I” Guy?
Much of this might sound circular—who is the "I" running the algorithm? The answer is that there's no separate “I”. When I say “I instruct my memory to recall movies,” I mean that one part of my neural circuitry (the part involved in conscious intention) triggers another part (the part responsible for memory retrieval). There's no homunculus, no little person inside doing the real deciding. The algorithm is me.
This is why I resist Sam's framing. Sam says my Wizard of Oz circuits weren't active “for reasons I can't possibly know and could not control.” But those reasons are neurological—they're part of me. When he says "your brain does something," he treats this as evidence that you didn't do it, as if you were separate from your brain, watching helplessly from the sidelines. But my brain doing it is me doing it. The deliberative algorithm running in my neurons is my free will. Or, to quote Eliezer Yudkowsky, thou art physics.
The algorithm involves both conscious and subconscious processes. Some steps happen outside awareness—like which movies pop into my head. But consciousness isn't merely observing the process; it's participating in it: setting goals, deciding on a course of action, evaluating options, vetoing bad ideas. I'm not positing a ghost in the machine. I'm saying the machine includes a component that does what we call "deliberation," and that component is part of the integrated system that is me.
Exhibit B: So, it’s an illusion?
Someone might say, “Ok, you’ve shown how the feeling of free will falls out of a deterministic process. So you’ve shown it’s an illusion, right?”
No! The deliberative algorithm is not just a post-hoc narrative layered on top of decisions made elsewhere; it is the causal process that produces the decision. The subjective feeling of choosing corresponds to the real computational work that the system performs.
If conscious deliberation were merely a spectator narration, then changing what I consciously attend to and consider would not change what I do. But it does. If you provide new reasons for my conscious deliberation—“don’t choose My Little Pony or we’ll all laugh at you”—I might come up with a different result.[8]
It’s certainly possible to fool oneself into thinking you had more control than you actually did. I’ve already admitted that I don’t have full introspective access to why my mind does exactly what it does. But if this is an illusion, it would require that something other than the deliberative algorithm determines the choice, while consciousness merely rationalizes afterward. This is not so; the algorithm is the cause. Conscious evaluation, memory retrieval, and reasoning are not epiphenomenal but instead are the steps by which the decision is made.
Exhibit C: Did you choose your preferences?
Did I choose my preferences? Mostly no, but they are still my preferences. I’ll explore this more later, but, for now, I’m happy to concede that I mostly didn't choose my taste in music, books, movies, or anything else. They were shaped by my genes, hormones, experiences, and countless other factors, none of which I selected from some prior vantage point. Puberty rewired my preferences without asking permission.
But this doesn't threaten free will as I've defined it (we’ll get to semantics later, I promise). The algorithm takes preferences as inputs and works with them. It doesn't require that you author those inputs from scratch.
The objection against identifying with my own preferences amounts to saying, “You didn't choose to be you, therefore you have no free will.” But this sets an impossible standard. To choose your own preferences, you'd need some prior set of preferences to guide the selection, and then you'd need to have chosen those, and so on, forever. The demand is incoherent. What remains is the thing people actually care about: that your choices flow from your values, through your reasoning, to your actions. That's free will. You can't choose to be someone else, but you can choose what to do as the person you are.
Exhibit D: What about those Libet Experiments?
What about those neuroscience experiments that seem to show decisions being made before conscious awareness? Don't these prove consciousness is just a passive witness?
The classic evidence here comes from Libet-style experiments (meta-analysis here), where brain activity (the “readiness potential”) appears before participants report awareness of their intention to move.[9]These findings are interesting, but they don't show that the entire deliberative algorithm I’ve described is epiphenomenal. When researchers detect early neural activity preceding simple motor decisions, they're detecting initial neural commitments in a task with no real stakes and no reasoning required. This doesn’t bypass conscious evaluation, simply because there's barely any evaluation to bypass.
In Sam’s movie example, the early “popping into consciousness” happens subconsciously, and I grant that. But the conscious evaluation, simulation, and selection that follows is still doing real computational work. The Libet experiments show consciousness isn't the first step, but they don't show it's causally inert. To establish that, we would need to see complex decisions where people weigh evidence, consider consequences, and change their minds, being fully determined before any conscious evaluation occurs.[10]
There are also more dramatic demonstrations, like experiments where transcranial magnetic stimulation (TMS) activates the motor cortex opposite to the one a participant intended to use, forcing the “wrong” hand to move. When asked why they moved that hand, participants say things like “I just changed my mind.” I’ve actually talked about these studies before. I agree that they show that consciousness can invent explanations for actions it didn't cause. But confabulation in artificial, forced-movement scenarios doesn't prove that deliberation is always post-hoc rationalization. It proves we can be fooled when experimenters hijack the system.
Exhibit E: Aren’t You Just the Conscious Witness of Your Thoughts?
Sam has repeatedly referred to our conscious experience as a mere witness to our actions. In his book, he said (my bolding):
I generally start each day with a cup of coffee or tea—sometimes two. This morning, it was coffee (two). Why not tea? I am in no position to know. I wanted coffee more than I wanted tea today, and I was free to have what I wanted. Did I consciously choose coffee over tea? No. The choice was made for me by events in my brain that I, as the conscious witness of my thoughts and actions, could not inspect or influence. Could I have “changed my mind” and switched to tea before the coffee drinker in me could get his bearings? Yes, but this impulse would also have been the product of unconscious causes. Why didn’t it arise this morning? Why might it arise in the future? I cannot know. The intention to do one thing and not another does not originate in consciousness—rather, it appears in consciousness, as does any thought or impulse that might oppose it.
[...]
I, as the conscious witness of my experience, no more initiate events in my prefrontal cortex than I cause my heart to beat.
He’s made similar arguments in his podcasts, such as Final Thoughts on Free Will (jump to 1:16:06 and listen for 1.5 minutes). In that episode, he responds to compatibilist philosophy by arguing that what “you” experience as conscious control is just being a conscious witness riding on top of unconscious neural causes, and calling all of that “you” (as compatibilists do) is a “bait-and-switch”. That is, compatibilists start with “you” in the intuitive sense—the conscious self—but then expand it to include all the unconscious processes you never experience or control. By that sleight of hand, Sam argues, compatibilists can say “you” chose freely, but only because they've redefined “you” to mean something the ordinary person wouldn't recognize. He concludes by saying, “The you that you take yourself to be isn’t in control of anything.”
I think this is a key crux of our disagreement. Sam sees consciousness as a mostly passive observer[11]. I think it’s an active participant, a working component of the deliberative algorithm. Contrary to his claim, I think it can initiate events in your prefrontal cortex AND influence your heartbeat.
Here's a simple demonstration: tell yourself to think about elephants for the next five seconds. Your conscious intention just shaped what happened in your prefrontal cortex. You don’t have complete control—it wouldn’t surprise me if a to-do list or a “did I turn off the stove?” trampled upon your elephantine pondering, but your conscious direction influenced events in your prefrontal cortex.
Of course, Sam would protest that the conscious intention to think about elephants arose from unconscious causes. This is true. But we need to distinguish origination (which I concede is unconscious) from governance. Even if the thought arose from the unconscious, it still went into the algorithm before you decided to act upon it. Therefore, you still had the ability to consciously deliberate, revise it if needed, or simply veto the whole idea.
I think Sam's analogy to heartbeats actually backfires. He means to show that consciousness is as powerless over thought as it is over cardiac rhythm. But notice that you can influence your heartbeat: imagine a frightening scenario vividly enough and your heart rate will increase. You can't stop your heart by willing it, but you can modulate it within a meaningful range.
I think this is a miniaturized version of a larger disagreement. Sam looks to the extremes and says, “You can’t choose what thoughts appear in your mind. You can’t stop your heart. You can’t inspect the rationale for your thoughts and actions. Looks bad for free will.” I look at the proximate areas and say, “You can choose to light up your elephant neural circuitry. You can choose to increase your heart rate. You can inspect the rationale for your thoughts and actions, albeit imperfectly. There’s plenty of free will here.” Your consciousness isn't omnipotent, but it isn't impotent either. It can modulate physiology, focus attention, and do real causal work while operating within constraints.
Sam is generally unimpressed with these sorts of claims. In his book, he quips: “Compatibilism amounts to nothing more than an assertion of the following creed: A puppet is free as long as he loves his strings.” But this gets the distinction backwards. A puppet would be unfree if the strings were pulled by an external controller, bypassing its algorithm. A person is free (in the compatibilist sense) when the “strings” are their own values, reasoning, and planning, and when the algorithm isn't being bypassed but is the thing doing the pulling.
I understand where Sam is coming from. I’ve said before that sometimes our executive function seems more like the brain's press secretary. But notice what a press secretary actually does. A pure figurehead would be someone who learns about decisions only after they're final. A real press secretary sits in on the meetings, shapes messaging strategy, and sometimes pushes back on policy because of how it will play. The question isn't whether consciousness has complete control, but whether it's contributing in the room when decisions get made.
Confabulation research shows that we sometimes invent explanations after the fact. It doesn't show that we always do, or that conscious reasoning never contributes. Again, the test is the counterfactual. You gave me a reason not to choose My Little Pony mid-deliberation, and it changed my decision. This means the conscious reasoning is doing real causal work, not just narration. That's compatible with also sometimes confabulating. We're imperfect reasoners, not mere witnesses.
Pathological Cases
Maybe a way to make the distinction between merely witnessing and being an active participant more clear is to talk about pathological cases. There are conditions where consciousness really does seem to be a mere witness, and, notably, we recognize them as pathologies:
Alien hand syndrome—Here’s how the Cleveland Clinic describes alien hand syndrome: “Alien hand syndrome occurs when your hand or limb (arm) acts independently from other parts of your body. It can feel like your hand has a mind of its own. [...] With this condition, you aren’t in control of what your hand does. Your hand doesn’t respond to your direction and performs involuntary actions or movements.” Here's an example from the Wikipedia page: “For example, one patient was observed putting a cigarette into her mouth with her intact, 'controlled' hand (her right, dominant hand), following which her left hand rose, grasped the cigarette, pulled it out of her mouth, and tossed it away before it could be lit by the right hand. The patient then surmised that 'I guess “he” doesn't want me to smoke that cigarette.'”
Epileptic automatisms—Neuropsychologist Peter Fenwick defined it as follows: “An automatism is an involuntary piece of behaviour over which an individual has no control. The behaviour is usually inappropriate to the circumstances, and may be out of character for the individual. It can be complex, co-ordinated and apparently purposeful and directed, though lacking in judgment. Afterwards the individual may have no recollection or only a partial and confused memory for his actions.”
Tourette syndrome—The paper Tourette Syndrome and Consciousness of Action says this: “Although the wish to move is perceived by the patient as involuntary, the decision to release the tic is often perceived by the patient as a voluntary capitulation to the subjective urge.”
How does any of this make sense if the non-pathological “you” is only a witness to actions? There would be no alien hand syndrome as it would all be alien. There could be no distinction between voluntary and involuntary behavior if it’s all involuntary to our consciousness. To me, these are all cases where consciousness isn’t able to play the active, deliberate role that it usually plays. What are these in Sam’s view?
Proximate vs Ultimate Authorship
A key distinction has been lurking in the background of this discussion, and it's time to make it explicit: the difference between proximate and ultimate authorship of our actions.
Proximate authorship means your deliberative algorithm was the immediate cause of an action. The decision ran through your conscious evaluation process: you weighed your options, considered the consequences, selected a course of action, and, afterwards, felt like you could (Could₁) have selected otherwise. In this sense, you authored the choice.
Ultimate authorship would mean you are the ultimate cause of your actions. This would mean that, somehow, the causal chain traces back to you and stops there.
Sam and I agree that no one has ultimate authorship. The causal chain does not stop with you. You did not choose to be you. Your deliberative algorithm—the very thing I'm calling “free will”—was itself shaped mostly by factors outside your control:
Your genes, which you didn't select
Your childhood environment, which you didn't choose
Your experiences, which are mostly a combination of events outside your control and the above, which you didn’t choose
This could go on and on. The causal chain stretches back through your parents, their parents, the evolution of the human brain, the formation of Earth, up to the Big Bang. As Carl Sagan put it, “to make an apple pie you must first invent the universe.” I have invented no universes; therefore, I have ultimate authorship over no apple pies (though I do have proximate authorship over many delicious ones, just for the record).
How to Make an Agent out of Clay
So how are we, without ultimate authorship, supposed to actually be anything? When does it make sense to think of ourselves as agents, with preferences we endorse, reasons we respond to, and a will of our own? In short, how do we become a “self”?
Earlier I said my preferences were my preferences in some meaningful way, but how can that be if I didn’t choose them? And even if I did choose them, I didn’t choose the process by which I chose them. And if, somehow, I chose that as well, we can just follow the chain back far enough and we'll reach something unauthored. That regress is exactly why ultimate authorship is impossible, and I’ve already conceded it.
But notice what the regress argument assumes: it gives all the credit to ultimate authorship and none to proximate authorship. By that standard, nothing we ordinarily call control or choice would count.
Consider a company as an example. Let’s say I make a bunch of decisions for my company. I say we’re going to build Product A and not Product B, we’re going to market it this way and not that way, and so on. In any common usage of the words, I clearly made those decisions—they were under my control. But did I, by Sam’s ultimate authorship standard? Well, the reason I wanted to build Product A is because I thought it would sell well. And that would generate revenue. And that would make the company more valuable. But, did I make the decision to set the goal of the company to be making money? Well, I wasn’t a founder of the company, so it wasn’t my idea to make a for-profit company in the first place. Therefore, by the standard of ultimate authorship, I had no control and made no decisions! The founders made every one of them when they decided to found a for-profit company. This, of course, is not how we think about decision-making and control.
What matters for agency isn’t whether your starting point was self-created; it’s whether the system can govern itself from the inside, whether it can reflect on its results and revise its own motivations over time.
Humans can evaluate their own evaluations. I can have competing desires and reason through them. I can want a cigarette but also not want to want cigarettes, and that second-order stance can reshape the first over time. That’s the feedback loop inside the decision-making system. The algorithm doesn’t just output actions; it can also adjust the weights it uses to produce future actions.
Here’s a real example from my life: I believe I’ve successfully convinced myself that I like broccoli. Years ago, I made a conscious decision to tell myself I really liked broccoli. I didn't hate it prior, but I wouldn’t have said I particularly enjoyed it. But I decided I'd be better off if I did, so I gathered all my anti-rationalist powers and told myself I enjoyed the taste. I ate it more often, and each time I told myself how much I was enjoying it. Within a couple of years, I realized I wasn't pushing anymore. I just liked broccoli. Frozen broccoli, microwaved with salt, pepper, and a little lemon juice, is now my go-to snack. And it’s delicious.
Now, we don't have ultimate authorship of either the first-order desire (disliking broccoli) or second-order desire (wanting to have a healthier diet), so who cares? But notice what happened here. This wasn't just a parameter being adjusted in some optimization process. It was me deciding what kind of person I wanted to be and reshaping my preferences to match. That’s me authoring at the proximate level and taking ownership of the kind of person I’m becoming. The broccoli preference became mine not because I authored it from scratch, but because I consciously endorsed and cultivated it. It coheres with who I take myself to be.
This matters because I want to show that, over time, humans can become coherent agents. I want to show that humans are a distinct category from just a pile of reflexes or a mere conscious witness to one's actions.
And this is why the regress to ultimate authorship doesn’t touch what matters. If “ownership” required self-creation, then no belief, value, or intention would ever count as yours either, because those, too, trace back to unchosen influences. But that’s not how we actually draw the line. We treat a preference as yours when it is integrated into your identity.
Note what this reveals about entities that can have free will. To reflect on your own desires you have to be able to represent them as your desires. You have to be able to take yourself as an object of evaluation and revise yourself over time. That requires a self-model robust enough to support second‑order evaluation: not just “I want X,” but “do I want to be the kind of person who wants X?”
You can see how the sense of agency develops in humans over time. It wouldn’t make sense to describe an infant as much of an agent. But over time, humans develop a sense of who they are and who they want to be. They can reflect on themselves and change accordingly. The algorithm can, to some degree, rewrite its own code in light of an identity it is actively shaping. This is another sense in which proximate authorship is “enough”. Not only can we run the algorithm, we can modify it.
That capacity for self-editing is a real boundary in nature. It separates agents from mere processes. A muscle spasm can't reflect on itself. A craving can't decide it would rather be a different kind of craving. But I can, and that's the distinction that matters when we ask whether someone acted freely.
Sam’s entire objection seems to boil down to the assumption that control requires ultimate authorship. But this assumption doesn’t hold.
Disputing Definitions
OK, some of this has gotten into semantics, so, in keeping with my pledge: I hereby dispute the definition.
As we’ve seen, when I say “I have free will,” I don’t mean I’m the ultimate, uncaused source of my decisions, untouched by genes, environment, or prior events. I mean I have the capacity to translate reasons, values, and goals into actions in a way that is responsive to evidence. Or, in short, to run the algorithm.
So why call this “free will”?
First, you can see how the feeling of free will falls out of this algorithm. When people say they “could have done otherwise,” they are feeling their choice-making algorithm at work, and, as I’ve shown, that algorithm really is at work. The phenomenon matches the feeling of free will, so I say it’s appropriate to call it that.
Second, I think this definition matches how people talk in everyday life. Consider the following:
“I could have made that shot.”
“You could have studied harder.”
“The defendant could have acted differently.”
“The car could have gone faster.”
“It could have rained.”
In all of these, “could have” means something like, “given the situation, a different outcome was within reach under slightly different conditions.”
For example, consider “I could have made that shot.” If I miss a half-court shot, I might say “I could have made that shot.” By that, I mean that, given my skill, if I tried again under similar conditions, it’s possible I could have made it. Making it is within my ability. If I try a full-court shot and the ball falls 20 feet short, then I probably just couldn’t have made it. I lack the physical capacity.
This is Could₁. It’s about alternative outcomes across nearby scenarios (e.g. I could have made that shot if the wind was a little bit different).
Could₁: Could have done otherwise if my reasons or circumstances were different.
Contrast that with what the sentence would mean if people were using Could₂. The sentence would be, “I could have made that shot even if everything about the past and the laws of nature were the same.” It says that, rewinding every atom in the universe and every law of physics, things could (Could₂) have gone differently. This is a completely different claim and it’s not what people mean when they use the word.
Could₂: Could have done otherwise even if we rewound the universe to the exact same state—same atoms, same brain state, same everything—and replayed it.
This is not some complex claim that relies on consciousness. I’m talking about basic standard usage of the word “could”. Here’s another example: Which do people mean by “The car could go faster”? Do they mean:
The car could (Could₁) go faster had I pressed harder on the accelerator? Or,
The car could (Could₂) go faster even if the accelerator remained exactly how it was?
Could₁ is simply the standard usage of the term. In addition, it’s how it’s used in ordinary moral or legal discussions.
Take it or Leave it?
The term “free will” is what computer scientist Marvin Minsky would call a “suitcase phrase”—people pack different meanings into it and call it the same thing. There are some definitions of free will that Sam and I would both jump up and down and say, “No! That does not happen.” Mainly, the notion that if we were to reset the universe’s clock back 30 seconds and put every atom back in its place, someone could (Could₂) choose to act differently. But there are also some definitions of “free will”, like the feeling of weighing your options, reasoning your way to a conclusion, and acting based on that reasoning, where we should jump up and down and say, “Yes! That’s a real thing.”
Sam looks at the range of meanings people attach to "free will," sees the metaphysical baggage, and concludes we're better off abandoning the term. I look at the same thing and see most ordinary usage pointing toward something defensible. When someone says "I chose to stay late at work," they're not claiming to have escaped the causal order of the universe or exercised some quantum soul-power. They're saying the deliberative algorithm ran: they considered leaving, weighed their reasons, and decided to stay. That's Could₁, and it's real.
Sam has an analogy for what he thinks compatibilists are doing. He compares it to claiming that Atlantis is real—it's just the island of Sicily. Sure, it lacks the ancient advanced civilization, didn't sink into the Atlantic, and isn't “greater in extent than Libya and Asia”, but hey, it’s an island! Compatibilists, he suggests, are performing the same sleight of hand: pointing to something real but mislabeling it with a term that implies much more.
Sam's analogy seems to imply that Could₂ is the defining feature of free will, and that I've discarded it while keeping the name. But I think this gets it backwards. As I said, when people say “I could have done otherwise,” they mostly mean Could₁. Admittedly, the free will I'm describing doesn't deliver everything the term has ever been associated with. There’s no ultimate authorship, no metaphysical Could₂. But consider what people actually use free will for. They use it to distinguish choice from compulsion, to ground praise and blame, to make sense of deliberation. Could₁ does all of that and Could₂ does none of it. The features I'm preserving aren't peripheral; they're the load-bearing components. People want Could₁ from their free will and Sam is demanding Could₂.
I don’t understand why he seems to place so much importance in ultimate authorship. He seems to think that without it, “free will” names nothing worth preserving. But ultimate authorship was never part of how we actually explain human behavior. We’re billions of years into a cause-and-effect universe. When we ask "Why did she do that?" we don't expect an answer that traces back to the initial conditions of the universe. We expect proximate causes—reasons, motives, deliberation.
Any time someone asks, “Why?” there is an unbroken chain of answers that could technically answer the question. There’s a sweet spot for good explanations for most questions, and it’s neither the ultimate cause nor the most proximate one, though it’s often much closer to the latter. Consider some examples:
Why did he lose the chess match?
Immediate (and useless) proximate cause: His king was checkmated. (Duh!)
Useful proximate cause: Because he left his rook undefended, lost it, and his position collapsed.
Ultimate cause: The Big Bang
Why did the team lose the football game?
Immediate (and useless) proximate cause: Because the other team scored more points. (Thanks, Dad. Wasn't funny the first ten times.)
Useful proximate cause: They couldn't stop the run.
Ultimate cause: Again, the Big Bang
The same applies to moral explanations. “Why did he betray his friend?” calls for an answer about motives, reasoning, and character, not about the initial conditions of the universe. We explain human action in terms of proximate causes because that's the level at which deliberation, and therefore responsibility, operates. Ultimate authorship was never doing any work in these explanations. Letting it go costs us almost nothing we actually use.
Free will by any other name would smell as sweet
It’s worth stepping back and asking, “How does a philosophical concept like ‘free will’ gain metaphysical legitimacy anyway?” We’re not going to find it like we would a physical object. When I say “free will exists”, I’m not saying we’re going to see it in a brain scan.
This is why I say this isn't just about disputing a definition. I'm making a stronger claim: My point is that any coherent account of agency, responsibility, and reasoning must posit something playing the free-will role. There must be some concept that distinguishes deliberated action from compulsion, reflex, or accident.
Without it, I think you’re forced into some strange positions. In his podcast Final Thoughts on Free Will, he treats someone being convinced by an argument as having the same freedom as being “pushed off a cliff and then claiming that I'm free to fall”. In the podcast Making Sense of Free Will, he makes no distinction between choosing orange juice and having a muscle spasm (see “Responding to Another Sam Harris Thought Experiment” in the appendix).
The distinction between being persuaded and being pushed, between choosing and spasming, isn't some folk illusion we should discard in light of modern science. These are natural categories. It's a distinction that carves reality at its joints. A spasm is an “open-loop” process: a signal fires, the muscle contracts, and no feedback mechanism checks whether this action serves your goals. Choosing juice is a “closed-loop” control system: an option is proposed, simulated against your preferences, evaluated, and executed only if it passes muster. These are fundamentally different mechanisms. One is responsive to reasons; the other isn't. If you told me “the orange juice is poisoned,” I'd choose differently. If you told my leg “don't jerk” while tapping my patellar tendon, it would jerk anyway.
This is what makes the choice mine in a way the spasm isn't. The choice responds to what reasons mean to me. This is the difference between acting and being acted upon. Sure, both events are determined by prior causes, but to not see these as differences in kind seems, frankly, bizarre.
Or consider coercion. When someone holds a gun to your head, we say your choice was constrained. Yes, you “gave” them your wallet, but not freely. What makes this different from an unconstrained choice? It's not that determinism was more true in the coercion case. It's that your algorithm was given artificially narrowed options by an external agent.
Free Will as the Ontological Minimum
When I talk about free will, I'm not positing anything magical or spooky. Free will, as I've described it, exists the way beliefs exist. If you shrunk down in a Magic School Bus you wouldn’t find beliefs stored in labeled containers in the brain. But beliefs are real, right? They're what we call a certain functional capacity of the brain. Free will is similar. It's the name for what's happening when a system weighs reasons, considers alternatives, and selects among them.
This is the minimal ontological commitment required to make sense of how we actually think about people. When we hold someone responsible, when we distinguish choice from compulsion, when we ask “why did you do that?”, we expect a reasons-based answer. Sam can call it something else if he likes. But he needs something to mark these distinctions, or his account of human action becomes incoherent. He simply has a free-will-shaped hole in his ontology.
I'm genuinely curious: from Sam’s perspective, do “beliefs”, “reasons”, “thinking”, and “agents” exist? We distinguish humans from thermostats by saying we respond to reasons while thermostats respond only to temperature. If reasons are real and can be causes of action, why not free will? It's the same kind of thing, a higher-level description of what certain physical systems do, irreducible not because it's made of magic, but because it captures patterns the lower level doesn't.
Why It Matters: Phenomenology, Incentives, Morality, Law
Why does any of this matter? How does my defense of free will cash out in terms of things that we care about? I’ll list five reasons why this matters:
1. Phenomenology. People have a strong intuitive sense of free will. Where does this feeling come from, and does it track something real?
2. Incentives and behavior. Can people respond to rewards, punishments, and social pressure? How does free will relate to deterrence and rehabilitation?
3**. Moral responsibility.** Are people moral agents? Can they be held responsible for their actions?
4. Hatred and retributive punishment. Does anyone deserve to suffer for what they've done?
5. Crime and punishment. How should the legal system treat offenders?
Let me address each in turn.
Phenomenology: The Feeling of Free Will
We have a persistent feeling that we could have done otherwise. Is this feeling tracking something real, or is it an illusion? The answer depends on which “could” we mean. For Could₁, the sense that we would have chosen differently had our reasons, evidence, or preferences been different, yes, that’s completely real. But for Could₂, the sense that we might have chosen differently with every atom in the universe held fixed, no, that's not real.
And, as I’ve argued, Could₁ is what the feeling of free will is actually about. This is what makes the algorithm-based account satisfying: it explains the phenomenology of free will without explaining it away. When you run through options, simulate outcomes, and select among them, you're not passively watching a movie of yourself deciding. You're experiencing your deliberative process at work. The feeling of choosing is the choosing. That's what free will feels like from the inside, and that's what free will is.
Incentives and Behavior
Here, Sam and I agree on the facts. People obviously respond to incentives. Stigmatizing drunk driving works. Offering bonuses improves performance. Punishment can deter crime. We shape behavior through incentives all the time.
I think Sam would argue that this doesn’t mean they have free will, just that their behavior responds to inputs. Fine, you could say that, but if you need a system that responds to reasons, weighs options, and updates based on consequences to explain human behavior, you've just described free will but are refusing to use the term. Incentives work because they feed directly into your deliberative algorithm. They change the weights, alter the utility calculations, and thus change behavior. This is why we can hold people accountable, offer rewards, impose consequences, and expect behavior change.
Moral Agency and Responsibility
I’ve claimed that we have proximate authorship but not ultimate authorship of our actions. Is this “enough” authorship for moral responsibility? I believe so. I believe being a moral agent is being the kind of entity whose decision-making can incorporate moral reasoning. This is a bit beyond the scope here, but the following are the types of things I would expect a moral agent to be able to do:
Represent itself as a persisting individual (a self-model)
Represent other entities as having welfare (i.e., as beings that can be benefited or harmed)
Consider the effects of its actions on others, including tradeoffs between self-interest and moral reasons
Have its own welfare at stake such that some outcomes can be better or worse for it
Be responsive to reasons in the counterfactual sense: new reasons can change what it does (i.e., run the algorithm)
Update its future behavior in light of reasons (e.g. criticism or reflection)
This is why we treat adults differently from infants, and humans differently from bears. It's not that adults have ultimate authorship and infants don't; it's that adults have proximate authorship, and their algorithm can incorporate moral reasoning. A bear that mauls someone isn't a moral agent. It doesn't think, “How would I feel if someone did this to me?”
There are degrees here, of course. A four-year-old has more moral agency than an infant, and less than an adult. Someone with severe cognitive impairment may have diminished moral agency. The question is always: to what extent can this entity's algorithm incorporate moral reasoning?
Moral Desert, Hatred, and Retributive Punishment
In addition to moral responsibility, there's the question of desert, of whether wrongdoers deserve to suffer as retribution for their actions. Here, Sam and I completely agree that they do not. To deserve retribution in that deep sense, someone would need ultimate authorship of their actions.
To see why, consider an example Sam gives: someone commits violence because a brain tumor is pressing on their amygdala. We recognize them as a victim of neurology, not a monster deserving punishment. But now replace the tumor with an abusive childhood, genetic predispositions toward impulsivity, or serotonin imbalances. At each step, we're still describing physical causes the person didn't choose. The distinction between “tumor” and “bad genes” is arbitrary—both are prior causes outside the person's control. It's brain tumors all the way down. There but for the grace of God go I.[12]
Moral desert simply requires a metaphysical freedom that people do not have.
Once you give up ultimate authorship, a certain kind of hatred has to go with it. You can't coherently hate someone as the ultimate author of their evil, as if they, from nothing, simply chose to be bad. That hatred requires the same metaphysical freedom that no one actually has.
Think about a bear that mauls someone. The bear causes harm, and we might kill it for public safety, but we don't hate the bear. It's not the kind of thing that could deserve retribution. The important part is recognizing that, without ultimate authorship, the same logic extends to humans. People who do terrible things are not deserving of suffering for its own sake. On this, Sam has been a tireless voice, and I appreciate his advocacy of this position.
This doesn't eliminate all meanings of “hate” entirely, just a particular kind. You can still hate your job, Mondays, and git merge conflicts. You can definitely still hate dealing with git merge conflicts for your job on Mondays. But notice this is a different kind of hate. There’s no sense in which you want Monday to “pay for what it's done.” It's about anticipating that you’ll have a bad experience with it and seeking to avoid it.
The same applies to people. You can recognize that someone's algorithm doesn't adequately weigh others' suffering, and you can avoid them accordingly. But there’s no need to view your enemies as self-created monsters deserving retributive punishment.
On this point, Sam wins. Perhaps if retributive justice were all I cared about, I would agree with him that we should consider free will an illusion. But free will does more work than that. It's deliberation doing real causal work. It grounds the distinction between choice and compulsion, makes sense of why incentives change behavior, and gives meaning to praise and blame. Retributive punishment is the one piece that genuinely requires ultimate authorship, and it's the one piece I'm happy to let go.
Crime and Punishment
What does this mean for crime and punishment? Does this mean we can't hold anyone responsible? No. Sam and I are aligned here. We can hold people responsible without blaming them for ultimate authorship. We can and should hold people responsible in a forward-looking sense: for deterrence, rehabilitation, and public safety. Courts still need to distinguish intentional action from accident, choices made with a sound mind from those made under coercion or insanity. My account of free will provides exactly that framework: Did the algorithm run normally, or was it bypassed (reflex), broken (insanity), distorted (addiction), or given severely constrained options (coercion)?
Sam and I agree that sometimes we must incarcerate people because they are dangerous to others. But we do so to mitigate harm and deter future crime, not to exact retributive justice upon them.
Final Cruxes
Sam is right that there's no ghost in the machine, (probably[13]) no soul pulling levers from outside the causal chain, no metaphysical Could₂ freedom. We agree more than we disagree. (In fact, Dan Dennett has called Sam “a compatibilist in everything but name!”). However, I wanted to compile what I see as the core cruxes of disagreement into a list. If Sam and I were to sit down and productively hash this out, here's where I think we'd need to focus:
1. Is conscious deliberation causally efficacious, or is it epiphenomenal narration? I say the algorithm is the decision-making process—consciousness is doing real computational work. Sam says consciousness is merely “witnessing” decisions made elsewhere.
2. Is there a meaningful categorical difference between deliberated actions and reflexes? I say yes—one runs through the algorithm, one bypasses it. Sam seems to collapse this distinction since both are “caused”. But if there's no difference between choosing orange juice and having a muscle spasm, something has gone wrong.
3. Is there a meaningful categorical difference between entities that can reflect on and revise their own decision-making versus those that cannot? A thermostat responds to temperature; a human can respond to reasons, evaluate their own preferences, and update their future behavior accordingly. I taught myself to like broccoli. I would like to see a thermostat do that. I can notice a bad habit and work to change it. This capacity for reflective self-modification seems like a real category that separates agents from mere processes. Does Sam recognize this as a meaningful distinction, or is this also collapsed because both are ultimately “caused”?
4. What should we think of pathologies where someone feels like a mere witness to their actions? To me, these seem like cases where the algorithm is damaged and consciousness isn’t able to play its active, deliberate role that it usually plays. I don’t know how Sam would describe these.
5. What lessons should we learn from the Libet-style experiments? Does it show that consciousness is post-hoc rationalization, or merely that consciousness isn't the initiating step while still doing causal work downstream?
6. What should we think about an entity that has proximate authorship but not ultimate authorship (as all of us do)? Is that sufficient for moral responsibility, control, praise, and blame? Sam seems to think that without ultimate authorship, "control" is illusory. I think proximate authorship is sufficient, and that demanding ultimate authorship sets an impossible standard. The implication would be no one has ever controlled anything.
7. What counts as “you”? When Sam says “your brain did it,” he treats this as evidence against free will, almost as if “you” were separate from your brain. I say my brain doing it is me doing it. The deliberative algorithm running in my neurons is my free will. We may simply have different intuitions about where to draw the boundary of the self and whether being moved by your own values counts as freedom or puppetry. Similarly, should you identify with yourself? Should you take credit for the person you've become? Should we make anything of a person’s ability to become a more coherent agent over time versus a pile of unauthored behaviors? I say “yes”.
8. What criteria must a metaphysical concept meet to earn its place? If beliefs, reasons, and agents qualify, what test do these pass that free will uniquely fails? Does Sam reject it simply because of the historical “Could₂ baggage” associated with it? For me, a concept earns its keep by leading to and aligning with other natural categories, and doing without them requires tap-dancing around the concept.
9. What do ordinary people mean by “could have done otherwise”? I claim everyday usage is Could₁: “I would have acted differently if my reasons or circumstances had been different.” Sam seems to think people intuitively mean Could₂: “I could have acted differently with every atom in the universe held fixed.”
10. Is “free will” worth preserving as a concept, or should we retire it? Sam looks at the metaphysical baggage and says we're better off abandoning the term. I look at what people actually use the concept for and say these are the load-bearing features. If we abandon the term, don’t we need something else to replace it? Doesn't any coherent account of agency require something playing the free-will role?
I say let's keep the term. Free will names something real: a process fully physical, fully determined by prior causes, and yet still you doing the choosing. The algorithm isn't an illusion overlaid on "mere" physics. It is the physics, operating at a functional level that matters for morality, law, and human experience.
So, Sam, what would it take to convince you? If the algorithmic account captures what people mean by free will, does the work we need it to do, and doesn't require any spooky metaphysics, what's left to object to besides the name?
Appendix
Responding to Another Sam Harris Thought Experiment
I want to go over another thought experiment that Sam gives to show how all of his objections don't threaten the notion of free will as I’ve described it. This is from the podcast Making Sense of Free Will. The thought experiment starts at 42 minutes in. The narrator makes a point that Sam has made many times, but it’s made clearly here, so I’ll use it. Here’s the setup:
Put yourself in a seat on an airplane. You’re a bit thirsty, and the beverage cart is making its way down the aisle to you. The flight attendant asks you what you’d like to drink. You see the choices on the cart: apple juice, orange juice, soda, water. You ponder things for a moment, make up your mind, and you ask for orange juice. After a few satisfying sips, you go for another and suddenly experience a muscle spasm in your arm. The movement causes some juice to spill on your neighbor’s pant leg.
The narrator (echoing Sam), argues that the selection of orange juice and the spilling of the juice aren't as different as they seem. Yes, the spasm feels like something done to you. But did you really "choose" the orange juice? Did you create your preference for it? The narrator makes the case:
Maybe you had a flash of memory of your grandmother’s home. She had an orange tree in the backyard. Nostalgia is why you chose the orange juice over the apple juice. Subjectively speaking, does this really seem like an example of free will? Even the contents of that story are filled with things you didn’t choose, like your grandparents, where their house was, the fact that they had an orange tree, or the fact that your parents took you there when it was fruiting, and so on. And, in any case, as Sam points out, you can’t account for why this memory occurred to you in the very moment the flight attendant came by. Those neurons happened to be online and ready to fire at that moment. And, apparently, the neurons that could have fired that would have delivered the catchy slogan of your favorite apple juice advertisement and pushed you in that direction, didn’t fire. And, more importantly, you can’t account for why this grandmother story moved you to choose orange juice, rather than, say, be bored by orange juice because you had it so much as a kid.
This might sound compelling until you apply the algorithmic account. Then each objection dissolves:
"The contents of that story are filled with things you didn't choose." Yes! You don't have omnipotent control of the universe. There are constraints on your choice. That's no barrier to free will.
"You can't account for why this memory occurred to you in the very moment the flight attendant came by." True—we don't have full introspection into our minds. But that's OK. That information was fed into the algorithm. Had another memory popped in, it would have been incorporated instead, and might have influenced the outcome. The thoughts that did occur were sufficient to run the algorithm. Free will is sometimes constrained by your subconscious.
"You can't account for why this grandmother story moved you to choose orange juice." I'd dispute this because you do have some degree of introspection, but it doesn't matter either way. You don't need full introspection for free will.
"Why did you choose orange juice?" Because of my utility function, which was used in the evaluation step of my algorithm. The fact that this preference traces back to childhood memories I didn't choose doesn't change the fact that the algorithm ran. I don’t need ultimate authorship over my preferences. From a compatibilist perspective, this is no objection to free will.
I usually refer to people I don't know personally by their last names, but I've been listening to Sam's podcast for over a decade, and calling him “Harris” just feels strange. So I use his first name out of the creepy, one-sided familiarity that comes with being a longtime listener. I mean no disrespect. ↩︎
Libertarian free will has nothing to do with economic libertarianism; it’s just an unfortunate namespace collision. ↩︎
Sam is more confident here. I say “probably” and “likely” because we’re talking about souls, and, if they’re real, we have close to no idea how they work. We’re in speculative territory here, so it’s good to be cautious. ↩︎
In his book Free Will, he says both that “Free will is an illusion” (as well as in this essay) and that there is no illusion of free will because “The illusion of free will is itself an illusion” (also said in this podcast). Parsing this is beyond the scope here. In all cases he’s consistent about arguing that we do not have free will, so that’s why I word it like that. ↩︎
This is a highly simplified version. The real version would have lots of error checking and correction at each layer (just like layers 2 and 4 in the OSI model, if you’re familiar with that). For example, the real first step would really be making sure I understood the question. I’m going to leave these out for simplicity. ↩︎
A utility function determines how much you value different outcomes by weighting your options according to your preferences. ↩︎
Again, this is highly simplified. It’s not necessarily linear like this. If I decide I don’t like any in the evaluation stage, I can just go back to querying my memory. Or, if I realize I don’t just want to name a movie but also name a movie that will show that I’m an interesting guy, I’ll edit the goal to include that. ↩︎
Because the real question is, are you talking about My Little Pony: The Movie or My Little Pony: A New Generation? ↩︎
For a dissenting opinion on readiness potential and whether we’re interpreting it correctly, see “What Is the Readiness Potential?” by Schurger et al. ↩︎
It’s worth noting that Libet was a compatibilist himself. In his paper “Do we have free will?”, he argues that “the conscious function could still control the outcome; it can veto the act. Free will is therefore not excluded.” ↩︎
For example, I quoted above where he says, “you as the conscious witness of your inner life are not making decisions. All you can do is witness decisions once they're made.” However, although Sam often refers to consciousness as a mere witness, he has also said that it does things. In his book Free Will, he says: ↩︎
For a detailed examination of this idea, I recommend the Radiolab episode Blame. ↩︎
Sorry, I just have to put in a “probably” here because it’s a statement about souls, which are quasi-metaphysical, so we really shouldn’t be too certain how they would work. ↩︎
There is something it feels like to make a choice. As I decide how to open this essay, I have the familiar sense that I could express these ideas in many ways. I weigh different options, imagine how each might land, and select one. This process of deliberation is what most people call "free will", and it feels undeniably real.
Yet some argue it’s an illusion. One prominent opponent of the concept of free will is the author, podcaster, and philosopher Sam Harris. He has written a book on free will, spoken about it in countless public appearances, and devoted many podcast episodes to it. He has also engaged with defenders of free will, such as a lengthy back-and-forth and podcast interview with the philosopher Dan Dennett.
This essay is my attempt to convince Sam[1]of free will in the compatibilist sense, the view that free will and determinism are compatible. Compatibilists like me hold that we can live in a deterministic universe, fully governed by the laws of physics, and still have a meaningful notion of free will.
In what follows, I'll argue that this kind of free will is real: that deliberation is part of the causal pathway that produces action, not a post-hoc story we tell ourselves. Consciousness isn't merely witnessing decisions made elsewhere, but is instead an active participant in the process. And while none of us chose the raw materials we started with, we can still become genuine agents: selves that reflect on their own values, reshape them over time, and act from reasons they endorse. My aim is to explore where Sam and I disagree and to offer an account of free will that is both scientifically grounded and faithful to what people ordinarily mean by the term.
A Pledge Before We Start
Before we get too deep, I want to take a pledge that I think everyone debating free will should take:
I say this, in part, to acknowledge that some of the difference is down to semantics, but also that there’s much more than that to explore. I’ll aim to be clear about when we are and are not arguing over definitions. In defining “free will”, I’ll start with the intuitive sense in which most people use the term and I’ll sharpen it later.
While we’re on definitions, we should also distinguish between two senses of “could”. Here are the two definitions:
Here's an example of each case:
Compatibilist “free will”, which is what I’m arguing for, is about Could₁, not Could₂.
Sam’s Position
Areas of Agreement
Let me start with a list of things Sam and I agree on. I know not everyone will agree on these points, but Sam and I do, so, fair warning, some of these I’m not going to discuss in detail. I’ve used direct quotes from Sam when possible. In other cases I’ve used my wording but I believe Sam would agree with it:
Sam’s Thought Experiment
Sam argues that we do not have free will.[4]In the podcast, “Final Thoughts on Free Will” (transcript here), he provides an excellent thought experiment explaining his position. I quote sections of it below, but if you’re interested, I recommend listening to it in his voice. (I find it quite soothing.) Click here to jump right to the thought experiment and listen for the next nine and a half minutes. But in case you don’t want to do that, here’s what he says (truncated for brevity):
Free Will As a Deliberative Algorithm
I wanted to see if I could write down my process of making a decision to see if I could “find the free will” in it. I wrote down the following algorithm. Note that it is not in any way The General Algorithm for Free WillTM, but merely the process I noticed myself following for this specific task. Here’s what it felt like to me:[5]
So, where does this algorithm leave me? It leaves me with a vivid sense that “I chose X, but I could have chosen Y”. I can recall simulating the possibilities, and feel like I could have selected any of them (assuming they were all valid movies). In this case, when I say “could”, I’m using Could₁: I could (Could₁) have selected differently, had my reasons or preferences been different. It’s this sense of having the ability to act otherwise that makes me feel like I have free will, and it falls directly out of this algorithm.
This was simply the algorithm for selecting a movie, but this general structure can be expanded for more complex situations. The goal doesn’t have to be a response or some immediate need, but can include higher-order goals like maintaining a diet, self-improvement, or keeping promises. The evaluation phase would be significantly more elaborate for more complex tasks, such as thinking about constraints, effects on other people, whether there’s missing information, and so on. Even committing to a decision might require more steps. I might ask myself, “Was this just an impulse? Do I really want to do this?” And, importantly, I can evaluate the algorithm itself: “Do I need to change a step, or add a new step somewhere?”
In short, I’m saying free will is this control process, implemented in a physical brain, that integrates goals, reasons, desires, and so on. Some steps are conscious, some aren't. What matters is that the system is actively working through reasons for action, not passively witnessing a foregone conclusion. (Perhaps there is already a difference in definition from Sam’s, but I want to put that aside for another moment to fully explain how I think about it, then we’ll get to semantics.)
So when someone asks, "Did you have free will in situation X?" translate it to: "Did your algorithm run?"
Constraints and Influences
Let me be clear about what I'm not claiming. My compatibilist free will doesn't require:
Freedom from constraint. Sam points out that saying “Wizard of Oz” was not an option if I didn’t think of it at the time, even if I know about the film. This is true. But free will doesn’t mean you can select any movie, or any movie you’ve seen, or even any movie you’ve seen that you could remember if you thought longer. It just means that the algorithm ran. You had the free will to decide how much thought to put into this task, you had the free will to decide you had thought of enough options, and you had the free will to select one.
Consider a more extreme case: someone puts a gun to your head and demands your wallet. Do you have any free will in this situation? Your options are severely constrained—you could fight back, but I wouldn’t recommend it. However, you can still run the algorithm, so you have some diminished, yet non-zero amount of free will in this case. For legal and moral reasons, it would likely not be enough to be considered responsible for your actions (depending on the specific details, as this is a question of degree).
In these scenarios, you have constrained choices. Constraints come in many forms: physical laws (you can’t choose to fly), your subconscious (Wizard of Oz just didn’t come to mind), other people (the gunman), time, resources, and so on. None of these eliminates free will, because free will isn't about having unlimited options; it's about running the deliberative algorithm with whatever options you do have.
Freedom from influence. Sam gives many examples of how our decisions are shaped by things we're unaware of, such as priming effects, childhood memories, and neurotransmitter levels. That's fine. Free will is running the algorithm, not being immune to influence. Your algorithm incorporates these influences. It isn’t supposed to ignore them.
Perfect introspection. You don't need complete understanding as to why certain movies popped into your head or why you weighed one option over another.
We have some level of introspection into what goes on inside our brains, though it’s certainly not perfect, or maybe even very good. We confabulate more than we'd like to admit and spend a lot of time rationalizing after the fact. But the question isn't whether you can accurately report your reasoning; it's whether reasoning occurred. The algorithm works even when you can't fully explain your own preferences.
Complete unpredictability. Free will doesn’t require unpredictability. If I offer you a choice between chocolate ice cream and a poke in the eye with a sharp stick, you'll pick the ice cream every time. That predictability doesn’t mean you lack free will; it just means the algorithm reached an obvious conclusion. The question isn’t about whether the results were predictable, but whether the deliberative control process served as a guide versus being bypassed.
I think these distinctions resolve many of the issues Sam brings up. To hear them, you can listen to the thought experiment 42 minutes into the podcast episode Making Sense of Free Will. If you have these clarifications in mind, you'll find that his objections don't threaten compatibilist free will after all. See “Responding to Another Sam Harris Thought Experiment” in the appendix for my walkthrough of that thought experiment.
Objections, Your Honor
Let's address some likely objections to this algorithmic account of free will.
Exhibit A: Who Is This “I” Guy?
Much of this might sound circular—who is the "I" running the algorithm? The answer is that there's no separate “I”. When I say “I instruct my memory to recall movies,” I mean that one part of my neural circuitry (the part involved in conscious intention) triggers another part (the part responsible for memory retrieval). There's no homunculus, no little person inside doing the real deciding. The algorithm is me.
This is why I resist Sam's framing. Sam says my Wizard of Oz circuits weren't active “for reasons I can't possibly know and could not control.” But those reasons are neurological—they're part of me. When he says "your brain does something," he treats this as evidence that you didn't do it, as if you were separate from your brain, watching helplessly from the sidelines. But my brain doing it is me doing it. The deliberative algorithm running in my neurons is my free will. Or, to quote Eliezer Yudkowsky, thou art physics.
The algorithm involves both conscious and subconscious processes. Some steps happen outside awareness—like which movies pop into my head. But consciousness isn't merely observing the process; it's participating in it: setting goals, deciding on a course of action, evaluating options, vetoing bad ideas. I'm not positing a ghost in the machine. I'm saying the machine includes a component that does what we call "deliberation," and that component is part of the integrated system that is me.
Exhibit B: So, it’s an illusion?
Someone might say, “Ok, you’ve shown how the feeling of free will falls out of a deterministic process. So you’ve shown it’s an illusion, right?”
No! The deliberative algorithm is not just a post-hoc narrative layered on top of decisions made elsewhere; it is the causal process that produces the decision. The subjective feeling of choosing corresponds to the real computational work that the system performs.
If conscious deliberation were merely a spectator narration, then changing what I consciously attend to and consider would not change what I do. But it does. If you provide new reasons for my conscious deliberation—“don’t choose My Little Pony or we’ll all laugh at you”—I might come up with a different result.[8]
It’s certainly possible to fool oneself into thinking you had more control than you actually did. I’ve already admitted that I don’t have full introspective access to why my mind does exactly what it does. But if this is an illusion, it would require that something other than the deliberative algorithm determines the choice, while consciousness merely rationalizes afterward. This is not so; the algorithm is the cause. Conscious evaluation, memory retrieval, and reasoning are not epiphenomenal but instead are the steps by which the decision is made.
Exhibit C: Did you choose your preferences?
Did I choose my preferences? Mostly no, but they are still my preferences. I’ll explore this more later, but, for now, I’m happy to concede that I mostly didn't choose my taste in music, books, movies, or anything else. They were shaped by my genes, hormones, experiences, and countless other factors, none of which I selected from some prior vantage point. Puberty rewired my preferences without asking permission.
But this doesn't threaten free will as I've defined it (we’ll get to semantics later, I promise). The algorithm takes preferences as inputs and works with them. It doesn't require that you author those inputs from scratch.
The objection against identifying with my own preferences amounts to saying, “You didn't choose to be you, therefore you have no free will.” But this sets an impossible standard. To choose your own preferences, you'd need some prior set of preferences to guide the selection, and then you'd need to have chosen those, and so on, forever. The demand is incoherent. What remains is the thing people actually care about: that your choices flow from your values, through your reasoning, to your actions. That's free will. You can't choose to be someone else, but you can choose what to do as the person you are.
Exhibit D: What about those Libet Experiments?
What about those neuroscience experiments that seem to show decisions being made before conscious awareness? Don't these prove consciousness is just a passive witness?
The classic evidence here comes from Libet-style experiments (meta-analysis here), where brain activity (the “readiness potential”) appears before participants report awareness of their intention to move.[9]These findings are interesting, but they don't show that the entire deliberative algorithm I’ve described is epiphenomenal. When researchers detect early neural activity preceding simple motor decisions, they're detecting initial neural commitments in a task with no real stakes and no reasoning required. This doesn’t bypass conscious evaluation, simply because there's barely any evaluation to bypass.
In Sam’s movie example, the early “popping into consciousness” happens subconsciously, and I grant that. But the conscious evaluation, simulation, and selection that follows is still doing real computational work. The Libet experiments show consciousness isn't the first step, but they don't show it's causally inert. To establish that, we would need to see complex decisions where people weigh evidence, consider consequences, and change their minds, being fully determined before any conscious evaluation occurs.[10]
There are also more dramatic demonstrations, like experiments where transcranial magnetic stimulation (TMS) activates the motor cortex opposite to the one a participant intended to use, forcing the “wrong” hand to move. When asked why they moved that hand, participants say things like “I just changed my mind.” I’ve actually talked about these studies before. I agree that they show that consciousness can invent explanations for actions it didn't cause. But confabulation in artificial, forced-movement scenarios doesn't prove that deliberation is always post-hoc rationalization. It proves we can be fooled when experimenters hijack the system.
Exhibit E: Aren’t You Just the Conscious Witness of Your Thoughts?
Sam has repeatedly referred to our conscious experience as a mere witness to our actions. In his book, he said (my bolding):
He’s made similar arguments in his podcasts, such as Final Thoughts on Free Will (jump to 1:16:06 and listen for 1.5 minutes). In that episode, he responds to compatibilist philosophy by arguing that what “you” experience as conscious control is just being a conscious witness riding on top of unconscious neural causes, and calling all of that “you” (as compatibilists do) is a “bait-and-switch”. That is, compatibilists start with “you” in the intuitive sense—the conscious self—but then expand it to include all the unconscious processes you never experience or control. By that sleight of hand, Sam argues, compatibilists can say “you” chose freely, but only because they've redefined “you” to mean something the ordinary person wouldn't recognize. He concludes by saying, “The you that you take yourself to be isn’t in control of anything.”
I think this is a key crux of our disagreement. Sam sees consciousness as a mostly passive observer[11]. I think it’s an active participant, a working component of the deliberative algorithm. Contrary to his claim, I think it can initiate events in your prefrontal cortex AND influence your heartbeat.
Here's a simple demonstration: tell yourself to think about elephants for the next five seconds. Your conscious intention just shaped what happened in your prefrontal cortex. You don’t have complete control—it wouldn’t surprise me if a to-do list or a “did I turn off the stove?” trampled upon your elephantine pondering, but your conscious direction influenced events in your prefrontal cortex.
Of course, Sam would protest that the conscious intention to think about elephants arose from unconscious causes. This is true. But we need to distinguish origination (which I concede is unconscious) from governance. Even if the thought arose from the unconscious, it still went into the algorithm before you decided to act upon it. Therefore, you still had the ability to consciously deliberate, revise it if needed, or simply veto the whole idea.
I think Sam's analogy to heartbeats actually backfires. He means to show that consciousness is as powerless over thought as it is over cardiac rhythm. But notice that you can influence your heartbeat: imagine a frightening scenario vividly enough and your heart rate will increase. You can't stop your heart by willing it, but you can modulate it within a meaningful range.
I think this is a miniaturized version of a larger disagreement. Sam looks to the extremes and says, “You can’t choose what thoughts appear in your mind. You can’t stop your heart. You can’t inspect the rationale for your thoughts and actions. Looks bad for free will.” I look at the proximate areas and say, “You can choose to light up your elephant neural circuitry. You can choose to increase your heart rate. You can inspect the rationale for your thoughts and actions, albeit imperfectly. There’s plenty of free will here.” Your consciousness isn't omnipotent, but it isn't impotent either. It can modulate physiology, focus attention, and do real causal work while operating within constraints.
Sam is generally unimpressed with these sorts of claims. In his book, he quips: “Compatibilism amounts to nothing more than an assertion of the following creed: A puppet is free as long as he loves his strings.” But this gets the distinction backwards. A puppet would be unfree if the strings were pulled by an external controller, bypassing its algorithm. A person is free (in the compatibilist sense) when the “strings” are their own values, reasoning, and planning, and when the algorithm isn't being bypassed but is the thing doing the pulling.
I understand where Sam is coming from. I’ve said before that sometimes our executive function seems more like the brain's press secretary. But notice what a press secretary actually does. A pure figurehead would be someone who learns about decisions only after they're final. A real press secretary sits in on the meetings, shapes messaging strategy, and sometimes pushes back on policy because of how it will play. The question isn't whether consciousness has complete control, but whether it's contributing in the room when decisions get made.
Confabulation research shows that we sometimes invent explanations after the fact. It doesn't show that we always do, or that conscious reasoning never contributes. Again, the test is the counterfactual. You gave me a reason not to choose My Little Pony mid-deliberation, and it changed my decision. This means the conscious reasoning is doing real causal work, not just narration. That's compatible with also sometimes confabulating. We're imperfect reasoners, not mere witnesses.
Pathological Cases
Maybe a way to make the distinction between merely witnessing and being an active participant more clear is to talk about pathological cases. There are conditions where consciousness really does seem to be a mere witness, and, notably, we recognize them as pathologies:
Here's an example from the Wikipedia page: “For example, one patient was observed putting a cigarette into her mouth with her intact, 'controlled' hand (her right, dominant hand), following which her left hand rose, grasped the cigarette, pulled it out of her mouth, and tossed it away before it could be lit by the right hand. The patient then surmised that 'I guess “he” doesn't want me to smoke that cigarette.'”
How does any of this make sense if the non-pathological “you” is only a witness to actions? There would be no alien hand syndrome as it would all be alien. There could be no distinction between voluntary and involuntary behavior if it’s all involuntary to our consciousness. To me, these are all cases where consciousness isn’t able to play the active, deliberate role that it usually plays. What are these in Sam’s view?
Proximate vs Ultimate Authorship
A key distinction has been lurking in the background of this discussion, and it's time to make it explicit: the difference between proximate and ultimate authorship of our actions.
Proximate authorship means your deliberative algorithm was the immediate cause of an action. The decision ran through your conscious evaluation process: you weighed your options, considered the consequences, selected a course of action, and, afterwards, felt like you could (Could₁) have selected otherwise. In this sense, you authored the choice.
Ultimate authorship would mean you are the ultimate cause of your actions. This would mean that, somehow, the causal chain traces back to you and stops there.
Sam and I agree that no one has ultimate authorship. The causal chain does not stop with you. You did not choose to be you. Your deliberative algorithm—the very thing I'm calling “free will”—was itself shaped mostly by factors outside your control:
This could go on and on. The causal chain stretches back through your parents, their parents, the evolution of the human brain, the formation of Earth, up to the Big Bang. As Carl Sagan put it, “to make an apple pie you must first invent the universe.” I have invented no universes; therefore, I have ultimate authorship over no apple pies (though I do have proximate authorship over many delicious ones, just for the record).
How to Make an Agent out of Clay
So how are we, without ultimate authorship, supposed to actually be anything? When does it make sense to think of ourselves as agents, with preferences we endorse, reasons we respond to, and a will of our own? In short, how do we become a “self”?
Earlier I said my preferences were my preferences in some meaningful way, but how can that be if I didn’t choose them? And even if I did choose them, I didn’t choose the process by which I chose them. And if, somehow, I chose that as well, we can just follow the chain back far enough and we'll reach something unauthored. That regress is exactly why ultimate authorship is impossible, and I’ve already conceded it.
But notice what the regress argument assumes: it gives all the credit to ultimate authorship and none to proximate authorship. By that standard, nothing we ordinarily call control or choice would count.
Consider a company as an example. Let’s say I make a bunch of decisions for my company. I say we’re going to build Product A and not Product B, we’re going to market it this way and not that way, and so on. In any common usage of the words, I clearly made those decisions—they were under my control. But did I, by Sam’s ultimate authorship standard? Well, the reason I wanted to build Product A is because I thought it would sell well. And that would generate revenue. And that would make the company more valuable. But, did I make the decision to set the goal of the company to be making money? Well, I wasn’t a founder of the company, so it wasn’t my idea to make a for-profit company in the first place. Therefore, by the standard of ultimate authorship, I had no control and made no decisions! The founders made every one of them when they decided to found a for-profit company. This, of course, is not how we think about decision-making and control.
What matters for agency isn’t whether your starting point was self-created; it’s whether the system can govern itself from the inside, whether it can reflect on its results and revise its own motivations over time.
Humans can evaluate their own evaluations. I can have competing desires and reason through them. I can want a cigarette but also not want to want cigarettes, and that second-order stance can reshape the first over time. That’s the feedback loop inside the decision-making system. The algorithm doesn’t just output actions; it can also adjust the weights it uses to produce future actions.
Here’s a real example from my life: I believe I’ve successfully convinced myself that I like broccoli. Years ago, I made a conscious decision to tell myself I really liked broccoli. I didn't hate it prior, but I wouldn’t have said I particularly enjoyed it. But I decided I'd be better off if I did, so I gathered all my anti-rationalist powers and told myself I enjoyed the taste. I ate it more often, and each time I told myself how much I was enjoying it. Within a couple of years, I realized I wasn't pushing anymore. I just liked broccoli. Frozen broccoli, microwaved with salt, pepper, and a little lemon juice, is now my go-to snack. And it’s delicious.
Now, we don't have ultimate authorship of either the first-order desire (disliking broccoli) or second-order desire (wanting to have a healthier diet), so who cares? But notice what happened here. This wasn't just a parameter being adjusted in some optimization process. It was me deciding what kind of person I wanted to be and reshaping my preferences to match. That’s me authoring at the proximate level and taking ownership of the kind of person I’m becoming. The broccoli preference became mine not because I authored it from scratch, but because I consciously endorsed and cultivated it. It coheres with who I take myself to be.
This matters because I want to show that, over time, humans can become coherent agents. I want to show that humans are a distinct category from just a pile of reflexes or a mere conscious witness to one's actions.
And this is why the regress to ultimate authorship doesn’t touch what matters. If “ownership” required self-creation, then no belief, value, or intention would ever count as yours either, because those, too, trace back to unchosen influences. But that’s not how we actually draw the line. We treat a preference as yours when it is integrated into your identity.
Note what this reveals about entities that can have free will. To reflect on your own desires you have to be able to represent them as your desires. You have to be able to take yourself as an object of evaluation and revise yourself over time. That requires a self-model robust enough to support second‑order evaluation: not just “I want X,” but “do I want to be the kind of person who wants X?”
You can see how the sense of agency develops in humans over time. It wouldn’t make sense to describe an infant as much of an agent. But over time, humans develop a sense of who they are and who they want to be. They can reflect on themselves and change accordingly. The algorithm can, to some degree, rewrite its own code in light of an identity it is actively shaping. This is another sense in which proximate authorship is “enough”. Not only can we run the algorithm, we can modify it.
That capacity for self-editing is a real boundary in nature. It separates agents from mere processes. A muscle spasm can't reflect on itself. A craving can't decide it would rather be a different kind of craving. But I can, and that's the distinction that matters when we ask whether someone acted freely.
Sam’s entire objection seems to boil down to the assumption that control requires ultimate authorship. But this assumption doesn’t hold.
Disputing Definitions
OK, some of this has gotten into semantics, so, in keeping with my pledge: I hereby dispute the definition.
As we’ve seen, when I say “I have free will,” I don’t mean I’m the ultimate, uncaused source of my decisions, untouched by genes, environment, or prior events. I mean I have the capacity to translate reasons, values, and goals into actions in a way that is responsive to evidence. Or, in short, to run the algorithm.
So why call this “free will”?
First, you can see how the feeling of free will falls out of this algorithm. When people say they “could have done otherwise,” they are feeling their choice-making algorithm at work, and, as I’ve shown, that algorithm really is at work. The phenomenon matches the feeling of free will, so I say it’s appropriate to call it that.
Second, I think this definition matches how people talk in everyday life. Consider the following:
In all of these, “could have” means something like, “given the situation, a different outcome was within reach under slightly different conditions.”
For example, consider “I could have made that shot.” If I miss a half-court shot, I might say “I could have made that shot.” By that, I mean that, given my skill, if I tried again under similar conditions, it’s possible I could have made it. Making it is within my ability. If I try a full-court shot and the ball falls 20 feet short, then I probably just couldn’t have made it. I lack the physical capacity.
This is Could₁. It’s about alternative outcomes across nearby scenarios (e.g. I could have made that shot if the wind was a little bit different).
Contrast that with what the sentence would mean if people were using Could₂. The sentence would be, “I could have made that shot even if everything about the past and the laws of nature were the same.” It says that, rewinding every atom in the universe and every law of physics, things could (Could₂) have gone differently. This is a completely different claim and it’s not what people mean when they use the word.
This is not some complex claim that relies on consciousness. I’m talking about basic standard usage of the word “could”. Here’s another example: Which do people mean by “The car could go faster”? Do they mean:
Could₁ is simply the standard usage of the term. In addition, it’s how it’s used in ordinary moral or legal discussions.
Take it or Leave it?
The term “free will” is what computer scientist Marvin Minsky would call a “suitcase phrase”—people pack different meanings into it and call it the same thing. There are some definitions of free will that Sam and I would both jump up and down and say, “No! That does not happen.” Mainly, the notion that if we were to reset the universe’s clock back 30 seconds and put every atom back in its place, someone could (Could₂) choose to act differently. But there are also some definitions of “free will”, like the feeling of weighing your options, reasoning your way to a conclusion, and acting based on that reasoning, where we should jump up and down and say, “Yes! That’s a real thing.”
Sam looks at the range of meanings people attach to "free will," sees the metaphysical baggage, and concludes we're better off abandoning the term. I look at the same thing and see most ordinary usage pointing toward something defensible. When someone says "I chose to stay late at work," they're not claiming to have escaped the causal order of the universe or exercised some quantum soul-power. They're saying the deliberative algorithm ran: they considered leaving, weighed their reasons, and decided to stay. That's Could₁, and it's real.
Sam has an analogy for what he thinks compatibilists are doing. He compares it to claiming that Atlantis is real—it's just the island of Sicily. Sure, it lacks the ancient advanced civilization, didn't sink into the Atlantic, and isn't “greater in extent than Libya and Asia”, but hey, it’s an island! Compatibilists, he suggests, are performing the same sleight of hand: pointing to something real but mislabeling it with a term that implies much more.
Sam's analogy seems to imply that Could₂ is the defining feature of free will, and that I've discarded it while keeping the name. But I think this gets it backwards. As I said, when people say “I could have done otherwise,” they mostly mean Could₁. Admittedly, the free will I'm describing doesn't deliver everything the term has ever been associated with. There’s no ultimate authorship, no metaphysical Could₂. But consider what people actually use free will for. They use it to distinguish choice from compulsion, to ground praise and blame, to make sense of deliberation. Could₁ does all of that and Could₂ does none of it. The features I'm preserving aren't peripheral; they're the load-bearing components. People want Could₁ from their free will and Sam is demanding Could₂.
I don’t understand why he seems to place so much importance in ultimate authorship. He seems to think that without it, “free will” names nothing worth preserving. But ultimate authorship was never part of how we actually explain human behavior. We’re billions of years into a cause-and-effect universe. When we ask "Why did she do that?" we don't expect an answer that traces back to the initial conditions of the universe. We expect proximate causes—reasons, motives, deliberation.
Any time someone asks, “Why?” there is an unbroken chain of answers that could technically answer the question. There’s a sweet spot for good explanations for most questions, and it’s neither the ultimate cause nor the most proximate one, though it’s often much closer to the latter. Consider some examples:
The same applies to moral explanations. “Why did he betray his friend?” calls for an answer about motives, reasoning, and character, not about the initial conditions of the universe. We explain human action in terms of proximate causes because that's the level at which deliberation, and therefore responsibility, operates. Ultimate authorship was never doing any work in these explanations. Letting it go costs us almost nothing we actually use.
Free will by any other name would smell as sweet
It’s worth stepping back and asking, “How does a philosophical concept like ‘free will’ gain metaphysical legitimacy anyway?” We’re not going to find it like we would a physical object. When I say “free will exists”, I’m not saying we’re going to see it in a brain scan.
This is why I say this isn't just about disputing a definition. I'm making a stronger claim: My point is that any coherent account of agency, responsibility, and reasoning must posit something playing the free-will role. There must be some concept that distinguishes deliberated action from compulsion, reflex, or accident.
Without it, I think you’re forced into some strange positions. In his podcast Final Thoughts on Free Will, he treats someone being convinced by an argument as having the same freedom as being “pushed off a cliff and then claiming that I'm free to fall”. In the podcast Making Sense of Free Will, he makes no distinction between choosing orange juice and having a muscle spasm (see “Responding to Another Sam Harris Thought Experiment” in the appendix).
The distinction between being persuaded and being pushed, between choosing and spasming, isn't some folk illusion we should discard in light of modern science. These are natural categories. It's a distinction that carves reality at its joints. A spasm is an “open-loop” process: a signal fires, the muscle contracts, and no feedback mechanism checks whether this action serves your goals. Choosing juice is a “closed-loop” control system: an option is proposed, simulated against your preferences, evaluated, and executed only if it passes muster. These are fundamentally different mechanisms. One is responsive to reasons; the other isn't. If you told me “the orange juice is poisoned,” I'd choose differently. If you told my leg “don't jerk” while tapping my patellar tendon, it would jerk anyway.
This is what makes the choice mine in a way the spasm isn't. The choice responds to what reasons mean to me. This is the difference between acting and being acted upon. Sure, both events are determined by prior causes, but to not see these as differences in kind seems, frankly, bizarre.
Or consider coercion. When someone holds a gun to your head, we say your choice was constrained. Yes, you “gave” them your wallet, but not freely. What makes this different from an unconstrained choice? It's not that determinism was more true in the coercion case. It's that your algorithm was given artificially narrowed options by an external agent.
Free Will as the Ontological Minimum
When I talk about free will, I'm not positing anything magical or spooky. Free will, as I've described it, exists the way beliefs exist. If you shrunk down in a Magic School Bus you wouldn’t find beliefs stored in labeled containers in the brain. But beliefs are real, right? They're what we call a certain functional capacity of the brain. Free will is similar. It's the name for what's happening when a system weighs reasons, considers alternatives, and selects among them.
This is the minimal ontological commitment required to make sense of how we actually think about people. When we hold someone responsible, when we distinguish choice from compulsion, when we ask “why did you do that?”, we expect a reasons-based answer. Sam can call it something else if he likes. But he needs something to mark these distinctions, or his account of human action becomes incoherent. He simply has a free-will-shaped hole in his ontology.
I'm genuinely curious: from Sam’s perspective, do “beliefs”, “reasons”, “thinking”, and “agents” exist? We distinguish humans from thermostats by saying we respond to reasons while thermostats respond only to temperature. If reasons are real and can be causes of action, why not free will? It's the same kind of thing, a higher-level description of what certain physical systems do, irreducible not because it's made of magic, but because it captures patterns the lower level doesn't.
Why It Matters: Phenomenology, Incentives, Morality, Law
Why does any of this matter? How does my defense of free will cash out in terms of things that we care about? I’ll list five reasons why this matters:
1. Phenomenology. People have a strong intuitive sense of free will. Where does this feeling come from, and does it track something real?
2. Incentives and behavior. Can people respond to rewards, punishments, and social pressure? How does free will relate to deterrence and rehabilitation?
3**. Moral responsibility.** Are people moral agents? Can they be held responsible for their actions?
4. Hatred and retributive punishment. Does anyone deserve to suffer for what they've done?
5. Crime and punishment. How should the legal system treat offenders?
Let me address each in turn.
Phenomenology: The Feeling of Free Will
We have a persistent feeling that we could have done otherwise. Is this feeling tracking something real, or is it an illusion? The answer depends on which “could” we mean. For Could₁, the sense that we would have chosen differently had our reasons, evidence, or preferences been different, yes, that’s completely real. But for Could₂, the sense that we might have chosen differently with every atom in the universe held fixed, no, that's not real.
And, as I’ve argued, Could₁ is what the feeling of free will is actually about. This is what makes the algorithm-based account satisfying: it explains the phenomenology of free will without explaining it away. When you run through options, simulate outcomes, and select among them, you're not passively watching a movie of yourself deciding. You're experiencing your deliberative process at work. The feeling of choosing is the choosing. That's what free will feels like from the inside, and that's what free will is.
Incentives and Behavior
Here, Sam and I agree on the facts. People obviously respond to incentives. Stigmatizing drunk driving works. Offering bonuses improves performance. Punishment can deter crime. We shape behavior through incentives all the time.
I think Sam would argue that this doesn’t mean they have free will, just that their behavior responds to inputs. Fine, you could say that, but if you need a system that responds to reasons, weighs options, and updates based on consequences to explain human behavior, you've just described free will but are refusing to use the term. Incentives work because they feed directly into your deliberative algorithm. They change the weights, alter the utility calculations, and thus change behavior. This is why we can hold people accountable, offer rewards, impose consequences, and expect behavior change.
Moral Agency and Responsibility
I’ve claimed that we have proximate authorship but not ultimate authorship of our actions. Is this “enough” authorship for moral responsibility? I believe so. I believe being a moral agent is being the kind of entity whose decision-making can incorporate moral reasoning. This is a bit beyond the scope here, but the following are the types of things I would expect a moral agent to be able to do:
This is why we treat adults differently from infants, and humans differently from bears. It's not that adults have ultimate authorship and infants don't; it's that adults have proximate authorship, and their algorithm can incorporate moral reasoning. A bear that mauls someone isn't a moral agent. It doesn't think, “How would I feel if someone did this to me?”
There are degrees here, of course. A four-year-old has more moral agency than an infant, and less than an adult. Someone with severe cognitive impairment may have diminished moral agency. The question is always: to what extent can this entity's algorithm incorporate moral reasoning?
Moral Desert, Hatred, and Retributive Punishment
In addition to moral responsibility, there's the question of desert, of whether wrongdoers deserve to suffer as retribution for their actions. Here, Sam and I completely agree that they do not. To deserve retribution in that deep sense, someone would need ultimate authorship of their actions.
To see why, consider an example Sam gives: someone commits violence because a brain tumor is pressing on their amygdala. We recognize them as a victim of neurology, not a monster deserving punishment. But now replace the tumor with an abusive childhood, genetic predispositions toward impulsivity, or serotonin imbalances. At each step, we're still describing physical causes the person didn't choose. The distinction between “tumor” and “bad genes” is arbitrary—both are prior causes outside the person's control. It's brain tumors all the way down. There but for the grace of God go I.[12]
Moral desert simply requires a metaphysical freedom that people do not have.
Once you give up ultimate authorship, a certain kind of hatred has to go with it. You can't coherently hate someone as the ultimate author of their evil, as if they, from nothing, simply chose to be bad. That hatred requires the same metaphysical freedom that no one actually has.
Think about a bear that mauls someone. The bear causes harm, and we might kill it for public safety, but we don't hate the bear. It's not the kind of thing that could deserve retribution. The important part is recognizing that, without ultimate authorship, the same logic extends to humans. People who do terrible things are not deserving of suffering for its own sake. On this, Sam has been a tireless voice, and I appreciate his advocacy of this position.
This doesn't eliminate all meanings of “hate” entirely, just a particular kind. You can still hate your job, Mondays, and git merge conflicts. You can definitely still hate dealing with git merge conflicts for your job on Mondays. But notice this is a different kind of hate. There’s no sense in which you want Monday to “pay for what it's done.” It's about anticipating that you’ll have a bad experience with it and seeking to avoid it.
The same applies to people. You can recognize that someone's algorithm doesn't adequately weigh others' suffering, and you can avoid them accordingly. But there’s no need to view your enemies as self-created monsters deserving retributive punishment.
On this point, Sam wins. Perhaps if retributive justice were all I cared about, I would agree with him that we should consider free will an illusion. But free will does more work than that. It's deliberation doing real causal work. It grounds the distinction between choice and compulsion, makes sense of why incentives change behavior, and gives meaning to praise and blame. Retributive punishment is the one piece that genuinely requires ultimate authorship, and it's the one piece I'm happy to let go.
Crime and Punishment
What does this mean for crime and punishment? Does this mean we can't hold anyone responsible? No. Sam and I are aligned here. We can hold people responsible without blaming them for ultimate authorship. We can and should hold people responsible in a forward-looking sense: for deterrence, rehabilitation, and public safety. Courts still need to distinguish intentional action from accident, choices made with a sound mind from those made under coercion or insanity. My account of free will provides exactly that framework: Did the algorithm run normally, or was it bypassed (reflex), broken (insanity), distorted (addiction), or given severely constrained options (coercion)?
Sam and I agree that sometimes we must incarcerate people because they are dangerous to others. But we do so to mitigate harm and deter future crime, not to exact retributive justice upon them.
Final Cruxes
Sam is right that there's no ghost in the machine, (probably[13]) no soul pulling levers from outside the causal chain, no metaphysical Could₂ freedom. We agree more than we disagree. (In fact, Dan Dennett has called Sam “a compatibilist in everything but name!”). However, I wanted to compile what I see as the core cruxes of disagreement into a list. If Sam and I were to sit down and productively hash this out, here's where I think we'd need to focus:
1. Is conscious deliberation causally efficacious, or is it epiphenomenal narration? I say the algorithm is the decision-making process—consciousness is doing real computational work. Sam says consciousness is merely “witnessing” decisions made elsewhere.
2. Is there a meaningful categorical difference between deliberated actions and reflexes? I say yes—one runs through the algorithm, one bypasses it. Sam seems to collapse this distinction since both are “caused”. But if there's no difference between choosing orange juice and having a muscle spasm, something has gone wrong.
3. Is there a meaningful categorical difference between entities that can reflect on and revise their own decision-making versus those that cannot? A thermostat responds to temperature; a human can respond to reasons, evaluate their own preferences, and update their future behavior accordingly. I taught myself to like broccoli. I would like to see a thermostat do that. I can notice a bad habit and work to change it. This capacity for reflective self-modification seems like a real category that separates agents from mere processes. Does Sam recognize this as a meaningful distinction, or is this also collapsed because both are ultimately “caused”?
4. What should we think of pathologies where someone feels like a mere witness to their actions? To me, these seem like cases where the algorithm is damaged and consciousness isn’t able to play its active, deliberate role that it usually plays. I don’t know how Sam would describe these.
5. What lessons should we learn from the Libet-style experiments? Does it show that consciousness is post-hoc rationalization, or merely that consciousness isn't the initiating step while still doing causal work downstream?
6. What should we think about an entity that has proximate authorship but not ultimate authorship (as all of us do)? Is that sufficient for moral responsibility, control, praise, and blame? Sam seems to think that without ultimate authorship, "control" is illusory. I think proximate authorship is sufficient, and that demanding ultimate authorship sets an impossible standard. The implication would be no one has ever controlled anything.
7. What counts as “you”? When Sam says “your brain did it,” he treats this as evidence against free will, almost as if “you” were separate from your brain. I say my brain doing it is me doing it. The deliberative algorithm running in my neurons is my free will. We may simply have different intuitions about where to draw the boundary of the self and whether being moved by your own values counts as freedom or puppetry. Similarly, should you identify with yourself? Should you take credit for the person you've become? Should we make anything of a person’s ability to become a more coherent agent over time versus a pile of unauthored behaviors? I say “yes”.
8. What criteria must a metaphysical concept meet to earn its place? If beliefs, reasons, and agents qualify, what test do these pass that free will uniquely fails? Does Sam reject it simply because of the historical “Could₂ baggage” associated with it? For me, a concept earns its keep by leading to and aligning with other natural categories, and doing without them requires tap-dancing around the concept.
9. What do ordinary people mean by “could have done otherwise”? I claim everyday usage is Could₁: “I would have acted differently if my reasons or circumstances had been different.” Sam seems to think people intuitively mean Could₂: “I could have acted differently with every atom in the universe held fixed.”
10. Is “free will” worth preserving as a concept, or should we retire it? Sam looks at the metaphysical baggage and says we're better off abandoning the term. I look at what people actually use the concept for and say these are the load-bearing features. If we abandon the term, don’t we need something else to replace it? Doesn't any coherent account of agency require something playing the free-will role?
I say let's keep the term. Free will names something real: a process fully physical, fully determined by prior causes, and yet still you doing the choosing. The algorithm isn't an illusion overlaid on "mere" physics. It is the physics, operating at a functional level that matters for morality, law, and human experience.
So, Sam, what would it take to convince you? If the algorithmic account captures what people mean by free will, does the work we need it to do, and doesn't require any spooky metaphysics, what's left to object to besides the name?
Appendix
Responding to Another Sam Harris Thought Experiment
I want to go over another thought experiment that Sam gives to show how all of his objections don't threaten the notion of free will as I’ve described it. This is from the podcast Making Sense of Free Will. The thought experiment starts at 42 minutes in. The narrator makes a point that Sam has made many times, but it’s made clearly here, so I’ll use it. Here’s the setup:
The narrator (echoing Sam), argues that the selection of orange juice and the spilling of the juice aren't as different as they seem. Yes, the spasm feels like something done to you. But did you really "choose" the orange juice? Did you create your preference for it? The narrator makes the case:
This might sound compelling until you apply the algorithmic account. Then each objection dissolves:
I usually refer to people I don't know personally by their last names, but I've been listening to Sam's podcast for over a decade, and calling him “Harris” just feels strange. So I use his first name out of the creepy, one-sided familiarity that comes with being a longtime listener. I mean no disrespect. ↩︎
Libertarian free will has nothing to do with economic libertarianism; it’s just an unfortunate namespace collision. ↩︎
Sam is more confident here. I say “probably” and “likely” because we’re talking about souls, and, if they’re real, we have close to no idea how they work. We’re in speculative territory here, so it’s good to be cautious. ↩︎
In his book Free Will, he says both that “Free will is an illusion” (as well as in this essay) and that there is no illusion of free will because “The illusion of free will is itself an illusion” (also said in this podcast). Parsing this is beyond the scope here. In all cases he’s consistent about arguing that we do not have free will, so that’s why I word it like that. ↩︎
This is a highly simplified version. The real version would have lots of error checking and correction at each layer (just like layers 2 and 4 in the OSI model, if you’re familiar with that). For example, the real first step would really be making sure I understood the question. I’m going to leave these out for simplicity. ↩︎
A utility function determines how much you value different outcomes by weighting your options according to your preferences. ↩︎
Again, this is highly simplified. It’s not necessarily linear like this. If I decide I don’t like any in the evaluation stage, I can just go back to querying my memory. Or, if I realize I don’t just want to name a movie but also name a movie that will show that I’m an interesting guy, I’ll edit the goal to include that. ↩︎
Because the real question is, are you talking about My Little Pony: The Movie or My Little Pony: A New Generation? ↩︎
For a dissenting opinion on readiness potential and whether we’re interpreting it correctly, see “What Is the Readiness Potential?” by Schurger et al. ↩︎
It’s worth noting that Libet was a compatibilist himself. In his paper “Do we have free will?”, he argues that “the conscious function could still control the outcome; it can veto the act. Free will is therefore not excluded.” ↩︎
For example, I quoted above where he says, “you as the conscious witness of your inner life are not making decisions. All you can do is witness decisions once they're made.” However, although Sam often refers to consciousness as a mere witness, he has also said that it does things. In his book Free Will, he says: ↩︎
For a detailed examination of this idea, I recommend the Radiolab episode Blame. ↩︎
Sorry, I just have to put in a “probably” here because it’s a statement about souls, which are quasi-metaphysical, so we really shouldn’t be too certain how they would work. ↩︎