The subsumption architecture for robotics invented by Rodney Brooks is based on the idea of connecting behavior to perception more directly, with fewer layers of processing and ideally no central processing at all. Its success, e.g. the Roomba, stands as proof that something akin to control theory can be used to generate complex agent-like behavior in the real world. In this post I'll try to give some convincing examples from literature and discuss a possible application to anti-akrasia.
We begin with Braitenberg vehicles. Imagine a dark flat surface with lamps here and there. Further imagine a four-wheeled kart with two light sensors at the front (left and right) and two independent motors connected to the rear wheels. Now connect the left light sensor directly to the right motor and vice versa. The resulting vehicle will seek out lamps and ram them at high speed. If you connect each sensor to the motor on its own side instead, the vehicle will run away from lamps, find a dark spot and rest there. If you use inverted (inhibitory) connectors from light sensors to motors, you get a car that finds lamps, approaches them and stops as if praying to the light.
Fast forward to a real world robot [PDF] built by Brooks and his team. The robot's goal is to navigate office space and gather soda cans. A wheeled base and a jointed hand with two fingers for grabbing. Let's focus on the grabbing task. You'd think the robot's computer should navigate the hand to what's recognized as a soda can and send out a grab instruction to fingers? Wrong. Hand navigation is implemented as totally separate from grabbing. In fact, grabbing is a dumb reflex triggered whenever something crosses an infrared beam between the fingers. The design constraint of separated control paths for different behaviors has given us an unexpected bonus: a human can hand a soda can to the robot which will grab it just fine. If you've ever interacted with toddlers, you know they work much the same way.
A recurrent theme in those designs is coordinating an agent's actions through the state of the world rather than an internal representation - in the words of Brooks, "using the world as its own model". This approach doesn't solve all problems - sometimes you do need to shut up and compute - but it goes surprisingly far, and biological evolution seems to have used it quite a lot: for example a moth spirals into the flame because it's trying to maintain a constant angle to the light direction, which works well for navigation when the light source is the moon.
Surprising insights arise when you start applying those ideas to yourself. I often take the metro from home to work and back. As a result I have two distinct visual recollections of each station along the way, corresponding to two directions of travel. (People who commute by car could relate to the same experience with visual images of the road.) Those visual recollections have formed associations to behavior that bypass the rational brain: if I'm feeling absent-minded, just facing the wrong direction can take me across the city in no time.
Now the takeaway related to akrasia that I've been testing for the last few days with encouraging results. Viewing your brain as a complete computer that you ought to modify from inside is an unnecessarily hard approach. Your brain plus your surroundings is the computer. A one-time act of changing your surroundings, physically going somewhere or rearranging stuff, does influence your behavior a lot - even if it shouldn't. Turn your head, change what you see, and you'll change yourself.