Publication Cover

Cognitive Neuroscience

Current Debates, Research & Reports
Volume 5, 2014 - Issue 2
1,534
Views
10
CrossRef citations to date
0
Altmetric
Commentaries

Active inference and agency

Abstract

I greatly enjoyed Seth’s compelling synthesis of sensorimotor contingencies and active inference. I would also like to thank Jim Hopkins for sending me the quote (below)—which speaks directly to the embodied nature of perceptual inference that underlies the perspectives reconciled in Seth (this issue). These perspectives include perception as hypothesis testing, affordance, and sensorimotor contingencies. This commentary briefly rehearses the fundaments of active inference and offers a formal basis for Seth’s key argument.

Each movement we make by which we alter the appearance of objects should be thought of as an experiment designed to test whether we have understood correctly the invariant relations of the phenomena before us, that is, their existence in definite spatial relations. (Helmholtz, 1971/1878)

Active inference tries to understand action and perception as working hand-in-hand to minimize a (free energy bound on) surprise. When surprise is quantified by prediction error we get predictive coding or processing. So why minimize surprise? The answer is simple: If we minimize surprise, we restrict ourselves to the embodied states that we expect to frequent. This single imperative has construct validity from several perspectives. For example, in statistics, surprise is the negative logarithm of Bayesian model evidence. This means that minimizing surprise is equivalent to Bayesian inference; hence, the Bayesian brain (Gregory, 1980; Dayan, Hinton, Neal, & Zemel, 1995). From the perspective of self-organization, autopoiesis, and allostasis, the long-term average of surprise is entropy. This means minimizing surprise allows biological agents to resist a natural tendency to disorder (Maturana & Varela, 1980; Sterling & Eyer, 1988). Finally, if we associate surprise with cost or negative value, then minimizing surprise maximizes expected value or utility (Friston et al., 2013).

To minimize prediction error, one clearly needs a (hierarchical) model that can generate predictions of sensory input. The hierarchical aspect is crucial here, in the sense that the nature of inferences or expectations is determined by the predictions they generate. For example, predictions may be limited to the visual domain, or could be proprioceptive predictions that engage classical reflex arcs. The key point here is that at some hierarchical level, expectations will furnish bilateral predictions of exteroceptive (sensory) and proprioceptive (motor) sensations. These amodal expectations can be construed as sensorimotor constructs—that rest upon sensorimotor contingencies for their acquisition or learning. In short, sensorimotor contingencies are a necessary part of hierarchical inference (Gibson, 1977; Noë, 2004). However, this does not mean that all expectations are sensorimotor; there are necessarily some perceptual expectations (e.g., qualia) that are purely sensory in nature (see Figure 1). Technically, these are sequestered from sensorimotor constructs by a Markov blanket (Pearl, 1988). This implies that not all perception is sensorimotor in nature (e.g., listening to music); for example, synesthetic concurrents that appear to be conditionally independent of beliefs about motor control:

However, the notion of mastery of SMC relevant to perceptual presence seems to suggest the involvement of more than just a generative model predicting ongoing sensorimotor flow. The incorporation of explicitly conditional (or even meta-conditional) aspect seems essential. (Seth, this issue)

Figure 1. Schematic of a hierarchical generative model illustrating the implicit partition of hidden states (that are inferred in terms of expectations by the brain). This partition emphasizes the distinction between hidden (sensorimotor) states that furnish descending predictions in both the exteroceptive (sensory) and proprioceptive (motor) domain.

As noted by Seth, prediction errors can be resolved by adjusting hierarchical expectations to improve the predictions (i.e., perception)—or sensations can be re-sampled to comply with predictions (i.e., action). This necessarily entails predictions of what would be sampled under allowable actions. In other words, our generative models have to entertain fictive or counterfactual outcomes to enable action to resolve prediction errors—or avoid surprises. This is a central tenet of active inference, as demonstrated in the counterfactual predictive processing that underlies oculomotor control and visual searches (Friston, Adams, Perrinet, & Breakspear, 2012). The same basic idea underlies recent formulations of decision making and choice, in which action is sampled from prior beliefs about (future) control states, given the current sensory state (Friston et al., 2013): This expression says that the least surprising policy or control sequence (over times ) minimizes the difference (Kullback-Leibler divergence) between the final outcomes , given the current state and a policy, and outcomes that are, a priori, anticipated. This formulation of prior beliefs about controlled outcomes is closely related to KL-control (Ortega & Braun, 2013; van den Broek, Wiegerinck, & Kappen, 2010) and can be related directly to classical utility theory (Friston et al., 2013).

From our point of view, it says two important things. First, it casts behavior explicitly in terms of prior beliefs about sensorimotor contingencies: “PPSMC requires that counterfactual predictions be explicitly incorporated as part of the priors in a hierarchical generative model.” (Seth, this issue)

Second, it shows that when expectations about sensory states do not inform beliefs about control, there is no mutual information between the sensory expectation and the final outcome: In other words, there are purely sensory expectations that do not inform future or counterfactual outcomes—and have no sensorimotor contingency. One can imagine quantifying the degree to which sensory beliefs inform controlled outcomes in terms of the mutual information between the current and future sensory state—a metric of counterfactual richness: “My specific claim is that the subjective veridicality (or perceptual presence) of normal perception depends precisely on the counterfactual richness of the corresponding generative models.” (Seth, this issue)

REFERENCES

  • Dayan, P., Hinton, G. E., & Neal, R., Zemel R.S.  (1995). The Helmholtz machine. Neural Computation, 7, 889904. doi:10.1162/neco.1995.7.5.889 [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Friston, K., Adams, R. A., Perrinet, L., & Breakspear, M. (2012). Perceptions as hypotheses: Saccades as experiments. Frontiers in Psychology, 3, 151. doi:10.3389/fpsyg.2012.00151 [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Friston, K., Schwartenbeck, P., FitzGerald, T., Moutoussis, M., Behrens, T., & Dolan, R. J. (2013). The anatomy of choice: Active inference and agency. Frontiers Human Neuroscience, 7, 598. doi:10.3389/fnhum.2013.00598 [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Gibson, J. J. (1977). The theory of affordances. In R. E. Shaw, & J. Bransford (Eds.), Perceiving, acting, and knowing: Toward an ecological psychology (pp. 6782). Hillsdale, NJ: Erlbaum. [Google Scholar]
  • Gregory, R. L. (1980). Perceptions as hypotheses. Philosophical Transactions of the Royal Society B: Biological Sciences, 290, 181197. doi:10.1098/rstb.1980.0090 [Crossref], [PubMed], [Web of Science ®][Google Scholar]
  • Helmholtz, H. (1878 (1971)). The facts of perception. In R. K. Middletown (Ed.), The selected writings of Hermann von Helmholtz (p. 384). Connecticut: Wesleyan University Press. [Google Scholar]
  • Maturana, H. R., & Varela, F. (1980). Autopoiesis: The organization of the living. In H. R. Maturana, & F. Varela (Ed.), Autopoiesis and cognition. Netherlands: Reidel: Dordrecht. [Crossref][Google Scholar]
  • Noë, A. (2004). Action in perception. Cambridge, MA: MIT press. [Google Scholar]
  • Ortega, P. A., & Braun, D. A. (2013). Thermodynamics as a theory of decision-making with information-processing costs. Proceedings of the Royal Society A: Mathematical, Physical and Engineering Sciences, 469, 2153. doi:1098/rspa.2012.0683 [Crossref], [Web of Science ®][Google Scholar]
  • Pearl, J. (1988). Probabilistic reasoning in intelligent systems: Networks of plausible inference. San Fransisco, CA: Morgan Kaufmann. [Google Scholar]
  • Sterling, P., & Eyer, J. (1988). Allostasis: A new paradigm to explain arousal pathology. In S. Fisher, & J. Reason (Eds.), Handbook of life stress, cognition and health (pp. 629649). New York: John Wiley & Sons. [Google Scholar]
  • van den Broek, J. L., Wiegerinck, W. A., & Kappen, H. J. (2010). Risk-sensitive path integral control. UAI, 6, 18. [Google Scholar]

Reprints and Permissions

Please note: We are unable to provide a copy of the article, please see our help page How do I view content?

To request a reprint or commercial or derivative permissions for this article, please click on the relevant link below.

Permission can also be obtained via Rightslink. For more information please visit our Permissions help page.