Fact-checked by Grok 2 weeks ago

Free energy principle

The free energy principle (FEP) is a theoretical in and theoretical that posits self-organizing biological systems, such as the , maintain and resist entropic disorder by minimizing variational , which serves as an upper bound on the (negative log probability) of sensory inputs relative to internal generative models. This minimization process unifies , learning, and under a single imperative: systems adapt by optimizing approximate posterior beliefs about environmental causes through hierarchical , effectively suppressing prediction errors to align sensory data with expectations. Proposed by Karl Friston in the mid-2000s, the FEP draws from statistical physics, , and variational methods to explain how living systems occupy low- states in a fluctuating , avoiding transitions that could lead to disorder or death. At its core, is formulated as F = \langle \ln q(\mu) - \ln p(\tilde{y}, \mu) \rangle, where q(\mu) is a recognition density approximating the true posterior over hidden causes \mu, and p(\tilde{y}, \mu) is the joint probability under a ; minimizing F bounds sensory and enables self-evidencing behaviors. This principle extends beyond the brain to any autopoietic system, encompassing , evolutionary , and even social phenomena, by framing as active inference—where agents sample the world to fulfill predictions rather than passively react. Key implications include as a neural implementation, where top-down predictions from higher cortical layers meet bottom-up errors, driving and to resolve uncertainties. The FEP has influenced fields like active inference models in and , offering a normative account of cognition that contrasts with reward-based by emphasizing intrinsic model evidence over extrinsic goals. Ongoing research explores its role in psychiatric disorders, such as , where aberrant precision weighting of predictions may underlie symptoms like hallucinations.

Introduction

Definition and core concepts

The free energy principle (FEP) is a theoretical framework proposing that self-organizing biological systems resist disorder by minimizing variational , which serves as a tractable upper bound on surprise—or self-information—thereby preserving their structural and functional integrity while maintaining . This minimization process ensures that systems remain in states they are optimized for, avoiding excessive unpredictability in their sensory exchanges with the . The principle frames as an inevitable consequence of thermodynamic imperatives, where actively model and interact with their surroundings to bound the improbability of their observations. Central to the FEP are concepts like , defined as the negative logarithm of the probability of sensory data under a 's internal , representing the degree of mismatch between expectations and actual inputs. Markov blankets further delineate boundaries, comprising sensory states that report external influences and active states that influence the , while shielding internal states from direct external access and enabling . These elements collectively a from its milieu, allowing it to infer causes of sensations and act to fulfill predictions. The FEP provides a normative account of across scales of , from cellular to neural processing in brains, positing that all such entities minimize to sustain viability through and environmental modulation. For instance, it explains how organisms anticipate and adapt to perturbations, ensuring long-term survival by resolving uncertainties in sensory streams. Originally proposed by Karl Friston in 2006 and elaborated in key publications through 2010, the framework integrates , action, and learning as unified processes under free energy minimization.

Historical development

The free energy principle (FEP) traces its conceptual origins to 19th-century , particularly Hermann von Helmholtz's theory of , which posited that involves the brain's probabilistic of sensory to infer the external . This idea laid early groundwork for viewing the brain as an inferential mechanism, a theme echoed in later FEP formulations. In the mid-20th century, emerged as a key influence, emphasizing feedback loops and self-regulation in complex systems to maintain stability against environmental perturbations. W. Ross Ashby's homeostat, developed in the 1940s, exemplified these principles through a device that autonomously adapted to disturbances via ultrastable reconfiguration, prefiguring notions of adaptive minimization in biological agents. Karl Friston formalized the FEP in the mid-2000s, initially applying variational free energy to model cortical responses and in data. His paper explicitly articulated the FEP as a unifying account of function, drawing on variational methods to bound or . This was expanded in a seminal review, which positioned the FEP as a comprehensive theory integrating , action, and learning through free energy minimization. Throughout the 2010s, the FEP evolved by incorporating frameworks, such as Rao and Ballard's 1999 model of hierarchical error minimization in , to explain neural implementations of inference. Active inference, an extension linking perception to action via expected , gained prominence in Friston's 2010 work on behavioral optimization. Key milestones included Friston's 2013 paper broadening the FEP to biological and , influencing interdisciplinary applications. Post-2020 refinements have addressed limitations in assuming steady-state conditions, extending the FEP to non-equilibrium in random dynamical systems for more general applicability to living processes. These developments build briefly on traditions while emphasizing variational approximations for practical computation.

Theoretical Foundations

Variational free energy

The variational free energy, denoted as F, is defined in the free energy principle as an approximation to the of sensory data, formulated as F[q(\mu)] = D_{\text{KL}}[q(\mu) \parallel p(\mu \mid s)] - \ln p(s), where q(\mu) is a variational approximating the true posterior p(\mu \mid s) over hidden states \mu, s represents sensory inputs, D_{\text{KL}} is the Kullback-Leibler divergence, and -\ln p(s) is the or self-information of the sensory data. This expression arises in variational applied to generative models, where the system posits a joint p(s, \mu) over observables and their hidden causes. The derivation of F as an upper bound on follows from the non-negativity of the KL divergence and applied to the log-evidence lower bound. Specifically, the satisfies \ln p(s) = \mathbb{E}_{p(\mu \mid s)}[\ln p(s)] \geq \mathbb{E}_{q(\mu)}[\ln p(s, \mu) - \ln q(\mu)], which rearranges to F[q(\mu)] \geq -\ln p(s), with equality when q(\mu) = p(\mu \mid s). This bound ensures that minimizing F provides a tractable for minimizing surprise, avoiding direct computation of the intractable posterior. The components of F reflect a between accuracy and : the KL term D_{\text{KL}}[q(\mu) \parallel p(\mu \mid s)] measures , quantifying the information gain or deviation from the true posterior (often decomposed further into divergence from priors), while -\ln p(s) relates to accuracy via the expected log-likelihood under the approximate posterior, \mathbb{E}_{q(\mu)}[\ln p(s \mid \mu)]. In this framework, biological or artificial systems employ hierarchical generative models to infer the causes of sensory inputs, encoding priors over hidden states in a top-down manner to predict s.

Free energy minimization

The free energy principle posits that adaptive systems, whether biological or artificial, minimize variational F through on its partial derivatives with respect to internal model parameters \theta, such that \frac{\partial F}{\partial \theta} = 0. This optimization process underpins both , by refining approximate posterior beliefs about hidden causes of sensory inputs, and learning, by updating the to better predict future observations. As a variational bound on —the negative log probability of sensory data under the true —minimizing F effectively reduces the between predicted and actual sensory states, enabling systems to maintain accurate internal representations of their environment. Systems achieve this minimization through two primary strategies. The perceptual strategy involves updating internal beliefs or states to align the model's predictions with incoming sensory data, effectively performing approximate without altering the external world. In contrast, the active strategy entails selecting that modify the environment or sensory inputs to better match the system's expectations, thereby resolving prediction errors through environmental reconfiguration rather than . These complementary approaches allow systems to navigate dynamically, with the choice between them depending on the relative costs of versus . The implications of free energy minimization extend to the and persistence of self-organizing systems, as it ensures long-term survival by systematically reducing expected over time. By bounding , minimization prevents the accumulation of improbable states that could lead to dissipative phase transitions or system dissolution, thereby promoting adaptive behaviors that sustain the system's operational integrity against entropic decay. This normative imperative frames not as isolated self-regulation but as a process of resilient exchange with the environment, where minimized corresponds to maximized model and enhanced . Underpinning this framework is the assumption that systems operate as ergodic random dynamical processes at non-equilibrium steady states, where trajectories revisit a bounded with a unique . ensures that time-averaged behaviors converge to ensemble averages, allowing to serve as a that drives the system toward low-entropy configurations while maintaining boundary integrity through a separating internal and external states. This steady-state condition is essential for persistence, as it confines states to a compact set resistant to external perturbations, with minimization actively enforcing the ergodic to avoid ergodicity-breaking transitions that would undermine adaptation.

Theoretical Connections

Bayesian inference

The free energy principle (FEP) implements by framing and learning as the optimization of hierarchical generative models that approximate the posterior over hidden causes of sensory data. Specifically, the FEP employs , where minimizing the variational free energy F provides an approximation to the posterior p(\mu | s) through an approximate q(\mu) = \arg\min_q D_{\text{KL}}[q(\mu) \| p(\mu | s)], with \mu representing the hidden states or causes and s the sensory inputs. This minimization ensures that q(\mu) closely matches the true posterior by reducing the Kullback-Leibler divergence, effectively bounding the surprise or negative log-evidence of the data under the . The variational free energy F thus serves as a tractable proxy for inference in complex, high-dimensional spaces where exact Bayesian updating is computationally infeasible. A core mechanism for this inference in the FEP is , which operates through hierarchical to propagate and minimize prediction errors across levels of a . In this framework, prediction errors are computed as \delta = s - g(\mu), where g(\mu) is the top-down prediction of sensory input generated from the current estimate of states \mu, and these errors iterative updates to the estimates via on the : \mu' = \mu - \frac{\partial F}{\partial \mu}. This process unfolds hierarchically, with bottom-up error signals ascending to higher levels to refine priors and top-down predictions descending to suppress errors at lower levels, enabling the system to iteratively refine its internal model of the world. thus realizes as a distributed, recurrent that balances sensory evidence with prior expectations without requiring centralized . The FEP further incorporates empirical Bayes to learn hierarchical priors from , allowing the construction of generative models that capture the statistical structure of the environment. Under empirical Bayes, higher-level parameters serve as hyperpriors that are themselves inferred from lower-level sensory , enabling the to adapt its expectations dynamically rather than relying on fixed priors. This approach supports the development of deep generative models where each level predicts the causes at the level below, fostering a unified of the world's causal hierarchies. Precision weighting plays a crucial role in modulating the influence of prediction errors during , where is defined as the inverse variance of noise in sensory or signals. In the FEP, \Pi weights errors such that \delta^\Pi = \Pi \delta, amplifying reliable (high-) signals and attenuating noisy (low-) ones to optimize the balance between and in Bayesian updating. This mechanism ensures that prioritizes evidence with low uncertainty, enhancing the accuracy of posterior approximations in hierarchical models.

Thermodynamics and self-organization

The free energy principle (FEP) provides a unifying framework for understanding how biological systems maintain order in non-equilibrium thermodynamic environments by minimizing variational free energy, which bounds the surprise or entropy associated with sensory inputs. This minimization process aligns with non-equilibrium thermodynamics, where systems resist the second law of thermodynamics' tendency toward disorder by actively dissipating energy to sustain low-entropy steady states. In particular, free energy minimization reduces the production of thermodynamic entropy, enabling systems to self-organize without external specification. This connection to echoes Ilya Prigogine's of dissipative structures, where open systems far from equilibrium form ordered patterns through irreversible processes that export to their surroundings. Under the FEP, biological agents achieve similar by encoding generative models of their environment, allowing them to predict and counteract surprising states that would otherwise increase internal . By minimizing , these systems conform to the principle of least action, optimizing energy dissipation while preserving structural integrity against environmental perturbations. Markov blankets formalize the thermodynamic boundaries of self-organizing s under the FEP, acting as semi-permeable interfaces that separate internal states from external influences while minimizing flows of across the boundary. These blankets ensure between internal and external dynamics, reducing fluxes that could destabilize the and thereby preserving its as a distinct in a fluctuating . For instance, in cellular , a cell's plasma membrane functions as a , where transmembrane proteins and ion channels regulate solute exchanges to minimize deviations from internal steady states, effectively implementing minimization through metabolic feedback loops that sustain low- configurations without relying on higher-level neural mechanisms.

Information theory

In the free energy principle (FEP), is defined as the self-information of a sensory state, quantified as the negative logarithm of its probability, -\ln p(s), where s represents the sensory input or state. This measure captures the improbability of an observed state under the system's , with higher indicating greater deviation from expected outcomes. The FEP posits that biological systems resist excessive to maintain their , as sustained high would imply maladaptive states. Variational free energy serves as an upper bound on , ensuring that its minimization approximates the suppression of self-information. More broadly, the expected free energy bounds the Shannon entropy H = -\int p(s) \ln p(s) \, ds, which represents the average over the distribution of sensory states. By minimizing expected free energy, systems effectively reduce in their sensory environment, aligning internal models with external realities without directly computing intractable posterior distributions. The minimization of free energy also relates to mutual information I(s; \mu), where \mu denotes the internal states or approximate posterior beliefs about sensory causes. This mutual information quantifies the shared information between sensory states and internal representations, and its maximization occurs equivalently through free energy minimization. In this framework, free energy minimization maximizes the evidence lower bound (ELBO), expressed as L = \mathbb{E}_q[\ln p(s \mid \mu)] - D_{\text{KL}}[q(\mu) \parallel p(\mu)], where q(\mu) is the approximate posterior, p(\mu) is the , and the first term reflects accuracy in reconstructing sensory while the second penalizes from priors. This bound provides a tractable surrogate for model evidence, facilitating inference by balancing representation fidelity and prior consistency. The FEP's free energy minimization embodies a trade-off akin to rate-distortion theory, where accuracy (fidelity to sensory data) is balanced against (descriptive cost of the internal model). can be decomposed as complexity minus accuracy, with measuring the relative between approximate and true posteriors, and accuracy capturing the expected log-likelihood of observations. This optimization compresses sensory information efficiently, minimizing distortion (error in representation) subject to a rate constraint (model simplicity), much like lossy coding in . In autopoietic systems under the FEP, Markov blankets enforce information closure by partitioning internal states from external ones, thereby reducing from the outer . The blanket—comprising sensory and active states—mediates all interactions, ensuring that internal dynamics evolve to minimize variational and thus external . This closure supports self-maintenance by confining information flows, allowing the system to treat external causes as effectively stationary while adapting to perturbations.

Perception and Action

Perceptual inference

In the free energy principle, perceptual inference is conceptualized as an online process that minimizes variational by updating internal beliefs about hidden causes of sensory data in . This occurs through the optimization of a density, which approximates the true posterior distribution over causes, achieved via on prediction errors—discrepancies between predicted and actual sensory inputs. These errors drive recurrent message-passing among neuronal populations, akin to schemes, enabling rapid adjustments to perceptual states without altering the underlying . The hierarchical structure of generative models underpins this , where top-down from higher levels provide contextual priors that meet bottom-up sensory signals at lower levels, resolving ambiguities in sensory data. Prediction errors propagate upwards to refine higher-level expectations, while top-down signals suppress errors at successive layers, ensuring a coherent across scales. This bidirectional flow, implemented through distinct forward and backward cortical connections, allows the to infer complex, temporally extended causes from noisy, continuous inputs. Categorization emerges from this framework as the formation of perceptual states from continuous sensory streams, facilitated by variational modes in the approximate posterior. Under approximations like the Laplace method, the recognition becomes unimodal, selecting a single mode that best explains the and effectively discretizing causes into categorical representations, such as object identities or temporal sequences. This leverages empirical priors from hierarchical models to group sensory regularities, promoting efficient encoding over exhaustive distributions. A illustrative example of perceptual inference failures is the rubber hand illusion, where synchronous visuotactile stimulation leads to the misattribution of a fake hand as one's own, reflecting an inability to resolve competing generative models due to ambiguous multisensory predictions. In this scenario, prediction errors from incongruent proprioceptive and visual cues are minimized by adopting a unified model incorporating the rubber hand, highlighting how variational inference prioritizes surprise reduction over veridicality when model evidence is balanced.

Active inference

Active inference formulates the selection of actions within the as the choice of that minimize expected free energy, thereby enabling agents to balance goal-directed behavior with uncertainty reduction. A \pi is defined as a sequence of actions over time, representing possible trajectories through the environment that an agent might pursue. The expected free energy G(\pi) under a \pi quantifies the anticipated surprise or divergence from the agent's , providing a scalar value for policy evaluation. The expected free energy is expressed as G(\pi) = \sum_{\tau > t} \mathbb{E}_{Q} [\ln P(o_{\tau}|s_{\tau}) - \ln Q(s_{\tau}|\pi)], where Q denotes the agent's approximate posterior beliefs, P is the , the sum is over future time steps \tau after current time t, s_{\tau} are environmental states, and o_{\tau} are sensory outcomes. This functional decomposes into two key terms: the (or pragmatic) component, which captures the expected from preferred outcomes and drives of known goals, and the (or epistemic) component, which measures in predictions and encourages . Policies are selected probabilistically via a over the negative expected , \pi^* = \softmax(-\beta [G](/page/G)(\pi)), where \beta is an inverse controlling the of choice; this mechanism ensures that agents prefer policies minimizing future surprise while incorporating precision weighting. The epistemic term in particular relates to , as its minimization promotes novelty-seeking behaviors that resolve about the , effectively driving epistemic . For instance, in simulated tasks, agents under active navigate ambiguous environments by selecting policies that first gather about resource locations () before exploiting them (), thereby minimizing long-term expected without exhaustive search. This approach integrates seamlessly with perceptual by using updated sensory beliefs to inform policy evaluation.

Neuroscience Applications

Learning and memory

In the free energy principle (FEP), perceptual learning emerges as a slow process that minimizes variational by updating hyperparameters, which encode prior expectations about the environment's . These updates occur through mechanisms akin to Hebbian , where synaptic strengths adjust to reduce prediction errors accumulated over multiple exposures, effectively refining the generative model's parameters to better anticipate sensory data. This hierarchical optimization ensures that higher-level priors constrain lower-level inferences, promoting efficient encoding of environmental regularities without rapid fluctuations. Empirical support for these learning processes comes from in vitro experiments with networks of rat cortical neurons, which demonstrate that neural activity and conform to free-energy minimization during tasks, quantitatively validating FEP predictions as of 2023. Memory formation under the FEP is conceptualized as the development and maintenance of generative models that simulate sensory trajectories. , in particular, captures sequences of states and transitions in these models, allowing agents to reconstruct past experiences by minimizing associated with temporal predictions. , meanwhile, sustains high-precision beliefs about current states during inference, enabling the temporary prioritization of relevant predictions to resolve in ongoing tasks. Synaptic plasticity provides the neurobiological substrate for these processes, with long-term potentiation (LTP) and (LTD) driven by prediction errors that act as gradients for free energy minimization. In this framework, error signals modulate synaptic weights such that coincident pre- and postsynaptic activity strengthens connections when predictions align with outcomes, mirroring Hebbian rules but grounded in variational Bayesian updates. This dependency ensures plasticity adapts the to minimize surprise over time-scales relevant to learning. An illustrative example is habit formation, where repeated actions flatten the expected free energy landscape, reducing epistemic value (information gain) while preserving pragmatic value (goal attainment). Through active , agents learn habitual policies by caching low-entropy trajectories that reliably minimize future free energy, transitioning from exploratory to exploitative behavior.

Attention and precision

In the free energy principle, emerges as a for optimizing the of errors during perceptual . , denoted as \psi, is formally defined as the inverse of the variance \sigma^2 of sensory or errors (\psi = 1/\sigma^2), which quantifies the reliability of incoming signals. This acts as a modulator on error-signaling neurons, amplifying responses from more trustworthy channels while downweighting unreliable ones, thereby refining the hierarchical minimization of variational . By dynamically adjusting this , the prioritizes information that best reduces , aligning perceptual processes with environmental demands. Salience within this framework bifurcates into endogenous and exogenous forms of control. Endogenous is goal-driven, originating from top-down priors that enhance toward task-relevant features, such as voluntarily shifting focus in a search task. In contrast, exogenous is stimulus-driven, triggered by bottom-up surprises like sudden visual onsets that capture involuntarily. This distinction allows the system to balance internal expectations with external perturbations, ensuring efficient in schemes. Neuromodulatory systems underpin these precision dynamics. Acetylcholine primarily tunes precision by increasing synaptic gain on prediction error units, promoting selective attention through sharpened sensory representations and associated gamma oscillations. Dopamine, meanwhile, modulates salience in error signaling, particularly by encoding the motivational value of predictions, which influences the weighting of errors in value-sensitive contexts. A key example of precision overload is the , where rapid successive stimuli lead to missed detection of a second target due to insufficient allocation in hierarchical ; the ongoing optimization for the first stimulus temporarily saturates the system's to update for subsequent inputs. This phenomenon highlights how constraints in temporal processing can disrupt , as the struggles to balance error minimization across layers.

Broader Extensions

Artificial intelligence and robotics

The free energy principle (FEP) has been applied in and to develop autonomous agents that minimize variational free energy through active inference, enabling adaptive perception-action loops in engineered systems. In robotics, active inference frameworks leverage the expected free energy (EFE) to balance exploitation and , allowing robots to select actions that reduce while pursuing goals in dynamic environments. Post-2020 implementations have demonstrated active on humanoid platforms like the , where generative models minimize prediction errors for body perception and reaching tasks, outperforming traditional methods in noisy sensory conditions. For instance, variational autoencoder-based controllers integrated with active have enabled the to perform visuomotor tasks, such as object tracking, by optimizing for epistemic foraging in real-world settings. These advancements highlight FEP's role in creating robust, self-calibrating robotic systems capable of without explicit reward signals. In AI alignment research, the FEP provides a foundation for safe AGI by framing alignment as surprise minimization, where agents inherently avoid catastrophic states that increase free energy, thus promoting systemic safety in multi-agent interactions. Discussions from 2023 to 2025 emphasize how EFE-driven policies can mitigate risks in agentic systems, ensuring robustness against misalignment by prioritizing predictive coherence over short-term gains. This approach has been proposed as a metric for evaluating AGI safety, reducing the likelihood of unintended harmful behaviors through variational bounds on risk. The FEP underpins curiosity-driven learning in () and variational autoencoders (VAEs), where intrinsic rewards are derived from hidden state uncertainty to encourage exploration in sparse-reward environments. A 2024 framework integrates FEP-based curiosity with deep , using KL divergence between predictive distributions as a reward signal to guide agents toward informative states without external supervision. This has enabled efficient solving of robotic tasks from pixel inputs via active inference, where planning optimizes long-horizon policies by treating action selection as variational inference over generative models. Such methods outperform standard baselines in adaptability, as seen in simulations of and . Recent 2025 models explore in through FEP-based , positing that "inner screens" of predictive processing enable self-referential by minimizing in hierarchical . These architectures simulate attentional shifts as precision-weighted updates, allowing systems to model their own "experiences" via EFE, with applications in generative agents that exhibit emergent . For example, implementations in demonstrate how FEP constrains attentional mechanisms to produce coherent, non-hallucinating outputs akin to human-like focus.

Biological and physical systems

The free energy principle (FEP) has been extended to intracellular processes, where it explains the of compartmentalization as a mechanism to minimize variational by preserving integrity. In 2024 studies, researchers demonstrated that systems governed by the FEP naturally induce the formation of intracellular compartments, such as organelles, through the minimization of at system boundaries, which effectively reduces by constraining internal states against external perturbations. This process amplifies at cellular boundaries, leading to self-organized structures that maintain non-equilibrium steady states essential for cellular persistence. For instance, bioelectric signaling within cells acts as a that propagates compartmentalization, aligning with the FEP's that adaptive systems states to bound . In , the FEP frames as a process that favors robust phenotypes capable of minimizing over generations, ensuring survival in fluctuating environments. Karl Friston's work posits that biological systems, including evolving populations, maintain their form through Markov blankets that partition internal and external states, with selection pressures acting to optimize phenotypes that resist and . This perspective, developed onward from Friston's foundational formulations, views as an implicit minimization, where phenotypes that robustly predict and adapt to environmental contingencies are selected, thereby perpetuating self-organizing traits across scales from molecules to organisms. Subsequent extensions reinforce that evolutionary dynamics align with the FEP by prioritizing variational bounds on phenotypic variability, promoting and adaptability without invoking teleological assumptions. The FEP also applies to physical systems, particularly non-equilibrium dissipative structures that minimize fluxes to sustain ordered states far from . In such systems, like dissipative particle dynamics, particles exchange energy and matter with their environment, with flows governed by on a free energy functional that bounds . This minimization ensures the persistence of structures, such as self-assembling colloids or reaction-diffusion patterns, by constraining trajectories to low-surprise attractors, akin to biological but in purely physical contexts. The principle's formulation for random dynamical systems highlights how these fluxes arise from ergodic partitioning, linking Prigogine's dissipative structures to variational inference without requiring biological specificity. A key example of the FEP in non-neural biology is bacterial chemotaxis, interpreted as active inference where cells minimize expected without cognitive processes. Recent extensions from 2023 to 2025 model bacterial navigation toward nutrients as under the FEP, with run-and-tumble motility adjusting policies to reduce sensory prediction errors and epistemic value. In , for instance, temporal and spatial gradient sensing embodies active inference by updating internal models of attractant concentrations, thereby bounding surprise through directed movement that preserves the cell's non-equilibrium state. This molecular implementation demonstrates how the FEP unifies and at the simplest biological scales, with mutual constraints from exteroceptive (chemical) and interoceptive (metabolic) signals driving adaptive behavior.

Criticisms and Debates

Falsifiability concerns

Critics have argued that the free energy principle (FEP) lacks falsifiability because it can retrospectively accommodate any observed behavior by framing it as an instance of free energy minimization, rendering it tautological rather than empirically testable. More recent critiques have amplified these concerns, portraying the FEP as a pseudoscientific construct that evades refutation through terminological maneuvers. In a 2025 preprint, Madhur Mangalam described the FEP as an "emperor's new pseudo-theory," arguing that core concepts like "surprise" are redefined so broadly that contradictory evidence—such as systems failing to minimize free energy—can be reinterpreted as successful minimization under alternative model specifications. Mangalam contended that this adaptability undermines the FEP's scientific status, as it permits proponents to claim universality without genuine vulnerability to empirical disconfirmation. Proponents of the FEP counter that such criticisms mischaracterize as a descriptive empirical when it is fundamentally normative, akin to physical laws like the principle of least action, which guide behavior without being falsifiable in observation. Karl Friston, the FEP's primary architect, has emphasized that provides a mathematical prescription for optimal inference and control, testable through specific, falsifiable predictions derived from it—such as the modulation of sensory precision weighting in perceptual tasks or the emergence of hierarchical in neural responses. These derivative models, rather than itself, bear the empirical burden, allowing targeted experiments to validate or refute FEP-based hypotheses. A related point of contention involves the empirical detectability of , which partition systems into internal, external, and boundary states under the FEP and are essential for defining autonomous agents. Critics argue that in complex, real-world biological systems, identifying these blankets is practically infeasible due to entangled interactions and incomplete observational data, potentially rendering the FEP's application unverifiable. For instance, Vicente Raja and colleagues in 2021 highlighted how the "Markov blanket trick"—imposing such partitions analytically—expands the FEP's scope to non-adaptive systems without clear empirical demarcation, raising questions about whether blankets can be robustly detected beyond idealized simulations. This debate underscores ongoing philosophical scrutiny over the FEP's boundaries as a scientific construct.

Empirical and conceptual challenges

The free energy principle (FEP) faces empirical gaps in its applications, primarily due to reliance on correlational evidence from techniques like fMRI rather than direct causal or behavioral tests. While FEP-inspired models have been fitted to fMRI data to explain perceptual inference and active control, these analyses often highlight predictive alignments without manipulating underlying mechanisms or verifying outcomes through targeted interventions. Independent empirical verification remains scarce, as the framework's mathematical complexity hinders straightforward testing against alternative hypotheses. For instance, experiments provide some quantitative support for FEP predictions on , but they underscore challenges in identifying generative models at cellular levels and ensuring refutability beyond descriptive fits. Recent consciousness models grounded in FEP, such as those integrating active for minimal phenomenal , lack robust behavioral validation as of 2025, though emerging whole-brain simulations have begun incorporating neural data from states to link predictions to observable correlates. These models propose that emerges from minimization in hierarchical predictive processing, yet they rely on theoretical simulations or phenomenological reports without fully linking to observable behavioral metrics like under or adaptive responses in dynamic tasks. This gap persists because FEP's abstraction allows flexible interpretations that fit diverse data but evade specific falsification through behavioral paradigms. Conceptually, the FEP's assumption of —positing that systems sample all states over time to reach a non-equilibrium —struggles in non-steady biological contexts like transient neural during learning or . Early formulations required ergodic flows for free energy bounds, but critiques highlight that real brains operate in path-dependent, non-ergodic regimes where historical contingencies disrupt convergence. Refinements from 2023 to 2025, including thermodynamic extensions to , address this by relaxing ergodicity for variational inference in fluctuating environments, though full resolution remains ongoing. Scaling FEP to real brains presents further hurdles, particularly in capturing hierarchical depth beyond simplified models. Computational implementations often limit layers to a few levels for tractability, whereas cortical hierarchies span multiple spatiotemporal scales, complicating precise mapping of precision-weighted predictions across depths. Integration with for multi-agent interactions, such as social inference, is incomplete, as FEP extensions to bounded-rational agents fail to fully incorporate incomplete information or strategic equilibria without approximations. Looking ahead, 2024–2025 advancements in expected () for emphasize the need for interdisciplinary empirical tests to bridge these gaps, such as combining neural recordings with behavioral assays in ecological settings. In , EFE-based shows promise but encounters analogous scaling issues in high-dimensional environments.

References

  1. [1]
  2. [2]
    The free-energy principle: a unified brain theory? - Nature
    Jan 13, 2010 · Karl Friston shows that different global brain theories all describe principles by which the brain optimizes value and surprise.
  3. [3]
    Life as we know it | Journal of The Royal Society Interface
    Sep 6, 2013 · This paper presents a heuristic proof (and simulations of a primordial soup) suggesting that life—or biological self-organization—is an ...Missing: book | Show results with:book
  4. [4]
    Predictive coding in the visual cortex: a functional interpretation of ...
    Using a hierarchical model of predictive coding, we show that visual cortical neurons with extra-classical RF properties can be interpreted as residual error ...
  5. [5]
    The free energy principle made simpler but not too simple
    Jun 19, 2023 · This paper provides a concise description of the free energy principle, starting from a formulation of random dynamical systems in terms of ...
  6. [6]
  7. [7]
    Free-energy and the brain - PMC - PubMed Central - NIH
    This paper has three sections. In the first, we lay out the theory behind the free-energy principle, starting from a selectionist standpoint and ending with ...
  8. [8]
    None
    ### Summary of the Free Energy Principle from the Paper
  9. [9]
    A Free Energy Principle for Biological Systems - PMC
    This paper describes a free energy principle that tries to explain the ability of biological systems to resist a natural tendency to disorder.
  10. [10]
    [PDF] The free-energy principle: a unified brain theory?
    Jan 13, 2010 · This paper introduces the central role of generative models and variational approaches to hierarchical self-supervised learning and relates this ...
  11. [11]
    None
    ### Summary of Predictive Coding under the Free-Energy Principle
  12. [12]
    Predictive coding under the free-energy principle - Journals
    May 12, 2009 · Predictive coding under the free-energy principle. Karl Friston. Karl Friston. k.friston@fil.ion.ucl.ac.uk · Google Scholar · Find this author ...Missing: Friston | Show results with:Friston
  13. [13]
    Attention, Uncertainty, and Free-Energy - PMC - PubMed Central - NIH
    These range from ordinary least squares to advanced variational deconvolution schemes (see Friston, 2008). For example, the schemes used to invert ...
  14. [14]
  15. [15]
  16. [16]
  17. [17]
    Generalised free energy and active inference - PMC - PubMed Central
    This is variational free energy (Beal 2003; Dayan et al. 1995; Friston 2003) which depends upon specifying a generative model of how data are caused. This ...Introduction · Table 1 · Comparison Of Active...<|control11|><|separator|>
  18. [18]
    [PDF] Active Inference: A Process Theory - Free Energy Principle
    This article describes a process theory based on active inference and be- lief propagation. Starting from the premise that all neuronal processing.
  19. [19]
    [PDF] Generalised free energy and active inference - bioRxiv
    Apr 23, 2018 · G π is the expected free energy, conditioned on a policy. It is ... Friston K, Buzsaki G (2016) The Functional Anatomy of Time: What ...
  20. [20]
    Scene Construction, Visual Foraging, and Active Inference - Frontiers
    Active inference assumes that the only self-consistent prior belief is that our actions will minimize free energy; in other words, we (believe we) will behave ...
  21. [21]
    Active Inference, epistemic value, and vicarious trial and error - PMC
    This section describes four simulations of VTE behavior using a Bayesian (Active Inference) formulation of choice and foraging (Friston et al. 2013, 2014, 2015) ...
  22. [22]
    None
    ### Summary of Perceptual Inference in the Free Energy Principle
  23. [23]
    A free energy principle for the brain - ScienceDirect.com
    In this paper, we show these perceptual processes are just one aspect of emergent behaviours of systems that conform to a free energy principle.Missing: original | Show results with:original
  24. [24]
    Active inference and learning - ScienceDirect.com
    This paper offers an active inference account of choice behaviour and learning. It focuses on the distinction between goal-directed and habitual behaviour.
  25. [25]
    Working memory, attention, and salience in active inference - Nature
    Nov 7, 2017 · We use active inference to demonstrate how attention and salience can be disambiguated in terms of message passing between populations of neurons.
  26. [26]
    Bayesian Mechanics of Synaptic Learning Under the Free-Energy ...
    The free-energy principle (FEP) advocated by Karl Friston explicates the local, recurrent, and self-supervised cognitive dynamics of the brain's higher-order ...
  27. [27]
    Caching mechanisms for habit formation in Active Inference - PMC
    In other words, agents can learn habits through minimising their expected free energy – and then select them from an augmented repertoire of policies, in a ...
  28. [28]
    The free-energy principle: a rough guide to the brain? - ScienceDirect
    A free energy principle for the brain. J. Physiol. Paris. (2006) ; The Bayesian brain: the role of uncertainty in neural coding and computation. Trends Neurosci.
  29. [29]
    Predictive coding under the free-energy principle - Journals
    May 12, 2009 · ... perceptual inference, eLife, 10.7554/eLife.11476, 5. Maes P (2016) ... Rubber Hand Illusion, Neural Plasticity, 10.1155/2016/8307175 ...
  30. [30]
    A key role of the hippocampal P3 in the attentional blink
    May 1, 2023 · The attentional blink (AB) refers to an impaired identification of target stimuli (T2), which are presented shortly after a prior target ...
  31. [31]
    How Active Inference Could Help Revolutionise Robotics - MDPI
    In this paper, we explain how active inference—a well-known description of sentient behaviour from neuroscience—can be exploited in robotics ...Missing: iCub post-
  32. [32]
  33. [33]
  34. [34]
  35. [35]
    A reply to Byrnes on the Free Energy Principle - AI Alignment Forum
    Mar 3, 2023 · And we don't know in what direction the AGI architecture will evolve (whether via self-modification or not). The FEP ontology might be one of ...Missing: minimization | Show results with:minimization
  36. [36]
    Intrinsic Rewards for Exploration without Harm from Observational ...
    May 13, 2024 · Based on the Free Energy Principle (FEP), this paper proposes hidden state curiosity, which rewards agents by the KL divergence between the ...Missing: variational autoencoders
  37. [37]
    How do inner screens enable imaginative experience? Applying the ...
    We present a model of imaginative experience that is compliant with the free-energy principle (FEP). ... Depression thus entails altered precision weighting ...
  38. [38]
    The Emperor's New Pseudo-Theory: How the Free Energy Principle ...
    The Emperor's New Pseudo-Theory: How the Free Energy Principle Ransacked Neuroscience ... Murphy et al. (2024). Natural language syntax. complies with the free-.
  39. [39]
    (PDF) The Markov Blanket Trick: On the Scope of the Free Energy ...
    Mar 25, 2021 · The free energy principle (FEP) has been presented as a unified ... Markov blankets were first proposed by Judea Pearl (1988) in the c ...Missing: falsifiability | Show results with:falsifiability<|control11|><|separator|>
  40. [40]
    Free energy: a user's guide | Biology & Philosophy
    Jul 20, 2022 · Second, we describe the historical trajectory of the framework and highlight its novel features (“A brief history of the free energy principle” ...
  41. [41]
    Experimental validation of the free-energy principle with in vitro ...
    Aug 7, 2023 · ... variational Bayes formation. a Schematics of the experimental setup ... Friston, K. J. The free-energy principle: a unified brain theory?
  42. [42]
  43. [43]
    [PDF] The Free Energy Principle: A Unifying Framework Connecting ...
    May 16, 2025 · Abstract. This paper deeply explores Karl Friston's Free Energy Principle (FEP) as a meta-theory with significant potential for unifying our ...
  44. [44]
    How particular is the physics of the free energy principle?
    The free energy principle (FEP) states that any dynamical system can be interpreted as performing Bayesian inference upon its surrounding environment.Missing: empirical detectability
  45. [45]
    Free energy and inference in living systems | Interface Focus
    Apr 14, 2023 · This study presents an FE minimization theory overarching the essential features of both the thermodynamic and neuroscientific FE principles.
  46. [46]
    An Overview of the Free Energy Principle and Related Research
    Apr 23, 2024 · This principle states that the processes of perception, learning, and decision making—within an agent—are all driven by the objective of “ ...
  47. [47]
    [PDF] Toward a Theory of Interactions Between Boundedly-Rational Agents
    We propose that free- energy-based models provide a promising framework for studying cooperation and conflict among boundedly-rational players with incomplete ...
  48. [48]
    [PDF] Expected Free Energy-based Planning as Variational Inference - arXiv
    EFE-based planning, grounded in the Free Energy Principle, minimizes Expected Free Energy (EFE) and is shown to arise from variational inference.