Abstract
Most theories of behavior posit that agents tend to maximize some form of reward or utility. However, animals very often move with curiosity and seem to be motivated in a reward-free manner. Here we abandon the idea of reward maximization and propose that the goal of behavior is maximizing occupancy of future paths of actions and states. According to this maximum occupancy principle, rewards are the means to occupy path space, not the goal per se; goal-directedness simply emerges as rational ways of searching for resources so that movement, understood amply, never ends. We find that action-state path entropy is the only measure consistent with additivity and other intuitive properties of expected future action-state path occupancy. We provide analytical expressions that relate the optimal policy and state-value function and prove convergence of our value iteration algorithm. Using discrete and continuous state tasks, including a high-dimensional controller, we show that complex behaviors such as “dancing”, hide-and-seek, and a basic form of altruistic behavior naturally result from the intrinsic motivation to occupy path space. All in all, we present a theory of behavior that generates both variability and goal-directedness in the absence of reward maximization.
Similar content being viewed by others
Introduction
Natural agents are endowed with a tendency to move, explore, and interact with their environment1,2. For instance, human newborns unintentionally move their body parts3, and 7–12-month-old infants spontaneously babble vocally4 and with their hands5. Exploration and curiosity are major drives for learning and discovery through information-seeking6,7,8. These behaviors seem to elude a simple explanation in terms of extrinsic reward maximization. However, intrinsic motivations, such as curiosity, push agents to visit new states by performing novel courses of action, which helps learning and the discovery of even larger rewards in the long run9,10. Therefore, it has been argued that exploration and curiosity could arise as a consequence of seeking extrinsic reward maximization by endowing agents with the necessary inductive biases to learn in complex and ever-changing natural environments11,12.
While most theories of rational behavior do posit that agents are reward maximizers13,14,15,16, very few of us would agree that the sole goal of living agents is maximizing money gains or food intake. Indeed, expressing excessive emphasis on these types of goals is usually seen as a sign of psychological disorders17,18. Further, setting a reward function by design as the goal of artificial agents is, more often than not, arbitrary14,19,20,21, leading to the recurrent problem faced by theories of reward maximization of defining what rewards are22,23,24,25,26. In some cases, like in artificial games, rewards can be unambiguously defined, such as number of collected points or wins27. However, in most situations defining rewards is task-dependent, non-trivial, and problematic. For instance, a vacuum cleaner robot could be designed to either maximize the weight or volume of dust collected, energy efficiency, or a weighted combination of them28. In more complex cases, companies can aim at maximizing profit, but without a suitable innovation policy, profit maximization can be self-defeating29. Even when a task is well-defined from the perspective of an external observer, the actions and states of natural agents are not always compatible with a deterministic maximization of rewards and are consistently found to be highly variable30,31,32. Natural agents are also sensitive to the uncertainty in or about the environment, behaving in ways that are inconsistent with strict reward maximization in familiar environments33,34. While there are isolated attempts to understand the function and mechanisms of behavioral variability35, and risk sensitivity in uncertain environments36,37,38, general principles for the importance of stochasticity in so-called goal-directed behavior are lacking.
Here, we abandon the idea that the goal is maximizing extrinsic rewards and that movement over space is a means to achieve this goal. Instead, we adopt the opposite view, inspired by the nature of our intrinsic drives: we propose that the objective is to maximally occupy action-state path space, understood in a broad sense, in the long term. We call this principle the maximum occupancy principle (MOP), which posits that the goal of agents is to generate all sorts of variable behaviors and visit, on average, as much space (action-state paths) as possible in the future. According to MOP, extrinsic rewards serve to obtain the energy necessary to move in order to occupy action-state space, they are not the goals per se. The usual exploration-exploitation tradeoff39, therefore, disappears: agents that seek to occupy space “solve” this issue naturally because they care about rewards only as a means to an end. Furthermore, in this sense, surviving is only preferred because it is needed to keep visiting action-state path space. Our theory provides a rational account of exploratory and curiosity-driven behavior where the problem of defining a reward function vanishes, and captures the variability of behavior40,41,42,43,44,45 by taking it as a principle.
In this work, we model a MOP agent interacting with the environment as a Markov decision process (MDP) where the intrinsic, immediate reward is the occupancy of the next action-state visited, which is largest when performing an uncommon action and visiting a rare state—there are no extrinsic rewards (i.e., no task is defined) that drive the agent. We show that (weighted) action-state path entropy is the only measure of occupancy consistent with additivity per time step, positivity, and smoothness. Due to the additivity property, the value of being in a state, defined as the expected future time-discounted action-state path entropy, can be written in the form of a Bellman equation, which has a unique solution that can be found with an iterative map. Following this entropy objective leads to agents that seek variability while being sensitive to the constraints imposed by the agent-environment interaction on the future path availability. We demonstrate in various simulated experiments with discrete and continuous state and action spaces that MOP generates complex behaviors that, to the human eye, look genuinely goal-directed and playful, such as hide-and-seek in a prey-predator problem, dancing of a cartpole, a basic form of altruism in an agent-and-pet example, and rich behaviors in a high-dimensional quadruped.
MOP builds over an extensive literature on entropy-regularized reinforcement learning (RL)46,47,48,49,50,51,52,53,54,55,56 or pure entropic objectives57,58,59,60,61,62. This body of work emphasizes the regularization benefits of entropy for learning, but extrinsic rewards still serve as the major drive of behavior, and arbitrary mixtures of action-state entropy are rarely considered56. Our work also relates to reward-free theories of behavior. These typically minimize prediction errors63,64,65,66,67,68, seek novelty69,70,71, or maximize data compression72, and therefore the major behavioral driver depends on the agent’s experience with the world. On the other hand, MOP agents find the action-states that lead to high future occupancy “interesting”, regardless of experience. There are two other approaches that sit closer to this description, one maximizing mutual information between actions and future states (empowerment, MPOW)20,73,74,75, and the other minimizing the distance between the actual and a desired state distribution (free energy principle, FEP)76,77. We show that both MPOW and FEP tend to collapse to deterministic policies with little behavioral variability. In contrast, MOP results in lively and seemingly goal-directed behavior by taking behavioral variability and the constraints of embodied agents as principles.
Results
Maximum occupancy principle
We model an agent as a finite action-state MDP in discrete time. The policy π describes the probability π(a∣s) of performing action a, from some set \({{{{{{\mathcal{A}}}}}}}(s)\), given that the agent is in state s at some time step, and \(p({s}^{{\prime} }| s,a)\) is the transition probability from s to a successor state \({s}^{{\prime} }\) in the next time step given that action a is performed. Starting at t = 0 in state s0, an agent performing a sequence of actions and experiencing state transitions τ ≡ (s0, a0, s1,...,at, st+1,...) gets a return defined as
with action and state weights α > 0 and β ≥ 0, respectively, and discount factor 0 < γ < 1. A larger return is obtained when, from st, a low-probability action at is performed and followed by a low-probability transition to a state st+1. Therefore, maximizing the return in Eq. (1) favors “visiting” action-states (at, st+1) with a low transition probability. From st+1, another low-probability action-state transition is preferred, and so on, such that low-probability trajectories τ are more rewarding than high-probability ones. Thus, the agent is pushed to visit action-states that are rare or “unoccupied”, implementing the intuitive notion of MOP. Due to the freedom to choose action at given state st and the uncertainty of the resulting next state st+1, apparent in Eq. (1), the term “action-states” used here is more natural than “state-actions”. We stress that this return is purely intrinsic, namely, there is no extrinsic reward that the agent seeks to maximize. We define intrinsic rewards as any reward signal that depends on the policy or the state transition probability, and therefore it can change with the course of learning as the policy is improved, or the environment is learned. An extrinsic reward is the complementary set of reward signals: any function R(s, a) that is both policy-independent and transition probability-independent, and therefore it does not change with the course of improving the policy or learning the state transition probability of the environment.
The agent is assumed to optimize the policy π to maximize the state-value Vπ(s), defined as the expected return
given the initial condition s0 = s and following policy π, that is, the expectation is over the at ~ π(⋅∣st) and st+1 ~ p(⋅∣st, at), t ≥ 0. In the last identity, we have rewritten the expectations of the terms in Eq. (1) as a discounted and weighted sum of action and successor state conditional entropies \({{{{{{\mathcal{H}}}}}}}(A| s)=-{\sum}_{a}\pi (a| s)\ln \pi (a| s)\) and \({{{{{{\mathcal{H}}}}}}}({S}^{{\prime} }| s,a)=-{\sum}_{{s}^{{\prime} }}p({s}^{{\prime} }| s,a)\ln p({s}^{{\prime} }| s,a)\), respectively, averaged over previous states and actions.
We define a MOP agent as the one that optimizes the policy to maximize the state value in Eq. (2). The entropy representation in Eq. (2) of MOP has several implications. First, agents prefer regions of state space that lead to a large number of successor states (Fig. 1a) or larger number of actions (Fig. 1b). Second, death (absorbing) states where only one action-state (i.e., “stay”) is available forever are naturally avoided by a MOP agent, as they promise zero future action and state entropy (Fig. 1c—hence absorbing states s+ have zero state-value, Vπ(s+) = 0). Therefore, our framework implicitly incorporates a survival instinct. Finally, regions of state space where there are “rewarding” states that increase the capacity of the agent to visit further action-states (such as filling an energy reservoir) are more frequently visited than others (Fig. 1d).
We found that maximizing the discounted action-state path entropy in Eq. (2) is the only reasonable way of formalizing MOP, as it is the only measure of action-state path occupancy in Markov chains consistent with the following intuitive conditions (see Methods and Supplemental Sec. A): if a path τ has probability p, visiting it results in an occupancy gain C(p) that (i) decreases with p and (ii) is first-order differentiable. Condition (i) implies that visiting a low-probability path increases occupancy more than visiting a high-probability path, and our agents should tend to occupy “unoccupied” path space; condition (ii) requires that the measure should be smooth. We also ask that (iii) the occupancy of paths, defined as the expectation of occupancy gains over paths given a policy, is the sum of the expected occupancies of their subpaths (additivity condition). This last condition implies that agents can accumulate occupancy over time by keeping visiting low-probability action-states, but the accumulation should be consistent with the Markov property of the decision process. These conditions are similar but not exactly the same as those used to derive Shannon’s information measure78 (see Methods and Supplemental Sec. A).
Optimal policy and state-value function
The state-value Vπ(s) in Eq. (2) can be recursively written using the values of successor states through the standard Bellman equation
where the sum is over the available actions a from state s, \({{{{{{\mathcal{A}}}}}}}(s)\), and over the successor states \({s}^{{\prime} }\) given the performed action at state s. The number of actions and state transitions available does not need to be the same for all states s. In particular, absorbing states s+ have only the “stay” action available, leading to zero action and state transition entropies, and therefore they have zero state-value Vπ(s+) = 0 regardless of the policy.
The optimal policy π* that maximizes the state value is defined as \({\pi }^{*}=\arg {\max }_{\pi }{V}_{\pi }\) and the optimal state value is
where the maximization is with respect to the {π(⋅∣⋅)} for all actions and states. To obtain the optimal policy, we first determine the critical points of the expected return Vπ(s) in Eq. (3) using Lagrange multipliers (see Methods and Supplemental Sec. B). The optimal state-value V*(s) is found to obey the non-linear self-consistency set of equations
where Z(s) is the partition function, defined by substitution, and the critical policy satisfies
Note that the optimal policy in MOP only depends on a single parameter, the ratio β/α and that the optimal state-value function is simply scaled by α. We find that the solution to the non-linear system of Eqs. (5) is unique and, moreover, the unique solution is the absolute maximum of the state-values over all policies (Supplemental Sec. C).
To determine the actual value function from such non-linear set of equations, we derive an iterative map, a form of value iteration that exactly incorporates the optimal policy at every step. Defining \({z}_{i}=\exp ({\alpha }^{-1}\gamma V({s}_{i}))\), pijk = p(sj∣si, ak) and \({{{{{{{\mathcal{H}}}}}}}}_{ik}={\alpha }^{-1}\beta {{{{{{\mathcal{H}}}}}}}({S}^{{\prime} }| {s}_{i},{a}_{k})\), Eq. (5) can be turned into the iterative map
for n ≥ 0 and with initial conditions \({z}_{i}^{(0)} > 0\). Here, the matrix with coefficients wik ∈ {0, 1} indicates whether action ak is available at state si (wik = 1) or not (wik = 0), and j extends over all states, with the understanding that if a state sj is not a possible successor from state si after performing action ak then pijk = 0. We find that the infinite series \({z}_{i}^{(n)}\) defined in Eq. (7) converges to a finite limit \({z}_{i}^{(n)}\to {z}_{i}^{\infty }\) regardless of the initial condition in the positive first orthant, and that \({V}^{*}({s}_{i})=\alpha {\gamma }^{-1}\ln {z}_{i}^{\infty }\) is the optimal state-value function, which solves Eq. (5) (Supplemental Sec. C). Iterative maps similar to Eq. (7) have been studied before46,79, subsequently shown to have uniqueness80 and convergence guarantees54,81 in the absence of state entropy terms. A summary of results and particular examples can be found in Supplemental Sec. D.
We note that in the definition of return in Eq. (2) we could replace the absolute action entropy terms \({{{{{{\mathcal{H}}}}}}}(A| s)\) by relative entropies of the form \(-{D}_{{{{{{{\rm{KL}}}}}}}}(\pi (a| s)| | {\pi }_{0}(a| s))={\sum}_{a}\pi (a| s)\ln ({\pi }_{0}(a| s)/\pi (a| s))\), as in KL-regularization46,50,55,79, but in the absence of any extrinsic rewards. In this case, one obtains an equation identical to (7) where the coefficients wik are simply replaced by π0(ak∣si), one-to-one. This apparently minor variation uncovers a major qualitative difference between absolute and relative action entropy objectives: as ∑kwik ≥ 1, absolute entropy-seeking favors visiting states with a large action accessibility, that is, where the sum ∑kwik and thus the argument of Eq. (7) tends to be largest. In contrast, as ∑kπ0(ak∣si) = 1, maximizing relative entropies provides no preference for states s with a large number of accessible actions \(| {{{{{{\mathcal{A}}}}}}}(s)|\). This happens even if the default policy is uniform in the actions, as then the immediate intrinsic return becomes \(-{D}_{{{{{{{\rm{KL}}}}}}}}(\pi (a| s)| | {\pi }_{0}(a| s))={{{{{{\mathcal{H}}}}}}}(A| s)-\ln | {{{{{{\mathcal{A}}}}}}}(s)|\), instead of \({{{{{{\mathcal{H}}}}}}}(A| s)\). The negative logarithm penalizes visiting states with large number of actions, which is the opposite goal to occupying action-state path space (see details in Supplemental Sec. F).
MOP agents quickly fill physical space
In very simple environments with high symmetry and little constraints, like open space, maximizing path occupancy amounts to performing a random walk that chooses at every step any available action with equal probability. However, in realistic environments where space is not homogeneous, where there are energetic limitations for moving, or where there are absorbing states, a random walk is no longer optimal. To illustrate how interesting behaviors arise from MOP in these cases, we first tested how a MOP agent moving in a 4-room and 4-food-sources environment (Fig. 2a) compares in occupying physical space to a random walker (RW) and to a reward-seeking agent (R agent) (see Methods for definitions of the agents). The definitions of the three agents are identical in most ways. They have nine possible movement actions, including not moving; they all have an internal state corresponding to the available energy, which reduces by one unit at every time step and gets increased by a fixed amount (food gain) whenever a food source is visited; and they can move as long as their energy is non-zero. The total state space is the Cartesian product between physical space and internal energy. The agents differ however in their objective function. The MOP agent has a reward-free objective and implements MOP by maximizing path action entropy, Eq. (2). In contrast, the R agent maximizes future discounted reward (in this case, food), and displays stochastic behavior through an ϵ-greedy action selection, with ϵ matched to the survival of the MOP agent (Supplemental Sec. E and Fig. E.2a). Finally, the RW is simply an agent that in each state takes a uniformly random action from the available actions at that state.
We find that the MOP agent generates behaviors that can be dubbed goal-directed and curiosity-driven (Supplementary Movie 1). First, by storing enough energy in its reservoir, the agent reaches far, entering the four rooms in the long term (Fig. 2b, left panel), and visiting every location of the arena except when food gain is small (Fig. 2d, blue line). In contrast, the R agent lingers over one of the food sources for most of the time (Fig. 2b, middle panel; Supplementary Movie 1). Although its ϵ-greedy action selection allows for brief exploration of other rooms, the R agent does not on average visit the whole arena (Fig. 2d, orange line). Finally, the RW dies before it has time to visit a large fraction of the physical space (Fig. 2b, right panel). These differences hold for a large range of food gains (Fig. 2d). The MOP agent, while designed to generate variability, is also capable of deterministic behavior: when its energy is low, it moves toward the food sources with little to no variability, a distinct mark of goal-directedness (Fig. 2c, corner spots show that only one action is considered by optimal policy).
We next considered a slightly more complex environment where actions in one of the rooms lead to uniformly stochastic transitions to any of the neighboring locations (noisy room—a spatial version of the noisy TV problem66,82). A stochastic region in the environment can reflect uncertainty about this region (e.g., due to agents with limited resources) or “true” noise in the environment. Regardless of the source, a MOP agent with β ≠ 0 will exhibit risk-sensitivity, i.e., preference or avoidance of this uncertainty15. In particular, MOP agents with β > 0 (see Eq. (2)) have a preference for stochastic state transitions and a priori they could get attracted and stuck in the noisy room, where actions do not have any predictable effect. Indeed, we see that for larger β, which measures the strength of the state entropy contribution to the agent’s objective, the attraction to the noisy room increases (Fig. 2e, right panel). However, MOP agents also care about future states, and thus getting stuck in regions where energy cannot be predictably obtained is avoided by sufficiently long-sighted agents, as shown by the reduction of the time spent in the noisy room with increasing γ (Fig. 2e; Supplemental Sec. E.3). This shows how MOP agents can tradeoff immediate with future action-state occupancy.
Hide and seek in a prey-predator interaction
More interesting behaviors arise from MOP in increasingly complex environments. To show this, we next considered a prey and a predator in a grid world with a safe area (a “home”) and a single food source (Fig. 3a). The prey (a “mouse”, gray up triangle) is the agent whose behavior is optimized by maximizing future action path entropy, while the predator (a “cat”, red down triangle) acts passively chasing the prey. The state of the agent consists of its location and energy level, but it also includes the predator’s location being accurately perceived. The prey can move as in the previous 4-room grid world and it also has a finite energy reservoir. For simplicity, we only considered a food gain equal to the size of the energy reservoir, such that the agent fully replenishes its reservoir each time it visits the food source. The predator has the same available actions as the agent and is attracted to it stochastically, i.e., actions that move the predator towards the agent are more probable than those that move it away from it (Supplemental Sec. E.4).
MOP generates complex behaviors, not limited to visiting the food source to increase the energy buffer and hide at home. In particular, the agent very often first teases the cat and then performs a clockwise rotation around the obstacle, which forces the cat to chase it around, leaving the food source free for harvest (Fig. 3a, arrows show an example; Supplementary Movie 2, MOP agent). Importantly, this behavior is not restricted to clockwise rotations, as the agent performs an almost equal number of counterclockwise rotations to free the food area (Fig. 3c, MOP agent, blue line). The variability of these rotations in the MOP agent is manifest in the lack of virtually any preferred directionality of movement in the arena at any single position. Indeed, arrows pointing toward several directions indicate that on average the mouse moves following different paths to get to the food source (Fig. 3b, MOP agent). Finally, the optimal value function and optimal policy show that the MOP agent can display deterministic behaviors as a function of internal state as well as distance to the cat (Fig. 3d): for instance, it prefers running away from the cat when energy is large (right), and it risks getting caught to avoid starvation if energy is small (left), both behaviors starkly opposite to stochastic actions.
The behavior of the MOP agent was compared with an R agent that receives a reward of 1 each time it is alive and 0 otherwise. To promote variable behavior in this agent as well, we implemented an ϵ-greedy action selection (Supplemental Sec. E.4), where ϵ was chosen to match the average lifetime of the MOP agent (Supplemental Fig. E.2b). The behavior of the R agent was strikingly less variable than that of the MOP agent, spending more time close to the food source (Fig. 3b, R agent). Most importantly, while the MOP agent performs an almost equal number of clock and counterclockwise rotations, the R agent strongly prefers the clockwise rotations, reaching 90% of all observed rotations (Supplementary Movie 3, R-agent; Fig. 3c, orange line). This shows that the R agent mostly exploits only one strategy to survive and displays a smaller behavioral repertoire than the MOP agent.
Dancing in an entropy-seeking cartpole
In the previous examples, complex behaviors emerge as a consequence of the presence of obstacles, predators, and limited food sources, but the actual dynamics of the agents are very coarse-grained. Here, we considered a system with physically realistic dynamics, the balancing cartpole83,84, composed of a moving cart with an attached pole free to rotate (Fig. 4a). The cartpole is assumed to reach an absorbing state when either it hits a border, or when the pole angle exceeds 36°. Thus, we consider a broad range of angles that makes the agents reach a larger state space than in standard settings85. We discretized the state space and used a linear interpolation to solve for the optimal value function in Eq. (4), and to implement the optimal policy in Eq. (6) (Supplemental Sec. E.5). The MOP agent widely occupies the horizontal position, and more strikingly it produces a wide variety of pole angles, constantly swinging sideways as if it were dancing (Supplementary Movie 4, MOP agent; Fig. 4b, c).
We compared the behavior of an MOP agent with that of an R agent that receives a reward of 1 for being alive and 0 otherwise. The R agent gets this reward regardless of the pole angle and cart position within the allowed broad ranges, so that behaviors of the MOP and R agents can be better compared without explicitly favoring in any of them any specific behavior, such as the upright pole position. As expected, the R agent maintains the pole close to the balanced position throughout most of a long episode (Fig. 4b, bottom), because it is the furthest to the absorbing states and thus the safest. Therefore, the R agent produces very little behavioral variability (Fig. 4c, right panel) and no movement that could be dubbed “dancing” (Supplementary Movie 4, R agent). Although both MOP and R agents use a similar strategy which keeps the pole pointing towards the center for substantial amounts of time (Fig. 4c, positive angles correlate with positive positions in both panels), the behavior of the R agent is qualitatively different and is best described as a bang-bang sort of control for which the angle is kept very close to zero while the cart is allowed to travel and oscillate around the origin, which is more apparent in the actual paths of the agent (see trajectories in phase space in Supplementary Movie 5). We also find that the R agent does not display much variability in state space even after using an ϵ-greedy action selection (Supplemental Fig. E.3, Supplementary Movie 6), with ϵ chosen to match average lifetimes between agents (Supplemental Fig. E.2c). This result showcases that the MOP agent exhibits the most appropriate sort of variability for a given average lifetime.
We finally introduced a slight variation to the environment, where the right half of the arena has stochastic state transitions, to showcase the ability of MOP to model risk-sensitive agents. Here, when agents choose an action (force) to be executed, a state transition in the desired direction occurs with probability 1 − η, and a transition corresponding to zero force occurs with probability η (Fig. 4d). Therefore, a MOP agent that seeks state entropy (β > 0) will show a preference for the right side, where there is in principle higher state entropy resulting from the stochastic transitions over more successor states than on the left side. Indeed, we find that MOP agents spend more time on the right side as β increases, regardless of the probability η (Fig. 4e). For fixed γ, spending more time on the right side can bring the life expectancy to decrease significantly depending on β and η (Supplemental Fig. E.2d, e). Interestingly, for β > 0 there is an optimal value of the noise η that maximizes the fraction of time spent on the right side (Fig. 4f), which is a form of stochastic resonance. Therefore, for different β, qualitatively different behaviors emerge as a function of the noise level η.
MOP agents can also seek entropy of others
Next, we considered an example where an agent seeks to occupy path space, which includes another agent’s location as well as its own. The agent can freely move (Fig. 5a; gray triangle) and open or close a fence by pressing a lever in a corner (blue triangle). The pet of the agent (green triangle) can freely move if the fence is open, but when the fence is closed the pet is confined to move in the region where it is currently located. The pet moves randomly at each step, but its available actions are restricted by its available space (Supplemental Sec. E.6).
To maximize action-state path entropy, the agent ought to tradeoff the state entropy resulting from letting the pet free with the action entropy resulting from using the open-close action when visiting the lever location. The optimal tradeoff depends on the relative strength of action and state entropies. In fact, when state entropy weighs as much as action entropy (α = β = 1), the fraction of time that the agent leaves the fence open is close to 1 (rightmost point in Fig. 5b) so that the pet is free to move (Fig. 5c, right panel; β = 1 MOP agent). However, when the state entropy has zero weight (α = 1, β = 0), the fraction of time that the fence remains open is close to 0.5 (leftmost point in Fig. 5b), and the pet remains confined to the right side for most of the time (Fig. 5c, left panel; β = 0 MOP agent), the region where it was initially placed. As a function of β the fraction of time the fence is open increases. Therefore, the agent gives more freedom to its pet, as measured by the pet’s state entropy, by curtailing its own action freedom, as measured by action entropy, thus becoming more “altruistic”.
MOP compared to other reward-free approaches
One important question is how MOP compares to other reward-free, motivation-driven theories of behavior. Here we focus on two popular approaches: empowerment and the FEP. In empowerment (MPOW)20,73,74,75 agents maximize the mutual information between n-step actions and the successor states resulting from them20,86, a measure of their capability to perform diverse courses of actions with predictable consequences. MPOW formulates behavior as greedy maximization of empowerment20,73, such that agents move to accessible states with the largest empowerment (maximal mutual information), and stay there with high probability.
We applied MPOW to the grid world and cartpole environments (Fig. 6). In the gridworld, MPOW agents (5-step MPOW, see Supplemental Sec. G.1) prefer states from where they can reach many distinct states, such as the middle of a room. However, due to energetic constraints, they also gravitate towards the food source when energy is low, and they alternate between these two locations ad nauseam (Fig. 6a, middle; Supplementary Movie 7). In the cartpole, MPOW agents (3-step MPOW73, see Supplemental Sec. G.1) favor the upright position because, being an unstable fixed point, it is the state with highest empowerment, as previously reported73,87. Given the unstable equilibrium, the MPOW agent gets close to it but needs to continuously adjust its actions when greedily maximizing empowerment (Fig. 6b, middle; Supplementary Movie 8). The paths traversed by MPOW agents in state space are highly predictable, and they are similar to the ones of the R agent (see Fig. 4c). The only source of stochasticity comes from the algorithm, which approximately calculates empowerment, and thus a more precise estimation of empowerment leads to even less variability.
In the FEP, agents seek to minimize the negative log probability, called surprise, of a subset of desired states via the minimization of an upper bound, called free energy. This minimization reduces behavioral richness by making a set of desired (homeostatic) states highly likely76,77, rendering this approach almost opposite to MOP. In a recent MDP formalization, FEP agents aim to minimize the (future) expected free energy (EFE)88, which equals the future cumulative KL divergence between the probability of states and the desired (target) probability of those states (see Supplemental Sec. G.2 for details). Even though this objective contains the standard exploration entropy term on state transitions77,89, we prove that the optimal policy is deterministic (see Supplemental Sec. G.2).
As a consequence, we find that in both the gridworld and cartpole environments, the behavior of the EFE agent (receding horizon H = 200) is much less variable than the MOP agent in general (Fig. 6a, right panel for the gridworld, Supplementary Movie 7; and b, right panel, for the cartpole, Supplementary Movie 8). The only source of action variability in the EFE agent is due to the degeneracy of the EFE, and thus behavior collapses to a deterministic policy as soon as the target distribution is not perfectly uniform (see Supplemental Sec. G.2.2 for details). We finally prove that under discounted infinite horizon and assuming a deterministic environment, the EFE agent is equivalent to a classical reward maximizer agent with reward R = 1 for all non-absorbing states and R = 0 for the absorbing states (Supplemental Sec. G.2). In conclusion, MOP generates much more variable behaviors than MPOW and FEP.
MOP in continuous and large action-state spaces
The examples so far can be solved exactly with numerical methods, without relying on function approximation of the value function or the policy, which could obscure the richness of the resulting behaviors. However, one important question is whether our approach scales up to large continuous action-state spaces where no exact solutions are available. To show that MOP generates rich behaviors even in high-dimensional agents, we simulated a quadruped via MuJoCo90 from Gymnasium91 without imposing any explicit fine-tuned reward function (Fig. 7a). The only externally imposed conditions are the absorbing states, which are reached when either the agent falls (given by the torso touching the ground), or the torso reaches a maximum height91.
We first trained the MOP agent by approximating the state-value function, Eq. (5), using the soft-actor critic (SAC) architecture49 with zero rewards, which corresponds to the case α = 1 and β = 0. The MOP agent learns to stabilize itself and walk around, sometimes jumping, spinning, and moving up and down the legs, without any instructions to do so (Supplementary Movie 9). The MOP agent exhibits variable and long excursions over state space (Fig. 7b, c blue) and displays a broad distribution of speeds (Fig. 7d, blue). We compared the MOP agent with an R agent that obtains a reward of R = 1 whenever it is alive and R = 0 when it reaches an absorbing state. As before, we add variability to the R agent with an ϵ-greedy action selection, adjusting ϵ so that the average lifetime of the R agent matched that of the MOP agent (Supplemental Fig. E.4a). In contrast to the MOP agent, the R agents exhibit much shorter excursions (Fig. 7b, c yellow) and a velocity distribution that peaks around zero, indicating prolonged periods spent with no translational movement (Fig. 7d, yellow). When visually compared, the behavior for MOP and R agents shows stark differences (Supplementary Movie 9).
While the MOP agent elicits variable behaviors, it is also capable of generating deterministic, goal-directed behaviors when needed. To show this, we added a food source in the arena and extended the state of the agent with its internal energy. Now the agent can also die of starvation when the internal energy hits zero (Fig. 7e). As expected, when the initial location of the MOP quadruped is far from the food source, it directly moves to the food source to avoid dying from starvation (Fig. 7f). After the food source is reached for the first time, the MOP quadruped generates random excursions away from the food source. During these two phases, the agent displays very different speed distributions (Fig. 7g), showing also quantitative differences in the way it moves (see a comparison with the R agent in Supplemental Fig. E.4, and Supplementary Movie 10).
Finally, we modified the environment by adding state transition noise of various magnitudes in one half of the arena (x > 0), while the other half remained deterministic. We find that the agent’s behavior is modulated by β, which controls the preference of state transition entropy (see details in Supplemental Sec. E.7). As expected, for fixed α and positive noise magnitude, MOP agents show increasing preference toward the noisy side as β increases (Supplemental Fig. E.5). However, as noise magnitude increases, and for fixed β, MOP agents tend to avoid the noisy side to prevent them from falling. This shows that MOP agents can exhibit approach and avoidance behaviors depending on the environment’s stochasticity and their β hyperparameter.
Discussion
Often, the success of agents in nature is not measured by the amount of reward obtained, but by their ability to expand in state space and perform complex behaviors. Here we have proposed that a major goal of intelligence is to “occupy path space”. Extrinsic rewards are thus the means to move and occupy action-state path space, not the goal of behavior. In an MDP setting, we have shown that the intuitive notion of path occupancy is captured by future action-state path entropy, and we have proposed that behavior is driven by the maximization of this intrinsic goal—the MOP. We have solved the associated Bellman equation and provided a convergent iterative map to determine the optimal policy. In several discrete and continuous state examples, we have shown that MOP, along with the agent’s constraints and dynamics, leads to complex behaviors that are not observed in other simple reward-maximizing agents. Quick filling of physical space by a moving agent, hide-and-seek behavior and variable escaping routes in a predator-prey example, dancing in a realistic cartpole dynamical system, altruistic behavior in an agent-and-pet duo and successful, vigorous movement in a high-dimensional quadruped are all behaviors that strike as being playful, curiosity-driven and energetic.
To the human eye, these behaviors look genuinely goal-directed, like approaching the food source when the energy level is low or escaping from the cat when it gets close to the mouse (see Figs. 2c and 3d). Although MOP agents do not have any extrinsically designed goal, like eating or escaping, they generate these deterministic, goal-directed behaviors whenever necessary so that they can keep moving in the future and maximize future path action-state entropy (see Supplemental Sec. H). These results show that the presence of internal states (e.g., energy) and absorbing states (e.g., having zero energy or being eaten) are critical for generating interesting behaviors, as getting close to different types of absorbing states triggers qualitatively different behaviors. This capability of adapting variability depending on internal states has been overlooked in the literature and is essential to obtaining the goal-directed behaviors we have shown here.
While all reward-maximizing principles of behavior predict that learning reduces variability and it collapses to zero once learning is deemed to have finished (i.e., there always exists an optimal deterministic policy92), MOP predicts that behavioral stochasticity persists even after learning (i.e., optimal policies under MOP are non-deterministic). This would imply that noise is promoted as long as it does not compromise the functioning of the agent, even after extensive experience with an environment. This prediction is supported by observations across different living organisms and at different levels of organization, including perception, behavior, neural circuits, and individual neurons. At the perceptual level, a paradigmatic example of the ubiquity of stochasticity is multistable perception, observed in humans93 and other animals94,95,96,97, a phenomenon consisting of a stochastic alternation between percepts that never stops, even though stimuli are simple and familiar. At the behavioral level, stochastic actions are observed in all kinds of living organisms. For example, foraging behavior in bacteria, plants, and animals is well-described by random walks and Lévy flights30,98,99, which remain valid descriptions even in familiar environments. Further, mice are found to perform random choices from time to time in simple binary tasks even after extended periods of training32. Neuroanatomically, emerging evidence indicates that specific neural circuits play a crucial role in generating behavioral variability. A striking example is the avian basal ganglia–forebrain circuit, which modulates the variability of the birds’ songs100,101, even after learning. Indeed, adult male birds with learned motor skills exhibit a higher level of variability when they sing alone, compared to during courtship102,103. At the neural level, single-neuron responses across sensory and motor areas are highly variable, and this variability persists after long periods of stimulation and adaptation104,105,106. All this evidence points out that, at least in part, behavior and its associated neural circuits and components are stochastic, regardless of the state of learning.
Not only do organisms seem to generate variability, but they are also sensitive to the state-transition uncertainty in or about the environment, a phenomenon known as risk sensitivity. Risk-seeking and risk-aversive behaviors are both found in humans33,107, non-human primates34,108, rodents, birds and other species38. While other works have looked at risk-sensitivity in RL, it is commonly associated to sensitivity to variations in extrinsic rewards36,37. In contrast, MOP can model risk-sensitivity by weighing the state-transition entropy with β. By adding stochastic regions to otherwise noiseless environments (four-room gridworld, cartpole, and quadruped), we showed that MOP can model risk-sensitive agents in the absence of extrinsic reward. For example, for the right combination of state entropy weight β, and lookahead horizon, controlled by γ, MOP agents could get stuck in a noisy TV, consistent with the observation that humans have a preference for noisy TVs under particular conditions109. However, it can also capture the avoidance of noisy TVs for sufficiently long sighted agents (see Fig. 2e). These observations and results show that MOP could provide a unified account for the stochastic nature of behavior and its risk sensitivity, offering the opportunity to model such seemingly disparate set of observations by modulating its action and state entropy contributions.
A set of algorithms related to MOP, known as empowerment, has also proposed using reward-free objectives as the goal of behavior20,73,75. In this approach, the mutual information between a sequence of actions and the final state is maximized. This makes empowerment agents prefer states where actions lead to large and predictable changes, such as unstable fixed points73. We have shown that one drawback is that empowered agents tend to remain close to those states without producing diverse behavioral repertoires (see Fig. 6b and Supplementary Movie 8), as it also happens in causal entropy approaches110. Another difference is that empowerment is not additive over paths because the mutual information of a path of actions with the path of states is not the sum of the per-step mutual information, and thus it cannot be formalized as a cumulative per-step objective (Supplemental Sec. I)73,75,81,111, in contrast to action-state path entropy. We note, however, that an approximation to empowerment having the desired additive property could be obtained from our framework by putting β < 0 in Eq. (2), such that more predictable state transitions are preferred. Similarly to empowerment, we have also shown that agents following the FEP76,88 collapse behavior to deterministic policies in known environments (see Fig. 6b and Supplementary Movie 8). Other reward-free RL settings and pure exploration objectives have been proposed in the past57,59,61,67,112,113,114,115, but this body of work typically investigates how to efficiently sample MDPs to construct near-optimal policies when reward functions are introduced in the exploitation phase. More importantly, this work differs from ours in that the goal-directedness that MOP displays entails behavioral variability at its core, even in known environments (see examples above). Finally, other overlapping reward-free approaches focus on the unsupervised discovery of skills, by encouraging diversity26,116,117,118. While the motivation is similar, they focus on skill-conditioned policies, whereas our work demonstrates that complex sequences of behaviors are possible working from the primitive actions of agents, although a possible future avenue for MOP is to apply it to temporally extended actions119. In addition, these works define tasks based on extrinsic rewards, whereas we have shown that internal state signals are sufficient to let agents define sub-tasks autonomously.
Our approach is conceptually different as well from hybrid approaches that combine extrinsic rewards with action entropy or KL regularization terms46,47,50,52,120 for two main reasons. First, entropy-seeking behavior does not pursue any form of extrinsic reward maximization. But most importantly, using KL-regularization using a default policy π0(a∣s) in our framework would be self-defeating. This is because the absolute action entropy terms \({{{{{{\mathcal{H}}}}}}}(A| s)\) in the expected return in Eq. (2) favor visiting states where a large set of immediate and future action-states are accessible. In contrast, using relative action entropy (KL) precludes this effect by normalizing the number of accessible actions, as we have shown above. Additionally, minimizing the KL divergence with a uniform default policy and without extrinsic rewards leads to an optimal policy that is uniform regardless of the presence of absorbing states, equivalent to a random walk, which shows that a pure KL objective does not lead to interesting behaviors (Supplemental Sec. F, Supplemental Fig. F.6). The idea of having a variable number of actions that depend on the state is consistent with the concept of affordance121. While we do not address the question of how agents get the information about the available actions, an option would be to use the notion of affordances as actions122. Secondly, while previous work has studied the performance benefits of either action49, state51,57 or equally weighted action-state53,123 steady-state entropies, our work proposes mixing them arbitrarily through path entropy, leading to a more general theory without any loss in mathematical tractability56.
We have also shown that MOP is scalable to high-dimensional problems and when the state-transition matrix is unknown, using the SAC architecture124 to approximate the optimal policy prescribed by MOP. Nevertheless, several steps remain to have a more complete MOP theory with learning. Previous related attempts have introduced Z-learning46,79 and G-learning125 using off-policy methods, so our results could be extended to learning following similar lines. Other possibilities are using transition estimators using counts or pseudo-counts69, or hashing70, for the learning of the transition matrices. One potential advantage of our framework is that, as entropy-seeking behavior obviates extrinsic rewards, those rewards do not need to be learned and optimized, and thus the learning problem reduces to transition matrices learning. In addition, modeling and injecting prior information could be particularly simple in our setting in view that intrinsic entropy rewards can be easily bounded before the learning process if action space is known. Therefore, initializing the state-value function to the lower or upper bounds of the action-state path entropy could naturally model pessimism or optimism during learning, respectively.
All in all, we have introduced MOP as a novel theory of behavior, which promises new ways of understanding goal-directedness without reward maximization, and that can be applied to artificial agents to discover by themselves ways of surviving and occupying action-state space.
Methods
Entropy measures the occupancy of action-state paths
We consider a time-homogeneous MDP with finite state set \({{{{{{\mathcal{S}}}}}}}\) and finite action set \({{{{{{\mathcal{A}}}}}}}(s)\) for every state \(s\in {{{{{{\mathcal{S}}}}}}}\). Henceforth, the action-state xj = (aj, sj) is any joint pair of one available action aj and one possible successor state sj that results from making that action under policy π ≡ {π(a∣s)} from the action-state xi = (ai, si). By assumption, the availability of action aj depends on the previous state si alone, not on ai. Thus, the transition probability from xi to xj in one-time step is pij = π(aj∣si)p(sj∣si, aj), where p(sj∣si, ai) is the conditional probability of transitioning from state si to sj given that action aj is performed. Although there is no dependence of the previous action ai on this transition probability, it is notationally convenient to define transitions between action-states.
We conceive of rational agents as maximizing future action-state path occupancy. Any measure of occupancy should obey the intuitive Conditions 1–4 listed below.
Intuitive Conditions for a measure of action-state path occupancy:
-
1.
Occupancy gain of action-state xj from xi is a function of the transition probability pij, C(pij)
-
2.
Performing a low probability transition leads to a higher occupancy gain than performing a high probability transition, that is, C(pij) decreases with pij
-
3.
The first order derivative \({C}^{{\prime} }({p}_{ij})\) is continuous for pij ∈ (0, 1)
-
4.
(Definition: the action-state occupancy of a one-step path from action-state xi is the expectation over occupancy gains of the immediate successor action-states, \({C}_{i}^{(1)}\equiv {\sum}_{j}{p}_{ij}C({p}_{ij})\)). The action-state occupancy of a two-step path is additive, \({C}_{i}^{(2)}\equiv {\sum}_{jk}{p}_{ij}{p}_{jk}C({p}_{ij}{p}_{jk})={C}_{i}^{(1)}+{\sum}_{j}{p}_{ij}{C}_{j}^{(1)}\) for any choice of the pij and initial xi
Condition 1 simply states that occupancy gain from an initial action-state is defined over the transition probabilities to successor action-states in a sample space. Condition 2 implies that performing a low-probability transition leads to a higher occupancy of the successor states than performing a high-probability transition. This is because performing a rare transition allows the agent to occupy a space that was left initially unoccupied. Condition 3 imposes smoothness of the measure.
In Condition 4 we have defined the occupancy of the successor action-states (one-step paths) in the Markov chain as the expected occupancy gain. Condition 4 is the central property, and it imposes that the occupancy of action-states paths with two steps can be broken down into a sum of the occupancies of action-states at each time step. Note that the action-state path occupancy can be written as
which imposes a strong condition on the function C(p). Note also that the sum ∑jkpijpjkC(pijpjk) extends the notion of action-state to a path of two consecutive action-states, each path having probability pijpjk due to the (time-homogeneous) Markov property. The last equality is an identity. While here we consider paths of length equal to 2, in the Supplementary Information we show that there is no difference in imposing additivity to paths of any fixed or random length (Corollary 2). The additivity of occupancy over paths is equivalent to time homogeneity, that is, all times count the same.
Theorem 1
\(C(p)=-k\ln p\) with k > 0 is the only function that satisfies Conditions 1–4.
Corollary 1
The entropy \({C}_{i}^{(1)}=-k{\sum}_{j}{p}_{ij}\ln {p}_{ij}\) is the only measure of action-state occupancy of successor action-states xj from xi with transition probabilities pij consistent with Conditions 1–4.
See Supplementary Information for the proof. We have found that entropy is the measure of occupancy. Shannon’s famous derivation of entropy as a measure of information78 uses similar elements, but some differences are worthy to be mentioned. First, our proof uses the notion of additivity of occupancy on MDPs of length two (our Condition 4), while Shannon’s notion of additivity uses sequences of random variables of arbitrary length (his Condition 3), and therefore his condition is in a sense stronger than ours. Second, our proof enforces continuous derivative of the measure, while Shannon enforces continuity of the measure, rendering our Condition 3 stronger. Finally, we enforce a specific form of the measure as an average over occupancy gains (our Condition 4 again), because it intuitively captures the notion of occupancy, while Shannon does not enforce this structure in his information measure.
Critical values and policies
Theorem 2
The critical values V c(s) of the expected returns Vπ(s) in equation (3) with respect to the policy probabilities π = {π(a∣s): a ∈ A(s), s ∈ S} obey
where \({{{{{{\mathcal{H}}}}}}}({S}^{{\prime} }| s,a)=-{\sum}_{{s}^{{\prime} }}p({s}^{{\prime} }| s,a)\ln p({s}^{{\prime} }| s,a)\) is the entropy of the successors of s after performing action a, and Z(s) is the partition function.
The critical points (critical policies) are
one per critical value, where the partition function Z(s) is the normalization constant.
Note that we simultaneously optimize ∣S∣ expected returns, one per state s, each with respect to the set of probabilities π = {π(a∣s): a ∈ A(s), s ∈ S}.
Proof
We first note that the expected return in Eq. (2) is continuous and has continuous derivatives with respect to the policy except at the boundaries (i.e., π(a∣s) = 0 for some action-state (a, s)). Choosing a state s, we first take partial derivatives with respect to π(a∣s) for each \(a\in {{{{{{\mathcal{A}}}}}}}(s)\) in both sides of Eq. (3) and then evaluate them at a critical point πc to obtain the condition
where we have defined the partial derivative at the critical point \(\frac{\partial {V}_{\pi }({s}^{{\prime} })}{\partial \pi (a| s)}{| }_{{\pi }^{c}}\equiv \lambda ({s}^{{\prime} },s)\) and used the fact that this partial derivative should be action-independent. To understand this, note that the critical policy should lie in the simplex ∑aπ(a∣s) = 1, π(a∣s) ≥ 0, and therefore the gradient of \({V}_{\pi }({s}^{{\prime} })\) with respect to the π(a∣s) at the critical policy should be along the normal to the constraint surface, i.e., the diagonal direction (hence, action-independent), or be zero. Indeed, the action-independence of the \(\lambda ({s}^{{\prime} },s)\) also results from interpreting them as Lagrange multipliers: \(\lambda ({s}^{{\prime} },s)\) is the Lagrange multiplier corresponding to the state-value function at \({s}^{{\prime} }\), \({V}_{\pi }({s}^{{\prime} })\), associated to the constraint ∑aπ(a∣s) = 1, π(a∣s) ≥ 0, defining the simplex where the probabilities {π(a∣s): a ∈ A(s)} lie.
Noticing that the last term of Eq. (10) does not depend on a, we can solve for the critical policy πc(a∣s) to obtain Eq. (9). Eq. (9) implicitly relates the critical policy with the critical value of the expected returns from each state s. Inserting the critical policy, Eq. (9), into Eq. (3), we get Eq. (8), which is an implicit non-linear system of equations exclusively depending on the critical values.
It is easy to verify that the partial derivatives of Vπ(s) in Eq. (3) with respect to \(\pi ({a}^{{\prime} }| {s}^{{\prime} })\) for \(s\, \ne \, {s}^{{\prime} }\) are
and thus they provide no additional constraint on the critical policy. □
We further prove in the Supplementary Information (Theorems 3 and 4) that the critical point is unique, corresponds to the optimal policy, and that the optimal state-value function can efficiently be obtained from the iterative mapping in Eq. (7).
Definition of agents
MOP agent
In all the experiments presented, we introduce the MOP agent. The objective function that this agent maximizes, in general, is Eq. (2). As described in the manuscript, the α and β parameters control the weights of action and next-state entropies to the objective function, respectively, but only their ratio β/α is important. Unless indicated otherwise, we always use α = 1, β = 0 for the experiments. It is important to note that if the environment is deterministic, then the next-state entropy \({{{{{{\mathcal{H}}}}}}}({S}^{{\prime} }| s,a)=-{\sum}_{{s}^{{\prime} }}p({s}^{{\prime} }| s,a)\ln p({s}^{{\prime} }| s,a)=0\), and therefore β does not change the optimal policy, Eq. (6).
The number of actions and state transitions available can depend on the s. In particular, we consider absorbing states s+, where only the “stay” action available. This construction leads to states with zero action and state transition entropies, and thus absorbing states have zero state-value Vπ(s+) = 0. While a reward function (see next section) tells an agent what needs to be accomplished, absorbing states tell a MOP agent what not to do. Thus, to occupy action-state path space, a MOP agent fabricates by itself a variety of goals and subgoals that are solely constrained by the presence of absorbing states and the environment-agent dynamics.
We have implemented the iterative map, Eq. (7), to solve for the optimal value, using \({z}_{i}^{(0)}=1\) for all i as initial condition. Theorem (3) ensures that this iterative map finds a unique optimal value regardless of the initial condition in the first orthant. To determine a degree of convergence, we compute the supremum norm between iterations,
where \({V}_{i}=\frac{\alpha }{\gamma }\log ({z}_{i})\), and the iterative map stops when δ < 10−3.
R agent
We also introduce a reward-maximizing agent in the usual RL sense. In this case, the reward is r = 1 for living and r = 0 when dying. In other words, this agent maximizes life expectancy. Additionally, to emphasize the typical reward-seeking behavior and avoid degenerate cases induced by the tasks, we introduced a small reward for the Four-room grid world (see Supplementary Information for details of the environments). In all other aspects, the modeling of the R agent is identical to the MOP agent. To allow for reward-maximizing agents to display some stochasticity, we used an ϵ-greedy policy, the best in the family of ϵ-soft policies14. At any given state, a random admissible action is chosen with probability ϵ, and the action that maximizes the value is chosen with probability 1 − ϵ. Given that the world models \(p({s}^{{\prime} }| s,a)\) are known and the environments are static, this ϵ-greedy policy does not serve the purpose of exploration (in the sense of learning), but only to inject behavioral variability. Therefore, we construct an agent with state-independent variability, whose value function satisfies the optimality Bellman equation for this ϵ-greedy policy,
where \(| {{{{{{\mathcal{A}}}}}}}(s)|\) is the number of admissible actions at state s. To solve for the optimal value in this Bellman equation, we perform value iteration14. The ϵ-greedy policy for the R agent is therefore given by
where ties in \(\arg \max\) are broken randomly. Note that if ϵ = 0, we obtain the usual greedy optimal policy that maximizes reward. The parameter ϵ is typically chosen so that average lifetimes for both R and MOP agents are matched.
Reporting summary
Further information on research design is available in the Nature Portfolio Reporting Summary linked to this article.
Data availability
The data generated in this study have been generated through a custom code126. The specific quadruped data analyzed in this study are available under request, although an equivalent dataset can be generated with the code provided.
Code availability
The code to generate the results and various figures is available as Python and Julia code along with guided notebooks to reproduce the figures126.
References
Ryan, R. M. & Deci, E. L. Intrinsic and extrinsic motivations: classic definitions and new directions. Contemp. Educ. Psychol. 25, 54–67 (2000).
Oudeyer, P.-Y., Kaplan, F. & Hafner, V. V. Intrinsic motivation systems for autonomous mental development. IEEE Trans. Evolut. Comput. 11, 265–286 (2007).
Adolph, K. E. & Berger, S. E. Motor development. in Handbook of Child Psychology Vol. 2. (Wiley Online Library, 2007).
MacNeilage, P. F. & Davis, B. L. On the origin of internal structure of word forms. Science 288, 527–531 (2000).
Petitto, L. A. & Marentette, P. F. Babbling in the manual mode: evidence for the ontogeny of language. Science 251, 1493–1496 (1991).
Dietrich, A. The cognitive neuroscience of creativity. Psychon. Bull. Rev. 11, 1011–1026 (2004).
Kidd, C. & Hayden, B. Y. The psychology and neuroscience of curiosity. Neuron 88, 449–460 (2015).
Gottlieb, J., Oudeyer, P.-Y., Lopes, M. & Baranes, A. Information-seeking, curiosity, and attention: computational and neural mechanisms. Trends Cogn. Sci. 17, 585–593 (2013).
Gittins, J., Glazebrook, K. & Weber, R. Multi-Armed Bandit Allocation Indices (John Wiley & Sons, 2011).
Averbeck, B. B. Theory of choice in bandit, information sampling and foraging tasks. PLoS Comput. Biol. 11, e1004164 (2015).
Doll, B. B., Simon, D. A. & Daw, N. D. The ubiquity of model-based reinforcement learning. Curr. Opin. Neurobiol. 22, 1075–1081 (2012).
Wang, M. Z. & Hayden, B. Y. Latent learning, cognitive maps, and curiosity. Curr. Opin. Behav. Sci. 38, 1–7 (2021).
Von Neumann, J. & Morgenstern, O.Theory of Games And Economic Behavior (Princeton University Press, 2007).
Sutton, R. S. et al. Introduction to Reinforcement Learning (MIT Press Cambridge, 1998).
Kahneman, D. & Tversky, A. Prospect theory: an analysis of decision under risk. in Handbook of the Fundamentals of Financial Decision Making: Part I 99–127 (World Scientific, 2013).
Silver, D., Singh, S., Precup, D. & Sutton, R. S. Reward is enough. Artif. Intell. 299, 103535 (2021).
Rash, C. J., Weinstock, J. & Van Patten, R. A review of gambling disorder and substance use disorders. Subst. Abus. Rehabil. 7, 3 (2016).
Ágh, T. et al. A systematic review of the health-related quality of life and economic burdens of anorexia nervosa, bulimia nervosa, and binge eating disorder. Eat. Weight Disord.-Stud. Anorex., Bulim. Obes. 21, 353–364 (2016).
McNamara, J. M. & Houston, A. I. The common currency for behavioral decisions. Am. Nat. 127, 358–378 (1986).
Klyubin, A. S., Polani, D. & Nehaniv, C. L. Empowerment: a universal agent-centric measure of control. In Proc. 2005 IEEE Congress on Evolutionary Computation Vol. 1 128–135 (IEEE, 2005).
Lehman, J. & Stanley, K. O. Abandoning objectives: evolution through the search for novelty alone. Evolut. Comput. 19, 189–223 (2011).
Singh, S., Lewis, R. L. & Barto, A. G. Where do rewards come from? In Proc. of the Annual Conference of the Cognitive Science Society 2601–2606 (Cognitive Science Society, 2009).
Zhang, T., Rosenberg, M., Zeyu, J., Perona, P. & Meister, M. Endotaxis: a universal algorithm for mapping, goal-learning, and navigation. eLife 12, RP84141 (2024).
Schmidhuber, J. A possibility for implementing curiosity and boredom in model-building neural controllers. In Proc. of the International Conference on Simulation of Adaptive Behavior: From Animals to Animats (eds. J. A. Meyer & S. W. Wilson) 222–227 (MIT Press/Bradford Books, 1991).
Hadfield-Menell, D., Milli, S., Abbeel, P., Russell, S. J. & Dragan, A. Inverse reward design. in Advances in Neural Information Processing Systems Vol. 30 (University of California, Berkeley, 2017).
Eysenbach, B., Gupta, A., Ibarz, J. & Levine, S. Diversity is all you need: Learning skills without a reward function. In 7th International Conference on Learning Representations, (ICLR, 2019).
Schrittwieser, J. et al. Mastering atari, go, chess and shogi by planning with a learned model. Nature 588, 604–609 (2020).
Asafa, T., Afonja, T., Olaniyan, E. & Alade, H. Development of a vacuum cleaner robot. Alex. Eng. J. 57, 2911–2920 (2018).
Kline, S. J. & Rosenberg, N. An overview of innovation. in Studies on Science and the Innovation Process: Selected Works of Nathan Rosenberg 173–203 (World Scientific, 2010).
Campeau, W., Simons, A. M. & Stevens, B. The evolutionary maintenance of Lévy flight foraging. PLOS Comput. Biol. 18, e1009490 (2022).
Beron, C. C., Neufeld, S. Q., Linderman, S. W. & Sabatini, B. L. Mice exhibit stochastic and efficient action switching during probabilistic decision making. Proc. Natl Acad. Sci. 119, e2113961119 (2022).
Pisupati, S., Chartarifsky-Lynn, L., Khanal, A. & Churchland, A. K. Lapses in perceptual decisions reflect exploration. eLife 10, e55490 (2021).
Braun, D. A., Nagengast, A. J. & Wolpert, D. Risk-sensitivity in sensorimotor control. in Frontiers in Human Neuroscience Vol. 5 https://www.frontiersin.org/articles/10.3389/fnhum.2011.00001 (Frontiers, 2011).
Hayden, B. Y., Heilbronner, S. R., Nair, A. C. & Platt, M. L. Cognitive influences on risk-seeking by rhesus macaques. Judgm. Decis. Mak. 3, 389–395 (2008).
Renart, A. & Machens, C. K. Variability in neural activity and behavior. Curr. Opin. Neurobiol. 25, 211–220 (2014).
Fei, Y., Yang, Z., Chen, Y., Wang, Z. & Xie, Q. Risk-sensitive reinforcement learning: near-optimal risk-sample tradeoff in regret. in Advances in Neural Information Processing Systems Vol. 33 22384–22395 https://proceedings.neurips.cc/paper/2020/hash/fdc42b6b0ee16a2f866281508ef56730-Abstract.html (Curran Associates, Inc., 2020).
Grau-Moya, J. et al. Beyond Bayes-optimality: meta-learning what you know you don’t know http://arxiv.org/abs/2209.15618 (2022).
Kacelink, A. & Bateson, M. Risky theories—the effects of variance on foraging decisions1. Am. Zool. 36, 402–434 (1996).
Wilson, R. C., Bonawitz, E., Costa, V. D. & Ebitz, R. B. Balancing exploration and exploitation with information and randomization. Curr. Opin. Behav. Sci. 38, 49–56 (2021).
Moreno-Bote, R., Knill, D. C. & Pouget, A. Bayesian sampling in visual perception. Proc. Natl Acad. Sci. 108, 12491–12496 (2011).
Recanatesi, S., Pereira-Obilinovic, U., Murakami, M., Mainen, Z. & Mazzucato, L. Metastable attractors explain the variable timing of stable behavioral action sequences. Neuron 110, 139–153 (2022).
Corver, A., Wilkerson, N., Miller, J. & Gordus, A. Distinct movement patterns generate stages of spider web building. Curr. Biol. 31, 4983–4997 (2021).
Dagenais, P., Hensman, S., Haechler, V. & Milinkovitch, M. C. Elephants evolved strategies reducing the biomechanical complexity of their trunk. Curr. Biol. 31, 4727–4737 (2021).
Mochol, G., Kiani, R. & Moreno-Bote, R. Prefrontal cortex represents heuristics that shape choice bias and its integration into future behavior. Curr. Biol. 31, 1234–1244 (2021).
Cazettes, F., Murakami, M., Renart, A. & Mainen, Z. F. Reservoir of decision strategies in the mouse brain. bioRxiv. (Cold Spring Harbor Laboratory, 2021).
Todorov, E. Efficient computation of optimal actions. Proc. Natl Acad. Sci. 106, 11478–11483 (2009).
Ziebart, B. D. Modeling purposeful adaptive behavior with the principle of maximum causal entropy (Carnegie Mellon University, 2010).
Haarnoja, T., Tang, H., Abbeel, P. & Levine, S. Reinforcement learning with deep energy-based policies. In Proc. International Conference on Machine Learning 1352–1361 (PMLR, 2017).
Haarnoja, T., Zhou, A., Abbeel, P. & Levine, S. Soft actor-critic: off-policy maximum entropy deep reinforcement learning with a stochastic actor. In Proc. International Conference on Machine Learning 1861–1870 (PMLR, 2018).
Schulman, J., Chen, X. & Abbeel, P. Equivalence between policy gradients and soft q-learning. arXiv preprint arXiv:1704.06440 (2017).
Neu, G., Jonsson, A. & Gómez, V. A unified view of entropy-regularized Markov decision processes. arXiv preprint arXiv:1705.07798 (2017).
Hausman, K., Springenberg, J. T., Wang, Z., Heess, N. & Riedmiller, M. Learning an embedding space for transferable robot skills. In Proc. International Conference on Learning Representations (ICLR, 2018).
Tishby, N. & Polani, D. Information theory of decisions and actions. in Perception-action Cycle, 601–636 (Springer, 2011).
Nachum, O., Norouzi, M., Xu, K. & Schuurmans, D. Bridging the gap between value and policy based reinforcement learning. in Advances in Neural Information Processing Systems Vol. 30 (Curran Associates Inc., 2017).
Galashov, A. et al. Information asymmetry in KL-regularized RL. In International Conference on Learning Representations (ICLR, 2019).
Grytskyy, D., Ramírez-Ruiz, J. & Moreno-Bote, R. A general Markov decision process formalism for action-state entropy-regularized reward maximization http://arxiv.org/abs/2302.01098 (2023).
Hazan, E., Kakade, S., Singh, K. & Van Soest, A. Provably efficient maximum entropy exploration. In Proc. International Conference on Machine Learning, 2681–2691 (PMLR, 2019).
Liu, H. & Abbeel, P. Behavior from the void: unsupervised active pre-training. Adv. Neural Inf. Process. Syst. 34, 18459–18473 (2021).
Mutti, M., Pratissoli, L. & Restelli, M. Task-agnostic exploration via policy gradient of a non-parametric state entropy estimate. In Proc. of the AAAI Conference on Artificial Intelligence Vol. 35, 9028–9036 (2021).
Seo, Y. et al. State entropy maximization with random encoders for efficient exploration. In Proc. International Conference on Machine Learning 9443–9454 (PMLR, 2021).
Zhang, C., Cai, Y., Huang, L. & Li, J. Exploration by maximizing Rényi entropy for reward-free RL framework. In Proc. of the AAAI Conference on Artificial Intelligence Vol. 35 10859–10867 (2021).
Amin, S., Gomrokchi, M., Satija, H., van Hoof, H. & Precup, D. A Survey of Exploration Methods in Reinforcement Learning. arXiv preprint arXiv:2109.00157 (2021).
Burda, Y., Edwards, H., Storkey, A. & Klimov, O. Exploration by random network distillation. In International Conference on Learning Representations (ICLR, 2018).
Achiam, J. & Sastry, S. Surprise-based intrinsic motivation for deep reinforcement learning. arXiv preprint arXiv:1703.01732 (2017).
Fountas, Z., Sajid, N., Mediano, P. & Friston, K. Deep active inference agents using Monte-Carlo methods. Adv. neural Inf. Process. Syst. 33, 11662–11675 (2020).
Burda, Y. et al. Large-Scale Study of Curiosity-Driven Learning. (ICLR, 2019).
Pathak, D., Agrawal, P., Efros, A. A. & Darrell, T. Curiosity-driven exploration by self-supervised prediction. In Proc. International Conference on Machine Learning 2778–2787 (PMLR, 2017).
Hafner, D. et al. Action and perception as divergence minimization http://arxiv.org/abs/2009.01791 (2022).
Bellemare, M. et al. Unifying count-based exploration and intrinsic motivation. Advances in Neural Information Processing Systems Vol. 29 (NIPS, 2016).
Tang, H. et al. #Exploration: a study of count-based exploration for deep reinforcement learning. in Advances in Neural Information Processing Systems Vol. 30 (2017).
Aubret, A., Matignon, L. & Hassas, S. An information-theoretic perspective on intrinsic motivation in reinforcement learning: a survey. Entropy 25, 327 (2023).
Schmidhuber, J. Driven by compression progress: a simple principle explains essential aspects of subjective beauty, novelty, surprise, interestingness, attention, curiosity, creativity, art, science, music, jokes. In Workshop on anticipatory behavior in adaptive learningsystems, 48–76 (Springer Berlin Heidelberg, 2009).
Jung, T., Polani, D. & Stone, P. Empowerment for continuous agent—environment systems. Adapt. Behav. 19, 16–39 (2011).
Still, S. & Precup, D. An information-theoretic approach to curiosity-driven reinforcement learning. Theory Biosci. 131, 139–148 (2012).
Mohamed, S. & Jimenez Rezende, D. Variational information maximisation for intrinsically motivated reinforcement learning. in Advances in Neural Information Processing Systems Vol. 28 (NIPS, 2015).
Friston, K., Kilner, J. & Harrison, L. A free energy principle for the brain. J. Physiol. 100, 70–87 (2006).
Buckley, C. L., Kim, C. S., McGregor, S. & Seth, A. K. The free energy principle for action and perception: a mathematical review. J. Math. Psychol. 81, 55–79 (2017).
Shannon, C. E. A mathematical theory of communication. Bell Syst. Tech. J. 27, 379–423 (1948).
Todorov, E. Linearly-solvable Markov decision problems. in Advances in Neural Information Processing Systems Vol. 19 (NIPS, 2006).
Rubin, J., Shamir, O. & Tishby, N. Trading value and information in MDPs. in Decision Making with Imperfect Decision Makers, 57–74 (Springer, 2012).
Leibfried, F., Pascual-Díaz, S. & Grau-Moya, J. A Unified Bellman Optimality principle combining reward maximization and empowerment. in Advances in Neural Information Processing Systems Vol. 32 https://proceedings.neurips.cc/paper_files/paper/2019/hash/13384ffc9d8bdb21c53c6f72d46f7866-Abstract.html (Curran Associates, Inc., 2019).
Schmidhuber, J. Curious model-building control systems. In Proc. International Joint Conference on Neural Networks 1458–1463 (1991).
Barto, A. G., Sutton, R. S. & Anderson, C. W. Neuronlike adaptive elements that can solve difficult learning control problems. in Proc. IEEE Transactions on Systems, Man, and Cybernetics 834–846 (IEEE, 1983).
Florian, R. V. Correct equations for the dynamics of the cart-pole system. in Center for Cognitive and Neural Studies (Coneural) (Citeseer, Romania, 2007).
Brockman, G. et al. Openai gym. arXiv preprint arXiv:1606.01540 (2016).
Blahut, R. Computation of channel capacity and rate-distortion functions. IEEE Trans. Inf. Theory 18, 460–473 (1972).
Klyubin, A. S., Polani, D. & Nehaniv, C. L. Keep your options open: an information-based driving principle for sensorimotor systems. PLOS ONE 3, e4018 (2008).
Da Costa, L., Sajid, N., Parr, T., Friston, K. & Smith, R. Reward maximization through discrete active inference. Neural Comput. 35, 807–852 (2023).
Tschantz, A., Millidge, B., Seth, A. K. & Buckley, C. L. Reinforcement learning through active inference. arXiv preprint arXiv:2002.12636 (2020).
Todorov, E., Erez, T. & Tassa, Y. Mujoco: a physics engine for model-based control. In Proc. 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems 5026–5033 (IEEE, 2012).
Towers, M. et al. Gymnasium https://zenodo.org/record/8127025 (2023).
Puterman, M. L. in Markov Decision Processes: Discrete Stochastic Dynamic Programming https://books.google.ca/books?hl=en&lr=&id=VvBjBAAAQBAJ&oi=fnd&pg=PT9&dq=markov+decision+processes:+discrete+stochastic+dynamic+programming+pdf&ots=rtgxtNPVQG&sig=ezUHyiSDwVSvQD0mVNk_LqRHTVE (John Wiley & Sons, 2014).
Wheatstone, C. XVIII. Contributions to the physiology of vision. Part the first. On some remarkable, and hitherto unobserved, phenomena of binocular vision. Philos. Trans. R. Soc. Lond. 128, 371–394 (1838).
Leopold, D., Maier, A. & Logothetis, N. K. Measuring subjective visual perception in the nonhuman primate. J. Conscious. Stud. 10, 115–130 (2003).
Pettigrew, J. D., Collin, S. P. & Ott, M. Convergence of specialised behaviour, eye movements and visual optics in the sandlance (Teleostei) and the chameleon (Reptilia). Curr. Biol. 9, 421–424 (1999).
Moreno-Bote, R., Shpiro, A., Rinzel, J. & Rubin, N. Alternation rate in perceptual bistability is maximal at and symmetric around equi-dominance. J. Vis. 10, 1–1 (2010).
Carter, O., van Swinderen, B., Leopold, D. A., Collin, S. & Maier, A. Perceptual rivalry across animal species. J. Comp. Neurol. 528, 3123–3133 (2020).
Ariel, G. et al. Swarming bacteria migrate by Lévy Walk. Nat. Commun. 6, 8396 (2015).
Reijers, V. C. et al. A Lévy expansion strategy optimizes early dune building by beach grasses. Nat. Commun. 10, 2656 (2019).
Doya, K. & Sejnowski, T. J. A novel reinforcement model of birdsong vocalization learning. in Advances in Neural Information Processing Systems Vol. 7 https://proceedings.neurips.cc/paper/1994/hash/0a113ef6b61820daa5611c870ed8d5ee-Abstract.html (MIT Press, 1994).
Aronov, D. & Fee, M. S. Natural changes in brain temperature underlie variations in song tempo during a mating behavior. PLOSOne 7, e47856 (2012).
Kao, M. H. & Brainard, M. S. Lesions of an avian basal ganglia circuit prevent context-dependent changes to song variability. J. Neurophysiol. 96, 1441–1455 (2006).
Woolley, S. C. & Doupe, A. J. Social context-induced song variation affects female behavior and gene expression. PLOS Biol. 6, e62 (2008).
Softky, W. R. & Koch, C. The highly irregular firing of cortical cells is inconsistent with temporal integration of random EPSPs. J. Neurosci. 13, 334–350 (1993).
Tolhurst, D. J., Movshon, J. A. & Dean, A. F. The statistical reliability of signals in single neurons in cat and monkey visual cortex. Vis. Res. 23, 775–785 (1983).
Shadlen, M. N. & Newsome, W. T. The variable discharge of cortical neurons: implications for connectivity, computation, and information coding. J. Neurosci. 18, 3870–3896 (1998).
Fishburn, P. C. & Kochenberger, G. A. Two-piece von Neumann–Morgenstern utility functions*. Decis. Sci. 10, 503–518 (1979).
Eisenreich*, B. R., Hayden, B. Y. & Zimmermann, J. Macaques are risk-averse in a freely moving foraging task. Sci. Rep. 9, 15091 (2019).
Modirshanechi, A., Lin, W.-H., Xu, H. A., Herzog, M. H. & Gerstner, W. The curse of optimism: a persistent distraction by novelty https://www.biorxiv.org/content/10.1101/2022.07.05.498835v2 (2022).
Wissner-Gross, A. D. & Freer, C. E. Causal entropic forces. Phys. Rev. Lett. 110, 168702 (2013).
Volpi, N. C. & Polani, D. Goal-directed empowerment: combining intrinsic motivation and task-oriented behaviour. in IEEE Transactions on Cognitive and Developmental Systems (IEEE, 2023).
Lee, L. et al. Efficient exploration via state marginal matching. arXiv preprint arXiv:1906.05274 (2019).
Jin, C., Krishnamurthy, A., Simchowitz, M. & Yu, T. Reward-free exploration for reinforcement learning. in Proc. International Conference on Machine Learning, 4870–4879 (PMLR, 2020).
Mutti, M. & Restelli, M. An intrinsically-motivated approach for learning highly exploring and fast mixing policies. In Proc. of the AAAI Conference on Artificial Intelligence Vol. 34, 5232–5239 (2020).
Eysenbach, B. & Levine, S. Maximum entropy RL (provably) solves some robust RL problems. In International Conference on Learning Representations (ICLR, 2022).
Gregor, K., Rezende, D. J. & Wierstra, D. Variational intrinsic control http://arxiv.org/abs/1611.07507 (2016). ArXiv:1611.07507 [cs].
Sharma, A., Gu, S., Levine, S., Kumar, V. & Hausman, K. Dynamics-aware unsupervised skill discovery. In Proc. International Conference on Learning Representations (2020). MAG ID: 2995736683 S2ID: ae3b2768b0a3c73410bce0d2ae03feaf01f6f864.
Park, S., Lee, K., Lee, Y. & Abbeel, P. Controllability-aware unsupervised skill discovery http://arxiv.org/abs/2302.05103 (2023).
Sutton, R. S., Precup, D. & Singh, S. Between MDPs and semi-MDPs: a framework for temporal abstraction in reinforcement learning. Artif. Intell. 112, 181–211 (1999).
Grau-Moya, J., Leibfried, F., Genewein, T. & Braun, D. A. Planning with information-processing constraints and model uncertainty in Markov decision processes. In Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD Proceedings, Part II 16 475–491 (Springer International Publishing, 2016).
Gibson, J. J.The Ecological Approach to Visual Perception: Classic Edition (Psychology Press, 2014).
Khetarpal, K., Ahmed, Z., Comanici, G., Abel, D. & Precup, D. What can I do here? A theory of affordances in reinforcement learning. In Proc. International Conference on Machine Learning, 5243–5253 https://proceedings.mlr.press/v119/khetarpal20a.html (PMLR, 2020).
Peters, J., Mulling, K. & Altun, Y. Relative entropy policy search. In Proc. of the AAAI Conference on Artificial Intelligence Vol. 24 1607–1612 https://ojs.aaai.org/index.php/AAAI/article/view/7727 (2010).
Haarnoja, T. et al. Soft actor-critic algorithms and applications http://arxiv.org/abs/1812.05905 (2019).
Fox, R., Pakman, A. & Tishby, N. Taming the noise in reinforcement learning via soft updates. In 32nd Conference on Uncertainty in Artificial Intelligence, 202–211 (AUAI, 2016).
Ramírez-Ruiz, J., Grytskyy, D., Mastrogiuseppe, C., Habib, Y. & Moreno-Bote, R. Complex behavior from intrinsic motivation to occupy action-state path space. https://zenodo.org/records/11401402.
Acknowledgements
This work is supported by the Howard Hughes Medical Institute (HHMI, ref 55008742), ICREA Academia 2022 and MINECO (Spain; BFU2017-85936-P) to R.M.-B., MINECO/ESF (Spain; PRE2018-084757) to J.R.-R, and AGAUR-FI ajuts from Generalitat de Catalunya/ESF (2024 FI-B3 00020) to C.M and (2023 FI-1 00245) to Y.H.
Author information
Authors and Affiliations
Contributions
J.R.-R. and R.M.-B. conceived the presented idea. J.R.-R., D.G., and R.M.B. developed the theory. J.R.-R., D.G., C.M., and Y.H. performed the computations and analyzed the data. All authors discussed the results and contributed to the final manuscript.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Peer review
Peer review information
Nature Communications thanks the anonymous, reviewer(s) for their contribution to the peer review of this work. A peer review file is available.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary information
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Ramírez-Ruiz, J., Grytskyy, D., Mastrogiuseppe, C. et al. Complex behavior from intrinsic motivation to occupy future action-state path space. Nat Commun 15, 6368 (2024). https://doi.org/10.1038/s41467-024-49711-1
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41467-024-49711-1
Comments
By submitting a comment you agree to abide by our Terms and Community Guidelines. If you find something abusive or that does not comply with our terms or guidelines please flag it as inappropriate.