Hubbry Logo
Differential gameDifferential gameMain
Open search
Differential game
Community hub
Differential game
logo
7 pages, 0 posts
0 subscribers
Be the first to start a discussion here.
Be the first to start a discussion here.
Differential game
Differential game
from Wikipedia

In game theory, differential games are dynamic games that unfold in continuous time, meaning players’ actions and outcomes evolve smoothly rather than in discrete steps,[1] and for which the rate of change of each state variable—like position, speed, or resource level—is governed by a differential equation. This distinguishes them from turn-based games (sequential games) like chess, focusing instead on real-time strategic conflicts.

Differential games are sometimes called continuous-time games, a broader term that includes them.[2] While the two overlap significantly, continuous-time games also encompass models not governed by differential equations, such as those with stochastic jump processes, where abrupt, unpredictable events introduce discontinuities

Early differential games, often inspired by military scenarios, modeled situations like a pursuer chasing an evader, such as a missile targeting an aircraft.[3] Today, they also apply to fields like economics and engineering, analyzing competition over resources or the control of moving systems.[4][5]

Connection to optimal control

[edit]

Differential games are related closely to optimal control problems. In an optimal control problem there is single control and a single criterion to be optimized; differential game theory generalizes this to two controls and two criteria, one for each player.[6] Each player attempts to control the state of the system so as to achieve its goal; the system responds to the inputs of all players.

History

[edit]

In the study of competition, differential games have been employed since a 1925 article by Charles F. Roos.[7] The first to study the formal theory of differential games was Rufus Isaacs, publishing a text-book treatment in 1965.[8] One of the first games analyzed was the 'homicidal chauffeur game'.

Random time horizon

[edit]

Games with a random time horizon are a particular case of differential games.[9] In such games, the terminal time is a random variable with a given probability distribution function. Therefore, the players maximize the mathematical expectancy of the cost function. It was shown that the modified optimization problem can be reformulated as a discounted differential game over an infinite time interval[10][11]

Applications

[edit]

Differential games have been applied to economics. Recent developments include adding stochasticity to differential games and the derivation of the stochastic feedback Nash equilibrium (SFNE). A recent example is the stochastic differential game of capitalism by Leong and Huang (2010).[12] In 2016 Yuliy Sannikov received the John Bates Clark Medal from the American Economic Association for his contributions to the analysis of continuous-time dynamic games using stochastic calculus methods.[13][14]

Additionally, differential games have applications in missile guidance[15][16] and autonomous systems.[17]

For a survey of pursuit–evasion differential games see Pachter.[18]

See also

[edit]

Notes

[edit]

Further reading

[edit]
[edit]
Revisions and contributorsEdit on WikipediaRead on Wikipedia
from Grokipedia
A differential game is a mathematical framework within that models strategic interactions among decision-makers whose actions influence the evolution of a governed by ordinary differential equations, typically in continuous time, where players seek to optimize their individual payoff functions under conflicting interests. Developed primarily during the mid-20th century at the , the theory was pioneered by Rufus Isaacs in the early 1950s and formalized in his seminal 1965 book, which applied it to scenarios involving warfare, pursuit-evasion, and . Key concepts include the distinction between zero-sum and noncooperative games, information structures such as open-loop and closed-loop strategies, and equilibrium solutions like Nash equilibria or saddle points, often solved using techniques from theory and viscosity solutions to Hamilton-Jacobi-Isaacs equations. Applications span diverse fields, including for air combat and traffic control, for resource management and models, for pursuit-evasion tasks, and for competitive decision-making in dynamic environments. The theory has evolved to encompass hybrid systems, elements, and multi-player extensions, maintaining its relevance in modern interdisciplinary research.

Fundamentals

Definition and Basic Principles

A differential game is a mathematical model that extends game theory to continuous-time dynamics, involving two or more decision-makers, or players, who interact strategically by selecting controls that influence the evolution of a shared system described by ordinary differential equations (ODEs). In this framework, the state of the system, denoted x(t)x(t), evolves continuously from an initial condition x(0)x(0) to a terminal time TT, with the dynamics governed by an equation of the form x˙(t)=f(t,x(t),u1(t),,un(t))\dot{x}(t) = f(t, x(t), u_1(t), \dots, u_n(t)), where ui(t)u_i(t) represents the control actions chosen by player ii from admissible sets UiU_i. This setup captures scenarios where players' decisions are made over time, anticipating the responses of others, in contrast to static games or discrete-time models. The basic principles of differential games revolve around a non-cooperative environment where players pursue conflicting objectives, typically aiming to optimize individual payoff functions that depend on the system's . Each player's payoff, such as Ji=gi(x(T))+0Tli(t,x(t),u1(t),,un(t))dtJ_i = g_i(x(T)) + \int_0^T l_i(t, x(t), u_1(t), \dots, u_n(t)) \, dt, measures their performance, where gig_i is a terminal and lil_i a running , and players select strategies—either open-loop (time-dependent) or feedback (state-dependent)—to maximize or minimize JiJ_i while considering opponents' actions. Time progresses continuously, allowing for instantaneous adjustments, and the interaction often leads to equilibria like equilibria, where no player can improve their payoff by unilaterally deviating from their strategy. These principles build on theory as a single-player precursor but incorporate strategic interdependence. Prerequisite to understanding differential games are core elements of adapted to continuous dynamics: payoffs quantify outcomes influenced by all players' choices, while strategies define how controls are selected based on available up to time tt, such as past states or full histories. In this context, players anticipate rational responses from others, fostering a balance of avoidance and strategic foresight without assuming joint optimization. A representative example is a simple two-player differential game where the state x(t)x(t) satisfies x˙(t)=u1(t)u2(t)\dot{x}(t) = u_1(t) - u_2(t), with x(0)=x0x(0) = x_0 and controls u1(t),u2(t)u_1(t), u_2(t) bounded in magnitude. Player 1 seeks to maximize J1=x(T)0Tu1(t)2dtJ_1 = x(T) - \int_0^T u_1(t)^2 \, dt, driving the state positively at minimal control cost, while player 2 aims to maximize J2=x(T)0Tu2(t)2dtJ_2 = -x(T) - \int_0^T u_2(t)^2 \, dt, countering with opposing effort; the resulting interplay determines the terminal state through continuous control adjustments.

Key Components

Differential games model conflicts between multiple decision-makers whose actions influence a shared dynamic . The core elements include the state of the , the controls available to each player, the objectives they seek to achieve, the available to them, and the temporal scope of the interaction. These components form the foundational structure for analyzing strategic interactions in continuous-time settings. The state variables describe the configuration of the at time tt, denoted as x(t)Rnx(t) \in \mathbb{R}^n, where nn is the dimension of the state space. The of the state is governed by a of the form x˙(t)=f(t,x(t),u1(t),,uN(t)),\dot{x}(t) = f(t, x(t), u^1(t), \dots, u^N(t)), with x(0)=x0x(0) = x_0, where f:[0,T]×Rn×i=1NUiRnf: [0, T] \times \mathbb{R}^n \times \prod_{i=1}^N U^i \to \mathbb{R}^n is a representing the , and UiRmiU^i \subset \mathbb{R}^{m_i} are compact control sets for each of the NN players. Control variables represent the actions or strategies of the players, typically as bounded measurable functions ui(t)Uiu^i(t) \in U^i for player ii. In noncooperative settings, strategies can be open-loop (functions of time only) or closed-loop (state-dependent feedback policies, such as ui(t)=γi(t,x(t))u^i(t) = \gamma^i(t, x(t))), allowing players to adapt to the evolving system. For two-player zero-sum games, the controls u1(t)u_1(t) and u2(t)u_2(t) directly oppose each other in influencing the state trajectory. Objective functions quantify the payoffs or costs for each player, driving their strategic choices. In a general NN-player differential game, player ii seeks to minimize (or maximize, depending on convention) their cost functional Ji(u1,,uN;x0)=0TLi(t,x(t),u1(t),,uN(t))dt+Φi(x(T)),J^i(u^1, \dots, u^N; x_0) = \int_0^T L^i(t, x(t), u^1(t), \dots, u^N(t)) \, dt + \Phi^i(x(T)), where LiL^i is the running cost and Φi\Phi^i is the terminal cost. In zero-sum two-player cases, the game simplifies to player 1 minimizing JJ while player 2 maximizes it, often with J2=J1J^2 = -J^1. These functionals capture competing interests, such as minimizing time to capture in pursuit-evasion scenarios. The information structure specifies what players know when selecting controls, typically assuming where all players observe the full state x(t)x(t) at each instant. This enables non-anticipative strategies, ensuring no player can react to future actions of others. Under , feedback strategies are common, mapping current states to controls without foresight. The defines the duration of the game, either fixed at a terminal time T<T < \infty, leading to finite-horizon problems with explicit endpoint costs, or infinite (T=T = \infty), often incorporating discounting factors like eρte^{-\rho t} in the integral to ensure convergence and model long-term interactions.

Historical Development

Early Contributions

The origins of differential games trace back to early 20th-century efforts in economics to model competitive dynamics using differential equations. In 1925, Charles F. Roos introduced the first known formulation resembling a differential game in his paper "A Mathematical Theory of Competition," where he analyzed a duopoly scenario as a two-player interaction governed by systems of differential equations representing firms' production decisions over time. This work treated economic rivalry as a non-cooperative contest, anticipating later game-theoretic extensions, though it lacked explicit saddle-point solutions or value functions. Roos expanded these ideas in his 1927 paper "A Dynamical Theory of Economics," further exploring dynamic equilibria in competitive markets through continuous-time models. Influences from broader game theory and control theory emerged in the following decades, providing conceptual groundwork without fully merging into continuous-time game frameworks. John von Neumann's 1928 minimax theorem established the foundations of zero-sum games, emphasizing optimal strategies in discrete settings that later inspired adaptations to differential structures. During the 1940s and 1950s, nascent ideas in optimal control, such as Richard Bellman's dynamic programming approach introduced in 1957, focused on single-agent trajectory optimization but highlighted parallels to multiplayer decision-making under uncertainty. These developments connected discrete game theory to continuous dynamics, yet stopped short of integrating opposing players' controls in a unified differential game context. World War II spurred informal mathematical applications to military tactics, particularly in antisubmarine warfare, where operations research teams analyzed evasion maneuvers to counter U-boat threats. Pioneered by figures like , these efforts employed probabilistic models and search theory to optimize convoy routing and detection strategies, effectively modeling pursuit-evasion scenarios without formal game-theoretic rigor. Such practical insights into adversarial motion predated systematic theory, influencing postwar advancements in dynamic games. These pre-1960s contributions laid essential groundwork, paving the way for Rufus Isaacs' formal synthesis at in the mid-1950s.

Isaacs and Modern Foundations

Rufus Isaacs laid the modern foundations of through his pioneering synthesis of and principles during his tenure at the . Joining RAND in 1948, Isaacs began developing the field in the early 1950s amid military-funded research focused on strategic conflicts like aerial combat and pursuit scenarios, which were central to Cold War defense analysis sponsored by the U.S. Air Force. His efforts culminated in the 1965 book Differential Games: A Mathematical Theory with Applications to Warfare and Pursuit, Control and Optimization, which formalized as dynamic optimization problems involving adversarial agents with continuous-time state evolutions governed by . Isaacs introduced core concepts such as the and min-max optimization over strategies, emphasizing applications to warfare where one player's gain equals the other's loss. A hallmark example in Isaacs' framework is the "homicidal chauffeur game," a pursuit-evasion problem where the evader maneuvers with simple circular motion at a fixed speed, while the pursuer employs car-like dynamics with bounded turning radius, highlighting the asymmetry in mobility that complicates optimal play. This illustration, first sketched in Isaacs' 1951 RAND report and elaborated in his book, demonstrated how differential games extend classical pursuit problems by incorporating realistic kinematic constraints and non-cooperative objectives. The RAND environment, with its emphasis on practical military simulations, fostered this integration, influencing subsequent work in missile guidance and tactical decision-making. Key advancements in the late 20th century addressed the mathematical challenges of Isaacs' formulations, particularly the nonlinear Hamilton-Jacobi-Isaacs equations defining the value function. In 1983, Michael Crandall and Pierre-Louis Lions introduced , a generalized notion of weak solutions for first-order partial differential equations that ensures uniqueness and stability even for non-smooth value functions typical in game settings. This breakthrough, building on earlier viscosity ideas for Hamilton-Jacobi equations, enabled rigorous analysis of differential games where classical differentiable solutions do not exist. Subsequent extensions in the 2000s and beyond refined the field for complex systems. Alexey Matveev and Andrey Savkin contributed to hybrid dynamical systems and robust navigation, applying differential game methods to problems like collision avoidance and state estimation under uncertainty, as detailed in their 2000 book Qualitative Theory of Hybrid Dynamical Systems with Applications to Hybrid Control Problems. More recently, Yuliy Sannikov advanced stochastic variants through continuous-time models using stochastic calculus, earning the 2016 John Bates Clark Medal for insights into dynamic contracting and principal-agent interactions in stochastic dynamic games. These developments solidified differential games as a versatile tool bridging deterministic control and probabilistic decision-making.

Mathematical Framework

Standard Formulation

A differential game is formally defined as a two-player zero-sum dynamic game in which the state evolution of a system is governed by ordinary differential equations influenced by the actions of two opposing players, one seeking to minimize the payoff and the other to maximize it. The standard setup involves a state vector x(t)Rnx(t) \in \mathbb{R}^n satisfying the dynamics x˙(t)=f(t,x(t),u(t),v(t)),x(0)=x0,\dot{x}(t) = f(t, x(t), u(t), v(t)), \quad x(0) = x_0, where u(t)URm1u(t) \in U \subseteq \mathbb{R}^{m_1} and v(t)VRm2v(t) \in V \subseteq \mathbb{R}^{m_2} are the control inputs available to the minimizing player (Player I) and the maximizing player (Player II), respectively, with UU and VV being compact convex sets. The objective functional, or payoff, is given by J(x0,u(),v())=0Tl(t,x(t),u(t),v(t))dt+g(x(T)),J(x_0, u(\cdot), v(\cdot)) = \int_0^T l(t, x(t), u(t), v(t)) \, dt + g(x(T)), where l:[0,T]×Rn×U×VRl: [0,T] \times \mathbb{R}^n \times U \times V \to \mathbb{R} is the running cost (continuous and bounded) and g:RnRg: \mathbb{R}^n \to \mathbb{R} is the terminal cost (continuous), with fixed terminal time T>0T > 0. Player I aims to choose controls to minimize JJ, while Player II aims to maximize it. Assumptions on ff, ll, and gg typically include continuity and conditions to ensure and of solutions to the dynamics. Strategies in differential games extend beyond open-loop controls u()u(\cdot), v()v(\cdot) to account for the dynamic interaction, incorporating feedback from the state and . A key class is non-anticipative strategies, where a strategy for Player I, denoted γ\gamma, maps the of Player II's actions up to time tt (but not future actions) to a control u(t)=γ(t,{v(s):0st})u(t) = \gamma(t, \{v(s) : 0 \leq s \leq t\}), ensuring no foresight of the opponent's future moves; the symmetric definition applies to Player II. Feedback strategies, a subclass of non-anticipative ones, are state-dependent: u(t)=u~(t,x(t))u(t) = \tilde{u}(t, x(t)) and v(t)=v~(t,x(t))v(t) = \tilde{v}(t, x(t)), often leading to closed-loop dynamics x˙(t)=f(t,x(t),u~(t,x(t)),v~(t,x(t)))\dot{x}(t) = f(t, x(t), \tilde{u}(t, x(t)), \tilde{v}(t, x(t))). These strategies are measurable and ensure well-defined trajectories. The value of the game quantifies its equilibrium under optimal play. The lower value is defined as V(x0)=infγsupδJ(x0,γ,δ)\underline{V}(x_0) = \inf_{\gamma} \sup_{\delta} J(x_0, \gamma, \delta), where the infimum is over Player I's non-anticipative strategies γ\gamma and the supremum over Player II's δ\delta; the upper value is V(x0)=supδinfγJ(x0,γ,δ)\overline{V}(x_0) = \sup_{\delta} \inf_{\gamma} J(x_0, \gamma, \delta). The game possesses a value V(x0)V(x_0) if V(x0)=V(x0)\underline{V}(x_0) = \overline{V}(x_0), in which case optimal strategies exist yielding this value. Existence is guaranteed under the Isaacs condition, which requires that the Hamiltonian H(t,x,λ,u,v)=l(t,x,u,v)+λf(t,x,u,v)H(t, x, \lambda, u, v) = l(t, x, u, v) + \lambda \cdot f(t, x, u, v) admits a saddle point for every t[0,T]t \in [0,T], xRnx \in \mathbb{R}^n, λRn\lambda \in \mathbb{R}^n, specifically minuUmaxvVH=maxvVminuUH\min_{u \in U} \max_{v \in V} H = \max_{v \in V} \min_{u \in U} H. This condition ensures the lower and upper values coincide and equal the viscosity solution of the associated Isaacs equation. A representative example is the linear-quadratic (LQ) differential game, where the dynamics are affine: x˙(t)=Ax(t)+Bu(t)+Cv(t)\dot{x}(t) = A x(t) + B u(t) + C v(t), x(0)=x0x(0) = x_0, with constant matrices ARn×nA \in \mathbb{R}^{n \times n}, BRn×m1B \in \mathbb{R}^{n \times m_1}, CRn×m2C \in \mathbb{R}^{n \times m_2}. The zero-sum payoff takes the quadratic form J(x0,u(),v())=0T[x(t)TQx(t)+u(t)TRu(t)v(t)TSv(t)]dt+x(T)TMx(T),J(x_0, u(\cdot), v(\cdot)) = \int_0^T \left[ x(t)^T Q x(t) + u(t)^T R u(t) - v(t)^T S v(t) \right] dt + x(T)^T M x(T), where Q0Q \geq 0, M0M \geq 0, R>0R > 0, S>0S > 0 are symmetric matrices, capturing Player I's quadratic cost penalized by uu and Player II's benefit from vv. Under the Isaacs condition (here, the Hamiltonian's saddle point exists due to the quadratic structure), the value V(x0)=x0TPx0V(x_0) = x_0^T P x_0 is quadratic, with PP solving a Riccati equation, and feedback strategies are linear: u(t,x)=K1(t)xu^*(t,x) = -K_1(t) x, v(t,x)=K2(t)xv^*(t,x) = K_2(t) x. This setup models applications like pursuit-evasion with quadratic performance metrics.

Isaacs' Method and Value Function

In differential games, Rufus Isaacs introduced a method based on dynamic programming to determine optimal strategies and the game's value by deriving a partial differential equation (PDE) satisfied by the value function. This approach extends the Hamilton-Jacobi-Bellman equation from optimal control to account for adversarial interactions between players. The value function V(t,x)V(t, x) is defined as V(t,x)=infγsupδJ(t,x,γ,δ)V(t, x) = \inf_{\gamma} \sup_{\delta} J(t, x, \gamma, \delta), where J(t,x,γ,δ)J(t, x, \gamma, \delta) is the cost functional for the minimizer (Player I using γ\gamma) against the maximizer's δ\delta, specifically J(t,x,γ,δ)=g(x(T))+tTl(s,x(s),γ(s),δ(s))dsJ(t, x, \gamma, \delta) = g(x(T)) + \int_t^T l(s, x(s), \gamma(s), \delta(s)) \, ds for fixed terminal time TT. It satisfies the terminal condition V(T,x)=g(x)V(T, x) = g(x). Under regularity assumptions, such as of the dynamics and payoffs, VV is continuous and provides the guaranteed payoff for the minimizer when both players play optimally. Isaacs' method proceeds backward in time using dynamic programming principles, constructing VV iteratively from the terminal condition to characterize saddle-point equilibria. The core of this approach is the Isaacs equation, a first-order nonlinear PDE given by Vt+minuUmaxvVH(t,x,V,u,v)=0,\frac{\partial V}{\partial t} + \min_{u \in U} \max_{v \in V} H(t, x, \nabla V, u, v) = 0, where the Hamiltonian is H(t,x,p,u,v)=l(t,x,u,v)+pf(t,x,u,v)H(t, x, p, u, v) = l(t, x, u, v) + p \cdot f(t, x, u, v). This equation arises from optimizing the Hamiltonian over opposing controls, reflecting the zero-sum nature; under the Isaacs condition, minumaxvH=maxvminuH\min_u \max_v H = \max_v \min_u H. For smooth cases, solutions to the Isaacs equation yield explicit optimal feedback strategies via the : u=argminuH(t,x,V,u,v)u^* = \arg\min_u H(t, x, \nabla V, u, v^*) and v=argmaxvH(t,x,V,u,v)v^* = \arg\max_v H(t, x, \nabla V, u^*, v), ensuring equilibrium play. In non-smooth scenarios, where classical solutions may fail due to discontinuities or non-convexity, viscosity solutions provide a robust framework, defining VV as the unique continuous solution satisfying the equation in a generalized using functions. This extension ensures and under mild assumptions, such as of control sets and continuity of ff and ll.

Relation to Optimal Control

Parallels in Formulation

Differential games build upon the foundational structures of theory by extending the single-player framework to scenarios involving multiple decision-makers with opposing objectives. In classical , a single agent seeks to minimize a functional J=t0tfl(x(t),u(t))dt+ϕ(x(tf))J = \int_{t_0}^{t_f} l(x(t), u(t)) \, dt + \phi(x(t_f)), subject to the state dynamics x˙(t)=f(x(t),u(t))\dot{x}(t) = f(x(t), u(t)), where xx is the state vector and uu is the control input. This problem is typically solved using the Hamilton-Jacobi-Bellman (HJB) equation, which characterizes the value function V(x,t)V(x,t) as satisfying Vt+minu[l(x,u)+xVf(x,u)]=0\frac{\partial V}{\partial t} + \min_u \left[ l(x,u) + \nabla_x V \cdot f(x,u) \right] = 0, with appropriate boundary conditions. The extension to differential games introduces multiple players, such as a minimizer and a maximizer, leading to a bifurcated Hamiltonian where the optimization becomes minumaxv\min_u \max_v (or vice versa) over the respective controls uu and vv, while sharing the same underlying dynamics x˙=f(x,u,v)\dot{x} = f(x, u, v). This adaptation preserves core elements of , including the application of for open-loop solutions, which yields necessary conditions via state-costate equations x˙=f(x,u,v)\dot{x} = f(x, u, v) and λ˙=Hx\dot{\lambda} = -\frac{\partial H}{\partial x}, along with transversality conditions at the terminal time, such as λ(tf)=xϕ(x(tf))\lambda(t_f) = \nabla_x \phi(x(t_f)). In this view, differential games can be interpreted as problems where one player's control acts as a disturbance to the other, maintaining the variational structure of the single-player case. Historically, Rufus Isaacs developed the theory of differential games in the mid-, drawing directly from Richard Bellman's dynamic programming approach introduced in the early , which provided the recursive optimization framework underpinning the HJB equation in continuous time. Isaacs' seminal RAND memoranda and subsequent book formalized these parallels, treating multiplayer interactions as natural generalizations of single-agent optimization while leveraging the same analytical tools for solution synthesis.

Extensions and Differences

Differential games extend the framework of by incorporating multiple decision-makers with conflicting objectives, leading to fundamental challenges absent in single-agent optimization. Unlike , where a single Hamiltonian function is minimized or maximized to derive necessary conditions via the Pontryagin maximum principle, differential games involve separate Hamiltonian functions for each player, reflecting their individual cost functionals. These Hamiltonians must be coupled through the , resulting in a system of coupled costate equations that intertwine the adjoint variables (costates) of all players, complicating the derivation of optimality conditions. For the game to possess a value, the Isaacs condition requires that the order of minimization and maximization over control sets can be interchanged, yielding a single Isaacs equation; without this condition, the value may not exist, as non-anticipative strategies fail to guarantee equilibrium. The adversarial nature of differential games further distinguishes them from , where controls are chosen to optimize against known dynamics or benign disturbances. In games, each player's control acts as a deliberate disturbance to the others, framing the problem as a worst-case where one agent's minimizes the maximum impact from opponents. This aligns with interpretations, where uncertainties or adversarial inputs are modeled as actions of a minimizing player seeking to maximize damage, ensuring guaranteed performance bounds such as H∞ norms that attenuate disturbances below a specified level. Consequently, solutions emphasize resilience against strategic opposition rather than mere efficiency. Solution complexities in differential games often manifest as value gaps, where the lower value (the maximum guaranteed payoff for the maximizer under optimal play) differs from the upper value (the minimum enforced loss for the minimizer), indicating the absence of a pure saddle-point equilibrium. This non-uniqueness arises from the multi-agent interactions and lack of convexity in joint control spaces, contrasting with optimal control's typical of a unique minimizer under standard assumptions. To address this, differential games prioritize synthesis—closed-loop strategies that depend on the current state—over open-loop solutions common in control, providing robustness to evolving adversarial responses; Isaacs' historical approach underscored this need for state-feedback policies to construct viable equilibria. For instance, while optimal control may yield a unique via open-loop controls, games may require approximate ε-saddle points, where strategies achieve near-equilibrium within a small deviation, as exact saddles do not exist without additional regularity conditions.

Special Cases and Variants

Pursuit-Evasion Games

Pursuit-evasion games represent a core in differential game theory, where a pursuer seeks to minimize the time required to capture an evader, while the evader aims to maximize this time or avoid capture altogether. These zero-sum games typically involve two players with state spaces governed by ordinary differential equations, bounded controls, and a terminal payoff defined by the capture condition. In the standard formulation, the pursuer's state xp\mathbf{x}_p evolves as x˙p=up\dot{\mathbf{x}}_p = \mathbf{u}_p with upvp\|\mathbf{u}_p\| \leq v_p, representing simple motion at bounded speed vpv_p, while the evader's state xe\mathbf{x}_e follows x˙e=ue\dot{\mathbf{x}}_e = \mathbf{u}_e with ueve\|\mathbf{u}_e\| \leq v_e, often with ve<vpv_e < v_p. Capture occurs when the distance xpxe<r\|\mathbf{x}_p - \mathbf{x}_e\| < r for some capture radius r>0r > 0, and the value of the game is the optimal time to terminal set under strategies. The relative dynamics can be analyzed in a frame fixed to the pursuer, reducing the problem to a single agent's motion relative to a moving target. A seminal example is the homicidal chauffeur game, introduced by Rufus Isaacs in 1965, modeling a car (pursuer) chasing a pedestrian (evader) in the plane. The pursuer maintains constant speed w=1w = 1 but has a minimum turning radius R=1R = 1, yielding dynamics x˙p=sinθ\dot{x}_p = \sin \theta, y˙p=cosθ\dot{y}_p = -\cos \theta, θ˙=u\dot{\theta} = u with u1/R|u| \leq 1/R; the evader has simple motion x˙e=v1\dot{x}_e = v_1, y˙e=v2\dot{y}_e = v_2 with v12+v22ϵ\sqrt{v_1^2 + v_2^2} \leq \epsilon
Add your contribution
Related Hubs
User Avatar
No comments yet.