STAR-LQG is a MATLAB framework for learning-based robust control in linear-quadratic (LQ) games. It enables a controlled agent to iteratively identify and adapt to unknown adversary strategies while ensuring robust stability and near-optimal performance under bounded disturbances.
This approach combines set-membership estimation with robust LQR design via linear matrix inequalities (LMIs) to achieve an
Traditional methods in multi-agent learning and differential games often require players to alternate between learning and control phases, which is unrealistic in dynamic environments where agents must continuously adapt in real time. Existing adaptive schemes frequently rely on restrictive assumptions, such as requiring all agents to follow the same adaptation law or neglecting external disturbances.
STAR-LQG addresses these limitations by:
- Learning adversary strategies online using set-membership estimation.
- Ensuring robust closed-loop stability for all plausible adversary strategies.
- Adapting policies in real time, even under unknown adversary behavior and bounded disturbances.
-
Maintaining convergence to a neighborhood-optimal solution, formalized as an
$\epsilon$ -Nash equilibrium.
The system dynamics are governed by a linear model:
where:
-
$\mathbf{x} \in \mathbb{R}^{n_x}$ : system state -
$u_1 \in \mathbb{R}^{n_{u_1}}$ : controlled agent input -
$u_2 \in \mathbb{R}^{n_{u_2}}$ : adversary input (unknown) -
$\mathbf{w}(t) \in \mathbb{R}^{n_x}$ : bounded external disturbance -
$\mathbf{A}$ ,$\mathbf{B}_1$ : known system matrices -
$\mathbf{B}_2$ : unknown adversary input matrix
Each player minimizes an infinite-horizon quadratic cost:
The controlled agent seeks a feedback gain
where
The set of admissible disturbances is modeled as a convex polytope:
- Infinite-horizon LQ game solver with analytical Nash equilibrium computation.
- Set-membership estimation of adversary strategies from online data.
- Robust LQR synthesis via LMIs to guarantee closed-loop stability.
- Real-time policy adaptation under unknown adversary behavior and disturbances.
- MATLAB implementation with simulations and visualization scripts.
git clone https://github.com/TUM-ITR/STAR-LQG.git
cd STAR-LQGEnsure you have MATLAB installed with:
- Optimization Toolbox
- Control System Toolbox
- CVX (for convex optimization)
STAR-LQG/
│
├── main/ # Main simulation scripts
├── functions/ # Core algorithms: LQ solvers, LS, set-membership
├── plots/ # Scripts to visualize simulation results
├── examples/ # Example systems and parameter setups
└── README.md