Differentiable Model Predictive Control on the GPU
Overview
Overall Novelty Assessment
The paper introduces a GPU-accelerated differentiable MPC solver combining sequential quadratic programming with a custom preconditioned conjugate gradient routine featuring tridiagonal preconditioning. It resides in the 'Preconditioned Conjugate Gradient Approaches' leaf, which contains only two papers total (including this one and one sibling). This is a notably sparse research direction within the broader taxonomy of 36 papers across 28 leaf nodes, suggesting the specific combination of GPU acceleration, differentiability, and PCG-based iterative solvers remains relatively underexplored compared to sampling-based or direct shooting methods.
The taxonomy reveals several neighboring directions: the sibling 'Condensed-Space Interior-Point Methods' leaf focuses on eliminating state variables rather than iterative PCG solvers, while the parallel 'Parallel Shooting and Direct Methods' branch employs primal-dual KKT solvers or multilevel SQP without emphasizing PCG preconditioning. Nearby 'Differentiable MPC for End-to-End Learning' nodes integrate MPC with actor-critic frameworks but do not necessarily prioritize iterative linear system solvers. The paper's position bridges gradient-based optimization infrastructure with learning-integrated applications, sitting at the intersection of solver design and end-to-end training pipelines.
Among 16 candidates examined across three contributions, none clearly refute the core claims. The GPU-accelerated differentiable optimization tool examined 7 candidates with 0 refutations; the tridiagonal PCG routine examined 1 candidate with 0 refutations; and the robust drifting application examined 8 candidates with 0 refutations. This limited search scope—focused on top-K semantic matches—suggests that within the examined literature, the specific combination of SQP, custom PCG preconditioning, and differentiability for learning tasks appears distinct. However, the small candidate pool means the analysis does not capture exhaustive prior work in iterative MPC solvers or GPU optimization.
Given the sparse taxonomy leaf and absence of refutations among 16 examined candidates, the work appears to occupy a relatively novel niche within GPU-accelerated MPC. The emphasis on tridiagonal preconditioning for differentiable SQP distinguishes it from both sampling-heavy methods and direct KKT approaches. Nonetheless, the limited search scope and small sibling set mean this assessment reflects positioning within a focused literature subset rather than comprehensive field coverage.
Taxonomy
Research Landscape Overview
Claimed Contributions
The authors introduce DiffMPC, a differentiable solver for model predictive control that runs efficiently on GPUs. It uses sequential quadratic programming combined with a custom preconditioned conjugate gradient routine that exploits the sparse-in-time structure of optimal control problems to enable parallelization over time steps.
The authors adapt and implement a PCG routine with tridiagonal preconditioning that solves the KKT linear systems by exploiting the block-tridiagonal structure of the Schur complement. This design enables parallelization over time steps and supports warm-starting, making it suitable for GPU execution.
The authors demonstrate DiffMPC on a reinforcement learning task for autonomous drifting under model mismatch. They use domain randomization over nonlinear dynamics to learn MPC cost and vehicle parameters, achieving robust drifting through water puddles on a Toyota Supra.
Core Task Comparisons
Comparisons with papers in the same taxonomy category
[2] Mpcgpu: Real-time nonlinear model predictive control through preconditioned conjugate gradient on the gpu PDF
Contribution Analysis
Detailed comparisons for each claimed contribution
GPU-accelerated differentiable optimization tool for MPC
The authors introduce DiffMPC, a differentiable solver for model predictive control that runs efficiently on GPUs. It uses sequential quadratic programming combined with a custom preconditioned conjugate gradient routine that exploits the sparse-in-time structure of optimal control problems to enable parallelization over time steps.
[3] Physics-guided neural network and GPU-accelerated nonlinear model predictive control for quadcopter PDF
[19] Multilevel parallel GPU implementation of SQP solvers for Nonlinear MPC PDF
[23] ReLU-QP: A GPU-Accelerated Quadratic Programming Solver for Model-Predictive Control PDF
[46] Parallel Shooting Sequential Quadratic Programming for Nonlinear MPC Problems PDF
[47] Structure-exploiting sequential quadratic programming for model-predictive control PDF
[48] Towards safe and tractable Gaussian process-based MPC: Efficient sampling within a sequential quadratic programming framework PDF
[49] DEEP FLEXQP: ACCELERATED NONLINEAR PRO PDF
Preconditioned conjugate gradient routine with tridiagonal preconditioning
The authors adapt and implement a PCG routine with tridiagonal preconditioning that solves the KKT linear systems by exploiting the block-tridiagonal structure of the Schur complement. This design enables parallelization over time steps and supports warm-starting, making it suitable for GPU execution.
[37] Implementation of a distributed parallel in time scheme using PETSc for a parabolic optimal control problem PDF
Application to robust drifting via domain randomization and RL
The authors demonstrate DiffMPC on a reinforcement learning task for autonomous drifting under model mismatch. They use domain randomization over nonlinear dynamics to learn MPC cost and vehicle parameters, achieving robust drifting through water puddles on a Toyota Supra.