Evangelos A. Theodorou’s research while affiliated with Georgia Institute of Technology and other places

What is this page?


This page lists works of an author who doesn't have a ResearchGate profile or hasn't added the works to their profile yet. It is automatically generated from public (personal) data to further our legitimate goal of comprehensive and accurate scientific recordkeeping. If you are this author and want this page removed, please let us know.

Publications (304)


Deep Generalized Schr\"odinger Bridges: From Image Generation to Solving Mean-Field Games
  • Preprint

December 2024

·

6 Reads

Guan-Horng Liu

·

Tianrong Chen

·

Evangelos A. Theodorou

Generalized Schr\"odinger Bridges (GSBs) are a fundamental mathematical framework used to analyze the most likely particle evolution based on the principle of least action including kinetic and potential energy. In parallel to their well-established presence in the theoretical realms of quantum mechanics and optimal transport, this paper focuses on an algorithmic perspective, aiming to enhance practical usage. Our motivated observation is that transportation problems with the optimality structures delineated by GSBs are pervasive across various scientific domains, such as generative modeling in machine learning, mean-field games in stochastic control, and more. Exploring the intrinsic connection between the mathematical modeling of GSBs and the modern algorithmic characterization therefore presents a crucial, yet untapped, avenue. In this paper, we reinterpret GSBs as probabilistic models and demonstrate that, with a delicate mathematical tool known as the nonlinear Feynman-Kac lemma, rich algorithmic concepts, such as likelihoods, variational gaps, and temporal differences, emerge naturally from the optimality structures of GSBs. The resulting computational framework, driven by deep learning and neural networks, operates in a fully continuous state space (i.e., mesh-free) and satisfies distribution constraints, setting it apart from prior numerical solvers relying on spatial discretization or constraint relaxation. We demonstrate the efficacy of our method in generative modeling and mean-field games, highlighting its transformative applications at the intersection of mathematical modeling, stochastic process, control, and machine learning.


Figure 1: Wall-clock time comparison. DeepDistributedQP, DistributedQP (ours) and OSQP on large-scale QPs.
Figure 3: The DeepDistributedQP architecture. The proposed framework relies on unrolling the DistributedQP optimizer as a supervised deep learning framework. In particular, we interpret its iterations (4)-(9) as sequential network layers and introduce learnable components (orange blocks) to facilitate reaching the desired accuracy after a predefined number of allowed iterations.
Figure 7: Left: Local vs shared policies. We showcase the advantage of learning local policies over shared ones. Right: Performance guarantees. The obtained generalization bounds guarantee the performance of DeepDistributedQP and its improvements over its standard optimization counterpart DistributedQP.
Training and testing details for DeepQP.
List of OSQP penalty parameters used in centralized experiments.
Deep Distributed Optimization for Large-Scale Quadratic Programming
  • Preprint
  • File available

December 2024

·

11 Reads

·

Hunter Kuperman

·

Alex Oshin

·

[...]

·

Evangelos A. Theodorou

Quadratic programming (QP) forms a crucial foundation in optimization, encompassing a broad spectrum of domains and serving as the basis for more advanced algorithms. Consequently, as the scale and complexity of modern applications continue to grow, the development of efficient and reliable QP algorithms is becoming increasingly vital. In this context, this paper introduces a novel deep learning-aided distributed optimization architecture designed for tackling large-scale QP problems. First, we combine the state-of-the-art Operator Splitting QP (OSQP) method with a consensus approach to derive DistributedQP, a new method tailored for network-structured problems, with convergence guarantees to optimality. Subsequently, we unfold this optimizer into a deep learning framework, leading to DeepDistributedQP, which leverages learned policies to accelerate reaching to desired accuracy within a restricted amount of iterations. Our approach is also theoretically grounded through Probably Approximately Correct (PAC)-Bayes theory, providing generalization bounds on the expected optimality gap for unseen problems. The proposed framework, as well as its centralized version DeepQP, significantly outperform their standard optimization counterparts on a variety of tasks such as randomly generated problems, optimal control, linear regression, transportation networks and others. Notably, DeepDistributedQP demonstrates strong generalization by training on small problems and scaling to solve much larger ones (up to 50K variables and 150K constraints) using the same policy. Moreover, it achieves orders-of-magnitude improvements in wall-clock time compared to OSQP. The certifiable performance guarantees of our approach are also demonstrated, ensuring higher-quality solutions over traditional optimizers.

Download

Dynamics Modeling using Visual Terrain Features for High-Speed Autonomous Off-Road Driving

November 2024

·

2 Reads

Rapid autonomous traversal of unstructured terrain is essential for scenarios such as disaster response, search and rescue, or planetary exploration. As a vehicle navigates at the limit of its capabilities over extreme terrain, its dynamics can change suddenly and dramatically. For example, high-speed and varying terrain can affect parameters such as traction, tire slip, and rolling resistance. To achieve effective planning in such environments, it is crucial to have a dynamics model that can accurately anticipate these conditions. In this work, we present a hybrid model that predicts the changing dynamics induced by the terrain as a function of visual inputs. We leverage a pre-trained visual foundation model (VFM) DINOv2, which provides rich features that encode fine-grained semantic information. To use this dynamics model for planning, we propose an end-to-end training architecture for a projection distance independent feature encoder that compresses the information from the VFM, enabling the creation of a lightweight map of the environment at runtime. We validate our architecture on an extensive dataset (hundreds of kilometers of aggressive off-road driving) collected across multiple locations as part of the DARPA Robotic Autonomy in Complex Environments with Resiliency (RACER) program. https://www.youtube.com/watch?v=dycTXxEosMk


Operator Splitting Covariance Steering for Safe Stochastic Nonlinear Control

November 2024

·

17 Reads

Most robotics applications are typically accompanied with safety restrictions that need to be satisfied with a high degree of confidence even in environments under uncertainty. Controlling the state distribution of a system and enforcing such specifications as distribution constraints is a promising approach for meeting such requirements. In this direction, covariance steering (CS) is an increasingly popular stochastic optimal control (SOC) framework for designing safe controllers via explicit constraints on the system covariance. Nevertheless, a major challenge in applying CS methods to systems with the nonlinear dynamics and chance constraints common in robotics is that the approximations needed are conservative and highly sensitive to the point of approximation. This can cause sequential convex programming methods to converge to poor local minima or incorrectly report problems as infeasible due to shifting constraints. This paper presents a novel algorithm for solving chance-constrained nonlinear CS problems that directly addresses this challenge. Specifically, we propose an operator-splitting approach that temporarily separates the main problem into subproblems that can be solved in parallel. The benefit of this relaxation lies in the fact that it does not require all iterates to satisfy all constraints simultaneously prior to convergence, thus enhancing the exploration capabilities of the algorithm for finding better solutions. Simulation results verify the ability of the proposed method to find higher quality solutions under stricter safety constraints than standard methods on a variety of robotic systems. Finally, the applicability of the algorithm on real systems is confirmed through hardware demonstrations.



Feedback Schr{\"o}dinger Bridge Matching

October 2024

·

7 Reads

Recent advancements in diffusion bridges for distribution transport problems have heavily relied on matching frameworks, yet existing methods often face a trade-off between scalability and access to optimal pairings during training. Fully unsupervised methods make minimal assumptions but incur high computational costs, limiting their practicality. On the other hand, imposing full supervision of the matching process with optimal pairings improves scalability, however, it can be infeasible in many applications. To strike a balance between scalability and minimal supervision, we introduce Feedback Schr\"{o}dinger Bridge Matching (FSBM), a novel semi-supervised matching framework that incorporates a small portion (less than 8% of the entire dataset) of pre-aligned pairs as state feedback to guide the transport map of non coupled samples, thereby significantly improving efficiency. This is achieved by formulating a static Entropic Optimal Transport (EOT) problem with an additional term capturing the semi-supervised guidance. The generalized EOT objective is then recast into a dynamic formulation to leverage the scalability of matching frameworks. Extensive experiments demonstrate that FSBM accelerates training and enhances generalization by leveraging coupled pairs guidance, opening new avenues for training matching frameworks with partially aligned datasets.



Second-Order Constrained Dynamic Optimization

September 2024

·

17 Reads

This paper provides an overview, analysis, and comparison of second-order dynamic optimization algorithms, i.e., constrained Differential Dynamic Programming (DDP) and Sequential Quadratic Programming (SQP). Although a variety of these algorithms has been proposed and used successfully, there exists a gap in understanding the key differences and advantages, which we aim to provide in this work. For constrained DDP, we choose methods that incorporate nolinear programming techniques to handle state and control constraints, including Augmented Lagrangian (AL), Interior Point, Primal Dual Augmented Lagrangian (PDAL), and Alternating Direction Method of Multipliers. Both DDP and SQP are provided in single- and multiple-shooting formulations, where constraints that arise from dynamics are encoded implicitly and explicitly, respectively. In addition to reviewing these methods, we propose a single-shooting PDAL DDP. As a byproduct of the review, we also propose a single-shooting PDAL DDP which is robust to the growth of penalty parameters and performs better than the normal AL variant. We perform extensive numerical experiments on a variety of systems with increasing complexity towards investigating the quality of the solutions, the levels of constraint violation, iterations for convergence, and the sensitivity of final solutions with respect to initialization. The results show that DDP often has the advantage of finding better local minima, while SQP tends to achieve better constraint satisfaction. For multiple-shooting formulation, both DDP and SQP can enjoy informed initial guesses, while the latter appears to be more advantageous in complex systems. It is also worth highlighting that DDP provides favorable computational complexity and feedback gains as a byproduct of optimization.


MPPI-Generic: A CUDA Library for Stochastic Optimization

September 2024

·

27 Reads

This paper introduces a new C++/CUDA library for GPU-accelerated stochastic optimization called MPPI-Generic. It provides implementations of Model Predictive Path Integral control, Tube-Model Predictive Path Integral Control, and Robust Model Predictive Path Integral Control, and allows for these algorithms to be used across many pre-existing dynamics models and cost functions. Furthermore, researchers can create their own dynamics models or cost functions following our API definitions without needing to change the actual Model Predictive Path Integral Control code. Finally, we compare computational performance to other popular implementations of Model Predictive Path Integral Control over a variety of GPUs to show the real-time capabilities our library can allow for. Library code can be found at: https://acdslab.github.io/mppi-generic-website/ .


Second Order Stein-Variational Dynamic Optimization

September 2024

·

6 Reads

We present a novel second-order trajectory optimization algorithm based on Stein Variational Newton's Method and Maximum Entropy Differential Dynamic Programming. The proposed algorithm, called Stein Variational Differential Dynamic Programming, is a kernel-based extension of Maximum Entropy Differential Dynamic Programming that combines the best of the two worlds of sampling-based and gradient-based optimization. The resulting algorithm avoids known drawbacks of gradient-based dynamic optimization in terms of getting stuck to local minima, while it overcomes limitations of sampling-based stochastic optimization in terms of introducing undesirable stochasticity when applied in online fashion. To test the efficacy of the proposed algorithm, experiments are performed for both trajectory optimization and model predictive control. The experiments include comparisons with unimodal and multimodal Maximum Entropy Differential Dynamic Programming as well as Model Predictive Path Integral Control and its multimodal and Stein Variational extensions. The results demonstrate the superior performance of the proposed algorithms and confirm the hypothesis that there is a middle ground between sampling and gradient-based optimization that is indeed beneficial for the purposes of dynamic optimization. This middle ground consists of different mechanisms that combine sampling with gradient-based optimization. In this paper, we investigate these different mechanisms and show their benefits in dealing with non-convex dynamic optimization problems found in trajectory optimization and model predictive control.


Citations (40)


... Since CS only concerns the first two moments of the system state, it is computationally tractable in comparison to density control problems. In addition, although the methodology originally focused on unconstrained linear systems, CS problems (CSPs) under nonlinear dynamics and chance constraints can be solved via iterative local approximations [17], [18], [19]. ...

Reference:

Operator Splitting Covariance Steering for Safe Stochastic Nonlinear Control
Distributed Model Predictive Covariance Steering
  • Citing Conference Paper
  • October 2024

... We also incorporate safety constraints in this framework, such as minimizing rollover risk. The specific cost function and the MPPI variation we use are described in [20]. The planning at this scale is optimized over a 5-second prediction horizon, requiring accurate and computationally efficient dynamics modeling for effective planning. ...

Low Frequency Sampling in Model Predictive Path Integral Control
  • Citing Article
  • May 2024

IEEE Robotics and Automation Letters

... Recent studies developed machine learning based methods that learn the underlying optimal policy through the use of neural networks. [4][5][6][7] Others such as Wasserstein Gradient Flow (WGF) algorithms include the use of proximal algorithms [8][9][10] in the context of Schrödinger Bridge problem and generative modeling 11 with a JKO numerical scheme, probabilistic density control using an approximate Perron-Frobenius operator, 12, 13 density steering based on power moments, 14 Gaussian Mixture Models 15 with convex optimization, denoising diffusion-based control 16 and Wasserstein Tube MPC by utilizing Wasserstein ambiguity sets 17 for distributionally robust optimization (DRO). Moreover, alternatives has also been proposed for designing robust controllers for stochastic optimal control problems. ...

Deep L1 Stochastic Optimal Control Policies for Planetary Soft Landing
  • Citing Article
  • March 2024

Journal of Guidance, Control, and Dynamics

... Patel et al. [2] presented an MPPI framework for USV control under wave disturbances using Froude-Krylov theory. While they showed robust control under environmental disturbances, their work focused on trajectory tracking rather than the specific challenges of docking maneuvers that we address. ...

Model-Predictive Path-Integral Control of an Unmanned Surface Vessel with Wave Disturbance
  • Citing Conference Paper
  • September 2023

... Given these challenges, decentralized and autonomous control methods present a more promising alternative. By distributing decision-making authority across multiple agents, each agent can independently adjust its behavior based on local information without relying on a central command, enhancing the system's flexibility and scalability [21,22]. In the field of unmanned aerial vehicles (UAVs), decentralized control methods have been successfully applied in navigation and formation control [23,24]. ...

Distributed Differential Dynamic Programming Architectures for Large-Scale Multiagent Control
  • Citing Article
  • December 2023

IEEE Transactions on Robotics

... С другой стороны, с точки зрения управления роботом избыточность его степеней свободы усложняет расчеты и увеличивает время принятия решения. Решение задач управления роботами в условиях неопределенности, когда требуется принятие решения в текущем моменте (например, по выбору траектории, обходу препятствия и т.п.), находится в центре современных научных исследований в области робототехники и автоматизации и управления [1][2][3][4][5][6][7][8][9]. ...

Differentiable Robust Model Predictive Control

... Furthermore, [24] presents a PDE-based optimal robotic swarm coverage control policy, while [25] focuses on deriving the optimal density steering laws for systems with multiplicative noise for the infinite horizon case. Lastly, [26] introduces a hierarchical clustering-based density steering algorithm tailored for applications of distributed largescale robotic networks. ...

Distributed Hierarchical Distribution Control for Very-Large-Scale Clustered Multi-Agent Systems

... As in our previous work [7], we learn a dynamics model for our planner, Model Predictive Path Integral (MPPI) [22], to generate optimal trajectories for the vehicle. At a high level, MPPI is a sampling-based planner, and it operates by sampling various trajectories, performing forward rollouts of the dynamics, and optimizing the trajectories based on the cost of these sampled rollouts. ...

A Multi-step Dynamics Modeling Framework For Autonomous Driving In Multiple Environments
  • Citing Conference Paper
  • May 2023

... Approaches that consider a non-cooperative setting, where each robot has its own reward function (representing a different task), typically tackle the problem within the framework of dynamic games by reasoning about the Nash equilibrium of the multi-robot system (e.g. [14], [18], [20]), or by leveraging multi-objective optimization (e.g. [12]). ...

MPOGames: Efficient Multimodal Partially Observable Dynamic Games