BACK TO INDEX

Publications about 'optimization problems'
Articles in journal or book chapters
  1. P. Mestres, J. Cortés, and E.D. Sontag. Neural network-based universal formulas for control. Systems and Control Letters, 2026. Note: Submitted. Also arXiv https://arxiv.org/abs/2505.24744. Keyword(s): machine learning, artificial intelligence, control-Lyapunov functions, control barrier functions, universal formulas, neural networks.
    Abstract:
    We study the problem of designing a controller that satisfies an arbitrary number of affine inequalities at every point in the state space. This is motivated by the use of guardrails in autonomous systems. Indeed, a variety of key control objectives, such as stability, safety, and input saturation, are guaranteed by closed-loop systems whose controllers satisfy such inequalities. Many works in the literature design such controllers as the solution to a state-dependent quadratic program (QP) whose constraints are precisely the inequalities. When the input dimension and number of constraints are high, computing a solution of this QP in real time can become computationally burdensome. Additionally, the solution of such optimization problems is not smooth in general, which can degrade the performance of the system. This paper provides a novel method to design a smooth controller that satisfies an arbitrary number of affine constraints. This why we refer to it as a universal formula for control. The controller is given at every state as the minimizer of a strictly convex function. To avoid computing the minimizer of such function in real time, we introduce a method based on neural networks (NN) to approximate the controller. Remarkably, this NN can be used to solve the controller design problem for any task with less than a fixed input dimension and number of affine constraints, and is completely independent of the state dimension. Additionally, we show that the NN-based controller only needs to be trained with datapoints from a compact set in the state space, which significantly simplifies the training process. Various simulations showcase the performance of the proposed solution, and also show that the NN-based controller can be used to warmstart an optimization scheme that refines the approximation of the true controller in real time, significantly reducing the computational cost compared to a generic initialization.


  2. D.D. Jatkar, M.A. Al-Radhawi, C. A. Voigt, and E.D. Sontag. Modeling and minimization of dCas9-induced competition in CRISPRi-based genetic circuits. bioRxiv, 2025. [WWW] [doi:10.1101/2025.11.05.686856] Keyword(s): CRISPRi, retroactivity, feedback, logic-circuit, synthetic biology.
    Abstract:
    Implementing logic functions in living cells is a fundamental area of interest among synthetic biologists. The goal of designing biochemical circuits in synthetic biology is to make modular and tractable systems that perform well with predictable behaviors. Developing formalisms towards the design of such systems has proven to be difficult with the diverse retroactive effects that appear with respect to the context of the cell. Repressor-based circuits have various applications in biosynthesis, therapeutics, and bioremediation. Particularly using CRISPRi, competition for components of the system (unbound dCas9) can affect the achievable dynamic range of repression. Moreover, the toxicity of dCas9 via non-specific binding inhibits high levels of expression and limits the performance of genetic circuits. In this work, we study the computation of Boolean functions through CRISPRi based circuits built out of NOT and NOR gates. We provide algebraic expressions that allow us to evaluate the steady-state behaviors of any realized circuit. Our mathematical analysis reveals that the effective non-cooperativity of any given gate is a major bottleneck for increasing the dynamic range of the outputs. Further, we find that under the condition of competition between promoters for dCas9, certain circuit architectures perform better than others depending on factors such as circuit depth, fan-in, and fan-out. We pose optimization problems to evaluate the effects engineerable parameter values to find regimes in which a given circuit performs best. This framework provides a mathematical template and computational library for evaluating the performance of repressor-based circuits with a focus on effective cooperativity.


  3. E.D. Sontag. Some remarks on gradient dominance and LQR policy optimization. arXiv 2507.10452, 2025. [PDF] [doi:https://doi.org/10.48550/arXiv.2507.10452] Keyword(s): gradient dominance, gradient flows, LQR, reinforcement learning, machine learning, artificial intelligence, optimal control.
    Abstract:
    Solutions of optimization problems, including policy optimization in reinforcement learning, typically rely upon some variant of gradient descent. There has been much recent work in the machine learning, control, and optimization communities applying the Polyak-Åojasiewicz Inequality (PLI) to such problems in order to establish an exponential rate of convergence (a.k.a. ``linear convergence'' in the local-iteration language of numerical analysis) of loss functions to their minima under the gradient flow. Often, as is the case of policy iteration for the continuous-time LQR problem, this rate vanishes for large initial conditions, resulting in a mixed globally linear / locally exponential behavior. This is in sharp contrast with the discrete-time LQR problem, where there is global exponential convergence. That gap between CT and DT behaviors motivates the search for various generalized PLI-like conditions, and this paper addresses that topic. Moreover, these generalizations are key to understanding the transient and asymptotic effects of errors in the estimation of the gradient, errors which might arise from adversarial attacks, wrong evaluation by an oracle, early stopping of a simulation, inaccurate and very approximate digital twins, stochastic computations (algorithm ``reproducibility''), or learning by sampling from limited data. We describe an ``input to state stability'' (ISS) analysis of this issue. We also discuss convergence and PLI-like properties of ``linear feedforward neural networks'' in feedback control. Much of the work described here was done in collaboration with Arthur Castello B. de Oliveira, Leilei Cui, Zhong-Ping Jiang, and Milad Siami. This is a short paper summarizing the slides presented at my keynote at the 2025 L4DC (Learning for Dynamics \& Control Conference) in Ann Arbor, Michigan, 05 June 2025. A partial bibliography has been added.


Conference articles
  1. L. Cui, Z.P. Jiang, and E. D. Sontag. Small-covariance noise-to-state stability of stochastic systems and its applications to stochastic gradient dynamics. In 2026 American Control Conference (ACC), 2026. Note: Submitted. Also arXiv:2509.24277. [PDF] [doi:https://doi.org/10.48550/arXiv.2509.24277] Keyword(s): noise to state stability, input to state stability, stochastic systems.
    Abstract:
    This paper studies gradient dynamics subject to additive stochastic noise, which may arise from sources such as stochastic gradient estimation, measurement noise, or stochastic sampling errors. To analyze the robustness of such stochastic gradient systems, the concept of small-covariance noise-to-state stability (NSS) is introduced, along with a Lyapunov-based characterization. Furthermore, the classical Polyak–Lojasiewicz (PL) condition on the objective function is generalized to the $\mathcal{K}$-PL condition via comparison functions, thereby extending its applicability to a broader class of optimization problems. It is shown that the stochastic gradient dynamics exhibit small-covariance NSS if the objective function satisfies the $\mathcal{K}$-PL condition and possesses a globally Lipschitz continuous gradient. This result implies that the trajectories of stochastic gradient dynamics converge to a neighborhood of the optimum with high probability, with the size of the neighborhood determined by the noise covariance. Moreover, if the $\mathcal{K}$-PL condition is strengthened to a $\mathcal{K}_\infty$-PL condition, the dynamics are NSS; whereas if it is weakened to a general positive-definite-PL condition, the dynamics exhibit integral NSS. The results further extend to objectives without globally Lipschitz gradients through appropriate step-size tuning. The proposed framework is further applied to the robustness analysis of policy optimization for the linear quadratic regulator (LQR) and logistic regression.


  2. M.K. Wafi, A.C.B de Oliveira, and E.D. Sontag. On the (almost) global exponential convergence of overparameterized policy optimization for the LQR problem. In 2026 American Control Conference (ACC), 2025. Note: Submitted. Also arXiv:2510.02140. [PDF] Keyword(s): machine learning, artificial intelligence, gradient dominance, gradient flows, LQR, reinforcement learning.
    Abstract:
    In this work we study the convergence of gradient methods for nonconvex optimization problems -- specifically the effect of the problem formulation to the convergence behavior of the solution of a gradient flow. We show through a simple example that, surprisingly, the gradient flow solution can be exponentially or asymptotically convergent, depending on how the problem is formulated. We then deepen the analysis and show that a policy optimization strategy for the continuous-time linear quadratic regulator (LQR) (which is known to present only asymptotic convergence globally) presents almost global exponential convergence if the problem is overparameterized through a linear feed-forward neural network (LFFNN). We prove this qualitative improvement always happens for a simplified version of the LQR problem and derive explicit convergence rates for the gradient flow. Finally, we show that both the qualitative improvement and the quantitative rate gains persist in the general LQR through numerical simulations.


  3. A.C.B de Oliveira, L. Cui, and E. D. Sontag. Remarks on the Polyak-Lojasiewicz inequality and the convergence of gradient systems. In Proc. 64th IEEE Conference on Decision and Control (CDC), 2025. Note: To appear. Extended version in arXiv:2503.23641. [PDF] [doi:https://doi.org/10.48550/arXiv.2503.23641] Keyword(s): machine learning, artificial intelligence, gradient dominance, gradient flows, LQR, reinforcement learning.
    Abstract:
    This work explores generalizations of the Polyak-Lojasiewicz inequality (PLI) and their implications for the convergence behavior of gradient flows in optimization problems. Motivated by the continuous-time linear quadratic regulator (CT-LQR) policy optimization problem -- where only a weaker version of the PLI is characterized in the literature -- this work shows that while weaker conditions are sufficient for global convergence to, and optimality of the set of critical points of the cost function, the "profile" of the gradient flow solution can change significantly depending on which "flavor" of inequality the cost satisfies. After a general theoretical analysis, we focus on fitting the CT-LQR policy optimization problem to the proposed framework, showing that, in fact, it can never satisfy a PLI in its strongest form. We follow up our analysis with a brief discussion on the difference between continuous- and discrete-time LQR policy optimization, and end the paper with some intuition on the extension of this framework to optimization problems with L1 regularization and solved through proximal gradient flows.


  4. C. Darken, M.J. Donahue, L. Gurvits, and E.D. Sontag. Rate of approximation results motivated by robust neural network learning. In COLT '93: Proceedings of the sixth annual conference on Computational learning theory, New York, NY, USA, pages 303-309, 1993. ACM Press. [doi:http://doi.acm.org/10.1145/168304.168357] Keyword(s): machine learning, artificial intelligence, machine learning, neural networks, optimization problems, approximation theory.



BACK TO INDEX




Disclaimer:

This material is presented to ensure timely dissemination of scholarly and technical work. Copyright and all rights therein are retained by authors or by other copyright holders.




Last modified: Thu Nov 6 10:10:57 2025
Author: sontag.


This document was translated from BibTEX by bibtex2html