Loading…
Type: Parallel Session clear filter
arrow_back View All Dates
Tuesday, July 22
 

10:30am PDT

Parallel Sessions 4A: Progress in Nonsmooth Optimization
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Progress in Nonsmooth Optimization
Chair: Feng Ruan
Cluster: Optimization Under Uncertainty and Data-driven Optimization

Talk 1: Subgradient Convergence Implies Subdifferential Convergence on Weakly Convex Functions: With Uniform Rates Guarantees
Speaker: Feng Ruan
Abstract: In nonsmooth, nonconvex stochastic optimization, understanding the uniform convergence of subdifferential mappings is crucial for analyzing stationary points of sample average approximations of risk as they approach the population risk. Yet, characterizing this convergence remains a fundamental challenge. This work introduces a novel perspective by connecting the uniform convergence of subdifferential mappings to that of subgradient mappings as empirical risk converges to the population risk. We prove that, for stochastic weakly-convex objectives, and within any open set, a uniform bound on the convergence of subgradients -- chosen arbitrarily from the corresponding subdifferential sets -- translates to a uniform bound on the convergence of the subdifferential sets itself, measured by the Hausdorff metric. Using this technique, we derive uniform convergence rates for subdifferential sets of stochastic convex-composite objectives. Our results do not rely on key distributional assumptions in the literature, which require the population and finite sample subdifferentials to be continuous in the Hausdorff metric, yet still provide tight convergence rates. These guarantees lead to new insights into the nonsmooth landscapes of such objectives within finite samples.

Talk 2: Variational Theory and Algorithms for a Class of Asymptotically Approachable Nonconvex Problems
Speaker: Ying Cui
Abstract: We investigate a class of composite nonconvex functions, where the outer function is the sum of univariate extended-real-valued convex functions and the inner function is the limit of difference-of-convex functions. A notable feature of this class is that the inner function can be merely lower semicontinuous instead of continuously differentiable. It covers a range of important yet challenging applications, including the composite value functions of nonlinear programs and the value-at-risk constraints. We propose an asymptotic decomposition of the composite function that guarantees epi-convergence to the original function, leading to necessary optimality conditions for the corresponding minimization problem. The proposed decomposition also enables us to design a numerical algorithm such that any accumulation point of the generated sequence, if exists, satisfies the newly introduced optimality conditions. These results expand on the study of so-called amenable functions introduced by Poliquin and Rockafellar in 1992, which are compositions of convex functions with smooth maps, and the prox-linear methods for their minimization.

Talk 3: Survey Descent: a Case-Study in Amplifying Optimization Research with Modern ML Workflows
Speaker: X.Y. Han
Abstract: Within the classic optimization, one learns that for strongly convex objectives that are smooth, gradient descent ensures linear convergence of iterates and objective values relative to the number of gradient evaluations. Nonsmooth objective functions are more challenging: existing solutions typically invoke cutting plane methods whose complexities are difficult to bound, leading to convergence guarantees that are sublinear in the cumulative number of gradient evaluations. We instead propose a multipoint generalization of the gradient descent called Survey Descent. In this method, one first leverages a one-time initialization procedure to gather a "survey" of points. Then, during each iteration of the method, the survey points are updated in parallel using a simple, four-line procedure inspired by gradient descent. Under certain regularity conditions, we prove that Survey Descent then achieves a desirable performance by converging linearly to the optimal solution in the nonsmooth setting. Despite being an nominally mathematical endeavor, we discuss how the development of Survey Descent was significantly accelerated by a frictionless computational workflow made possible by tools from modern machine learning (ML); how this model of applying new ML workflows to solve open questions in optimization and applied probability could amplify the researchers' productivity; practical computational bottlenecks that could hinder this integration; and what tools are needed to overcome those obstacles.

Speakers
FR

Feng Ruan

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
YC

Ying Cui

Name: Dr. Slothington "Slow Convergence" McNapfaceTitle: Distinguished Professor of Continuous Optimization & Energy MinimizationAffiliation: The Lush Canopy Institute of Sluggish AlgorithmsBio:Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
avatar for X.Y. Han

X.Y. Han

Assistant Professor, Chicago Booth
Name: Prof. X.Y. HanTitle: Assistant Professor of Operations Management and Applied AIAffiliation: Chicago BoothBio:X.Y. Han is an assistant professor of Operations Management and Applied Artificial Intelligence at the University of Chicago, Booth School of Business. His research... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 101 3501 Trousdale Pkwy, 101, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4B: Optimization for Neural Network Pruning and Quantization
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Optimization for Neural Network Pruning and Quantization
Chair: Lin Xiao
Cluster: Optimization Applications (Communication, Energy, Health, ML, ...)

Talk 1: Understanding Neural Network Quantization and its Robustness Against Data Poisoning Attacks
Speaker: Yiwei Lu
Abstract: Neural network quantization, exemplified by BinaryConnect (BC) and its variants, has become a standard approach for model compression. These methods typically employ the sign function in the forward pass, with various approximations used for gradient computation during backpropagation. While effective, these techniques often rely on heuristics or "training tricks" that lack theoretical grounding. This talk explores the optimization perspective of these quantization methods, introducing forward-backward quantizers as a principled framework. We present ProxConnect++ (PC++), a generalization that encompasses existing quantization techniques and provides automatic theoretical guarantees. Furthermore, we reveal an unexpected benefit of neural network quantization: enhanced robustness against data poisoning attacks.

Talk 2: Convex Regularizations for Pruning- and Quantization-Aware Training of Neural Networks
Speaker: Lin Xiao
Abstract: We present a convex regularization approach for pruning- and quantization-aware training of deep neural networks. While it is well-known that group Lasso can induce structured pruning, we show that convex, piece-wise affine regularizations (PAR) can effectively induce quantization. Previous work have limited success in practice due to the challenge of integrating structured regularization with stochastic gradient methods. We derive an aggregate proximal stochastic gradient method (AProx) that can successfully produce desired pruning and quantization results. Moreover, we establish last-iterate convergence of the method, which better supports the computational practice than the classical theory of average-iterate convergence.

Talk 3: Quantization through Piecewise-Affine Regularization: Optimization and Statistical Guarantees
Speaker: Jianhao Ma
Abstract: Optimization problems involving discrete or quantized variables can be very challenging due to the combinatorial nature of the design space. We show that (coordinate-wise) piecewise-affine regularization (PAR) can effectively induce quantization in the optimization variables. PAR provides a flexible modeling and computational framework for quantization based on continuous and convex optimization. In addition, for linear regression problems, we can approximate $\ell_1$- and squared $\ell_2$-regularizations using different parameterizations of PAR, and obtain statistical guarantees that are similar to those of Lasso and ridge regression, all with quantized regression variables.

Speakers
YL

Yiwei Lu

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
LX

Lin Xiao

Lin Xiao is a Research Scientist at Facebook AI Research (FAIR) in Seattle, Washington. He received BE from Beijing University of Aeronautics and Astronautics (Beihang University) and PhD from Stanford University, and was a postdoctoral fellow in the Center for the Mathematics of... Read More →
JM

Jianhao Ma

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 201 3501 Trousdale Pkwy, 201, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4C: Methods for Large-Scale Nonlinear Optimization IV
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Methods for Large-Scale Nonlinear Optimization IV
Chair: Albert Berahas
Cluster: Nonlinear Optimization

Talk 1: High Probability Analysis for Negative Curvature Methods with Probabilistic Oracles
Speaker: Wanping Dong
Abstract: We consider a negative curvature method for continuous nonlinear nonconstrained optimization problems under a stochastic setting where the function values, gradients, and Hessian (products) are available only through inexact probabilistic oracles. Our goal is to develop algorithms that have high probabilistic second-order convergence and affordable complexity so that they can be used for large-scale problems. We introduce general conditions on the probabilistic oracles and propose a method that dynamically chooses between negative curvature and descent steps. We derive a high probability tail bound on the iteration complexity of the algorithm and show improvements compared to our previous negative curvature method. A practical variant is implemented to illustrate the power of the proposed algorithm.

Talk 2: Retrospective Approximation for Stochastic Constrained Problems Using Sequential Quadratic Programming
Speaker: Shagun Gupta
Abstract: Sequential Quadratic Programming (SQP) is one of the state-of-the-art algorithms used to solve deterministic constrained nonlinear optimization problems. In recent years, the framework has been extended to solve deterministic equality and inequality constrained problems with stochastic objective functions. In response to the challenges posed by stochasticity, various schemes have been incorporated into SQP algorithms to adapt key parameters, such as step size and merit parameter, from the deterministic setting to the stochastic setting. These include stochastic line search, Lipschitz constant estimation, and Hessian averaging. In our work, we leverage SQP algorithms within the innovative framework of Retrospective Approximation. This framework introduces a novel approach to solving stochastic constrained problems by allowing the SQP algorithm to solve a series of subsampled deterministic subproblems. Each deterministic subproblem is solved not to optimality, but to a specified accuracy with an increasing sample size for the subsampled deterministic problems. This strategic decoupling of stochasticity from the SQP algorithm proves instrumental, enabling the utilization of legacy deterministic solvers. Thus, by decoupling stochasticity, the Retrospective Approximation framework facilitates the integration of legacy deterministic solvers, reducing requirements for hyper-parameter tuning in stochastic settings. We provide theoretical convergence requirements for the increase in the subsampling batch size and required solution accuracy for deterministic subproblems. We also conduct numerical experiments to showcase the utilization of legacy deterministic solvers for stochastic constrained problems.

Talk 3: Stochastic Second-order Inexact Augmented Lagrangian Framework for Nonconvex Expectation Constrained Optimization
Speaker: Yash Kumar
Abstract: In this talk, we present methods for solving stochastic nonconvex optimization problems where both the objective function and the constraints are expectations of stochastic functions. We consider an inexact Augmented Lagrangian framework for solving these problems, employing stochastic second-order methods for the subproblems instead of first-order methods. This framework ensures convergence to second-order stationary points instead of approximate first-order stationary points. Furthermore, these methods do not require access to full Hessians but only Hessian-vector products, which are typically twice the computational cost of gradients. We provide convergence guarantees for the stochastic second-order inexact Augmented Lagrangian framework, along with total computational complexity guarantees for various second-order subproblem solvers. Numerical experiments on constrained machine learning classification problems demonstrate the efficiency of the proposed framework.

Speakers
AB

Albert Berahas

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
WD

Wanping Dong

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
SG

Shagun Gupta

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
YK

Yash Kumar

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 208 3501 Trousdale Pkwy, 208, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4D: Recent Advances in Large-scale Optimization II
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Recent Advances in Large-scale Optimization II
Chair: Salar Fattahi
Cluster: Nonlinear Optimization

Talk 1: Distributionally Robust Optimization via Iterative Algorithms in Continuous Probability Space
Speaker: Yao Xie
Abstract: We consider a minimax problem motivated by distributionally robust optimization (DRO) when the worst-case distribution is continuous, leading to significant computational challenges due to the infinite-dimensional nature of the optimization problem. Recent research has explored learning the worst-case distribution using neural network-based generative models to address these computational challenges but lacks algorithmic convergence guarantees. This paper bridges this theoretical gap by presenting an iterative algorithm to solve such a minimax problem, achieving global convergence under mild assumptions and leveraging technical tools from vector space minimax optimization and convex analysis in the space of continuous probability densities. In particular, leveraging Brenier's theorem, we represent the worst-case distribution as a transport map applied to a continuous reference measure and reformulate the regularized discrepancy-based DRO as a minimax problem in the Wasserstein space. Furthermore, we demonstrate that the worst-case distribution can be efficiently computed using a modified Jordan-Kinderlehrer-Otto (JKO) scheme with sufficiently large regularization parameters for commonly used discrepancy functions linked to the radius of the ambiguity set. Additionally, we derive the global convergence rate and quantify the total number of subgradient and inexact modified JKO iterations required to obtain approximate stationary points. These results are potentially apply to nonconvex and nonsmooth scenarios, with broad relevance to modern machine learning applications.

Talk 2: Federated Natural Policy Gradient and Actor Critic Methods for Multi-task Reinforcement Learning
Speaker: Yuejie Chi
Abstract: Federated reinforcement learning (RL) enables collaborative decision making of multiple distributed agents without sharing local data trajectories. In this work, we consider a multi-task setting, in which each agent has its own private reward function corresponding to different tasks, while sharing the same transition kernel of the environment. Focusing on infinite-horizon Markov decision processes, the goal is to learn a globally optimal policy that maximizes the sum of the discounted total rewards of all the agents in a decentralized manner, where each agent only communicates with its neighbors over some prescribed graph topology. We develop federated vanilla and entropy-regularized natural policy gradient (NPG) methods in the tabular setting under softmax parameterization, where gradient tracking is applied to estimate the global Q-function to mitigate the impact of imperfect information sharing. We establish non-asymptotic global convergence guarantees under exact policy evaluation, where the rates are nearly independent of the size of the state-action space and illuminate the impacts of network size and connectivity. To the best of our knowledge, this is the first time that near dimension-free global convergence is established for federated multi-task RL using policy optimization. We further go beyond the tabular setting by proposing a federated natural actor critic (NAC) method for multi-task RL with function approximation, and establish its finite-time sample complexity taking the errors of function approximation into account.

Talk 3: Invariant Low-Dimensional Subspaces in Gradient Descent for Learning Deep Networks
Speaker: Qing Qu
Abstract: Over the past few years, an extensively studied phenomenon in training deep networks is the implicit bias of gradient descent towards parsimonious solutions. In this work, we first investigate this phenomenon by narrowing our focus to deep linear networks. Through our analysis, we reveal a surprising "law of parsimony" in the learning dynamics when the data possesses low-dimensional structures. Specifically, we show that the evolution of gradient descent starting from orthogonal initialization only affects a minimal portion of singular vector spaces across all weight matrices. In other words, the learning process happens only within a small invariant subspace of each weight matrix, even though all weight parameters are updated throughout training. This simplicity in learning dynamics could have significant implications for both efficient training and a better understanding of deep networks. First, the analysis enables us to considerably improve training efficiency by taking advantage of the low-dimensional structure in learning dynamics. We can construct smaller, equivalent deep linear networks without sacrificing the benefits associated with the wider counterparts. Moreover, we demonstrate the potential implications for efficient training deep nonlinear networks. Second, it allows us to better understand deep representation learning by elucidating the progressive feature compression and discrimination from shallow to deep layers. The study paves the foundation for understanding hierarchical representations in deep nonlinear networks.

Speakers
SF

Salar Fattahi

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
YX

Yao Xie

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
YC

Yuejie Chi

Name: Dr. Slothington "Slow Convergence" McNapfaceTitle: Distinguished Professor of Continuous Optimization & Energy MinimizationAffiliation: The Lush Canopy Institute of Sluggish AlgorithmsBio:Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
QQ

Qing Qu

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 210 3501 Trousdale Pkwy, 210, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4E: Random subspace algorithms in optimization
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Random subspace algorithms in optimization
Chair: Pierre-Louis Poirion
Cluster: Optimization For Data Science

Talk 1: Zeroth-order Random Subspace AAlgorithm for Non-smooth Convex Optimization
Speaker: Akiko Takeda
Abstract: Zeroth-order optimization, which does not use derivative information, is one of the significant research areas in the field of mathematical optimization and machine learning. Although various studies have explored zeroth-order algorithms, one of the theoretical limitations is that oracle complexity depends on the dimension, i.e., on the number of variables, of the optimization problem. In this talk, we propose a zeroth-order random subspace algorithm by combining a gradient-free algorithm (specifically, Gaussian randomized smoothing with central differences) with random projection to reduce the dependency of the dimension in oracle complexity. The algorithm has a local convergence rate independent of the original dimension under some local assumptions.

Talk 2: Random matrix to solve large linear system
Speaker: Jiaming Yang
Abstract: Large matrices arise in real-world applications in the areas of machine learning, data analysis and optimization: from the representation of massive datasets with high dimensional features, to the first and second-order derivatives of an objective function that needs to be optimized. How can we capture the intrinsic patterns and physics efficiently and accurately? As an interdisciplinary research area that roots in theoretical computer science (TCS), random matrix theory (RMT), and recently thrives in scientific computing and machine learning, Randomized Numerical Linear Algebra (RNLA) offers a promising avenue to alleviate computational challenges by introducing randomness to large-scale problems, with the guarantee of creating efficient approximate solutions that retain high accuracy. Specifically, randomized subspace methods are one of the most popular methods that are well established in theory but only start flourishing in the area of optimization recently. In my recent work [Derezinski,Yang, STOC 2024] and [Derezinski, Musco, Yang, SODA 2025], we focus on the problem of solving a linear system and develope different types of stochastic optimization algorithms. Our algorithms provably achieve better time complexity results, and are linear in the input matrix size when we assume that it has a flat tail. As one of our key techniques, we construct a low- rank Nystrom approximation with sparse random sketching, resulting in an easy-to-construct preconditioner with the effective guarantee from the randomized subspace theory.

Talk 3: Random Subspace Homogenized Trust Region Method
Speaker: Pierre-Louis Poirion
Abstract: We proposes the Random Subspace Homogenized Trust Region (RSHTR) method with the best theoretical guarantees among random subspace algorithms for nonconvex optimization. Furthermore, under rank-deficient conditions, RSHTR converge to a second-order stationary point quadratically. Experiments on real-world datasets verify the benefits of RSHTR.

Speakers
AT

Akiko Takeda

University of Tokyo/RIKEN
Akiko Takeda received the Doctor of Science degree in information science from the Tokyo Institute of Technology, Japan, in 2001. She is currently a professor in the Department of Mathematical Informatics, The University of Tokyo, and the team leader of Continuous Optimization Team... Read More →
avatar for Jiaming Yang

Jiaming Yang

PhD Student, University of Michigan
PP

Pierre-Louis Poirion

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 212 3501 Trousdale Pkwy, 212, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4F: Contextual Stochastic Optimization under Streaming Data and Decision Dependency
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Contextual Stochastic Optimization under Streaming Data and Decision Dependency
Chair: Guzin Bayraksan
Cluster: Optimization Under Uncertainty and Data-driven Optimization

Talk 1: Residuals-Based Contextual Distributionally Robust Optimization with Decision-Dependent Uncertainty
Speaker: Xian Yu
Abstract: We consider a residuals-based distributionally robust optimization model, where the underlying uncertainty depends on both covariate information and our decisions. We adopt regression models to learn the latent decision dependency and construct a nominal distribution (thereby ambiguity sets) around the learned model using empirical residuals from the regressions. Ambiguity sets can be formed via the Wasserstein distance, a sample robust approach, or with the same support as the nominal empirical distribution (e.g., phi-divergences), where both the nominal distribution and the radii of the ambiguity sets could be decision- and covariate-dependent. We provide conditions under which desired statistical properties, such as asymptotic optimality, rates of convergence, and finite sample guarantees, are satisfied. Via cross-validation, we devise data-driven approaches to find the best radii for different ambiguity sets, which can be decision-(in)dependent and covariate-(in)dependent. Through numerical experiments, we illustrate the effectiveness of our approach and the benefits of integrating decision dependency into a residuals-based DRO framework.

Talk 2: Distribution-Free Algorithms for Predictive Stochastic Programming in the Presence of Streaming Data
Speaker: Suvrajeet Sen
Abstract: This work studies a fusion of concepts from stochastic programming and nonparametric statistical learning in which data is available in the form of covariates interpreted as predictors and responses. Such models are designed to impart greater agility, allowing decisions under uncertainty to adapt to the knowledge of predictors (leading indicators). This work studies two classes of methods: one of the methods may be classified as a first-order method, whereas the other studies piecewise linear approximations. In addition, our study incorporates several non-parametric estimation schemes, including k nearest neighbors (kNN) and other standard kernel estimators. Our computational results demonstrate that the new algorithms outperform traditional approaches which were not designed for streaming data applications requiring simultaneous estimation and optimization. (This work was performed as part of the first author's doctoral dissertation.)

Talk 3: An Alternating Optimization Method for Contextual Distributionally Robust Optimization under Streaming Data
Speaker: Guzin Bayraksan
Abstract: We consider data-driven decision-making that incorporates a prediction model within the 1-Wasserstein distributionally robust optimization (DRO) given joint observations of uncertain parameters and covariates using regression residuals in a streaming-data setting. In this setting, additional data become available and allow decisions to adapt to the growing knowledge of the underlying uncertainty. The ambiguity set shrinks as more data is observed. We propose an efficient online optimization method for this streaming-data contextual DRO setting, which iteratively alternates between optimizing the decision and determining the worst-case distribution. We analyze the asymptotic convergence properties of this algorithm and establish dynamic regret bounds to certify the performance of online solutions. Through numerical experiments, we validate our theoretical findings and demonstrate that our approach significantly enhances computational efficiency while maintaining high solution quality under streaming data.

Speakers
XY

Xian Yu

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
SS

Suvrajeet Sen

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
GB

Guzin Bayraksan

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 156 3518 Trousdale Pkwy, 156, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4G: Analysis and Design of First Order Methods for Convex Optimization
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Analysis and Design of First Order Methods for Convex Optimization
Chair: Bartolomeo Stellato
Cluster: Optimization For Data Science

Talk 1: Toward a grand unified theory of accelerations in optimization and machine learning
Speaker: Ernest Ryu
Abstract: Momentum-based acceleration of first-order optimization methods, first introduced by Nesterov, has been foundational to the theory and practice of large-scale optimization and machine learning. However, finding a fundamental understanding of such acceleration remains a long-standing open problem. In the past few years, several new acceleration mechanisms, distinct from Nesterov’s, have been discovered, and the similarities and dissimilarities among these new acceleration phenomena hint at a promising avenue of attack for the open problem. In this talk, we discuss the envisioned goal of developing a mathematical theory unifying the collection of acceleration mechanisms and the challenges that are to be overcome.

Talk 2: Data-driven performance estimation of first-order methods
Speaker: Jisun Park
Abstract: We introduce a data-driven approach to analyze the probabilistic performance of first-order optimization algorithms. Combining Wasserstein Distributionally Robust Optimization to the performance estimation framework, we derive probabilistic performance guarantees to a wide range first-order methods. We show that our method is able to achieve significant reductions in conservatism compared to classical worst-case performance analysis tools.

Talk 3: Exact Verification of First-Order Methods for Quadratic Optimization via Mixed-Integer Programming
Speaker: Vinit Ranjan
Abstract: We present a mixed-integer programming based framework to exactly verify the convergence of first-order methods for parametric convex quadratic and linear optimization. We formulate the verification problem as a mathematical optimization problem where we maximize the infinity norm of the fixed-point residual at the last iteration subject to constraints on the parameters and initial iterates. Our approach uses affine and piecewise affine steps to exactly represent a wide range of gradient, projection, and proximal steps. We scale the mixed-integer formulation using advanced bound tightening and strong formulations for the piecewise affine steps. Numerical examples show orders of magnitude lower worst-case residuals that more closely match the practical convergence.

Speakers
BS

Bartolomeo Stellato

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
ER

Ernest Ryu

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JP

Jisun Park

Postdoctoral Research Fellow, Princeton University
Name: Dr. Slothington "Slow Convergence" McNapfaceTitle: Distinguished Professor of Continuous Optimization & Energy MinimizationAffiliation: The Lush Canopy Institute of Sluggish AlgorithmsBio:Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 114 3501 Trousdale Pkwy, 114, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4H: Advances in Network Optimization and Cooperative Learning
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Advances in Network Optimization and Cooperative Learning
Chair: Cesar A Uribe
Cluster: Multi-agent Optimization and Games

Talk 1: Optimally Improving Cooperative Learning in a Social Settin
Speaker: Shahrzad Haddadan
Abstract: We consider a cooperative learning scenario where a collection of networked agents with individually owned classifiers update their predictions, for the same classification task, through communication or observations of each other’s predictions. Clearly if highly influential vertices use erroneous classifiers, there will be a negative effect on the accuracy of all the agents in the network. We ask the following question: how can we optimally fix the prediction of a few classifiers so as maximize the overall accuracy in the entire network. To this end we consider an aggregate and an egalitarian objective function. We show a polynomial time algorithm for optimizing the aggregate objective function, and show that optimizing the egalitarian objective function is NP-hard. Furthermore, we develop approximation algorithms for the egalitarian improvement. The performance of all of our algorithms are guaranteed by mathematical analysis and backed by experiments on synthetic and real data.

Talk 2: The engineering potential of fish research: swimming upstream to new solutions
Speaker: Daniel Burbano
Abstract: Millions of years of evolution have endowed animals with refined and elegant mechanisms to orient and navigate in complex environments. Elucidating the underpinnings of these processes is of critical importance not only in biology to understand migration and survival but also for engineered network systems to aid the development of bio-inspired algorithms for estimation and control. Particularly interesting is the study of fish navigation where different cues, such as vision and hydrodynamics are integrated and fed back to generate locomotion. Little is known, however, about the information pathways and the integration process underlying complex navigation problems. This talk will discuss recent advances in data-driven mathematical models based on potential flow theory, stochastic differential equations, and control theory describing fish navigation. In addition, we will discuss how biological insights gained from this research can be applied to robot navigation with zero-order optimization and estimation and control problems in network systems

Talk 3: An Optimal Transport Approach for Network Regression
Speaker: Alex Zalles
Abstract: We study the problem of network regression, where one is interested in how the topology of a network changes as a function of Euclidean covariates. We build upon recent developments in generalized regression models on metric spaces based on Fr\'echet means and propose a network regression method using the Wasserstein metric. We show that when representing graphs as multivariate Gaussian distributions, the network regression problem requires the computation of a Riemannian center of mass (i.e., Fr\'echet means). Fr\'echet means with non-negative weights translates into a barycenter problem and can be efficiently computed using fixed point iterations. Although the convergence guarantees of fixed-point iterations for the computation of Wasserstein affine averages remain an open problem, we provide evidence of convergence in a large number of synthetic and real-data scenarios. Extensive numerical results show that the proposed approach improves existing procedures by accurately accounting for graph size, topology, and sparsity in synthetic experiments. Additionally, real-world experiments using the proposed approach result in higher Coefficient of Determination () values and lower mean squared prediction error (MSPE), cementing improved prediction capabilities in practice.

Speakers
SH

Shahrzad Haddadan

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
DB

Daniel Burbano

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
AZ

Alex Zalles

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 116 3501 Trousdale Pkwy, 116, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4I: Optimization for Large Language Models and Kernels
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Optimization for Large Language Models and Kernels
Chair: Ming Yin
Cluster: Optimization Applications (Communication, Energy, Health, ML, ...)

Talk 1: Optimizing for a Proxy Reward in RLHF
Speaker: Banghua Zhu
Abstract: Reinforcement Learning from Human Feedback (RLHF) has become an important technique in post-training of Larger Language Models (LLM). During RLHF, one usually first trains a reward model from human preference data, and then optimizes the LLM for the proxy reward signal predicted by the reward model. In this talk, I'll discuss what makes a good reward model for RLHF from both theoretical and empirical observations.

Talk 2: Self-Play Preference Optimization for Language Model Alignment
Speaker: Yue Wu
Abstract: In this paper, we propose a self-play-based method for language model alignment, which treats the problem as a constant-sum two-player game aimed at optimizing the model to approximate the Nash equilibrium. Our approach, dubbed SPPO, is based on a new alignment objective derived from L2 regression. Interestingly, this new objective has a deep connection with the KL-regularized policy gradient and natural gradient methods, and can guarantee the convergence to the optimal solution. In our experiments, this theoretically motivated objective turns out highly effective. By leveraging a small pre-trained preference model, SPPO can obtain a highly-aligned model without additional external supervision from human or stronger language models.

Talk 3: Learning Counterfactual Distributions via Kernel Nearest Neighbors
Speaker: Kyuseong Choi
Abstract: Consider a setting with multiple units (e.g., individuals, cohorts, geographic locations) and outcomes (e.g., treatments, times, items), where the goal is to learn a multivariate distribution for each unit-outcome entry, such as the distribution of a user's weekly spend and engagement under a specific mobile app version. A common challenge is the prevalence of missing not at random data---observations are available only for certain unit-outcome combinations---where the observed distributions can be correlated with properties of distributions themselves, i.e., there is unobserved confounding. An additional challenge is that for any observed unit-outcome entry, we only have a finite number of samples from the underlying distribution. We tackle these two challenges by casting the problem into a novel distributional matrix completion framework and introduce a kernel-based distributional generalization of nearest neighbors to estimate the underlying distributions. By leveraging maximum mean discrepancies and a suitable factor model on the kernel mean embeddings of the underlying distributions, we establish consistent recovery of the underlying distributions even when data is missing not at random and positivity constraints are violated. Furthermore, we demonstrate that our nearest neighbors approach is robust to heteroscedastic noise, provided we have access to two or more measurements for the observed unit-outcome entries—a robustness not present in prior works on nearest neighbors with single measurements.

Speakers
BZ

Banghua Zhu

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
YW

Yue Wu

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
KC

Kyuseong Choi

PhD, Cornell Tech
Tuesday July 22, 2025 10:30am - 11:45am PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 100 3518 Trousdale Pkwy, 100, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4J: Dynamic Optimization: Deterministic and Stochastic Continuous-time Models I
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Dynamic Optimization: Deterministic and Stochastic Continuous-time Models I
Chair: Cristopher Hermosilla
Cluster: Multi-agent Optimization and Games

Talk 1: A Minimality Property of the Value Function in Optimal Control over the Wasserstein Space
Speaker: Cristopher Hermosilla
Abstract: In this talk we study an optimal control problem with (possibly) unbounded terminal cost in the space of Borel probability measures with finite second moment. We consider the metric geometry associated with the Wasserstein distance, and a suitable weak topology rendering this space locally compact. In this setting, we show that the value function of a control problem is the minimal viscosity supersolution of an appropriate Hamilton-Jacobi-Bellman equation. Additionally, if the terminal cost is bounded and continuous, we show that the value function is the unique viscosity solution of the Hamilton-Jacobi-Bellman equation.

Talk 2: Principal-Multiagents problem in continuous-time
Speaker: Nicolás Hernández
Abstract: We study a general contracting problem between the principal and a finite set of competitive agents, who perform equivalent changes of measure by controlling the drift of the output process and the compensator of its associated jump measure. In this setting, we generalize the dynamic programming approach developed by Cvitanić, Possamaï, and Touzi (2017) and we also relax their assumptions. We prove that the problem of the principal can be reformulated as a standard stochastic control problem in which she controls the continuation utility (or certainty equivalent) processes of the agents. Our assumptions and conditions on the admissible contracts are minimal to make our approach work. We also present a smoothness result for the value function of a risk–neutral principal when the agents have exponential utility functions. This leads, under some additional assumptions, to the existence of an optimal contract.

Talk 3: Unbounded viscosity solutions of Hamilton-Jacobi equations in the 2-Wasserstein space
Speaker: Othmane Jerhaoui
Abstract: In this talk, we study unbounded viscosity solutions of Hamilton-Jacobi equations in the 2-Wasserstein space over the Euclidean space. The notion of viscosity is defined by taking test functions that are locally Lipschitz and can be respresented as a difference of two geodesically semiconvex functions. First, We establish a comparison result for a general Hamiltonian sat- isfying mild hypotheses. Then, we will discuss well-posedness of a class of Hamilton-Jacobi equations with a Hamiltonian arising from classical mechanics.

Speakers
CH

Cristopher Hermosilla

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
NH

Nicolás Hernández

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
OJ

Othmane Jerhaoui

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 102 3501 Trousdale Pkwy, 102, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4K: Convex approaches to SDP
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Convex approaches to SDP
Chair: Richard Zhang
Cluster: Conic and Semidefinite Optimization

Talk 1: Fitting Tractable Convex Sets to Support Function Data
Speaker: Venkat Chandrasekaran
Abstract: The geometric problem of estimating an unknown compact convex set from evaluations of its support function arises in a range of scientific and engineering applications. Traditional approaches typically rely on estimators that minimize the error over all possible compact convex sets; in particular, these methods do not allow for much incorporation of prior structural information about the underlying set and the resulting estimates become increasingly more complicated to describe as the number of measurements available grows.  We address both of these shortcomings by describing a new framework for estimating tractably specified convex sets from support function evaluations.  Along the way, we also present new bounds on how well arbitrary convex bodies can be approximated by elements from structured non-polyhedral families of convex sets.  Our numerical experiments highlight the utility of our framework over previous approaches in settings in which the measurements available are noisy or small in number as well as those in which the underlying set to be reconstructed is non-polyhedral. (Joint work with Yong Sheng Soh and Eliza O'Reilly.)

Talk 2: Iterative methods for primal-dual scalings in conic optimization
Speaker: Lieven Vandenberghe
Abstract: A central element in the design of primal-dual interior-point methods for conic optimization is the definition of a suitable primal-dual scaling or metric. The talk will discuss simple iterative methods for computing primal-dual scalings. We will consider the Nesterov-Todd scaling for symmetric cones and extensions to sparse positive semidefinite matrix cones.

Talk 3: Generalized Cuts and Grothendieck Covers: a Primal-Dual Approximation Framework Extending the Goemans--Williamson Algorithm
Speaker: Nathan Benedetto Proenca
Abstract: We provide a primal-dual framework for randomized approximation algorithms utilizing semidefinite programming (SDP) relaxations. Our framework pairs a continuum of APX-complete problems including MaxCut, Max2Sat, MaxDicut, and more generally, Max-Boolean Constraint Satisfaction and MaxQ (maximization of a positive semidefinite quadratic form over the hypercube) with new APX-complete problems which are stated as convex optimization problems with exponentially many variables. These new dual counterparts, based on what we call Grothendieck covers, range from fractional cut covering problems (for MaxCut) to tensor sign covering problems (for MaxQ). For each of these problem pairs, our framework transforms the randomized approximation algorithms with the best known approximation factors for the primal problems to randomized approximation algorithms for their dual counterparts with reciprocal approximation factors which are tight with respect to the Unique Games Conjecture. For each APX-complete pair, our algorithms solve a single SDP relaxation and generate feasible solutions for both problems which also provide approximate optimality certificates for each other. Our work utilizes techniques from areas of randomized approximation algorithms, convex optimization, spectral sparsification, as well as Chernoff-type concentration results for random matrices.

Speakers
RZ

Richard Zhang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
VC

Venkat Chandrasekaran

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
LV

Lieven Vandenberghe

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
NB

Nathan Benedetto Proenca

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 118 3501 Trousdale Pkwy, 118, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4L: First-order methods for nonsmooth optimization
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: First-order methods for nonsmooth optimization
Chair: Digvijay Boob
Cluster: Nonlinear Optimization

Talk 1: Efficient Subgradient Method for Minimizing Nonsmooth Maximal Value Functions
Speaker: Hanyang Li
Abstract: We consider the minimization of a class of nonsmooth maximal value functions that are piecewise-smooth. Recent implementable Goldstein-style subgradient methods for general Lipschitz functions involve computationally intensive inner loops to approximate the descent direction. In this paper, we introduce a novel subgradient method that eliminates the sophisticated inner loops by employing a regularization technique, significantly improving computational efficiency for minimizing nonsmooth maximal value functions.

Talk 2: Dimension dependence in nonconvex optimization
Speaker: Guy Kornowski
Abstract: Optimization problems that arise in modern machine learning are often neither smooth nor convex, and typically high-dimensional. In this talk we will discuss the intricate role of the dimension in such problems, and compare it to smooth optimization. We will see that some approaches lead to significant gaps in dimension dependencies, yet sometimes these can be eliminated altogether. In particular, we will examine fundamental concepts such as stationarity, smoothing, and zero-order optimization, and show they exhibit exponential, polynomial, and no such gaps, respectively.

Talk 3: On the Sample Complexity of Imitation Learning for Smoothed Model Predictive Control
Speaker: Swati Padmanabhan
Abstract: Recent work in imitation learning has shown that having an expert controller that is both suitably smooth and stable enables stronger guarantees on the performance of the learned controller. However, constructing such smoothed expert controllers for arbitrary systems remains challenging, especially in the presence of input and state constraints. As our primary contribution, we show how such a smoothed expert can be designed for a general class of systems using a log-barrier-based relaxation of a standard Model Predictive Control (MPC) optimization problem. At the crux of this theoretical guarantee on smoothness is a new lower bound we prove on the optimality gap of the analytic center associated with a convex Lipschitz function, which we hope could be of independent interest. We validate our theoretical findings via experiments, demonstrating the merits of our smoothing approach over randomized smoothing.

Speakers
DB

Digvijay Boob

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
HL

Hanyang Li

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
avatar for Guy Kornowski

Guy Kornowski

Weizmann Institute of Science
Talk title:Dimension dependence in nonconvex optimizationBio:Guy Kornowski is a PhD student at the Weizmann Institute of Science, advised by Prof. Ohad Shamir. During his PhD he interned at Apple ML Research, where he worked with Kunal Talwar and Vitaly Feldman. His research focuses... Read More →
SP

Swati Padmanabhan

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 119 3501 Trousdale Pkwy, 119, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4M: Bilevel Optimization for Inverse Problems Part 2
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Bilevel Optimization for Inverse Problems Part 2
Chair: Juan Carlos de los Reyes
Cluster: PDE-constrained Optimization

Talk 1: Linesearch-enhanced inexact forward–backward methods for bilevel optimization
Speaker: Marco Prato
Abstract: Bilevel optimization problems arise in various real-world applications, often being characterized by the impossibility of having the exact objective function and its gradient available. Developing mathematically sound optimization methods that effectively handle inexact information is crucial for ensuring reliable and efficient solutions. In this talk we propose a line-search based algorithm for solving a bilevel optimization problem, where the approximate gradient and function evaluation obeys an adaptive tolerance rule. Our method is based on implicit differentiation under some standard assumptions, and its main novelty with respect to similar approaches is the well posed, inexact line-search procedure using only approximate function values and adaptive accuracy control. This work is partially supported by the PRIN project 20225STXSB, under the National Recovery and Resilience Plan (NRRP) funded by the European Union - NextGenerationEU.

Talk 2: Tensor train solution to uncertain optimization problems with shared sparsity penalty
Speaker: Akwum Onwunta
Abstract: We develop first- and second-order numerical optimization methods to solve non-smooth optimization problems featuring a shared sparsity penalty, constrained by differential equations with uncertainty. To alleviate the curse of dimensionality, we use tensor product approximations. To handle the non-smoothness of the objective function, we introduce a smoothed version of the shared sparsity objective. We consider both a benchmark elliptic PDE constraint and a more realistic topology optimization problem. We demonstrate that the error converges linearly in iterations and the smoothing parameter and faster than algebraically in the number of degrees of freedom, consisting of the number of quadrature points in one variable and tensor ranks. Moreover, in the topology optimization problem, the smoothed shared sparsity penalty reduces the tensor ranks compared to the unpenalized solution. This enables us to find a sparse high-resolution design under a high-dimensional uncertainty.

Talk 3: TBD
Speaker: Juan Carlos de los Reyes
Abstract: TBD

Speakers
avatar for Marco Prato

Marco Prato

Associate Professor, Università di Modena e Reggio Emilia, Italy
Name: Dr. Marco PratoTitle: Associate Professor in Numerical AnalysisAffiliation: Department of Physics, Informatics and Mathematics, University of Modenanad Reggio Emilia, ItalyBio:Dr. Marco Prato was born in Varazze, on the Italian Riviera, in 1980. He received the MSc Degree and... Read More →
JC

Juan Carlos de los Reyes

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 157 3518 Trousdale Pkwy, 157, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4N: Derivative-free stochastic optimization methods I
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Derivative-free stochastic optimization methods I
Chair: Stefan Wild
Cluster: Derivative-free Optimization

Talk 1: Improving the robustness of zeroth-order optimization solvers
Speaker: Stefan Wild
Abstract: Zeroth-order optimization solvers are often deployed in settings where little information regarding a problem's conditioning or noise level is known. An ideal solver will perform well in a variety of challenging settings. We report on our experience developing adaptive algorithms, which leverage information learned online to adapt critical algorithmic features. We illustrate our approach in trust-region-based reduced-space methods and show how trained policies can even be deployed effectively in nonstationary cases, where the noise seen changes over the decision space.

Talk 2: A noise-aware stochastic trust-region algorithm using adaptive random subspaces
Speaker: Kwassi Joseph Dzahini
Abstract: We introduce ANASTAARS, a noise-aware stochastic trust-region algorithm using adaptive random subspace strategies, that is effective not only for low- and moderate-dimensional cases, but also for potentially high-dimensional ones. The proposed method achieves scalability by optimizing random models that approximate the objective within low-dimensional affine subspaces, thereby significantly reducing per-iteration costs in terms of function evaluations. These subspaces and their dimension are defined via Johnson--Lindenstrauss transforms such as those obtained from Haar-distributed orthogonal random matrices. In contrast to previous work involving random subspaces with fixed dimensions, ANASTAARS introduces an adaptive subspace selection strategy. Instead of generating a completely new poised set of interpolation points at each iteration, the proposed method updates the model by generating only a few or even a single new interpolation point, reusing past points (and their corresponding function values) from lower-dimensional subspaces in such a way that the resulting set remains poised. This approach not only introduces a novel way to reduce per-iteration costs in terms of function evaluations, but also avoids constructing random models in fixed-dimension subspaces, resulting in a more efficient and optimization process through the use of adaptive subspace models. Furthermore, to address the observation that model-based trust-region methods perform optimally when the signal-to-noise ratio is high, ANASTAARS incorporates a strategy from noise-aware numerical optimization literature by utilizing an estimate of the noise level in each function evaluation. The effectiveness of the method is demonstrated through numerical experiments conducted on problems within the "quantum approximate optimization algorithm" (QAOA) framework.

Talk 3: A consensus-based global optimization method with noisy objective function
Speaker: Greta Malaspina
Abstract: Consensus based optimization is a derivative-free particles-based method for the solution of global optimization problems. Several versions of the method have been proposed in the literature, and different convergence results have been proved, with varying assumptions on the regularity of the objective function and the initial distribution of the particles. However, all existing results assume the objective function to be evaluated exactly at each iteration of the method. In this work, we extend the convergence analysis of a discrete-time CBO method to the case where only a noisy stochastic estimator of the objective function can be computed at a given point. In particular we prove that under suitable assumptions on the oracle’s noise, the expected value of the mean squared distance of the particles from the solution can be made arbitrarily small in a finite number of iterations. Some numerical results showing the impact of noise are also given.

Speakers
SW

Stefan Wild

Lawrence Berkeley National Laboratory
Name: Dr. Slothington "Slow Convergence" McNapfaceTitle: Distinguished Professor of Continuous Optimization & Energy MinimizationAffiliation: The Lush Canopy Institute of Sluggish AlgorithmsBio:Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
KJ

Kwassi Joseph Dzahini

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
GM

Greta Malaspina

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 256 3518 Trousdale Pkwy, 256, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4O: Feasible and infeasible methods for optimization on manifolds I
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Feasible and infeasible methods for optimization on manifolds I
Chair: Bin Gao
Cluster: Optimization on Manifolds

Talk 1: A double tracking method for optimization with decentralized generalized orthogonality constraints
Speaker: Xin Liu
Abstract: We consider the decentralized optimization problems with generalized orthogonality constraints, where both the objective function and the constraint exhibit a distributed structure. Such optimization problems, albeit ubiquitous in practical applications, remain unsolvable by existing algorithms in the presence of distributed constraints. To address this issue, we convert the original problem into an unconstrained penalty model by resorting to the recently proposed constraint-dissolving operator. However, this transformation compromises the essential property of separability in the resulting penalty function, rendering it impossible to employ existing algorithms to solve. We overcome this difficulty by introducing a novel algorithm that tracks the gradient of the objective function and the Jacobian of the constraint mapping simultaneously. The global convergence guarantee is rigorously established with an iteration complexity. To substantiate the effectiveness and efficiency of our proposed algorithm, we present numerical results on both synthetic and real-world datasets.

Talk 2: A projected gradient descent algorithm for ab initio fixed-volume crystal structure relaxation
Speaker: Yukuan Hu
Abstract: This paper is concerned with ab initio crystal structure relaxation under a fixed unit cell volume, which is a step in calculating the static equations of state and forms the basis of thermodynamic property calculations for materials. The task can be formulated as an energy minimization with a determinant constraint. Widely used line minimization-based methods (e.g., conjugate gradient method) lack both efficiency and convergence guarantees due to the nonconvex nature of the determinant constraint as well as the significant differences in the curvatures of the potential energy surface with respect to atomic and lattice components. To this end, we propose a projected gradient descent algorithm named PANBB. It is equipped with (i) search direction projections for lattice vectors, (ii) distinct curvature-aware initial trial step sizes for atomic and lattice updates, and (iii) a nonrestrictive line minimization criterion as the stopping rule for the inner loop. It can be proved that PANBB favors theoretical convergence to equilibrium states. Across a benchmark set containing 223 structures from various categories, PANBB achieves average speedup factors of approximately 1.41 and 1.45 over the conjugate gradient method and direct inversion in the iterative subspace implemented in off-the-shelf simulation software, respectively. Moreover, it normally converges on all the systems, manifesting its robustness. As an application, we calculate the static equations of state for the high-entropy alloy AlCoCrFeNi, which remains elusive owing to 160 atoms representing both chemical and magnetic disorder and the strong local lattice distortion. The results are consistent with the previous calculations and are further validated by experimental thermodynamic data.

Talk 3: Efficient optimization with orthogonality constraints via random submanifold approach
Speaker: Andi Han
Abstract: Optimization problems with orthogonality constraints are commonly addressed using Riemannian optimization, which leverages the geometric structure of the constraint set as a Riemannian manifold. This method involves computing a search direction in the tangent space and updating via a retraction. However, the computational cost of the retraction increases with problem size. To improve scalability, we propose a method that restricts updates to random submanifolds, reducing per-iteration complexity. We introduce two submanifold selection strategies and analyze the convergence for nonconvex functions, including those satisfying the Riemannian Polyak–Łojasiewicz condition, as well as for stochastic optimization problems. The approach generalizes to quotient manifolds derived from the orthogonal manifold.

Speakers
BG

Bin Gao

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
XL

Xin Liu

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
YH

Yukuan Hu

Postdoctoral Fellow, CERMICS, École des Ponts, IP Paris
AH

Andi Han

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 258 3518 Trousdale Pkwy, 258, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4P: Recent Advances in Conic Optimization and Machine Learning
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Recent Advances in Conic Optimization and Machine Learning
Chair: Godai Azuma
Cluster: Conic and Semidefinite Optimization

Talk 1: Procrustean Differential Privacy: A Parameter-Scalable Method for Privacy-Preserving Collaborative Learning
Speaker: Keiyu Nosaka
Abstract: Privacy-preserving collaborative learning enables multiple parties to jointly train machine learning models without directly sharing sensitive data. While approaches such as Federated Learning and Homomorphic Encryption offer robust privacy guarantees, they often incur significant computational and communication overhead as model complexity increases. In contrast, multiplicative perturbation techniques promise enhanced efficiency; however, they are typically hampered by increased privacy risks from collusion or by a reliance on extensive deep learning-based training to achieve satisfactory performance. In this work, we introduce a novel framework that bridges these extremes by integrating the analytic Gaussian mechanism of differential privacy with the Generalized Orthogonal Procrustes Problem. Our method delivers adjustable privacy–performance trade-offs through tunable differential privacy parameters, allowing practitioners to balance protection and efficiency according to specific application needs. We substantiate our approach with theoretical guarantees and numerical analyses that evaluate its performance across varying privacy levels, data dimensions, and numbers of collaborating parties.

Talk 2: Facial Structure of Copositive and Completely Positive Cones over a Second-Order Cone
Speaker: Mitsuhiro Nishijima
Abstract: A copositive cone over a second-order cone is the closed convex cone of real symmetric matrices whose associated quadratic forms are nonnegative over the given second-order cone. In this talk, we classify the faces of those copositive cones and their duals (i.e., completely positive cones over a second-order cone), and investigate their dimension and exposedness properties. Then we compute two parameters related to chains of faces of both cones. At the end, we discuss some possible extensions of the results with a view toward analyzing the facial structure of general copositive and completely positive cones.

Talk 3: Exact Semidefinite Relaxations for Safety Verification of Neural Network
Speaker: Godai Azuma
Abstract: We study the accuracy of DeepSDP which is proposed as a semidefinite programming (SDP) based method to measure the safety and the robustness of given neural networks by guaranteeing the bounds of their outputs. The dual problem of the DeepSDP is in fact an SDP relaxation of quadratic constraints representing ReLU activation functions. In this talk, we investigate the exactness of the DeepSDP by using exactness conditions for the general SDP relaxation so that the estimated robustness is improved on. We also discuss our assumptions and some results on the accuracy.

Speakers
KN

Keiyu Nosaka

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
MN

Mitsuhiro Nishijima

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
GA

Godai Azuma

Assistant Professor, Aoyama Gakuin University
Name: Dr. Slothington "Slow Convergence" McNapfaceTitle: Distinguished Professor of Continuous Optimization & Energy MinimizationAffiliation: The Lush Canopy Institute of Sluggish AlgorithmsBio:Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 106 3501 Trousdale Pkwy, 106, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4Q: Moment-SOS Hierarchy: From Theory to Computation in the Real World (I)
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Moment-SOS Hierarchy: From Theory to Computation in the Real World (I)
Chair: Heng Yang
Cluster: Conic and Semidefinite Optimization

Talk 1: Polynomial Matrix Optimization, Matrix-Valued Moments, and Sparsity
Speaker: Jie Wang
Abstract: This talk will introduce the matrix version of the moment-SOS hierarchy for solving polynomial matrix optimization problems. Various types of sparsities will be discussed to improve the scalability of the approach.

Talk 2: Practical non-symmetric cone programming algorithms for sums-of-squares optimization
Speaker: Dávid Papp
Abstract: Sums-of-squares relaxations of polynomial optimization problems are usually solved using off-the-shelf semidefinite programming (SDP) methods, which often leads to at least one of the following two problems when the relaxation order (that is, the degree of the SOS polynomials) is high: (1) the time and space complexity of SDP algorithms grow too fast to be practical; (2) the numerical conditioning of the SDPs is prohibitive to obtain accurate solutions. The talk focuses on how both can be mitigated using polynomial interpolation and replacing all-purpose semidefinite programming algorithms with non-symmetric cone programming methods that can optimize directly over sums-of-squares cones. Generalizations to other nonnegativity certificates (SONC, SAGE) will also be briefly discussed.

Talk 3: A squared smoothing Newton Method for semidefinite programming
Speaker: Ling Liang
Abstract: This paper proposes a squared smoothing Newton method via the Huber smoothing function for solving semidefinite programming problems (SDPs). We first study the fundamental properties of the matrix-valued mapping defined upon the Huber function. Using these results and existing ones in the literature, we then conduct rigorous convergence analysis and establish convergence properties for the proposed algorithm. In particular, we show that the proposed method is well-defined and admits global convergence. Moreover, under suitable regularity conditions, i.e., the primal and dual constraint nondegenerate conditions, the proposed method is shown to have a superlinear convergence rate. To evaluate the practical performance of the algorithm, we conduct extensive numerical experiments for solving various classes of SDPs. Comparison with the state-of-the-art SDP solvers demonstrates that our method is also efficient for computing accurate solutions of SDPs.

Speakers
avatar for Heng Yang

Heng Yang

Assistant Professor, Harvard University
Assistant Professor at Harvard University working on polynomial optimization and semidefinite programming.
JW

Jie Wang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
DP

Dávid Papp

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
LL

Ling Liang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 214 3501 Trousdale Pkwy, 214, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4R: Variational Analysis: Theory and Applications II
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Variational Analysis: Theory and Applications II
Chair: Walaa Moursi
Cluster: Nonsmooth Optimization

Talk 1: TBA
Speaker: Stephen Simons
Abstract: TBA

Talk 2: TBA
Speaker: Jon Vanderwerff
Abstract: TBA

Talk 3: TBA
Speaker: Gerald Beer
Abstract: TBA

Speakers
WM

Walaa Moursi

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
avatar for Stephen Simons

Stephen Simons

Professor Emeritus, University of California, Santa Barbara
Name: Dr. Stephen SimonsTitle: Professor Emeritus, University of California, Santa BarbaraAffiliation: University of California, Santa Barbara, Department of MathematicsBio:I am interested in the theory of monotone operators on general Banach space.
JV

Jon Vanderwerff

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
GB

Gerald Beer

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 154 3518 Trousdale Pkwy, 154, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4S: Recent Advances on PDE-constrained optimization packages and libraries : Part I
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Recent Advances on PDE-constrained optimization packages and libraries : Part I
Chair: Noemi Petra
Cluster: Computational Software

Talk 1: Empowering Numerical Optimization Across Disciplines with the Rapid Optimization Library (ROL)
Speaker: Denis Ridzal
Abstract: The Rapid Optimization Library (ROL) is a versatile, high-performance C++ library designed to address the complex demands of numerical optimization in various scientific and engineering disciplines. As an open-source effort through the Trilinos Project, ROL provides an extensive collection of state-of-the-art optimization algorithms capable of handling any application, hardware architecture, and problem size. This talk introduces ROL's key features, including its abstract linear algebra interface for universal applicability, modern algorithms for smooth, constrained, stochastic, risk-aware, and nonsmooth optimization, and its unique PDE-OPT application development kit for PDE-constrained optimization. Additionally, ROL offers an easy-to-use Python interface, which enhances its accessibility and usability across a wider range of applications and user communities. We highlight ROL's successful uses in fields ranging from electrodynamics and fluid dynamics to super-resolution imaging and machine learning. ROL's design philosophy, emphasizing vector abstractions, matrix-free interfaces, and a comprehensive suite of optimization algorithms, positions it as an important tool for researchers seeking to push the boundaries of numerical optimization. Authors: Robert Baraldi, Brian Chen, Aurya Javeed, Drew Kouri, Denis Ridzal, Greg von Winckel, Radoslav Vuchkov

Talk 2: CLAIRE: Constrained Large Deformation Diffeomorphic Image Registration
Speaker: Andreas Mang
Abstract: We present numerical methods for optimal control problems governed by geodesic flows of diffeomorphisms. Our work focuses on designing efficient numerical methods and fast computational kernels for high-performance computing platforms. We aim to compute a flow map that establishes spatial correspondences between two images of the same scene, modeled as geodesic flows of diffeomorphisms. The related optimization problem is non-convex and non-linear, leading to high-dimensional, ill-conditioned systems. Our solvers use advanced algorithms for rapid convergence and employ adjoint-based, second-order methods for numerical optimization. We provide results from real and synthetic data to assess convergence rates, time to solution, accuracy, and scalability of our methods. We also discuss strategies for preconditioning the Hessian and showcase results from our GPU-accelerated software package termed CLAIRE, which is optimized for clinically relevant problems and scales to hundreds of GPUs on modern architectures.

Talk 3: PyOED: An Extensible Suite for Data Assimilation and Model-Constrained Optimal Design of Experiments
Speaker: Ahmed Attia
Abstract: This talk gives a high-level overview of PyOED, a highly extensible scientific package that enables rapid development, testing, and benchmarking of model-based optimal experimental design (OED) methods for inverse problems. PyOED brings together variational and Bayesian data assimilation (DA) algorithms for inverse problems, optimal design of experiments, and novel optimization, statistical, and machine learning methods, into an integrated extensible research environment. PyOED is continuously being expanded with a plethora of Bayesian inference, DA, and OED methods as well as new scientific simulation models, observation error models, priors, and observation operators. These pieces are added such that they can be permuted to enable developing and testing OED methods in various settings of varying complexities. Authors: Ahmed Attia (ANL), Abhijit Chowdhary (NCSU), Shady Ahmed (PNNL)

Speakers
NP

Noemi Petra

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
AA

Ahmed Attia

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 155 3518 Trousdale Pkwy, 155, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4T: Parametric Optimization Problems
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Parametric Optimization Problems
Chair: Xiaoqi Yang
Cluster: Fixed Points and Variational Inequalities

Talk 1: Recent stability results in parametric optimization
Speaker: Xiaoqi Yang
Abstract: Recover bounds/relative calmness for sparse optimization models are important for algorithmic convergence analysis in machine learning and compressed sensing. Some recent results in sparse optimization obtained by restricted isometry property and restricted eigenvalue conditions will be reviewed. Lipschitz-like property and Lipschitz continuity are at the core of stability analysis. A projectional coderivative of set-valued mappings will be discussed and and be applied to obtain a complete characterization for a set-valued mapping to have the Lipschitz-property relative to a closed and convex set.

Talk 2: Lipschitz continuity of solution multifunctions of extended l_1 regularization problems
Speaker: Kaiwen Meng
Abstract: In this paper we obtain a verifiable sufficient condition for a polyhedral multifunction to be Lipschitz continuous on its domain. We apply this sufficient condition to establish the Lipschitz continuity of the solution multifunction for an extended l_1 regularization problem with respect to the regularization parameter and the observation parameter under the assumption that the data matrix is of full row rank.

Talk 3: Existence of a solution for stochastic multistage vector variational inequality problems
Speaker: Shengkun Zhu
Abstract: In this talk, we will discuss basic form and extensive forms of a stochastic vector variational inequality problem. We will apply KKM theorem to prove the existence of a solution under the monotonicity and hemi-continuity assumptions.

Speakers
XY

Xiaoqi Yang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
avatar for Kaiwen Meng

Kaiwen Meng

Professor, Southwestern University of Finance and Economics
Name: Dr. Kaiwen MENGTitle: Professor of Continuous OptimizationAffiliation: Southwestern University of Finance and EconomicsBio:Dr. Kaiwen MENG is a Fun Fact:  
SZ

Shengkun Zhu

Name: Dr. Slothington "Slow Convergence" McNapfaceTitle: Distinguished Professor of Continuous Optimization & Energy MinimizationAffiliation: The Lush Canopy Institute of Sluggish AlgorithmsBio:Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 158 3518 Trousdale Pkwy, 158, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4U: Computational Frameworks and Applications: Hybridizing Continuous and Discrete Optimization in Practice
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Computational Frameworks and Applications: Hybridizing Continuous and Discrete Optimization in Practice
Chair: Jordan Jalving & Marta D'Elia
Cluster: Interplay Between Continuous and Discrete Optimization

Talk 1: Optimal Conceptual Design using Generalized Disjunctive Programming in IDAES/Pyomo
Speaker: E. Soraya Rawlings
Abstract: The need to design more sustainable and energy-efficient processes requires the ability to efficiently pose and solve optimization problems with both discrete and continuous decisions. Traditionally, design optimization problems have been expressed as mixed-integer nonlinear optimization problems (MINLP) using algebraic modeling languages (AMLs; e.g., AIMMS, AMPL, GAMS, JuMP, Pyomo, etc.). A challenge with AMLs is that, not only they combine the structure of the optimization problem with binary variables, but also lack libraries for fast development of process models. An alternative approach is to use an extended mathematical programming environment that provides supplemental capabilities. Relevant in conceptual design are environments with the ability to construct hierarchical models, similar to what is supported in process simulation environments, and the ability to express logical restrictions explicitly in the model. If the environment incorporates both abilities, it can support the construction of design superstructures or topologies that represent all possible combinations of process configurations that we would like to consider in the design process. In this work, we present the application of one such environment to design separation and integrated energy systems. We leverage the open source IDAES platform, which supports the construction of hierarchical process models in Pyomo. We then leverage Generalized Disjunctive Programming to construct design superstructures and systematically convert them to MINLP models that can be solved with standard MINLP solvers. References [1] M. L. Bynum, G. A. Hackebeil, W. E. Hart, C. D. Laird, B. L. Nicholson, J. D. Siirola, J.-P. Watson and D. L. Woodruff, Pyomo - Optimization Modeling in Python, 3rd ed., vol. 67, Springer International Publishing, 2021. [2] Lee, A., Ghouse, J. H., Eslick, J.C., Laird, C.D., Siirola, J.D., Zamarripa, M.A., Gunter, D., Shinn, J. H., Dowling, A. W., Bhattacharyya, D., Biegler, L. T., Burgard, A. P., & Miller, D.C., “The IDAES process modeling framework and model library—Flexibility for process simulation and optimization,” Journal of Advanced Manufacturing and Processing, vol. 3, no. 3, pp. 1-30, 2021. https://doi.org/10.1002/amp2.10095 Disclaimer Sandia National Laboratories is a multimission laboratory managed and operated by National Technology and Engineering Solutions of Sandia, LLC., a wholly owned subsidiary of Honeywell International, Inc., for the U.S. Department of Energy's National Nuclear Security Administration under contract DE-NA-0003525. This paper describes objective technical results and analysis. Any subjective views or opinions that might be expressed in the paper do not necessarily represent the views of the U.S. Department of Energy or the United States Government.

Talk 2: Bridging Continuous and Discrete Optimization in Julia
Speaker: Thibaut Cuvelier
Abstract: TBA

Talk 3: Computation Design: Integrating Implicit Modeling, Meshless Simulation, and Machine Learning for Optimization with Real-World Efficiency
Speaker: Todd Mcdevitt
Abstract: Across every industry, engineers face ever-shrinking time-to-market demands, often needing more time to optimize their designs fully. Traditional CAD and simulation tools fall short of enabling practical usage of optimization and generative design due to persistent bottlenecks in geometry regeneration and meshing. These limitations prevent engineers from fully leveraging the design space while meeting project timelines. In this presentation, we introduce a new paradigm in mechanical design optimization that integrates implicit modeling, meshless simulation methods, and machine learning to address these challenges. Through diverse use cases across industries, we demonstrate how these techniques unlock the practical use of optimization while adhering to project timelines. We will also compare the computational costs of surrogate models versus the direct evaluation of objective functions with the original high-fidelity model. Attendees will gain insights into the practical deployment of optimization workflows in industrial, fast-paced, multi-disciplinary settings.

Speakers
avatar for Thibaut Cuvelier

Thibaut Cuvelier

Software engineer, operations research, Google
Thibaut Cuvelier is currently a software engineer at Google Research, in the Operations Research team. He received a PhD in telecommunications from CentraleSupélec (université Paris-Saclay, France). He is currently working on applications of operations research and reinforcement... Read More →
TM

Todd Mcdevitt

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 108 3501 Trousdale Pkwy, 108, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4V: Recent advances in open-source continuous solvers
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Recent advances in open-source continuous solvers
Chair: Charlie Vanaret
Cluster: Computational Software

Talk 1: A new parallel interior point solver for HiGHS
Speaker: Filippo Zanetti
Abstract: HiGHS is open-source software for linear, mixed-integer and quadratic optimization. It includes an interior point method based on an iterative Krylov solver, that has attracted much attention, in particular from the energy modelling community. In this talk, the latest news on the development of a new interior point solver, based on a direct factorization, are presented. The talk discusses multiple approaches that have been considered and highlights the improvements compared to the existing solver. Issues regarding regularization, accuracy, dealing with dense columns and parallelization are also mentioned.

Talk 2: Uno, a modern Lagrange-Newton solver for nonconvex constrained optimization
Speaker: Charlie Vanaret
Abstract: Derivative-based iterative methods for nonlinearly constrained nonconvex optimization usually share common algorithmic components, such as strategies for computing a descent direction and mechanisms that promote global convergence. Based on this observation, we introduce an abstract framework with eight building blocks that describes most derivative-based iterative methods and unifies their workflows. We then present Uno, a Lagrange-Newton solver that implements our abstract framework and allows the automatic generation of various strategy combinations with no programming effort from the user. Uno is meant to (1) organize mathematical optimization strategies into a coherent hierarchy; (2) offer a wide range of efficient and robust methods that can be compared for a given instance; (3) reduce the cost of development and maintenance of multiple optimization solvers; and (4) enable researchers to experiment with novel optimization strategies while leveraging established subproblem solvers and interfaces to modeling languages. We demonstrate that Uno is highly competitive against state-of-the-art solvers such as filterSQP, IPOPT, SNOPT, MINOS, LANCELOT, LOQO and CONOPT. Uno is available as open-source software under the MIT license at https://github.com/cvanaret/Uno

Talk 3: Solving Direct Optimal Control Problems in Real-Time: A Byrd-Omojokun Funnel Solver for acados
Speaker: David Kiessling
Abstract: We present a nonlinear optimization SQP-type method within the acados software framework for solving direct optimal control in real time. acados is a versatile software framework that implements fast solvers exploiting the specific optimal control problem structure using tailored linear algebra and numerical integration methods. This talk focuses on a novel SQP implementation using a funnel to drive global convergence. We present a novel implementation of a Byrd-Omojokun method avoiding infeasible subproblems. Additionally, we will present numerical results comparing our solver against other state-of-the art optimization solvers.

Speakers
avatar for Filippo Zanetti

Filippo Zanetti

University of Edinburgh
Name: Dr. Filippo ZanettiTitle: HiGHS optimization developerAffiliation: HiGHS - University of EdinburghBio:I am working on developing an interior point solver for LP and convex QP for HiGHS. My main interests are in interior point methods and their implementation, numerical linear... Read More →
avatar for Charlie Vanaret

Charlie Vanaret

Argonne & ZIB
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 110 3501 Trousdale Pkwy, 110, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4W: Nonsmooth Optimization - Theory and Algorithms
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Nonsmooth Optimization - Theory and Algorithms
Chair: Shaoning Han
Cluster: nan

Talk 1: Analysis of a Class of Heaviside Composite Minimization Problems
Speaker: Shaoning Han
Abstract: The minimization of nonclosed functions is a difficult topic that has been minimally studied. Among such functions is a Heaviside composite function that is the composition of a Heaviside function with a possibly nonsmooth multivariate function. Unifying a statistical estimation problem with hierarchical selection of variables and a sample average approximation of composite chance constrained stochastic programs, a Heaviside composite optimization problem is one whose objective and constraints are defined by sums of possibly nonlinear multiples of such composite functions. In this talk, we first present difficulties of tackling such problems using existing optimization techniques. Then we establish stationarity conditions for the problem, and introduce a sufficient condition, termed local convex-like property, under which the proposed stationary point is a local minimizer. Finally, we briefly discuss solution strategies via lifting and reformulation techniques. For details of this work, please refer to "Han S, Cui Y, Pang J-S (2024) Analysis of a class of minimization problems lacking lower semicontinuity. Mathematics of Operations Research."

Talk 2: Non-smooth stochastic gradient descent using smoothing functions
Speaker: Jingfu Tan
Abstract: In this talk, we address stochastic optimization problems involving a composition of a non-smooth outer function and a smooth inner function, a formulation frequently encountered in machine learning and operations research. To deal with the non-differentiability of the outer function, we approximate the original non-smooth function using smoothing functions, which are continuously differentiable and approach the original function as a smoothing parameter goes to zero (at the price of increasingly higher Lipschitz constants). The proposed smoothing stochastic gradient method iteratively drives the smoothing parameter to zero at a designated rate. We establish convergence guarantees under strongly convex, convex, and nonconvex settings, proving convergence rates that match known results for non-smooth stochastic compositional optimization. In particular, for convex objectives, smoothing stochastic gradient achieves a~$1/T^{1/4}$ rate in terms of the number of stochastic gradient evaluations. We further show how the general compositional and finite sum compositional problems (widely used frameworks in large-scale machine learning and risk-averse optimization) fit the assumptions needed for the rates (unbiased gradient estimates, bounded second moments, and accurate smoothing errors). We will present numerical results indicating that smoothing stochastic gradient descent is a competitive method for certain classes of problems.

Talk 3: (Canceled) Convergence analysis of the 9th Chebyshev Method for Nonconvex, Nonsmooth Optimization Problems
Speaker: Jiabao Yang
Abstract: Ushiyama-Sato-Matsuo (2022) (hereafter referred to as [USM]) showed that some optimization methods can be regarded as numerical analysis methods for ordinary differential equations. [USM] proposed the 9th Chebyshev method, an optimization method that allows a larger step width than the steepest descent method, under the assumption that the objective function is convex and differentiable. Observing the left edge of the stability domain, we see the 9th Chebyshev method should allow about 78 times larger steps than the steepest descent method. On the other hand, the 9th Chebyshev method requires 9 evaluations of the gradient per iteration, while the steepest descent method requires only one. Considering most calculation time is spent on the gradient, we expect that the 9th Chebyshev method converges 78/9 = 8.7 times faster. However, there are many cases in the world where differentiable, smooth, convex, etc. cannot be used. For example, problems in image analysis and voice recognition require the use of nondifferentiable, nonsmooth, and nonconvex functions. The steepest descent method or the 9th Chebyshev method proposed by [USM] assumes that the objective function is convex and differentiable, and thus cannot be applied to problems with an objective function that is not necessarily differentiable. For nonconvex optimization problems, finding a good local optimal solution is a realistic goal, and Riis-Ehrhardt-Quispel-Scholieb (2022) (hereafter referred to as [REQS]) introduced an alternative concept of differentiation called “Clarke subdifferential” for functions that are not always differentiable, and proposed an optimization method that can be applied to objective functions that are not necessarily differentiable and convex. Although the objective function is not assumed to be differentiable or convex, such as being locally Lipschitz continuous, [REQS] prove that the optimization method converges to the set of Clarke stationary points defined in the framework of Clarke subdifferential under the condition that some assumptions are satisfied. In this talk, based on the proof method of [REQS], we propose a method that combines the method of [REQS] and the method of [USM] and prove that the proposed method converges to the set of Clarke stationary points under the same objective function assumption as [REQS]. Refernces: 1, Kansei Ushiyama, Shun Sato, Takayasu Matsuo, "Deriving efficient optimization methods based on stable explicit numerical methods", JSIAM Letters Vol.14 (2022) pp.29–32 2, Erlend S.Riis, Matthias J.Ehrhardt, G.R.W.Quispel, Carola-Bibiane Schonlieb, "A Geometric Integration Approach to Nonsmooth, Nonconvex Optimisation", Foundations of Computational Mathematics (2022) 22:1351–1394

Speakers
SH

Shaoning Han

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JT

Jingfu Tan

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JY

Jiabao Yang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 112 3501 Trousdale Pkwy, 112, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4X: Algorithms for Nonconvex Problems
Tuesday July 22, 2025 10:30am - 11:45am PDT
Session: Algorithms for Nonconvex Problems
Chair: Yulin Peng
Cluster: nan

Talk 1: NEW ALGORITHMS FOR HARD OPTIMIZATION problems PROBLEMS.
Speaker: Aharonl Ben-Tal
Abstract: NEW ALGORITHMS FOR HARD (NONCONVEX) OPTIMIZATION PROBLEMS. The problems addressed in this talk are: (1) Max of convex function (2) Max of max of convex function (3) Max of Difference of convex functions. Almost all existing algorithms for such problems suffer from might be called “the curse of obtaining a good starting point”. In our algorithms a starting point is computed by employing only tractable methods for convex problems. The core algorithm on which the algorithms for problems (2) and (3) are based, is the COMAX algorithm developed for problem (1), See Ben-Tal, A. and Roos E., "An Algorithm for Maximizing a Convex Function Based on its Minimizer". INFORMS Journal on Computing Volume: 34, Number: 6 (November-December 2022): 3200-

Talk 3: Conditional Infimum, Hidden Convexity and the S-Procedure
Speaker: Michel De Lara

Speakers
AB

Aharonl Ben-Tal

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
MD

Michel De Lara

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 10:30am - 11:45am PDT
Taper Hall (THH) 215 3501 Trousdale Pkwy, 215, Los Angeles, CA 90089

10:30am PDT

Parallel Sessions 4Y
Tuesday July 22, 2025 10:30am - 11:45am PDT
Tuesday July 22, 2025 10:30am - 11:45am PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 200 3518 Trousdale Pkwy, 200, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5A: Methods for Large-Scale Nonlinear Optimization V
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Methods for Large-Scale Nonlinear Optimization V
Chair: Raghu Bollapragada
Cluster: Nonlinear Optimization

Talk 1: The Interior-Point Frank-Wolfe Method for Minimizing Smooth Convex Functionals over Linearly Constrained Probability Spaces
Speaker: Di Yu
Abstract: Several important and large problem classes including optimal experimental design, emergency volunteer response, spectral optimal transport, and certain queuing problems can be posed as that of minimizing a smooth convex functional over the space of compactly supported probability measures subject to a linear constraint map. We introduce the interior-point Frank-Wolfe (IPFW) method for solving such problems, where a sequence of barrier problems is constructed as a way to handle the constraints, each of which is solved to increasing accuracy using a Frank-Wolfe method. Importantly, the Frank-Wolfe sub-problems are shown to have a ``closed-form'' solution expressed in terms of the constraint functionals and the influence function of the objective. The sequence of probability measures resulting from the IPFW framework is shown to converge in a certain sense to the correct solution, along with a complexity calculation. The problem and the proposed solution is illustrated through examples.

Talk 2: Local convergence of adaptively regularized tensor methods
Speaker: Karl Welzel
Abstract: Tensor methods are methods for unconstrained continuous optimization that can incorporate derivative information of up to order p > 2 by computing a step based on the pth-order Taylor expansion at each iteration. The most important among them are regularization-based tensor methods which have been shown to have optimal worst-case iteration complexity of finding an approximate minimizer. Moreover, as one might expect, this worst-case complexity improves as p increases, highlighting the potential advantage of tensor methods. Still, the global complexity results only guarantee pessimistic sublinear rates, so it is natural to ask how local rates depend on the order of the Taylor expansion p. In the case of functions that are uniformly convex of order q (ones that around the minimizer x* grow like the distance to x* to the qth power) and a fixed regularization parameter, the answer is given in a paper by Doikov and Nesterov from 2022: we get (p/(q-1))th-order local convergence of function values and gradient norms, if p > q-1. In particular, when the Hessian is positive definite at the minimizer (q=2) we get pth-order convergence, but also when the Hessian is singular at x* (q > 2) superlinear convergence (compared to Newton's linear convergence) is possible as long as enough derivatives are available. The value of the regularization parameter in their analysis depends on the Lipschitz constant of the pth derivative. Since this constant is not usually known in advance, adaptive regularization methods are more practical. We extend the local convergence results to locally uniformly convex functions and fully adaptive methods. We discuss how for p > 2 it becomes crucial to select the "right" minimizer of the regularized local model in each iteration to ensure all iterations are eventually successful. Counterexamples show that in particular the global minimizer of the subproblem is not suitable in general. If the right minimizer is used, the (p/(q-1))th-order local convergence is preserved, otherwise the rate stays superlinear but with an exponent arbitrarily close to one depending on the algorithm parameters.

Talk 3: Noise-Aware Sequential Quadratic Programming for Equality Constrained Optimization with Rank-Deficient Jacobians
Speaker: Jiahao Shi
Abstract: We propose and analyze a sequential quadratic programming algorithm for minimizing a noisy nonlinear smooth function subject to noisy nonlinear smooth equality constraints. The algorithm uses a step decomposition strategy and, as a result, is robust to potential rank-deficiency in the constraints, allows for two different step size strategies, and has an early stopping mechanism. Under linear independence constraint qualifications, convergence is established to a neighborhood of a first-order stationary point, where the radius of the neighborhood is proportional to the noise level in the objective function and constraints. Moreover, in the rank deficient setting, convergence to a neighborhood of an infeasible stationary point is established. Numerical experiments demonstrate the efficiency and robustness of the proposed method.

Speakers
RB

Raghu Bollapragada

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
DY

Di Yu

Name: Dr. Slothington "Slow Convergence" McNapfaceTitle: Distinguished Professor of Continuous Optimization & Energy MinimizationAffiliation: The Lush Canopy Institute of Sluggish AlgorithmsBio:Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
KW

Karl Welzel

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JS

Jiahao Shi

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 101 3501 Trousdale Pkwy, 101, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5B: Continuous and Discrete Optimization
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Continuous and Discrete Optimization
Chair: Marcia Fampa
Cluster: Interplay Between Continuous and Discrete Optimization

Talk 1: Stochastic Iterative Methods for Linear Problems with Adversarial Corruption
Speaker: Jamie Haddock
Abstract: Stochastic iterative methods, like stochastic gradient descent or the randomized Kaczmarz method, have gained popularity in recent times due to their amenability to large-scale data and distributed computing environments. This talk will focus on variants of the randomized Kaczmarz methods developed for problems with significant or adversarial corruption present in the problem-defining data. This type of corruption arises frequently in applications like medical imaging, sensor networks, error correction, and classification of mislabeled data. We will focus on recent results for linear feasibility problems and tensor regression problems.

Talk 2: Volume formulae for the convex hull of the graph of a $n$-monomial in some special cases
Speaker: Emily Speakman
Abstract: The spatial branch-and-bound algorithmic framework, used for solving non-convex mixed-integer nonlinear optimization problems, relies on obtaining quality convex approximations of the non-convex substructures in a problem formulation. A common example is a simple monomial, $y=x_1x_2, \dots, x_n$, defined over the box $[a_1, b_1] \times [a_2, b_2] \times \dots \times [a_n, b_n] \subseteq \R^n$. There are many choices of convex set that could be used to approximate this solution set, with the (polyhedral) convex hull giving the “tightest” or best possible outer approximation. By computing the volume of the convex hull, we obtain a measure that can be used to evaluate other convex outer approximations in comparison. In previous work, we have obtained a formula for the volume of the convex hull of the graph of a trilinear monomial (i.e., $n=3$) in terms of the $6 = 2n$ box parameters. Here, we seek to extend our work to the case of general $n$ by making additional assumptions on the box domain. In particular, we assume that only $k$ variables have a non-zero lower bound. In this work, we consider $k=1,2,3$, and conjecture the volume of the convex hull in each case. Moreover, we provide a proof for the case $k=1$.

Talk 3: New hierarchies for disjoint bilinear programs
Speaker: Mohit Tawarmalani
Abstract: Disjoint bilinear programs are mathematical optimization problems involving minimization of a bilinear function over a Cartesian product of polytope. In this paper, we iteratively obtain, in closed-form, rational functions that are barycentric coordinates of successively tighter outer-approximations of a polytope. This procedure converges in m steps, where m is the number of constraints describing the polytope. Using this procedure, we construct a finite hierarchy of relaxations that in m steps describes the convex hull of bilinear functions over the feasible region providing a linear reformulation for disjoint bilinear programming.

Speakers
JH

Jamie Haddock

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
ES

Emily Speakman

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
MT

Mohit Tawarmalani

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 201 3501 Trousdale Pkwy, 201, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5C: AI Meets Optimization (Part 2)
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: AI Meets Optimization (Part 2)
Chair: Wotao Yin
Cluster: Optimization for Emerging Technologies (LLMs, Quantum Computing, ...)

Talk 1: TBD
Speaker: Jan Drgona
Abstract: TBD

Talk 2: TBD
Speaker: Jialin Liu
Abstract: TBD

Talk 3: TBD
Speaker: Junchi Yan
Abstract: TBD

Speakers
JD

Jan Drgona

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
avatar for Jialin Liu

Jialin Liu

Assistant Professor, University of Central Florida
Name: Dr. Slothington "Slow Convergence" McNapfaceTitle: Distinguished Professor of Continuous Optimization & Energy MinimizationAffiliation: The Lush Canopy Institute of Sluggish AlgorithmsBio:Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JY

Junchi Yan

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 208 3501 Trousdale Pkwy, 208, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5D: Recent advances in algorithms for large-scale optimization (I)
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Recent advances in algorithms for large-scale optimization (I)
Chair: Xudong Li
Cluster: Computational Software

Talk 1: Infeasible Model Analysis In the Optverse Solver
Speaker: Zirui Zhou
Abstract: Isolating an Irreducible Infeasible Subset (IIS) of constraints is the best way to analyze an infeasible optimization model. The OptVerse solver incorporates very fast algorithms for this purpose. The LP analyzer takes advantage of the presolver to isolate a small subset of constraints for conventional analysis, whether or not the presolve detects infeasibility. The MIP analyzer uses new techniques to very quickly find an IIS that includes the integrality restrictions. Experimental results are given.

Talk 2: HOT: An Efficient Halpern Accelerating Algorithm for Optimal Transport Problems
Speaker: Yancheng Yuan
Abstract: This talk introduces an efficient HOT algorithm for solving the optimal transport (OT) problems with finite supports. We particularly focus on an efficient implementation of the HOT algorithm for the case where the supports are in $\mathbb{R}^2$ with ground distances calculated by $L_2^2$-norm. Specifically, we design a Halpern accelerating algorithm to solve the equivalent reduced model of the discrete OT problem. Moreover, we derive a novel procedure to solve the involved linear systems in the HOT algorithm in linear time complexity. Consequently, we can obtain an $\varepsilon$-approximate solution to the optimal transport problem with $M$ supports in $O(M^{1.5}/\varepsilon)$ flops, which significantly improves the best-known computational complexity. We further propose an efficient procedure to recover an optimal transport plan for the original OT problem based on a solution to the reduced model, thereby overcoming the limitations of the reduced OT model in applications that require the transport map. We implement the HOT algorithm in PyTorch and extensive numerical results show the superior performance of the HOT algorithm compared to existing state-of-the-art algorithms for solving the OT problems.

Talk 3: DNNLasso: Scalable Graph Learning for Matrix-Variate Data
Speaker: Meixia Lin
Abstract: We consider the problem of jointly learning row-wise and column-wise dependencies of matrix-variate observations, which are modelled separately by two precision matrices. Due to the complicated structure of Kronecker-product precision matrices in the commonly used matrix-variate Gaussian graphical models, a sparser Kronecker-sum structure was proposed recently based on the Cartesian product of graphs. However, existing methods for estimating Kronecker-sum structured precision matrices do not scale well to large scale datasets. In this work, we introduce DNNLasso, a diagonally non-negative graphical lasso model for estimating the Kronecker-sum structured precision matrix, which outperforms the state-of-the-art methods by a large margin in both accuracy and computational time.

Speakers
XL

Xudong Li

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
ZZ

Zirui Zhou

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
YY

Yancheng Yuan

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
ML

Meixia Lin

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 210 3501 Trousdale Pkwy, 210, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5E: Advances in Mixed-Integer Optimization
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Advances in Mixed-Integer Optimization
Chair: Alberto Del Pia
Cluster: Interplay Between Continuous and Discrete Optimization

Talk 1: Convexification Techniques For Logical Implication Constraints Involving Cardinality Requirements
Speaker: Jean-Philippe Richard
Abstract: Cardinality requirements and implications between groups of distinct variables are pervasive in applications and are often modeled through the use of integer programming techniques. We describe a general constructive scheme that allows for the convex hull of sets involving logical implication constraints relating the cardinality of groups of variables to be derived in a higher dimensional space. We also discuss aspects of projecting the formulations. We provide simple illustrative applications of the scheme, which subsume existing results in the literature.

Talk 2: Transfer Theorems in Mixed-Integer Convex Optimization
Speaker: Phillip Kerger
Abstract: In this talk, we will present two lines of work that explore the transferability of results between different settings in optimization. First, we will show how performance guarantees from noise-free convex optimization can be adapted to the stochastic setting, even when mixed-integer variables are present. This is achieved through a black-box transfer approach that applies broadly to first-order methods. Second, we will discuss how complexity results from continuous convex optimization can be extended to the mixed-integer setting, which leads to new lower bounds under various oracles, such as those with partial first-order information. Such black-box approaches are especially appealing to have results in easier-to-analyze cases immediately transfer to more complex ones. Finally, some remaining open questions will be discussed.

Talk 3: Extended formulations for some class of Delta-modular IPs
Speaker: Luze Xu
Abstract: Conforti et al. give a compact extended formulation for a class of bimodular-constrained integer programs, namely those that model the stable set polytope of a graph with no disjoint odd cycles. We extend their techniques to design compact extended formulations for the integer hull of translated polyhedral cones whose constraint matrix is strictly $\Delta$-modular and has rows that represent a cographic matroid. Our work generalizes the important special case from Conforti et al. concerning 4-connected graphs with odd cycle transversal number at least 4. We also discuss the necessity of our assumptions. This is joint work with Joseph Paat and Zach Walsh.

Speakers
JR

Jean-Philippe Richard

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
PK

Phillip Kerger

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
LX

Luze Xu

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 212 3501 Trousdale Pkwy, 212, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5F: Adaptive Methods
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Adaptive Methods
Chair: Ilyas Fatkhullin
Cluster: Optimization For Data Science

Talk 1: The Price of Adaptivity in Stochastic Convex Optimization
Speaker: Oliver Hinder
Abstract: We prove impossibility results for adaptivity in non-smooth stochastic convex optimization. Given a set of problem parameters we wish to adapt to, we define a ``price of adaptivity'' (PoA) that, roughly speaking, measures the multiplicative increase in suboptimality due to uncertainty in these parameters. When the initial distance to the optimum is unknown but a gradient norm bound is known, we show that the PoA is at least logarithmic for expected suboptimality, and double-logarithmic for median suboptimality. When there is uncertainty in both distance and gradient norm, we show that the PoA must be polynomial in the level of uncertainty. Our lower bounds nearly match existing upper bounds, and establish that there is no parameter-free lunch. En route, we also establish tight upper and lower bounds for (known-parameter) high-probability stochastic convex optimization with heavy-tailed and bounded noise, respectively.

Talk 2: Adaptive Online Learning and Optimally Scheduled Optimization
Speaker: Ashok Cutkosky
Abstract: In this talk I will describe some recent advances in online learning, and how these advances result in improved algorithms for stochastic optimization. We will first describe new online optimization algorithms that achieve optimal regret with neither prior knowledge of Lipschitz constants nor bounded domain assumptions, which imply stochastic optimization algorithms that perform as well as SGD with an optimally-tuned learning rate. We will then survey new and improved conversions from online to stochastic optimization that shed light on heuristic learning rate schedules popular in practice, and illustrate how this analysis allows us to begin investigation into identifying an optimal schedule of learning rates. This is in contrast to most literature on adaptive stochastic optimization that typically seeks to compete only with a single fixed learning rate. We will conclude by highlighting open problems in both online and stochastic optimization.

Talk 3: Unveiling the Power of Adaptive Methods Over SGD: A Parameter-Agnostic Perspective
Speaker: Junchi Yang
Abstract: Adaptive gradient methods are popular in optimizing modern machine learning models, yet their theoretical benefits over vanilla Stochastic Gradient Descent (SGD) remain unclear. We examines the convergence of SGD and adaptive methods when their hyperparameters are set without knowledge of problem-specific parameters. First, for smooth functions, we compare SGD to well-known adaptive methods like AdaGrad, Normalized SGD with Momentum (NSGD-M), and AMSGrad. While untuned SGD attains the optimal convergence rate, it comes at the expense of an unavoidable exponential dependence on the smoothness constant. In contrast, several adaptive methods reduce this exponential dependence to polynomial. Secondly, for a broader class of functions characterized by (L0, L1) smoothness, SGD fail without proper tuning. We show NSGD-M achieves a near-optimal rate, despite an exponential dependence on the L1 constant, which we show is unavoidable for a family of normalized momentum methods.

Speakers
IF

Ilyas Fatkhullin

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
OH

Oliver Hinder

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
AC

Ashok Cutkosky

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JY

Junchi Yang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 156 3518 Trousdale Pkwy, 156, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5G: Recent Advances in Large-scale Optimization III
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Recent Advances in Large-scale Optimization III
Chair: Salar Fattahi
Cluster: Nonlinear Optimization

Talk 1: Convexification of a class of optimization problems with step functions
Speaker: Andres Gomez
Abstract: We study nonlinear optimization problems with discontinuous step functions. This class of problems arises often in statistical problems, modeling correct/incorrect predictions with linear classifiers, fairness and nearly isotonic regression. We discuss special classes of problems where convex hulls can be computed explicitly, and discuss algorithmic implementations of the convexifications. Our computational results indicate a good statistical estimators can be obtained from the optimal solutions of the convex relaxations, and that exact methods can be substantially improved in some cases.

Talk 2: AGDA+: Proximal Alternating Gradient Descent Ascent Method with a Nonmonotone Adaptive Step-Size Search for Nonconvex Minimax Problems
Speaker: Xuan Zhang
Abstract: We consider double-regularized nonconvex-strongly concave (NCSC) minimax problems of the form (P) : minx maxy g(x) + f (x,y) − h(y), where g, h are closed convex, f is L-smooth in (x, y) and strongly concave in y. We propose a proximal alternating gradient descent ascent method AGDA+ that can adaptively choose nonmonotone primal-dual stepsizes to compute an approximate stationary point for (P) without requiring the knowledge of the global Lipschitz constant L. Using a nonmonotone step-size search (backtracking) scheme, AGDA+ stands out by its ability to exploit the local Lipschitz structure and eliminates the need for precise tuning of hyper-parameters. AGDA+ achieves the optimal iteration complexity of O(ε-2) and it is the first step-size search method for NCSC minimax problems that require only O(1) calls to ∇f per backtracking iteration. We also discuss how AGDA+ can easily be extended to search for μ as well. The numerical experiments demonstrate its robustness and efficiency.

Talk 3: Towards Weaker Variance Assumptions for Stochastic Optimization
Speaker: Ahmet Alacaoglu
Abstract: We revisit a classical assumption for analyzing stochastic gradient algorithms where the squared norm of the stochastic subgradient (or the variance for smooth problems) is allowed to grow as fast as the squared norm of the optimization variable. We contextualize this assumption in view of its inception in the 1960s, its seemingly independent appearance in the recent literature, its relationship to weakest-known variance assumptions for analyzing stochastic gradient algorithms, and its relevance even in deterministic problems for non-Lipschitz nonsmooth convex optimization. We build on and extend a connection recently made between this assumption and the Halpern iteration. For convex nonsmooth, and potentially stochastic, optimization we provide horizon-free algorithms with last-iterate rates. For problems beyond simple constrained optimization, such as convex problems with functional constraints, we obtain rates for optimality measures that do not require boundedness of the feasible set.

Speakers
SF

Salar Fattahi

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
AG

Andres Gomez

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
XZ

Xuan Zhang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
AA

Ahmet Alacaoglu

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 114 3501 Trousdale Pkwy, 114, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5H: Preference Robust Optimization
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Preference Robust Optimization
Chair: Wenjie Huang
Cluster: Optimization Under Uncertainty and Data-driven Optimization

Talk 1: conformal inverse optimization
Speaker: Erick Delage
Abstract: TBA

Talk 2: Preference robust optimization with quasi-concave choice functions in multi-attribute decision making: characterization and computation
Speaker: Wenjie Huang
Abstract: TBA

Talk 3: Adaptive Preference Elicitation in Preference Robust CPT-Based Shortfall
Speaker: Sainan Zhang
Abstract: TBA

Speakers
ED

Erick Delage

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
avatar for Wenjie Huang

Wenjie Huang

The University of Hong Kong
Dr. Wenjie Huang is in Department of Data and Systems Engineering and Musketeers Foundation Institute of Data Science, The University of Hong Kong (HKU). He received B.S. degree in Industrial Engineering and Management, with minor in Economics, from Shanghai Jiao Tong University in... Read More →
SZ

Sainan Zhang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 116 3501 Trousdale Pkwy, 116, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5I: Dynamic Optimization: Deterministic and Stochastic Continuous-time Models II
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Dynamic Optimization: Deterministic and Stochastic Continuous-time Models II
Chair: Cristopher Hermosilla
Cluster: Multi-agent Optimization and Games

Talk 1: Average Cost Problems Subject to State Constraints
Speaker: Nathalie Khalil
Abstract: Control systems with unknown parameters provide a natural framework for modeling uncer- tainties in various applications. In this work, we focus on pathwise state-constrained optimal control problems where these unknown parameters affect the system’s dynamics, cost function, endpoint constraint, and state constraint. The objective is to minimize a cost criterion expressed in integral form, the so-called “average cost”, with the cost function evaluated relative to a refer- ence probability measure defined over the set of unknown parameters. For this class of problems, we derive necessary optimality conditions. By using an average cost criterion, this approach offers an alternative to traditional minimax or robust optimization methods.

Talk 2: Degeneration in Optimal Control Problems with Non-regular Mixed Constraints
Speaker: Karla Cortez
Abstract: In this talk we will discuss the emergence of the degeneration phenomenon in the necessary conditions derived in recent literature on optimal control problems with non-regular mixed constraints. We will examine how the lack of regularity in these constraints can lead to trivial multipliers, hindering the applicability of classical optimality conditions. To address this issue, we will present non-degeneration conditions that ensure the existence of non-trivial multipliers in these problems and we will illustrate their potential through examples and preliminary results. The talk will conclude with a discussion on future directions for extending and validating these findings.

Talk 3: Uniqueness of Multipliers in Optimal Control Problems
Speaker: Jorge Becerril
Abstract: In this talk, we explore conditions for the uniqueness of multipliers in optimal control problems across different settings. We start with the simplest case involving only isoperimetric constraints, highlighting the connection between uniqueness and key concepts such as regularity and normality. Next, we examine optimal control problems with regular mixed constraints, focusing on piecewise continuous controls. In this context, the continuity assumption allows us to relate uniqueness to the observability of a certain no-input linear system. Under additional assumptions, using results from viability theory, we can establish a one-to-one correspondence between the set of Lagrange multipliers and the initial value of certain set-valued function. Lastly, we discuss ongoing efforts to extend these findings to the case of essentially bounded controls.

Speakers
CH

Cristopher Hermosilla

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
avatar for Nathalie Khalil

Nathalie Khalil

Researcher, University of Porto, Portugal
I am a Researcher and Lecturer at the Department of Electrical Engineering and Computer Science of the Faculty of Engineering at the University of Porto. I specialize in optimal control theory and its applications to engineering challenges.I have over 10 years of research experience... Read More →
KC

Karla Cortez

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JB

Jorge Becerril

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 100 3518 Trousdale Pkwy, 100, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5J: Models and Algorithms for Optimization with Rare Events
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Models and Algorithms for Optimization with Rare Events
Chair: Anirudh Subramanyam
Cluster: Optimization Under Uncertainty and Data-driven Optimization

Talk 1: Decision-Making under Extreme Risks: Configuring Optimization Algorithms for Rare-Event Optimization
Speaker: Henry Lam
Abstract: We consider stochastic optimization where the goal is not only to optimize an average-case objective, but also mitigate the occurrence of rare but catastrophic events. This problem, which is motivated from emerging applications such as safe AI, requires an integration of variance reduction methods into sampling-based optimization algorithms in order to attain sufficient solution accuracy. However, we explain how natural variance-reduction-optimization integration, even executed in an adaptive fashion studied by recent works, encounters fundamental challenges. On a high level, the challenge arises from the extreme sensitivity of tail-based objectives with respect to the decision variables, which renders the failure of traditional Lipschitz-based analyses. We offer some potential remedies and supporting numerical results.

Talk 2: Risk-Aware Path Integral Diffusion to Sample Rare Events
Speaker: Michael Chertkov
Abstract: TBD

Talk 3: Scaling Scenario-Based Chance-Constrained Optimization under Rare Events
Speaker: Jaeseok Choi
Abstract: Chance-constrained optimization is a suitable modeling framework for mitigating extreme event risk in many practical settings. The scenario approach is a popular solution method for chance-constrained problems, due to its straightforward implementation and ability to preserve problem structure. However, for safety-critical applications where violating constraints is nearly unacceptable, the scenario approach becomes computationally infeasible due to the excessively large sample sizes it demands. We address this limitation with a new yet straightforward decision-scaling technique. Our method leverages large deviation principles and relies on only mild nonparametric assumptions about the underlying uncertainty distributions. The method achieves an exponential reduction in sample size requirements compared to the classical scenario approach for a wide variety of constraint structures, while also guaranteeing feasibility with respect to the uncertain constraints. Numerical experiments spanning engineering and management applications show that our decision-scaling technique significantly expands the scope of problems that can be solved both efficiently and reliably.



Speakers
AS

Anirudh Subramanyam

Assistant Professor, Pennsylvania State University
Name: Dr. Slothington "Slow Convergence" McNapfaceTitle: Distinguished Professor of Continuous Optimization & Energy MinimizationAffiliation: The Lush Canopy Institute of Sluggish AlgorithmsBio:Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
HL

Henry Lam

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
MC

Michael Chertkov

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 102 3501 Trousdale Pkwy, 102, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5K: First-order methods for nonsmooth and constrained optimization problems
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: First-order methods for nonsmooth and constrained optimization problems
Chair: Digvijay Boob
Cluster: Nonlinear Optimization

Talk 1: Linearly Convergent Algorithms for Nonsmooth Problems with Unknown Smooth Pieces
Speaker: Zhe Zhang
Abstract: Non-smoothness is a major bottleneck to efficient optimization. In the absence of smoothness, the theoretical convergence rates drop from linear to sublinear for convex programs, and become orders of magnitude worse for nonconvex programs. Moreover, this huge is known to be unimprovable in general. We focus on mild, structured non-smooth problems: piecewise smooth (PWS) functions whose domain can be partitioned into subsets such that restricted to each subset the function is smooth. PWS functions cover ReLU functions, L1-penalties, and min-max saddle point problems as special cases. In particular, we study convex PWS functions for which we present globally linear convergent methods under the quadratic growth condition; as a corollary, we improve the iteration complexity for solving weakly convex PWS problems by orders of magnitude. Importantly, our method does not require any knowledge about individual smooth pieces, and is thus applicable even to general non-smooth programs exhibiting local PWS structure.

Talk 2: Primal-Dual Algorithm with Last iterate Convergence Guarantees for Stochastic Convex Optimization Problems
Speaker: Mohammad Khalafi
Abstract: We provide the first method that obtains the best-known convergence rate guarantees on the last iterate for stochastic composite nonsmooth convex function-constrained optimization problems. Our novel and easy-to-implement algorithm is based on the augmented Lagrangian technique and uses a new linearized approximation of constraint functions, leading to its name, the Augmented Constraint Extrapolation (Aug-ConEx) method. We show that Aug-ConEx achieves convergence rate in the nonsmooth stochastic setting without any strong convexity assumption and for the same problem with strongly convex objective function. While optimal for nonsmooth and stochastic problems, the Aug-ConEx method also accelerates convergence in terms of Lipschitz smoothness constants to and in the aforementioned cases, respectively. To our best knowledge, this is the first method to obtain such differentiated convergence rate guarantees on the last iterate for a composite nonsmooth stochastic setting without additional factors. We validate the efficiency of our algorithm by comparing it with a state-of-the-art algorithm through numerical experiments.

Talk 3: An Optimal Method for Minimizing Heterogeneously Smooth and Convex Compositions
Speaker: Aaron Zoll
Abstract: This talk will discuss a universal, optimal algorithm for convex minimization problems of the composite form f(x) + h(g_1(x), ..., g_m(x)). We allow each component function g_i(x) to independently range from being nonsmooth Lipschitz to smooth and from convex to strongly convex, described by notions of Holder continuous gradients and uniform convexity. We note that although the objective is built from a heterogeneous combination of components, it does not necessarily possess any smoothness, Lipschitzness, or any favorable structural properties. Regardless, our proposed sliding accelerated gradient method converges at least as fast as the optimal rate guarantees in terms of oracle access to (sub)gradients of each g_i seperately. Furthermore, given access to an estimate of the initial distance to optimal, we provide a “mostly parameter-free” variant. As a key application, fixing h as a nonpositive indicator function, this model readily captures functionally constrained minimization f(x) subject to g_i(x) \leq 0. Our algorithm and analysis are directly inspired by the Q-analysis technique developed for such smooth constrained minimization by Zhang and Lan. Our theory recovers their accelerated guarantees and extends them to benefit from heterogeneously smooth and convex constraints.

Speakers
DB

Digvijay Boob

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
ZZ

Zhe Zhang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
MK

Mohammad Khalafi

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
AZ

Aaron Zoll

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 118 3501 Trousdale Pkwy, 118, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5L: Optimization for Machine Learning and AI
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Optimization for Machine Learning and AI
Chair: Tianbao Yang
Cluster: Optimization Applications (Communication, Energy, Health, ML, ...)

Talk 1: Model Developmental Safety: A Constrained Optimization Approach
Speaker: Gang Li
Abstract: In this talk, we introduce introduce model developmental safety as a guarantee of a learning system such that in the model development process the new model should strictly preserve the existing protected capabilities of the old model while improving its performance on target tasks. To ensure the model developmental safety, we present a safety-centric framework by formulating the model developmental safety as data-dependent constraints. Under this framework, we study how to develop a pretrained vision-language model (aka the CLIP model) for acquiring new capabilities or improving existing capabilities of image classification. We propose an efficient constrained optimization algorithm with theoretical guarantee and use its insights to finetune a CLIP model with task-dependent heads for promoting the model developmental safety. Our experiments on improving vision perception capabilities on autonomous driving and scene recognition datasets demonstrate the efficacy of the proposed approach.

Talk 2: Provable Accelerated Convergence of Nesterov’s Momentum for Deep ReLU Neural Networks
Speaker: Jasper Liao
Abstract: Current state-of-the-art analyses on the convergence of gradient descent for training neural networks focus on characterizing properties of the loss landscape, such as the Polyak-Lojaciewicz (PL) condition and the restricted strong convexity. While gradient descent converges linearly under such conditions, it remains an open question whether Nesterov’s momentum enjoys accelerated convergence under similar settings and assumptions. In this work, we consider a new class of objective functions, where only a subset of the parameters satisfies strong convexity, and show Nesterov’s momentum achieves acceleration in theory for this objective class. We provide two realizations of the problem class, one of which is deep ReLU networks, which constitutes this work as the first that proves an accelerated convergence rate for non-trivial neural network architectures.

Talk 3: In-processing Methods for Training Partially Fair Predictive Models Based on Difference-of-Convex Constraints
Speaker: Qihang Lin
Abstract: Fairness in machine learning has become a critical concern, particularly in high-stakes applications. Existing approaches often focus on achieving full fairness across all score ranges generated by predictive models, ensuring fairness in both high and low-scoring populations. However, this stringent requirement can compromise predictive performance and may not align with the practical fairness concerns of stakeholders. In this work, we propose a novel framework for building partially fair machine learning models, which enforce fairness within a specific score range of interest, such as the middle range where decisions are most contested, while maintaining flexibility in other regions. We introduce two statistical metrics to rigorously evaluate partial fairness within a given score range, such as the top 20%–60% of scores. To achieve partial fairness, we propose an in-processing method by formulating the model training problem as constrained optimization with difference-of-convex constraints, which can be solved by an inexact difference-of-convex algorithm (IDCA). We provide the complexity analysis of IDCA for finding a nearly KKT point. Through numerical experiments on real-world datasets, we demonstrate that our framework achieves high predictive performance while enforcing partial fairness where it matters most.

Speakers
TY

Tianbao Yang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
GL

Gang Li

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JL

Jasper Liao

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
QL

Qihang Lin

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 119 3501 Trousdale Pkwy, 119, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5M: Recent progresses in derivative-free optimization II
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Recent progresses in derivative-free optimization II
Chair: Jeffrey Larson
Cluster: Derivative-free Optimization

Talk 1: Some Available Derivative (SAD) Optimization
Speaker: Jeffrey Larson
Abstract: Many practical optimization problems involve objective functions where gradient information is unavailable or expensive to compute for certain variables but is readily available for other variables. This talk presents an optimization method that effectively incorporates available partial gradient information with state-of-the-art derivative-free optimization methods. We introduce a new algorithm tailored to this setting, demonstrate its effectiveness through numerical experiments, and discuss its application to optimizing fusion stellarator designs, where coil parameter gradients are accessible, but plasma parameter gradients are not.

Talk 2: Barycenter of weight coefficient region of least norm updating quadratic models with vanishing trust-region radius
Speaker: Pengcheng Xie
Abstract: Derivative-free optimization problems, where objective function derivatives are unavailable, can be addressed using local quadratic models within a trust-region algorithm. We propose a model updating approach when high accuracy is demanded, such as when the trust-region radius vanishes. Our approach uses the barycenter of a particular coefficient region and is shown to be advantageous in numerical results.

Speakers
JL

Jeffrey Larson

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
avatar for Pengcheng Xie

Pengcheng Xie

Postdoctoral Scholar, Lawrence Berkeley National Laboratory
Pengcheng Xie works for Lawrence Berkeley National Laboratory in the United States as a postdoctoral scholar working with Dr. Stefan M. Wild, specializing in computational mathematics, mathematical optimization, machine learning, and numerical analysis. He holds a PhD from the Chinese... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 157 3518 Trousdale Pkwy, 157, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5N: Optimization on Riemannian manifolds and stratified sets (1/2)
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Optimization on Riemannian manifolds and stratified sets (1/2)
Chair: Guillaume Olikier
Cluster: Optimization on Manifolds

Talk 1: First-order optimization on stratified sets
Speaker: Guillaume Olikier
Abstract: This talk considers the problem of minimizing a differentiable function with locally Lipschitz continuous gradient on a closed subset of a Euclidean vector space that can be partitioned into finitely many smooth submanifolds. The partition is called a stratification and the submanifolds are called the strata. Under suitable assumptions on the stratification, first-order methods are proposed that generate a sequence in the set whose accumulation points are guaranteed to be Bouligand stationary. These methods combine retracted line searches along descent directions selected in the Bouligand tangent cone and projections onto the strata. Examples of a set satisfying the assumptions include the algebraic variety of real matrices of upper-bounded rank and its intersection with the cone of symmetric positive-semidefinite matrices. On these sets, Bouligand stationarity is the strongest necessary condition for local optimality.

Talk 2: The Effect of Smooth Parametrizations on Nonconvex Optimization Landscapes
Speaker: Joe Kileel
Abstract: Given a constrained optimization problem, we often tackle it by choosing a parameterization of the constraint set and then optimizing over the parameters. This lets us to approach optimization problems over manifolds or stratified sets through problems over simpler sets, such as Euclidean space without constraints. For example, if we need to optimize a real-valued cost over bounded-rank matrices, then we can parameterize the domain using a low-rank factorization (e.g., the SVD) and then optimize over the factors which are unconstrained. Alternatively, in deep learning when optimizing over the function space represented by a neural network, we have parameterized the space by the weights and biases in the network. In these situations, a natural question is: does the choice of parameterization affect the nonconvex optimization landscape? And: are some parameterizations “better” than others? In this talk, I'll present a geometric framework to formalize such questions and new analysis tools to help answer them. The theory will be applied to several examples, including the aforementioned ones as well as optimization problems from tensor decomposition and semidefinite programming. Based on joint works with Eitan Levin (Caltech), Nicolas Boumal (EPFL), and Chris Criscitiello (EPFL).

Talk 3: Geodesic convexity of polar decomposition
Speaker: Foivos Alimisis
Abstract: In this talk, we will analyze a hidden convexity-like structure for the polar decomposition problem in the orthogonal group. This turns out to be similar to convexity-like structures that have been discovered for the symmetric eigenvalue problem. We will talk about the theoretical, but as well practical implications of this structure in polar decomposition problems under uncertainty.

Speakers
GO

Guillaume Olikier

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JK

Joe Kileel

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
FA

Foivos Alimisis

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 256 3518 Trousdale Pkwy, 256, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5O: Global Optimization Theory and Algorithms
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Global Optimization Theory and Algorithms
Chair: Sergiy Butenko
Cluster: Global Optimization

Talk 1: Solving bilevel polynomial optimization by disjunctive decomposition
Speaker: Suhan Zhong
Abstract: We consider bilevel polynomial optimization problems whose lower level constraints are linear on lower level variables. We show that such a bilevel program can be reformulated as a disjunctive program using partial Lagrange multiplier expressions. Each branch problem of this disjunctive program can be efficiently solved by polynomial optimization techniques. We give necessary and sufficient conditions for solutions of branch problems to be global optimizer of original bilevel optimization, and sufficient conditions for the local optimality.

Talk 2: Adaptive Bilevel Knapsack Interdiciton 
Speaker: Jourdain Lamperski
Abstract: We consider adaptive bilevel knapsack interdiction. A leader aims to interdict items that a follower aims to fill their knapsack with. The leader does not have full information, as captured by a finite number of possible realizations of the follower. The leader can adapt to uncertainty by executing one of a budgeted number of precomputed interdiction policies.  We propose exact and approximate solution methods. In particular, we present theoretical performance guarantees for the approximate solution methods and evalutate their empirical performance against the exact methods through computational experiments. 

Talk 3: Continuous Approaches to Cluster-Detection Problems in Networks
Speaker: Sergiy Butenko
Abstract: We discuss continuous formulations for several important cluster-detection problems in networks. More specifically, the problems of interested are formulated as quadratic, cubic, or higher-degree polynomial optimization problems subject to linear constraints. The proposed formulations are used to develop analytical bounds as well as effective algorithms for some of the problems. Moreover, a novel hierarchy of nonconvex continuous reformulations of optimization problems on networks is discussed.

Speakers
SZ

Suhan Zhong

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JL

Jourdain Lamperski

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
SB

Sergiy Butenko

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 258 3518 Trousdale Pkwy, 258, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5P: Feasible and infeasible methods for optimization on manifolds II
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Feasible and infeasible methods for optimization on manifolds II
Chair: Bin Gao
Cluster: Optimization on Manifolds

Talk 1: A decentralized proximal gradient tracking algorithm for composite optimization on Riemannian manifolds
Speaker: Lei Wang
Abstract: This talk focuses on minimizing a smooth function combined with a nonsmooth regularization term on a compact Riemannian submanifold embedded in the Euclidean space under a decentralized setting. Typically, there are two types of approaches at present for tackling such composite optimization problems. The first, subgradient-based approaches, rely on subgradient information of the objective function to update variables, achieving an iteration complexity of $O(\epsilon^{-4}\log^2(\epsilon^{-2}))$. The second, smoothing approaches, involve constructing a smooth approximation of the nonsmooth regularization term, resulting in an iteration complexity of $O(\epsilon^{-4})$. This paper proposes a proximal gradient type algorithm that fully exploits the composite structure. The global convergence to a stationary point is established with a significantly improved iteration complexity of $O(\epsilon^{-2})$. To validate the effectiveness and efficiency of our proposed method, we present numerical results from real-world applications, showcasing its superior performance compared to existing approaches.

Talk 2: A low-rank augmented Lagrangian method for SDP-RLT relaxations of mixed-binary quadratic programs
Speaker: Di Hou
Abstract: The mixed-binary quadratic program (MBQP) with both equality and inequality constraints is a well-known NP-hard problem that arises in various applications. In this work, we focus on two relaxations: the doubly nonnegative (DNN) relaxation and the SDP-RLT relaxation, which combines the Shor relaxation with a partial first-order reformulation-linearization technique (RLT). We demonstrate the equivalence of these two relaxations by introducing slack variables. Furthermore, we extend RNNAL—a globally convergent Riemannian augmented Lagrangian method (ALM) originally developed for solving DNN relaxations—to handle SDP-RLT relaxations. RNNAL penalizes the inequality constraints while keeping the equality constraints in the ALM subproblems. By applying a low-rank decomposition in each ALM subproblem, the feasible region is transformed into an algebraic variety with advantageous geometric properties for us to apply a Riemannian gradient descent method. Our algorithm can efficiently solve general semidefinite programming (SDP) problems, including relaxations for quadratically constrained quadratic programming (QCQP). Extensive numerical experiments confirm the efficiency of the proposed method.

Talk 3: An improved unconstrained approach for bilevel optimization
Speaker: Nachuan Xiao
Abstract: In this talk, we focus on the nonconvex-strongly-convex bilevel optimization problem (BLO). In this BLO, the objective function of the upper-level problem is nonconvex and possibly nonsmooth, and the lower-level problem is smooth and strongly convex with respect to the underlying variable. We show that the feasible region of BLO is a Riemannian manifold. Then we transform BLO to its corresponding unconstrained constraint dissolving problem (CDB), whose objective function is explicitly formulated from the objective functions in BLO. We prove that BLO is equivalent to the unconstrained optimization problem CDB. Therefore, various efficient unconstrained approaches, together with their theoretical results, can be directly applied to BLO through CDB. We propose a unified framework for developing subgradient-based methods for CDB. Remarkably, we show that several existing efficient algorithms can fit the unified framework and be interpreted as descent algorithms for CDB. These examples further demonstrate the great potential of our proposed approach.

Speakers
BG

Bin Gao

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
LW

Lei Wang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
DH

Di Hou

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
NX

Nachuan Xiao

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 106 3501 Trousdale Pkwy, 106, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5Q: Moment-SOS Hierarchy: From Theory to Computation in the Real World (II)
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Moment-SOS Hierarchy: From Theory to Computation in the Real World (II)
Chair: Jie Wang
Cluster: Conic and Semidefinite Optimization

Talk 1: Bregman primal--dual first-order method and application to sparse semidefinite programming
Speaker: Xin Jiang
Abstract: We discuss the centering problem in large-scale semidefinite programming with sparse coefficient matrices. The logarithmic barrier function for the cone of positive semidefinite completable sparse matrices is used as the distance-generating kernel. For this distance, the complexity of evaluating the Bregman proximal operator is shown to be roughly proportional to the cost of a sparse Cholesky factorization. This is much cheaper than the standard proximal operator with Euclidean distances, which requires an eigenvalue decomposition. Then primal-dual proximal algorithm with Bregman distances are applied to solve large-scale sparse semidefinite programs efficiently.

Talk 2: Moment-SOS hierarchies for variational problems and PDE control
Speaker: Giovanni Fantuzzi
Abstract: Moment-SOS hierarchies are an established tool to compute converging sequences of lower bounds on the global minimum of finite-dimensional polynomial optimization problems. In this talk, I will show that they can be combined with finite-element discretizations to give a "discretize-then-relax" framework for solving two classes of infinite-dimensional problems: minimization problems for nonconvex integral functionals over functions from a Sobolev space, and some optimal control problems for nonlinear partial differential equations. For each class of problems, I will discuss conditions ensuring that this "discretize-then-relax" framework produces converging approximations (sometimes, bounds) to the global minimum and to the corresponding optimizer. Gaps between theory and practice will be illustrated by means of examples.

Talk 3: Inexact Augmented Lagrangian Methods for Semidefinite Optimization: Quadratic Growth and Linear Convergence
Speaker: Feng-Yi Liao
Abstract: Augmented Lagrangian Methods (ALMs) are widely employed in solving constrained optimizations, and some efficient solvers are developed based on this framework. Under the quadratic growth assumption, it is known that the dual iterates and the Karush–Kuhn–Tucker (KKT) residuals of ALMs applied to semidefinite programs (SDPs) converge linearly. In contrast, the convergence rate of the primal iterates has remained elusive. In this talk, we resolve this challenge by establishing new quadratic growth and error-bound properties for primal and dual SDPs under the strict complementarity condition. Our main results reveal that both primal and dual iterates of the ALMs converge linearly contingent solely upon the assumption of strict complementarity and a bounded solution set. This finding provides a positive answer to an open question regarding the asymptotically linear convergence of the primal iterates of ALMs applied to semidefinite optimization. 

Speakers
JW

Jie Wang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
XJ

Xin Jiang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
GF

Giovanni Fantuzzi

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 214 3501 Trousdale Pkwy, 214, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5R: Variational Analysis: Theory and Applications III
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Variational Analysis: Theory and Applications III
Chair: Walaa Moursi
Cluster: Nonsmooth Optimization

Talk 1: TBA
Speaker: Sedi Bartz
Abstract: TBA

Talk 2: TBA
Speaker: Taeho Yoon
Abstract: TBA

Talk 3: TBA
Speaker: Scott Lindstrom
Abstract: TBA

Speakers
WM

Walaa Moursi

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
SB

Sedi Bartz

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
avatar for TaeHo Yoon

TaeHo Yoon

Postdoctoral fellow, Johns Hopkins University
Hi, I'm a Rufus Isaacs Postdoctoral Fellow at Johns Hopkins University, Department of Applied Mathematics & Statistics.I work on optimization with general interest in topics including acceleration, minimax optimization, variational inequality, fixed point problems, learning in multiplayer... Read More →
SL

Scott Lindstrom

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 154 3518 Trousdale Pkwy, 154, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5S: Semidefinite Relaxations in Inference Problems
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Semidefinite Relaxations in Inference Problems
Chair: Yong Sheng Soh
Cluster: Conic and Semidefinite Optimization

Talk 1: On the exactness of SDP relaxation for quadratic assignment problem
Speaker: Shuyang Ling
Abstract: Quadratic assignment problem (QAP) is a fundamental problem in combinatorial optimization and finds numerous applications in operation research, computer vision, and pattern recognition. However, it is a very well-known NP-hard problem to find the global minimizer to the QAP. In this work, we study the semidefinite relaxation (SDR) of the QAP and investigate when the SDR recovers the global minimizer. In particular, we consider the two input matrices satisfy a simple signal-plus-noise model, and show that when the noise is sufficiently smaller than the signal, then the SDR is exact, i.e., it recovers the global minimizer to the QAP. It is worth noting that this sufficient condition is purely algebraic and does not depend on any statistical assumption of the input data. We apply our bound to several statistical models such as correlated Gaussian Wigner model. Despite the sub-optimality in theory under those models, empirical studies show the remarkable performance of the SDR. Our work could be the first step towards a deeper understanding of the SDR exactness for the QAP.

Talk 2: Exactness Conditions for Semidefinite Relaxations of the Quadratic Assignment Problem
Speaker: Junyu Chen
Abstract: The Quadratic Assignment Problem (QAP) is an important discrete optimization instance that encompasses many well-known combinatorial optimization problems, and has applications in a wide range of areas such as logistics and computer vision. The QAP, unfortunately, is NP-hard to solve. To address this difficulty, a number of semidefinite relaxations of the QAP have been developed. These techniques are known to be powerful in that they compute globally optimal solutions in many instances, and are often deployed as sub-routines within enumerative procedures for solving QAPs. In this paper, we investigate the strength of these semidefinite relaxations. Our main result is a deterministic set of conditions on the input matrices -- specified via linear inequalities -- under which these semidefinite relaxations are exact. Our result is simple to state, in the hope that it serves as a foundation for subsequent studies on semidefinite relaxations of the QAP as well as other related combinatorial optimization problems. As a corollary, we show that the semidefinite relaxation is exact under a perturbation model whereby the input matrices differ by a suitably small perturbation term. One technical difficulty we encounter is that the set of dual feasible solutions is not closed. To circumvent these difficulties, our analysis constructs a sequence of dual feasible solutions whose objective value approaches the primal objective, but never attains it.

Talk 3: Inference of planted subgraphs in random graphs
Speaker: Cheng Mao
Abstract: Suppose that we are given a graph G obtained from randomly planting a template graph in an Erdős–Rényi graph. Can we detect the presence and recover the location of the planted subgraph in G? This problem is well understood in the case where the template graph is a clique or a dense subgraph but has received less attention otherwise. We consider several settings where the planted subgraph is the power of a Hamiltonian cycle or a random geometric graph which arises in multiple applications. The difficulty of the detention or recovery problem primarily lies in the combinatorial nature of the associated optimization problem. To tackle the problem, we study computationally efficient methods such that subgraph counting and spectral methods which leverage the higher triangle density of the planted subgraph.

Speakers
YS

Yong Sheng Soh

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
SL

Shuyang Ling

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JC

Junyu Chen

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
CM

Cheng Mao

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 155 3518 Trousdale Pkwy, 155, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5T: Optimization for Approximation, Estimation, and Control
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Optimization for Approximation, Estimation, and Control
Chair: Martin Andersen
Cluster: Optimization Applications (Communication, Energy, Health, ML, ...)

Talk 1: Fast and Certifiable Nonconvex Optimal Control with Sparse Moment-SOS Relaxations
Speaker: Heng Yang
Abstract: Direct methods for optimal control, also known as trajectory optimization, is a workhorse for optimization-based control in robotics and beyond. Nonlinear programming with engineered initializations has been the de-facto approach for trajectory optimization, which however, can suffer from undesired local optimality. In this talk, I will first show that, using the machinery of sparse moment and sums-of-squares (SOS) relaxations, many nonconvex trajectory optimization problems can be solved to certifiable global optimality. That is, globally optimal solutions of the original nonconvex problems can be computed by solving convex semidefinite programs (SDPs) together with optimality certificates. I will then present a specialized SDP solver implemented in CUDA (C++) and runs in GPUs that exploits the structures of the problems to solve the convex SDPs at a scale far beyond existing solvers.

Talk 2: Optimal diagonal preconditioner and how to find it
Speaker: Zhaonan Qu
Abstract: Preconditioning has long been a staple technique in optimization, often applied to reduce the condition number of a matrix and speed up the convergence of algorithms. Although there are many popular preconditioning techniques in practice, most lack guarantees on reductions in condition number. Moreover, the degree to which we can improve over existing heuristic preconditioners remains an important practical question. In this talk, we discuss the problem of optimal diagonal preconditioning that achieves maximal reduction in the condition number of any full-rank matrix by scaling its rows and/or columns. We reformulate the problem as a quasi-convex optimization problem and design interior point method to solve it with O(log(1/ϵ)) iteration complexity. Next, we specialize to one-sided optimal diagonal preconditioning problems, and demonstrate that they can be formulated as standard dual SDP problems. Based on the SDP formulation, several computational techniques are applicable, and can greatly accelerate finding a good diagonal preconditioner with theoretical guarantees. Our experiments suggest that optimal diagonal preconditioners can significantly improve upon existing heuristic-based diagonal preconditioners at reducing condition numbers and speeding up iterative methods.

Talk 3: Power System State Estimation by Phase Synchronization and Eigenvectors
Speaker: Iven Guzel
Abstract: To estimate accurate voltage phasors from inaccurate voltage magnitude and complex power measurements, the standard approach is to iteratively refine a good initial guess using the Gauss-Newton method. But the nonconvexity of the estimation makes the Gauss-Newton method sensitive to its initial guess, so human intervention is needed to detect convergence to plausible but ultimately spurious estimates. This paper makes a novel connection between the angle estimation subproblem and phase synchronization to yield two key benefits: (1) an exceptionally high quality initial guess over the angles, known as a spectral initialization; (2) a correctness guarantee for the estimated angles, known as a global optimality certificate. These are formulated as sparse eigenvalue-eigenvector problems, which we efficiently compute in time comparable to a few Gauss-Newton iterations. Our experiments on the complete set of Polish, PEGASE, and RTE models show, where voltage magnitudes are already reasonably accurate, that spectral initialization provides an almost-perfect single-shot estimation of angles from moderately noisy bus power measurements (i.e. pairs of PQ measurements), whose correctness becomes guaranteed after a single Gauss--Newton iteration. For less accurate voltage magnitudes, the performance of the method degrades gracefully; even with moderate voltage magnitude errors, the estimated voltage angles remain surprisingly accurate.

Speakers
avatar for Heng Yang

Heng Yang

Assistant Professor, Harvard University
Assistant Professor at Harvard University working on polynomial optimization and semidefinite programming.
ZQ

Zhaonan Qu

Postdoc, Columbia University
My research interests are at the intersection of econometrics, operations research, and machine learning, with a focus on causal inference, optimization, choice modeling, and networks. I leverage novel connections between these topics to investigate foundational and policy-relevant... Read More →
IG

Iven Guzel

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 158 3518 Trousdale Pkwy, 158, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5U: Quantum Computing and Optimization
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Quantum Computing and Optimization
Chair: Reuben Tate
Cluster: Optimization for Emerging Technologies (LLMs, Quantum Computing, ...)

Talk 1: Advances in Warm-Started QAOA
Speaker: Reuben Tate
Abstract: Over a decade ago, the Quantum Approximate Optimization Algorithm (QAOA) was developed by Farhi et al. for solving certain problems in combinatorial optimization; the algorithm is a variational algorithm that involves the tuning of continuous circuit parameters. Five years ago, Tate et al. and Egger et al. independently developed a warm-start variant that involves using classically-obtained information to construct a warm-started initial quantum state. In this talk, we will go over recent developments in regards to Warm-Started QAOA from both a theoretical, empirical, and algorithmic-design perspective.

Talk 2: Investigating Alternative Metrics of Performance for QAOA
Speaker: Sean Feeney
Abstract: In modern optimization tasks, combinatorial optimization problems (COPs) are paramount in fields such as logistics, power systems, and circuit design. Quantum Approximate Optimization Algorithm (QAOA) has emerged as a promising quantum-classical variational approach for tackling these challenges, yet it often offers limited worst-case approximation guarantees. This work focuses on a recently introduced variant, Warm-Start QAOA, which leverages classical solutions as biased initial states in hopes of boosting the odds of finding high-quality solutions. We conduct an extensive numerical study of Warm-Start QAOA on 3-regular Max-Cut instances. Compared to theoretical lower bounds established by Tate et al. for single-round QAOA, our empirical results consistently indicate better performance across a range of tilt angles. However, this apparent advantage does not extend beyond the classical warm-start solution itself when the QAOA parameters are optimized solely by maximizing the expected objective value. This outcome highlights a pressing need to look beyond standard expectation-value metrics, as they may not capture the subtle elements required for surpassing strong classical baselines in practical settings. To address this gap, we introduce the Better Solution Probability (BSP) metric, which shifts the optimization target from maximizing expectation value performance to maximizing the probability of exceeding a given classical warm-start cut. Our simulations show that BSP-optimized Warm-Start QAOA frequently locates better solutions at non-trivial tilt angles, positions that outperform both the purely classical approach at zero tilt and standard QAOA at 90°, and does so with a non-vanishing probability. Notably, the BSP objective remains feasible to optimize in real-world conditions because it does not rely on a priori knowledge of the optimal solution. By exploring objective functions beyond expectation value, this study offers new insights into how hybrid quantum-classical methods can be enhanced for complex optimization tasks, paving the way for more robust algorithm design and a clearer path toward quantum advantage.

Talk 3: Quantum Hamiltonian Descent Algorithms for Nonlinear Optimization
Speaker: Yufan Zheng
Abstract: Nonlinear optimization is a vibrant field of research with wide-ranging applications in engineering and science. However, classical algorithms often struggle with local minima, limiting their effectiveness in tackling nonconvex problems. In this talk, we explore how quantum dynamics can be exploited to develop novel quantum optimization algorithms. Specifically, we introduce Quantum Hamiltonian Descent (QHD), a quantum optimization algorithm inspired by the connection between accelerated gradient descent and Hamiltonian mechanics. We will discuss the theoretical properties of QHD, including its global convergence in nonlinear and nonconvex optimization, along with strong empirical and theoretical evidence of its advantage over classical algorithms. Additionally, we will present an open-source implementation of the algorithm (named QHDOPT) and demonstrate its real-world applications.

Speakers
RT

Reuben Tate

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
SF

Sean Feeney

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 108 3501 Trousdale Pkwy, 108, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5V: Newton-ish and Higher-Order Methods
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Newton-ish and Higher-Order Methods
Chair: Nick Tsipinakis
Cluster: nan

Talk 1: Multilevel Regularized Newton Methods with Fast Convergence Rates
Speaker: Nick Tsipinakis
Abstract: We introduce new multilevel methods for solving large-scale unconstrained optimization problems. Specifically, the philosophy of multilevel methods is applied to Newton-type methods that regularize the Newton sub-problem using second-order information from a coarse (low dimensional) sub-problem. The new regularized multilevel methods provably converge from any initialization point and enjoy faster convergence rates than Gradient Descent. In particular, for arbitrary functions with Lipschitz continuous Hessians, we show that their convergence rate interpolates between the rate of Gradient Descent and that of the cubic Newton method. If, additionally, the objective function is assumed to be convex, then the proposed method converges with the fast $\mathcal{O}(k^{-2})$ rate. Hence, since the updates are generated using a coarse model in low dimensions, the theoretical results of this paper significantly speed-up the convergence of Newton-type or preconditioned gradient methods in practical applications. Preliminary numerical results suggest that the proposed multilevel algorithms are significantly faster than current state-of-the-art methods. [1] K. Mishchenko, Regularized Newton method with global convergence, SIAM Journal on Optimization, 33 (2023), pp. 1440–1462. [2] N. Doikov and Y. Nesterov, Gradient regularization of Newton method with Bregman dis-tances, Mathematical programming, 204 (2024), pp. 1–25. [3] N. Tsipinakis and P. Parpas, A multilevel method for self-concordant minimization, Journal of Optimization Theory and Applications, (2024), pp. 1–51.

Talk 2: First-ish Order Methods: Hessian-aware Scalings of Gradient Descent
Speaker: Oscar Smee
Abstract: Gradient descent is the primary workhorse for optimizing large-scale problems in machine learning. However, its performance is highly sensitive to the choice of the learning rate. A key limitation of gradient descent is its lack of natural scaling, which often necessitates expensive line searches or heuristic tuning to determine an appropriate step size. In this paper, we address this limitation by incorporating Hessian information to scale the gradient direction. By accounting for the curvature of the function along the gradient, our adaptive, Hessian-aware scaling method ensures a local unit step size guarantee, even in nonconvex settings. Near a local minimum that satisfies the second-order sufficient conditions, our approach achieves linear convergence with a unit step size. We show that our method converges globally under a significantly weaker version of the standard Lipschitz gradient smoothness assumption. Even when Hessian information is inexact, the local unit step size guarantee and global convergence properties remain valid under mild conditions. Finally, we validate our theoretical results empirically on a range of convex and nonconvex machine learning tasks, showcasing the effectiveness of the approach. Preprint: https://arxiv.org/abs/2502.03701

Speakers
NT

Nick Tsipinakis

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
OS

Oscar Smee

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 110 3501 Trousdale Pkwy, 110, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5W: Systems of Quadratic Equations/Inequalities
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Systems of Quadratic Equations/Inequalities
Chair: Ruey-Lin Sheu
Cluster: nan

Talk 1: On the implementation issue of the non-homogeneous strict Finsler and the non-homogeneous Calabi theorems
Speaker: Min-Chi Wang
Abstract: Let f(x) and g(x) be two real quadratic functions defined on ℝⁿ. The existence of solutions to the system of (in)equality constraints [f(x) = 0, g(x) = 0] and [f(x) = 0, g(x) ≤ 0] has rarely been studied in the literature. Recently, new results have emerged, called the non-homogeneous strict Finsler Lemma and the non-homogeneous Calabi Theorem, which provide necessary and sufficient conditions for the solvability of the above two systems of (in)equality constraints. This talk aims to simplify the conditions to facilitate easier implementation. Numerical results show that our proposed method is efficient in computation.

Talk 2: A QP1QC approach for deciding whether or not two quadratic surfaces intersect
Speaker: Ting-Tsen Lin
Abstract: Given two $n$-variate quadratic functions $f(x)=x^TAx+2a^Tx+a_0, g(x)=x^TBx+2b^Tx+b_0$, we are interested in knowing whether or not the two hypersurfaces $\{x\in \mathbb{R}^n: f(x)=0\}$ and $\{x\in \mathbb{R}^n: g(x)=0\}$ intersect with each other. There are two aspects to look at this problem. In one respect, the famous Finsler-Calabi theorem (1936-1964) asserts that, if $n\ge3$ and $f,g$ are quadratic forms, $f=g=0$ has no common solution other than the trivial one $x=0$ if and only if there exists a positive definite matrix pencil $\alpha A+\beta B\succ0.$ The result is in general not true for non-homogeneous quadratic functions. On the other hand, Levin (c. late 1970) tried to directly solve the intersection curve of $\{x\in \mathbb{R}^n: f(x)=0\}$ and $\{x\in \mathbb{R}^n: g(x)=0\},$ but it turned out to be way too ambitious. In this paper, we show that, by incorporating with the information about the unboundedness and the un-attainability of several (at most 4) quadratic programming problems with one single quadratic constraint (QP1QC), the answer as to whether or not $f(x)=x^TAx+2a^Tx+a_0=0$ and $g(x)=x^TBx+2b^Tx+b_0=0$ intersect can be successfully determined. References: E. Calabi, Linear systems of real quadratic forms, Proceedings of the American Mathematical Society, 15 (1964), pp. 844–846. P. Finsler, Über das Vorkommen definiter und semidefiniter Formen in Scharen quadratischer Formen, Commentarii Mathematici Helvetici, 9 (1936), pp. 188–192. J. Z. Levin, A parametric algorithm for drawing pictures of solid objects composed of quadric surfaces, Communications of the ACM, 19 (1976), pp. 555–563. J. Z. Levin, Mathematical models for determining the intersections of quadric surfaces, Computer Graphics and Image Processing, 11 (1979), pp. 73–87.

Talk 3: Last two pieces of puzzle for un-solvability of a system of two quadratic (in)equalities
Speaker: Ruey-Lin Sheu
Abstract: Given two quadratic functions f(x) = x^{T}Ax+2a^{T}x+a_0 and g(x) = x^{T}Bx+ 2b^{T }x+b_0, each associated with either the strict inequality (< 0); non-strict inequality (≤ 0); or the equality (= 0), it is a fundamental question to ask whether or not the joint system {x ∈ R^n: f(x) ⋆ 0} and {x ∈ R^n: g(x)#0}, where ⋆ and # can be any of {
Speakers
MW

Min-Chi Wang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
TL

Ting-Tsen Lin

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
RS

Ruey-Lin Sheu

Professor, National Cheng Kung University
Name: Dr. Slothington "Slow Convergence" McNapfaceTitle: Distinguished Professor of Continuous Optimization & Energy MinimizationAffiliation: The Lush Canopy Institute of Sluggish AlgorithmsBio:Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 112 3501 Trousdale Pkwy, 112, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5X: Modeling, Solvers, and Quantum
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Session: Modeling, Solvers, and Quantum
Chair: Giacomo Nannicini
Cluster: nan

Talk 1: Making Some Smooth Problems with Derivative Jumps Easier to Express
Speaker: David Gay
Abstract: Some optimization problems involve an objective or constraints with derivative jumps. The jumps may be due to absolute values or the minimum or maximum of two or more expressions or to the piecewise linearization of a nonlinear function. There are multiple ways to deal with such situations. For example, Griewank and Walther propose using regularization supplied by overloading the relevant expressions. Of interest in this talk is using binary variables to decide which pieces are currently relevant. In the AMPL modeling language, this is implicit when one uses the piecewise-linear notation. This talk will present examples and discuss plans to automatically convert abs(...), min(...), and max}(...) to piecewise-linear forms, enabling efficiently solving by various solvers.

Talk 2: The quantum central path method
Speaker: Giacomo Nannicini
Abstract: Abstract: We propose a new quantum algorithm for solving linear optimization problems by quantum-mechanical simulation of the central path. While interior point methods follow the central path with an iterative algorithm that works with successive linearizations of the perturbed KKT conditions, we perform a single simulation working directly with the nonlinear complementarity equations. Our approach is inspired by the Newton barrier flow studied in the 80s by several authors, and goes beyond those studies by giving a full description of a quantum algorithm that simulates the corresponding dynamical system. The algorithm has a worst case complexity with favorable scaling in the problem dimension compared to state-of-the-art methods, but worse scaling in the precision and the size of the solution. We hope that with further advancements, this method could pave the way for an end-to-end quantum speedup for linear optimization.

Talk 3: Modeling and algorithmic framework for complex optimization problems and equilibrium problems
Speaker: Olivier Huber
Abstract: Algebraic modeling languages (AML) have shaped the problem structure considered in numerical optimization, and indirectly have an impact on the instances coming from applications and the ones considered by solvers. However, problems not fitting the classical AML format are abundantly found in applications. For instance, nonsmoothness commonly arises in (multistage) stochastic programming problems with coherent risk measures. Bilevel or multilevel models feature multiple optimization problems, and so do Nash equilibrium problems. A given application can feature a combination of the above challenge. Consider the decentralized electricity market, where the weather affect the production level of some producers. This leads to a Nash equilibrium problem with multistage risk-averse agents. To capture these challenging but structured models, we propose a modeling framework based on a directed acyclic graph (DAG). The nodes are of two type: the first one subsumes classical optimization problems and variational inequalities, while the second one indicates a Nash behavior among its children nodes. The directed arcs capture the interactions between optimization problems. This modeling approach is implemented via annotation and extends an AML modeling power. Model transformations are defined over the DAG structure to convert part or all of the model into a form amenable to computations by existing solvers. Many instances captured by this modeling paradigm fall outside of the scope for robust solvers, an algorithmic framework to solve these instances, for instance by decomposition or approximation, is presented. An implementation of these ideas is present in ReSHOP, a reformulation solver for hierarchical optimization problems.

Speakers
DG

David Gay

Name: Dr. David M GayTitle: chief scientistAffiliation: AMPL OptimizationBio:Math major at the University of Michigan with senior year (1970-71) at the Albert-Ludwigs-Universitaet, Freiburg im Breisgau, Germany. Grad student at Cornell 1971-74, PhD in Computer Science (1975). Asst... Read More →
OH

Olivier Huber

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Taper Hall (THH) 215 3501 Trousdale Pkwy, 215, Los Angeles, CA 90089

1:15pm PDT

Parallel Sessions 5Y
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Tuesday July 22, 2025 1:15pm - 2:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 200 3518 Trousdale Pkwy, 200, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6A: Adaptive Stochastic Gradient Methods
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Adaptive Stochastic Gradient Methods
Chair: Lin Xiao
Cluster: Nonlinear Optimization

Talk 1: The Road Less Scheduled
Speaker: Aaron Defazio
Abstract: Schedule-Free learning algorithms allow for the training of models in an any-time fashion, without compromising on speed, memory or final test metrics. I will dive into the details of how Schedule-Free learning works and show how it provides further quality-of-life improvements to practitioners, and provide details of our winning entry to the AlgoPerf algorithmic efficiency optimization challenge that used Schedule-Free AdamW.

Talk 2: Analyzing AdaGrad Under Anisotropic Smoothness Assumptions
Speaker: Yuxing Liu
Abstract: Adaptive gradient methods have demonstrated remarkable success for training large-scale deep neural networks. However, the theoretical understanding of these methods, particularly in the large batch size regime (which is commonly used in practice), remains limited. In this talk, we aim to address this gap by introducing a generalized anisotropic smoothness assumption that better reflects the behavior of modern neural network training. Our theoretical analysis reveals that AdaGrad achieves provably faster convergence compared to standard gradient methods, even when large batch sizes are employed. These results provide valuable theoretical insights into the practical efficacy of adaptive gradient methods.

Talk 3: A Novel Approach to Loss Landscape Characterization without Over-Parametrization
Speaker: Antonio Orvieto
Abstract: Modern machine learning heavily depends on the effectiveness of optimization techniques. While deep learning models have achieved remarkable empirical results in training, their theoretical underpinnings remain somewhat elusive. Ensuring the convergence of optimization methods requires imposing specific structures on the objective function, which often do not hold in practice. One prominent example is the widely recognized Polyak-Lojasiewicz (PL) inequality, which has garnered considerable attention in recent years. However, validating such assumptions for deep neural networks entails substantial and often impractical levels of over-parametrization. In order to address this limitation, we propose a novel class of functions that can characterize the loss landscape of modern deep models without requiring extensive over-parametrization and can also include saddle points. Crucially, we prove that gradient-based optimizers possess theoretical guarantees of convergence under this assumption. Finally, we validate the soundness of our assumption through both theoretical analysis and empirical experimentation across a diverse range of deep learning models.

Speakers
LX

Lin Xiao

Lin Xiao is a Research Scientist at Facebook AI Research (FAIR) in Seattle, Washington. He received BE from Beijing University of Aeronautics and Astronautics (Beihang University) and PhD from Stanford University, and was a postdoctoral fellow in the Center for the Mathematics of... Read More →
AD

Aaron Defazio

Research Scientist, Meta Platforms, Inc.
Aaron Defazio is a Research Scientist at Meta on the Fundamental AI Research Team, specializing in the field of optimization algorithms for machine learning. Aaron holds a PhD in Computer Science from ANU (Australian National University) and has a rich background in research, having... Read More →
YL

Yuxing Liu

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
AO

Antonio Orvieto

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 101 3501 Trousdale Pkwy, 101, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6B: Machine Learning Algorithms
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Machine Learning Algorithms
Chair: Mark Schmidt
Cluster: Optimization Applications (Communication, Energy, Health, ML, ...)

Talk 1: Leveraging Variable Sparsity to Refine Pareto Stationarity in Multi-Objective Optimization
Speaker: Yaoliang Yu
Abstract: Gradient-based multi-objective optimization (MOO) is essential in modern machine learning, with applications in e.g., multi-task learning, federated learning, algorithmic fairness and reinforcement learning. In this work, we first reveal some limitations of Pareto stationarity, a widely accepted first-order condition for Pareto optimality, in the presence of sparse function-variable structures. Next, to account for such sparsity, we propose a novel solution concept termed Refined Pareto Stationarity (RPS), which we prove is always sandwiched between Pareto optimality and Pareto stationarity. We give an efficient partitioning algorithm to automatically mine the function-variable dependency and substantially trim non-optimal Pareto stationary solutions. Then, we show that gradient-based descent algorithms in MOO can be enhanced with our refined partitioning. In particular, we propose Multiple Gradient Descent Algorithm with Refined Partition (RP-MGDA) as an example method that converges to RPS, while still enjoying a similar per-step complexity and convergence rate. Lastly, we validate our approach through experiments on both synthetic examples and realistic application scenarios where distinct function-variable dependency structures appear. Our results highlight the importance of exploiting function-variable structure in gradient-based MOO and provide a seamless enhancement to existing approaches.

Talk 2: High-dimensional Optimization with Applications to Compute-Optimal Neural Scaling Laws
Speaker: Courtney Paquette
Abstract: Given the massive scale of modern ML models, we now only get a single shot to train them effectively. This restricts our ability to test multiple architectures and hyper-parameter configurations. Instead, we need to understand how these models scale, allowing us to experiment with smaller problems and then apply those insights to larger-scale models. In this talk, I will present a framework for analyzing scaling laws in stochastic learning algorithms using a power-law random features model, leveraging high-dimensional probability and random matrix theory. I will then use this scaling law to address the compute-optimal question: How should we choose model size and hyper-parameters to achieve the best possible performance in the most compute-efficient manner?

Talk 3: A Robustness Metric for Distribution Shifts
Speaker: John Duchi
Abstract: We revisit the stability of optimizers in statistical estimation and stochastic optimization problems, but instead of providing guarantees on the stability of the minimizers themselves, we investigate what shifts to the underlying data-generating process perturb solutions the most. To do so, we develop some new mathematical tools for stability analyses, with guarantees beyond typical differentiable problems. We also make connections with statistical hypothesis testing and discovery, showing how these new results provide certificates of validity---or potential invalidity---of statistical estimate.

Speakers
MS

Mark Schmidt

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
YY

Yaoliang Yu

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
CP

Courtney Paquette

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JD

John Duchi

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 201 3501 Trousdale Pkwy, 201, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6C: AI Meets Optimization (Part 1)
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: AI Meets Optimization (Part 1)
Chair: Wotao Yin
Cluster: Optimization for Emerging Technologies (LLMs, Quantum Computing, ...)

Talk 1: TBD
Speaker: Bartolomeo Stellato
Abstract: TBD

Talk 2: Differentiating through Solutions to Optimization Problems in Decision-Focused Learning
Speaker: Howard Heaton
Abstract: Many real-world problems can be framed as optimization problems, for which well-established algorithms exist. However, these problems often involve key parameters that are not directly observed. Instead, we typically have access to data that is correlated with these parameters, though the relationships are complex and difficult to describe explicitly. This challenge motivates the integration of machine learning with optimization: using machine learning to predict the hidden parameters and optimization to solve the resultant problem. This integration is known as decision-focused learning. In this talk, I will introduce decision-focused learning, with a particular focus on differentiating through solutions to optimization problems and recent advances in effectively scaling these computations.

Talk 3: TBD
Speaker: Ferdinando Fioretto
Abstract: TBD

Speakers
BS

Bartolomeo Stellato

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
FF

Ferdinando Fioretto

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 208 3501 Trousdale Pkwy, 208, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6D: Special Session in Honor of Suvrajeet Sen: Nonsmooth Methods in Stochastic Programming
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Special Session in Honor of Suvrajeet Sen: Nonsmooth Methods in Stochastic Programming
Chair: Junyi Liu
Cluster: Optimization Under Uncertainty and Data-driven Optimization

Talk 1: Improving Dimension Dependence in Zeroth-Order Schemes via Exponentially Shifted Gaussian Smoothing
Speaker: Uday Shanbhag
Abstract: Smoothing-enabled zeroth-order (ZO) methods for nonsmooth convex stochastic optimization have assumed increasing relevance. A shortcoming of such schemes is the dimension dependence in the complexity guarantees, a concern that impedes truly large-scale implementations. We develop a novel exponentially- shifted Gaussian smoothing (esGS) gradient estimator by leveraging a simple change-of-variable argument. The moment bounds of the (esGS) estimator are characterized by a muted dependence on dimension. When the (esGS) estimator is incorporated within a ZO framework, the resulting iteration complexity bounds are reduced to O(n\epsilon^{-2}) from O(n^2 \epsilon^{-2}), the latter being the best available for the existing two-point estimator with Gaussian smoothing. This is joint work with Mingrui Wang and Prakash Chakraborty.

Talk 2: Decomposition-based algorithm for infinite horizon stochastic programs on finite-state Markov chains
Speaker: Shuotao Diao
Abstract: Infinite horizon stochastic programs on finite-state Markov chains can be expressed as a semi-contractive model. In this work, we provide a unifying view of the conditions ensuring the existence of fixed points of the infinite horizon stochastic programs. We further develop a decomposition-based method to solve the infinite horizon stochastic programs with convergence guarantee.

Talk 3: Adaptive Sampling-based Nonconvex and Nonsmooth approaches for Stochastic Programs with Implicitly Decision-dependent Uncertainty
Speaker: Junyi Liu
Abstract: We consider a class of stochastic programming problems where the implicitly decision-dependent random variable follows a nonparametric regression model with heteroskedastic error. We develop an adaptive sampling-based algorithm that integrates the simulation scheme and statistical estimates to construct sampling-based surrogate functions in a way that the simulation process is guided by the algorithmic procedure. We establish the nonasymptotic convergence analysis in terms of $(\epsilon, \delta)$-nearly stationarity in expectation under variable proximal parameters and batch sizes that leads to superior convergence rate. Furthermore, we show that the proposed adaptive simulation scheme embedded in the sampling-based algorithm leads to better error control of sampling-based surrogate functions and thus enhance the stability and efficiency of the sampling-based algorithm, which are further evidenced by numerical results.

Speakers
US

Uday Shanbhag

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
SD

Shuotao Diao

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JL

Junyi Liu

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 210 3501 Trousdale Pkwy, 210, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6E: Recent advances in algorithms for large-scale optimization (II)
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Recent advances in algorithms for large-scale optimization (II)
Chair: Xudong Li
Cluster: Computational Software

Talk 1: Alternating minimization for distributionally robust principal component pursuit
Speaker: Xudong Li
Abstract: Recently, the square root principal component pursuit (SRPCP) model has garnered significant research interest. It is shown in the literature that the SRPCP model guarantees robust matrix recovery with a universal, constant penalty parameter. While its statistical advantages are well-studied, the computational aspects remain unexplored. In this talk, we focus on developing efficient algorithms for solving the SRPCP problem. Specifically, we propose a tuning-free alternating minimization (AM) algorithm, where each iteration involves subproblems with semi-closed updating rules. Additionally, we introduce techniques based on the variational formulation of the nuclear norm and BM decomposition to further accelerate the AM method. Extensive numerical experiments confirm the efficiency and robustness of our algorithms.

Talk 2: A new dual semismooth Newton method for polyhedral projections
Speaker: Chao Ding
Abstract: We propose a dual semismooth Newton method for computing the orthogonal projection onto a given polyhedron, a fundamental optimization problem that serves as a critical building block for numerous important applications, e.g., financial risk management, statistics, and machine learning. Classical semismooth Newton methods typically depend on subgradient regularity assumptions for achieving local superlinear or quadratic convergence. Our approach, however, marks a significant breakthrough by demonstrating that it is always possible to identify a point where the existence of a nonsingular generalized Jacobian is guaranteed, regardless of any regularity conditions. Furthermore, we explain this phenomenon and its relationship with the weak strict Robinson constraint qualification (W-SRCQ) from the perspective of variational analysis. Building on this theoretical advancement, we develop an inexact semismooth Newton method with superlinear convergence for solving the polyhedral projection problem.

Talk 3: A Proximal DC Algorithm for Sample Average Approximation of Chance Constrained Programming
Speaker: Rujun Jiang
Abstract: Chance constrained programming (CCP) refers to a type of optimization problem with uncertain constraints that are satisfied with at least a prescribed probability level. In this work, we study the sample average approximation (SAA) method for chance constraints, which is an important approach to CCP in the data-driven setting where only a sample of multiple realizations of the random vector in the constraints is available. The SAA method approximates the underlying distribution with an empirical distribution over the available sample. Assuming that the functions in the chance constraints are all convex, we reformulate the SAA of chance constraints into a difference-of-convex (DC) form. Additionally, by assuming the objective function is also a DC function, we obtain a DC constrained DC program. To solve this reformulation, we propose a proximal DC algorithm and show that the subproblems of the algorithm are suitable for off-the-shelf solvers in some scenarios. Moreover, we not only prove the subsequential and sequential convergence of the proposed algorithm but also derive the iteration complexity for finding an approximate Karush-Kuhn-Tucker point. To support and complement our theoretical development, we show via numerical experiments that our proposed approach is competitive with a host of existing approaches.

Speakers
XL

Xudong Li

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
CD

Chao Ding

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
avatar for Rujun Jiang

Rujun Jiang

Rujun Jiang is currently an associate professor at School of Data Science, Fudan University. He received his Bachelor degree in Mathematics in 2012 from University of Science and Technology of China. He then obtained his PhD degree in 2016 in The Chinese University of Hong Kong under... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 212 3501 Trousdale Pkwy, 212, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6F: Optimization for improving privacy and alignment for LLMs
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Optimization for improving privacy and alignment for LLMs
Chair: Mingyi Hong
Cluster: Optimization for Emerging Technologies (LLMs, Quantum Computing, ...)

Talk 1: multi-step preference optimization via two-player markov games
Speaker: Volkan Cevher
Abstract: Reinforcement Learning from Human Feedback (RLHF) has been highly success- ful in aligning large language models with human preferences. While prevalent methods like DPO have demonstrated strong performance, they frame interactions with the language model as a bandit problem, which limits their applicability in real-world scenarios where multi-turn conversations are common. Additionally, DPO relies on the Bradley-Terry model assumption, which does not adequately capture the non-transitive nature of human preferences. In this paper, we address these challenges by modeling the alignment problem as a two-player constant-sum Markov game, where each player seeks to maximize their winning rate against the other across all steps of the conversation. Our approach Multi-step Preference Optimization (MPO) is built upon the natural actor-critic framework (Peters & Schaal, 2008). We further develop MPO based on the optimistic online gradient descent algorithm (Rakhlin & Sridharan, 2013; Joulani et al., 2017). Theoretically, we provide a rigorous analysis for both algorithms on convergence and show that 0MPO requires O(ϵ−1) policy updates to converge to an ϵ-approximate Nash equi- librium. We also validate the effectiveness of our method through experiments on the multi-turn conversations dataset in MT-bench-101.

Talk 2: Getting More Juice Out of the SFT Data: Reward Learning from Human Demonstration via Bilevel Optimization Improves LLM Alignment
Speaker: Mingyi Hong
Abstract: Aligning human preference and value is an important requirement for contemporary foundation models. State-of-the-art techniques such as Reinforcement Learning from Human Feedback (RLHF) often consist of two stages: 1) supervised fine-tuning (SFT), where the model is fine-tuned by learning from human demonstration data; 2) Preference learning, where preference data is used to learn a reward model, which is in turn used by a reinforcement learning (RL) step to fine-tune the model. Such reward model serves as a proxy to human preference, and it is critical to guide the RL step towards improving the model quality. In this work, we argue that the SFT stage significantly benefits from learning a reward model as well. Instead of using the human demonstration data directly via supervised learning, we propose to leverage an Inverse Reinforcement Learning (IRL) and bilevel optimization technique to simultaneously build an reward model and a policy model. This approach leads to new SFT algorithms that are not only efficient to implement, but are robust to the presence of low-quality supervised learning data. Moreover, we discover a connection between the proposed IRL based approach, and a recent line of works called Self-Play Fine-tune (SPIN). Theoretically, we show that the proposed algorithms converge to the stationary solutions of the IRL problem. Empirically, we align 1B and 7B models using proposed methods and evaluate them on a reward benchmark model and the HuggingFace Open LLM Leaderboard. The proposed methods show significant performance improvement over existing SFT approaches. Our results indicate that it is beneficial to leverage reward learning throughout the entire alignment process.

Talk 3: Pre-training Differentially Private Models with Limited Public Data
Speaker: Xinwei Zhang
Abstract: The superior performance of large foundation models relies on the use of massive amounts of high-quality data, which often contain sensitive, private, and copyrighted material that requires formal protection. While differential privacy (DP) is a prominent method to gauge the degree of security provided to the models, its application is commonly limited to the model fine-tuning stage due to the performance degradation when DP is applied during the pre-training stage. Consequently, DP is yet incapable of protecting a substantial portion of the data used during the initial pre-training process. In this work, we provide a theoretical understanding of the efficacy of DP training by analyzing the improvement of per-iteration loss through the lens of the Hessian matrix for large neural networks. We make a key observation that DP optimizers' performance degradation can be significantly mitigated by the use of limited public data, which leads to a novel DP continual pre-training strategy. Empirically, using only 10\% of public data, our strategy can achieve DP accuracy of 41.5% on ImageNet-21k (with =8), as well as non-DP accuracy of 55.7% and 60.0% on downstream tasks Places365 and iNaturalist-2021, respectively, on par with state-of-the-art standard pre-training and substantially outperforming existing DP pre-trained models.

Speakers
VC

Volkan Cevher

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
MH

Mingyi Hong

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
XZ

Xinwei Zhang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 156 3518 Trousdale Pkwy, 156, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6G: Advances in Data-driven Optimization
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Advances in Data-driven Optimization
Chair: Rui Gao & Daniel Kuhn
Cluster: Optimization Under Uncertainty and Data-driven Optimization

Talk 1: A Class of Interpretable and Decomposable Multi-period Convex Risk Measures
Speaker: Luhao Zhang
Abstract: Multi-period risk measures evaluate the risk of a stochastic process by assigning it a scalar value. A desirable property of these measures is dynamic decomposition, which allows the risk evaluation to be expressed as a dynamic program. However, many widely used risk measures, such as Conditional Value-at-Risk, do not possess this property. In this work, we introduce a novel class of multi-period convex risk measures that do admit dynamic decomposition. Our proposed risk measure evaluates the worst-case expectation of a random outcome across all possible stochastic processes, penalized by their deviations from a nominal process in terms of both the likelihood ratio and the outcome. We show that this risk measure can be reformulated as a dynamic program, where, at each time period, it assesses the worst-case expectation of future costs, adjusting by reweighting and relocating the conditional nominal distribution. This recursive structure enables more efficient computation and clearer interpretation of risk over multiple periods.

Talk 2: An MILP-Based Solution Scheme for Factored and Robust Factored Markov Decision Processes
Speaker: Man-Chung Yue
Abstract: Factored Markov decision processes (MDPs) are a prominent paradigm within the artificial intelligence community for modeling and solving large-scale MDPs whose rewards and dynamics decompose into smaller, loosely interacting components. Through the use of dynamic Bayesian networks and context-specific independence, factored MDPs can achieve an exponential reduction in the state space of an MDP and thus scale to problem sizes that are beyond the reach of classical MDP algorithms. However, factored MDPs are typically solved using custom-designed algorithms that can require meticulous implementations and considerable fine-tuning. In this paper, we propose a mathematical programming approach to solving factored MDPs. In contrast to existing solution schemes, our approach leverages off-the-shelf solvers, which allows for a streamlined implementation and maintenance; it effectively capitalizes on the factored structure present in both state and action spaces; and it readily extends to the largely unexplored class of robust factored MDPs, whose transition kernels are only known to reside in a pre-specified ambiguity set. Our numerical experiments demonstrate the potential of our approach.

Talk 3: A Deep Learning Approach to Multistage Stochastic Programming
Speaker: Rui Gao
Abstract: Multistage stochastic programming problems are challenging due to the curse of dimensionality. We introduce a practical algorithm for solving multistage stochastic programming in high dimensions, leveraging neural networks to parameterize the policy. The proposed algorithm demonstrates effectiveness in terms of both accuracy and speed across a variety of problems.

Speakers
LZ

Luhao Zhang

Name: Dr. Luhao ZhangTitle: Assistant ProfessorAffiliation: Deparment of Applied Mathematics and Statistics, Johns Hopkins University
RG

Rui Gao

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 114 3501 Trousdale Pkwy, 114, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6H: Semidefinite programming and applications
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Semidefinite programming and applications
Chair: Georgina Hall
Cluster: Conic and Semidefinite Optimization

Talk 1: A Matrix Generalization of the Goemans-Williamson Algorithm for Orthogonality Constrained Optimization Problems
Speaker: Ryan Cory-Wright
Abstract: A central research question in optimization concerns developing algorithms that solve non-convex problems to provable near optimality in a practically tractable amount of time. One popular two-step methodology called ``relax-and-round'' successfully addresses many well-behaved non-convex problems. In 1995, Goemans and Williamson proposed a polynomial-time relax-and-round algorithm for approximately solving Max-Cut problems by rounding their semidefinite relaxations. In this work, we propose a matrix generalization of their approach which applies to orthogonal matrices where U^\top U=I, as opposed to binary variables where z^2=1, and study the quality of this approach in both theory and practice. Our approach has applications in low-rank matrix completion and sparse PCA with multiple PC problems, among others.

Talk 2: Generalized Ellipsoids
Speaker: Cemil Dibek
Abstract: Ellipsoids are fundamental in applied and computational mathematics, featuring in optimization, control, convex geometry, and statistics due to their geometric and computational properties. In this talk, we introduce a new family of symmetric convex bodies called generalized ellipsoids of degree d (GE-ds), with ellipsoids corresponding to the case of d=0. Generalized ellipsoids (GEs) retain many geometric, algebraic, and algorithmic properties of ellipsoids, while also being tractable to search for and optimize over. We show that one can search for GEs of a given degree by solving a semidefinite program whose size grows only linearly with dimension, and that every GE has a semidefinite representation whose size depends linearly on both its dimension and degree. In terms of expressiveness, we demonstrate that every symmetric full-dimensional polytope and every intersection of co-centered ellipsoids can be represented exactly as a GE-d. Using this result, we show that every symmetric convex body can be approximated arbitrarily well by a GE-d. We also present applications of GEs in areas such as portfolio optimization, stability analysis of switched linear systems, and robust optimization.

Talk 3: TBD
Speaker: Georgina Hall
Abstract: TBD

Speakers
RC

Ryan Cory-Wright

Assistant Professor of Analytics and Operations, Imperial College London
I am an Assistant Professor in the Analytics and Operations Group at Imperial College Business School (ICBS) since July 2023, affiliated with the Imperial-X initiative on interdisciplinary AI/ML.My research focuses on optimization, machine learning, statistics, and their application in business analytics. I am particularly interested in broadening the scope of optimization to address practical problems that current methods cannot solve... Read More →
CD

Cemil Dibek

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
GH

Georgina Hall

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 116 3501 Trousdale Pkwy, 116, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6I: Advances in Conic Optimization
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Advances in Conic Optimization
Chair: Timotej Hrga
Cluster: Conic and Semidefinite Optimization

Talk 1: Completely positive reformulations for sparse quadratic optimization
Speaker: Bo Peng
Abstract: In this talk, we explore the completely positive reformulation of quadratic optimization problems involving the $\ell_0$ quasinorm, which is well-known for promoting solution sparsity. Compared to existing approaches in the literature, we propose novel, more compact relaxations that are proven to be exact under milder conditions. Furthermore, we demonstrate the exactness of a decomposed completely positive relaxation by leveraging inherent sparse patterns of the model. A numerical study is conducted to compare double nonnegative relaxations derived from these reformulations. Extensive numerical experiments illustrate the quality of the resulting bounds while maintaining tractability and scalability.

Talk 2: Connectivity via convexity: Bounds on the edge expansion in graphs
Speaker: Timotej Hrga
Abstract: Convexification techniques have gained increasing interest over the past decades. In this work, we apply a recently developed convexification technique for fractional programs by He, Liu and Tawarmalani (2024) to the problem of determining the edge expansion of a graph. Computing the edge expansion of a graph is a well-known, difficult combinatorial problem that seeks to partition the graph into two sets such that a fractional objective function is minimized. We give a formulation of the edge expansion as a completely positive pro- gram and propose a relaxation as a doubly non-negative program, further strengthened by cutting planes. Additionally, we develop an augmented Lagrangian algorithm to solve the doubly non-negative program, obtaining lower bounds on the edge expansion. Numerical results confirm that this relaxation yields strong bounds and is computationally efficient, even for graphs with several hundred vertices.

Talk 3: Facial Reduction in BiqCrunch
Speaker: Ian Sugrue
Abstract: BiqCrunch is a binary quadratic solver that uses a semidefinite bounding procedure with a branch-and-bound framework to get exact solutions to combinatorial optimization problems. Since its latest official update, BiqCrunch has undergone further development, most notably the inclusion of the dimensionality reduction process known as facial reduction. We will discuss the details of how facial reduction is implemented, the numerical results that support this inclusion, as well as future development plans for BiqCrunch.

Speakers
BP

Bo Peng

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
TH

Timotej Hrga

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 100 3518 Trousdale Pkwy, 100, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6J: Continuous and discreate optimization solving real problems
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Continuous and discreate optimization solving real problems
Chair: Julio C. Góez
Cluster: Interplay Between Continuous and Discrete Optimization

Talk 1: Active constraint indicators in interior point algorithms for conic optimization
Speaker: Alexander Guldemond
Abstract: In conic optimization, determining which constraints are active at a given solution can be crucial information for the user. In this talk, we discuss an approach used by Mosek to predict active conic constraints within the framework of interior point algorithms. It is well known that in the case of the nonnegative orthant, active sets can be predicted by comparing $dx_i / x_i$ to $ds_i / s_i$ for each primal dual variable pair. We extend this technique to generic conic constraints, including non-symmetric cones. At each iteration, we leverage information from both the current iterate and the computed search direction to identify likely active conic constraints. We will present theoretical insights, computational experiments, and discuss the implications for large-scale optimization problems.

Talk 2: A Mixed-Integer Conic Program for the Multi-Pursuer Moving-Target Traveling Salesmen Problem
Speaker: Allen George Philip
Abstract: The Moving-Target Traveling Salesman Problem (MT-TSP) seeks to find a shortest path for a pursuer, that starts at a depot, visits a set of moving targets, and returns to the depot. This problem is motivated by various practical applications such as monitoring and surveillance, intercepting hostile UAVs, and motion planning for industrial robots. We consider the Multi-Pursuer Moving-Target Traveling Salesman Problem (MP-MT-TSP) which generalizes the MT-TSP to include multiple pursuers. We introduce a new Mixed-Integer Conic Program (MICP) for MP-MT-TSP where targets move along piecewise-linear paths. We compare our formulation with the current state-of-the-art MICP for the MP-MT-TSP, and present experimental results demonstrating significant improvements over the state-of-theart in both runtime and optimality ga

Talk 3: Sparse Polynomial Optimization for Water Networks
Speaker: Olga Heijmans-Kuryatnikova
Abstract: In this work we exploit sparsity in polynomial optimization for the valve setting problem in water networks. The valve setting problem consists of many small subproblems involving few variables and monomials and connected with sparse linking constraints. The problem exhibits correlative sparsity (subsets of variables appearing separately in constraints), term sparsity (few monomials present in the problem), and ideal sparsity (size reduction due to equality constraints). We suggest a new simple SDP relaxation that uses all types of sparsity to reach a trade-off between the bound quality and running times. We compare it with the existing sparse SDP relaxations and report numerical results on four water networks ranging in size from 4 to about 2000 nodes. The approach extends to electricity and gas network problems.

Speakers
JC

Julio C Góez

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
AG

Alexander Guldemond

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
AG

Allen George Philip

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
OH

Olga Heijmans-Kuryatnikova

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 102 3501 Trousdale Pkwy, 102, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6K: Applications To Signal Processing
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Applications To Signal Processing
Chair: Robert Bassett
Cluster: Optimization Applications (Communication, Energy, Health, ML, ...)

Talk 1: Stochastic Optimal Search with Signals
Speaker: Jefferson Huang
Abstract: We consider the problem of optimally searching for a target located at one of the nodes of a network. On each time step, the searcher receives a (possibly unreliable) signal indicating where the target is on the network. Starting with the case of searching on a line, we explore the structure of search policies that optimally make use of the signals. 

Talk 2: Sparse signal reconstruction for over-dispersed low-photon count imaging
Speaker: Roummel Marcia
Abstract: Low-photon count imaging is typically modeled by Poisson statistics. This discrete probability distribution model assumes that the mean and variance of a signal are equal. In the presence of greater variability in a dataset than what is expected, the negative binomial distribution is a suitable over-dispersed alternative to the Poisson distribution. In this work, we present an optimization framework for reconstructing sparse signals in these over-dispersed low-count settings.

Talk 3: Decentralized Sensor Network Localization via Customized Proximal Splitting
Speaker: Peter Barkley
Abstract: We apply recent advances in the design of custom proximal splitting algorithms to build a decentralized algorithm for solving the node-based SDP relaxation of the sensor network localization problem using noisy distance data. We implement the algorithm using MPI. We then explore the performance of various algorithm design options, and compare the decentralized algorithm with other decentralized approaches, including a graph-based decentralized ADMM.

Speakers
RB

Robert Bassett

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JH

Jefferson Huang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
RM

Roummel Marcia

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
avatar for Peter Barkley

Peter Barkley

US Navy
I am a naval officer and a PhD student in Operations Research at the Naval Postgraduate School. My research interests include decentralized optimization, scheduling, and machine learning. In previous roles in the Navy, I’ve flown the P-3C Orion as a flight instructor and mission... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 118 3501 Trousdale Pkwy, 118, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6L: Alternative and Hybrid Algorithms in Quantum Computing for Optimization and Applications
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Alternative and Hybrid Algorithms in Quantum Computing for Optimization and Applications
Chair: Xiu Yang
Cluster: Optimization for Emerging Technologies (LLMs, Quantum Computing, ...)

Talk 1: Unleashed from Constrained Optimization: Quantum Computing for Quantum Chemistry Employing Generator Coordinate Method
Speaker: Bo Peng
Abstract: Hybrid quantum-classical approaches offer potential solutions to quantum chemistry problems, yet they also introduce challenges. These challenges include addressing the barren plateau and ensuring the accuracy of the ansatze, which often manifest as constrained optimization problems. In this work, we explore the interconnection between constrained optimization and generalized eigenvalue problems through a unique class of Givens rotations. These rotations frequently serve as disentangled unitary coupled cluster building blocks constituting the ansatze in variational quantum eigensolver (VQE) and adaptive derivative-assembled pseudo-Trotter VQE (ADAPT-VQE) simulations. Herein, we employ Givens rotations to construct non-orthogonal, overcomplete many-body generating functions, projecting the system Hamiltonian into a working subspace. The resulting generalized eigenvalue problem is proven to generate rigorous lower bounds to the VQE/ADAPT-VQE energies, effectively circumventing the barren plateau issue and the heuristic nature of numerical minimizers in standard VQE processes. For practical applications, we further propose an adaptive scheme for the robust construction of many-body basis sets using these Givens rotations, emphasizing a linear expansion that balances accuracy and efficiency. The effective Hamiltonian generated by our approach would also facilitate the computation of excited states and evolution, laying the groundwork for more sophisticated quantum simulations in chemistry.

Talk 2: Quantum DeepONet: Neural operators accelerated by quantum computing
Speaker: Lu Lu
Abstract: In the realm of mathematics, engineering, and science, constructing models that reflect real- world phenomena requires solving partial differential equations (PDEs) with different param- eters. Recent advancements in DeepONet, which learn mappings between infinite-dimensional function spaces, promise efficient evaluations of PDE solutions for new parameter sets in a single forward pass. However, classical DeepONet entails quadratic time complexity concerning input dimensions during evaluation. Given the progress in quantum algorithms and hardware, we propose utilizing quantum computing to accelerate DeepONet evaluations, resulting in time complexity that is linear in input dimensions. Our approach integrates unary encoding and orthogonal quantum layers to facilitate this process. We benchmark our Quantum DeepONet using a variety of equations, including the first-order linear ordinary differential equation, advection equation, and Burgers' equation, demonstrating the method’s efficacy in both ideal and noisy conditions. Furthermore, we show that our quantum DeepONet can also be informed by physics, minimizing its reliance on extensive data collection. We expect Quantum DeepONet to be particularly advantageous in applications in outer loop problems which require to explore parameter space and solving the corresponding PDEs, such as forward uncertainty propagation and optimal experimental design.

Talk 3: Koopman Linearization for Optimization in Quantum Computing
Speaker: Xiu Yang
Abstract: Nonlinearity presents a significant challenge in developing quantum algorithms involving differential equations, prompting the exploration of various linearization techniques, including the well-known Carleman Linearization. Instead, this paper introduces the Koopman Spectral Linearization method tailored for nonlinear autonomous ordinary differential equations. This innovative linearization approach harnesses the interpolation methods and the Koopman Operator Theory to yield a lifted linear system. It promises to serve as an alternative approach that can be employed in scenarios where Carleman Linearization is traditionally applied. Numerical experiments demonstrate the effectiveness of this linearization approach for several commonly used nonlinear ordinary differential equations. Hence, it enables a special design of gradient-descent type of method based on the technique called Schrodingerization that is used to solve linear differential equations on quantum computers.

Speakers
BP

Bo Peng

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
LL

Lu Lu

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
XY

Xiu Yang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 119 3501 Trousdale Pkwy, 119, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6M: Data-Driven Decision and Control
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Data-Driven Decision and Control
Chair: Guannan Qu
Cluster: Optimization For Data Science

Talk 1: Robustness in data-driven control: Requirement? An opportunity
Speaker: Laixi Shi
Abstract: Reinforcement learning (RL), which strives to learn desirable sequential decisions based on trial-and-error interactions with an unknown environment, has achieved remarkable success recently in a variety of domains including games and large language model alignment. While standard RL has been heavily investigated recently, a policy learned in an ideal, nominal environment might fail catastrophically when the deployed environment is subject to small changes in task objectives or adversarial perturbations, especially in high-stake applications such as robotics and clinical trials. This talk concerns the central issues of sample efficiency and model robustness in reinforcement learning (RL) to reduce the sim-to-real gap in practice. We adopt the framework of distributionally robust Markov decision processes (RMDPs), aimed at learning a policy that optimizes the worst-case performance when the deployed environment falls within a prescribed uncertainty set around the nominal MDP. Despite recent efforts, the sample complexity of RMDPs remained mostly unsettled regardless of the uncertainty set in use. It was unclear if distributional robustness bears any statistical consequences when benchmarked against standard RL. Somewhat surprisingly, our results uncover that RMDPs are not necessarily easier or harder to learn than standard MDPs. The statistical consequence incurred by the robustness requirement depends heavily on the size and shape of the uncertainty set. In addition, we break down the sample barrier of robust RL in offline setting by providing the first provable near-optimal algorithm for offline robust RL that can learn under simultaneous model uncertainty and limited historical datasets.

Talk 2: Learning and Control in Countable State Spaces
Speaker: Rayadurgam Srikant
Abstract: We will consider policy optimization methods in reinforcement learning where the state space is countably infinite. The motivation arises from control problems in communication networks and matching markets. Specifically, we consider the popular Natural Policy Gradient (NPG) algorithm, which has been studied in the past only under the assumptions that the cost is bounded and the state space is finite, neither of which holds for the aforementioned control problems. Assuming a Lyapunov drift condition, which is naturally satisfied in some cases and can be satisfied in other cases at a small cost in performance, we design a state-dependent step-size rule which dramatically improves the performance of NPG for our intended applications. In addition to experimentally verifying the performance improvement, we also theoretically show that the iteration complexity of NPG can be made independent of the size of the state space. The key analytical tool we use is the connection between NPG stepsizes and the solution to Poisson’s equation. In particular, we provide policy-independent bounds on the solution to Poisson’s equation, which are then used to guide the choice of NPG stepsizes.

Talk 3: Distributionally Robust Control via Optimal Transport
Speaker: Liviu Aolaritei
Abstract: In this talk I will challenge the standard uncertainty models, i.e., robust (norm-bounded) and stochastic (one fixed distribution, e.g., Gaussian), and propose to model uncertainty in dynamical systems via Optimal Transport (OT) ambiguity sets. I will then show that OT ambiguity sets are analytically tractable: they propagate easily and intuitively through linear and nonlinear (possibly corrupted by noise) maps, and the result of the propagation is again an OT ambiguity set or can be tightly upper bounded by one. In the context of dynamical systems, this allows to consider multiple sources of uncertainty (e.g., initial condition, additive noise, multiplicative noise) and to capture in closed-form, via an OT ambiguity set, the resulting uncertainty in the state at any future time. The resulting OT ambiguity sets are also computationally tractable, and can be directly employed in various distributionally robust control formulations that can optimally trade between safety and performance.

Speakers
LS

Laixi Shi

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
RS

Rayadurgam Srikant

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
LA

Liviu Aolaritei

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 157 3518 Trousdale Pkwy, 157, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6N: The Role of Constraints in the Computational Complexity and Solvability of Optimization Problems
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: The Role of Constraints in the Computational Complexity and Solvability of Optimization Problems
Chair: Tatjana Chavdarova
Cluster: Fixed Points and Variational Inequalities

Talk 1: On the Role of Constraints in the Complexity of Min-Max Optimization
Speaker: Gabriele Farina
Abstract: We investigate the role of constraints in the computational complexity of min-max optimization. The work of Daskalakis, Skoulakis, and Zampetakis [2021] was the first to study min-max optimization through the lens of computational complexity, showing that min-max problems with nonconvex-nonconcave objectives are PPAD-hard. However, their proof hinges on the presence of joint constraints between the maximizing and minimizing players. The main goal of this paper is to understand the role of these constraints in min-max optimization. The first contribution of this paper is a fundamentally new proof of their main result, which improves it in multiple directions: it holds for degree 2 polynomials, it is essentially tight in the parameters, and it is much simpler than previous approaches, clearly highlighting the role of constraints in the hardness of the problem. Second, we show that with general constraints (i.e., the min player and max player have different constraints), even convex-concave min-max optimization becomes PPAD-hard. Along the way, we also provide PPAD-membership of a general problem related to quasi-variational inequalities, which has applications beyond our problem.

Talk 2: A First Order Primal-Dual Method for Solving Constrained Variational Inequalities
Speaker: Tatjana Chavdarova
Abstract: We introduce an interior-point method to solve constrained variational inequality (cVI) problems. By combining primal-dual and interior-point techniques, we develop a family of first-order methods called the ADMM-based interior-point method for constrained VIs (ACVI). This approach enables the solution of cVIs with complex constraints. We establish convergence guarantees for ACVI under the assumption that the operator is monotone (not necessarily L-Lipschitz) and that its subproblems can be solved exactly. Specifically, we achieve a convergence rate of O(1/sqrt(K)) for the gap function of the last iterate, matching the known lower bound. To address cases where subproblems cannot be solved analytically, we propose an inexact version of ACVI, which leverages a warm-starting technique for the subproblems, taking advantage of the fact that these subproblems change only slightly between iterations. When the operator is monotone, L-Lipschitz, and the errors in solving the subproblems decrease at appropriate rates, we demonstrate that the same convergence rate holds for the gap function. To the best of our knowledge, this is the first first-order interior-point method for general cVI problems with a global convergence guarantee.

Talk 3: The Computational Complexity of Finding Stationary Points in Non-convex Optimization
Speaker: Manolis Zampetakis
Abstract: Finding approximate stationary points, i.e., points where the gradient is approximately zero, of non-convex but smooth objective functions over unrestricted d-dimensional domains is one of the most fundamental problems in classical non-convex optimization. Nevertheless, the computational and query complexity of this problem are still not well understood when the dimension d of the problem is independent of the approximation error. In this paper, we show the following computational and query complexity results: 1. The problem of finding approximate stationary points over unrestricted domains is PLS-complete. 2. For d = 2, we provide a zero-order algorithm at a lower bound that shows that this algorithm achieves the optimal query complexity in both the constrained and the unconstrained setting. 3. Combining our results with recent results in complexity theory we show that finding approximate KKT points in constrained optimization is reducible to finding approximate stationary points in unconstrained optimization but the converse is impossible.

Speakers
GF

Gabriele Farina

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
TC

Tatjana Chavdarova

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
MZ

Manolis Zampetakis

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 256 3518 Trousdale Pkwy, 256, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6O: First-order methods for nonsmooth and constrained optimization - II
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: First-order methods for nonsmooth and constrained optimization - II
Chair: Zhe (Jimmy) Zhang
Cluster: Nonlinear Optimization

Talk 1: A Single-Loop Spider-Type Stochastic Subgradient Method for Nonconvex Nonsmooth Expectation Constrained Optimization
Speaker: Wei Liu
Abstract: Many real-world problems involve complex nonconvex functional constraints and large datasets, necessitating efficient stochastic methods for solving stochastic optimization problems. A majority of existing works assume no constraints or easy-to-project constraints. In this paper, we consider nonconvex stochastic optimization problems with nonconvex expectation constraints. We construct an unconstrained exact penalty model with a new penalized function for the expectation constraints that share the same stationary points as the original problem. To solve this problem, we present a single-loop spider-type stochastic subgradient first-order method, which utilizes the derivatives of f and g and the function value of g at each iteration. Under certain regularity conditions (weaker than Slater-type constraint qualification in existing works), we establish an oracle complexity result of O(ϵ−4) to reach a point that is ϵ\epsilonϵ-close to an ϵ\epsilonϵ-KKT point of the original problem in expectation, matching the lower bound for such tasks. As important applications, we apply our method to two fairness-constrained problems, demonstrating that it is at least 9 times faster than state-of-the-art algorithms, including switching subgradient methods and inexact proximal point methods.

Talk 2: Monotone Variational Inequality problem under Generalized Conditions
Speaker: Digvijay Boob
Abstract: We consider the well-known variational inequality (VI) problem on monotone operators under a novel Lipschitz type condition and possibly unbounded feasible set. This new class of problems cover various class of problems, including convex function constrained problem, convex semi-infinite constrained problem, general robust optimization problem, and function-constrained variational inequalities among others - covering wide class of problems as a specific case. We show that when the problem is generalized smooth, our method converges at the rate of O(1/K) in terms of Minty-gap criterion appropriately defined for unbounded sets. For strongly monotone generalized smooth problems, we show linear convergence. For generalized nonsmooth and stochastic VI problems, we show convergence at the rate of O(1/\sqrt{K}). To our best knowledge, this is the first time such problems are addressed, especially in the context of variational inequality problems.

Talk 3: Stochastic Block-Wise Iterative Methods for Training Feasibility-Based Neural Architectures
Speaker: Manish Krishan
Abstract: We present a unifying framework for finitely convergent stochastic block-wise iterative methods, tailored for training set-feasibility-based neural network architectures. We apply this disciplined approach to selecting both the architecture and its corresponding training algorithm for classification and regression tasks.

Speakers
avatar for Wei Liu

Wei Liu

research scholar, Rensselaer Polytechnic Institute
Name: Dr. Slothington "Slow Convergence" McNapfaceTitle: Distinguished Professor of Continuous Optimization & Energy MinimizationAffiliation: The Lush Canopy Institute of Sluggish AlgorithmsBio:Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
DB

Digvijay Boob

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
MK

Manish Krishan

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 258 3518 Trousdale Pkwy, 258, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6P: Recent Advances in PDE-Constrained Optimizatin
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Recent Advances in PDE-Constrained Optimizatin
Chair: Michael Ulbrich
Cluster: PDE-constrained Optimization

Talk 1: Probabilistic state constraints for optimal control problems under uncertainty
Speaker: Caroline Geiersbach
Abstract: In this talk, we discuss optimal control problems subject to random state constraints, where we distinguish between the chance-constrained case and the almost sure formulation. We highlight some of the difficulties in the infinite-dimensional setting, which is of interest in physics-based models where a control belonging to a Banach space acts on a system described by a partial differential equation (PDE) with random inputs or parameters. We study the setting in which the obtained state should be bounded uniformly over the physical domain with high probability, or even probability one. We apply our results to a model with a random elliptic PDE, where the randomness is induced by the right-hand side. For the chance-constrained setting, this structure allows us to obtain an explicit representation for the Clarke subdifferential of the probability function using the spherical radial decomposition of Gaussian random vectors. This representation is used for the numerical solution in a discretize-then-optimize approach. For the almost sure setting, we use a Moreau-Yosida regularization and solve a sequence of regularized problems in an optimize-then-discretize approach. The solutions are compared, providing insights for the development of further algorithms.

Talk 2: Image Registration in Non-Reflexive Banach Spaces
Speaker: Johannes Haubner
Abstract: We consider image registration as an optimal control problem using an optical flow formulation, i.e., we aim to solve an optimization problem that is governed by a linear hyperbolic transport equation. In order to be able to transform the image of a square into the image of a circle and ensure bi-Lipschitz continuity of the transformation, we aim for $W^{1,\infty}$-regularity of the velocity that parametrizes the transformation. This leads to an optimization problem in the non-reflexive Banach space $W^{1,\infty}$. We introduce relaxations of the optimization problem involving smoothed maximum and minimum functions and the Orlicz space $L_{\exp}$. To derive well-posedness results for the relaxed optimization problem, we revisit and establish new existence and uniqueness results for the linear hyperbolic transport equations. We present limit considerations and discuss differentiability.

Talk 3: Numerical methods for shape optimal shpe design of fluid-structure interaction problems
Speaker: Michael Ulbrich
Abstract: We consider the method of mappings for performing shape optimization for unsteady fluid–structure interaction (FSI) problems. The main focus is on the numerical implementation. Our modelling of the optimization problem is guided by several theoretical aspects, such as regularity requirements on the transformations and connections to differential geometric aspects. We discretize the problem in a way such that exact discrete gradients can be computed. This in combination with the way we represent shapes allows for an efficent application of general purpose optimization solvers. Our numerical tests focus on problems derived from an FSI benchmark to validate our implementation, which builds on FEniCS, dolfin-adjoint, and IPOPT. The method is used to optimize parts of the outer boundary and the interface.

Speakers
CG

Caroline Geiersbach

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
JH

Johannes Haubner

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
MU

Michael Ulbrich

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 106 3501 Trousdale Pkwy, 106, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6Q: Optimization on Riemannian manifolds and stratified sets (2/2)
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Optimization on Riemannian manifolds and stratified sets (2/2)
Chair: Guillaume Olikier
Cluster: Optimization on Manifolds

Talk 1: A space-decoupling framework for optimization on bounded-rank matrices with orthogonally invariant constraints
Speaker: Bin Gao
Abstract: Imposing additional constraints on low-rank optimization has garnered growing interest recently. However, the geometry of coupled constraints restricts the well-developed low-rank structure and makes the problem nonsmooth. In this paper, we propose a space-decoupling framework for optimization problems on bounded-rank matrices with orthogonally invariant constraints. The "space-decoupling" is reflected in several ways. Firstly, we show that the tangent cone of coupled constraints is the intersection of the tangent cones of each constraint. Secondly, we decouple the intertwined bounded-rank and orthogonally invariant constraints into two spaces, resulting in optimization on a smooth manifold. Thirdly, we claim that implementing Riemannian algorithms is painless as long as the geometry of additional constraint is known a prior. In the end, we unveil the equivalence between the original problem and the reformulated problem. The numerical experiments validate the effectiveness and efficiency of the proposed framework.

Talk 2: A Riemannian rank-adaptive method for tensor completion in the tensor-train format
Speaker: Charlotte Vermeylen
Abstract: Riemannian rank-adaptive methods (RRAMs) are state-of-the-art methods that aim to optimize a continuously differentiable function on a low-rank variety, a problem appearing notably in low-rank matrix or tensor completion. The RRAMs that are developed for the set of bounded rank matrices iteratively optimize over smooth fixed-rank manifolds, starting from a low initial rank. They increase the rank by performing a line search along a descent direction selected in the tangent cone to the variety. This direction can be the projection of the negative gradient onto the tangent cone but does not need to; for instance, the RRAM developed by Gao and Absil uses the projection of the negative gradient onto the part of the tangent cone that is normal to the tangent space. Additionally, they decrease the rank based on a truncated SVD when the algorithm converges to an element of a lower-rank set. This is possible because the manifold is not closed. We aim to generalize these RRAMs to the tensor-train (TT) format. In this format, only the RRAM by Steinlechner is known to us from the literature. This method is developed for high-dimensional tensor completion and has a random rank update mechanism in the sense that each TT-rank is increased subsequently by one by adding a small random term in the TTD of the current best approximation such that the cost function does not increase much. No rank reduction step is included which makes the algorithm prone to overfitting. We improve this RRAM by including a method to increase the TT-rank based on an approximate projection of the negative gradient onto the tangent cone to the variety. The tangent cone is the set of all tangent vectors to the variety. The approximate projection ensures that the search direction is sufficiently gradient related. Furthermore, a method is proposed to determine how much the rank should be increased for the low-rank tensor completion problem (LRTCP). Lastly, a method to decrease the rank is proposed, which is necessary when the iterate comes close to a lower-rank set. This is possible because, as for the manifold of fixed-rank matrices, the manifold of fixed-rank TTDs is not closed.

Talk 3: Low-rank optimization on Tucker tensor varieties
Speaker: Renfeng Peng
Abstract: In the realm of tensor optimization, the low-rank Tucker decomposition is crucial for reducing the number of parameters and saving storage. In this talk, we delve into the geometry and optimization methods for Tucker tensor varieties---the set of tensors with bounded Tucker rank---which is notably more intricate than the well-explored matrix varieties. We give an explicit parametrization of the tangent cone of Tucker tensor varieties and leverage its geometry to develop provable gradient-related line-search methods for optimization on Tucker tensor varieties. In practice, low-rank tensor optimization suffers from the difficulty of choosing a reliable rank parameter. To this end, we incorporate the established geometry and propose a Tucker rank-adaptive method that aims to identify an appropriate rank. Numerical experiments on tensor completion reveal that the proposed methods are in favor of recovering performance over other state-of-the-art methods. This is joint work with Bin Gao (AMSS) and Ya-xiang Yuan (AMSS).

Speakers
GO

Guillaume Olikier

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
BG

Bin Gao

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
CV

Charlotte Vermeylen

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
avatar for Renfeng Peng

Renfeng Peng

Ph. D. student, Academy of Mathematics and Systems Science, Chinese Academy of Sciences
Name: Renfeng PengTitle: Low-rank optimization on Tucker tensor varietiesAffiliation: Academy of Mathematics and Systems Science, Chinese Academy of SciencesBio:Renfeng Peng is a Ph.D. student of the Optimization Group at Academy of Mathematics and Systems Science (AMSS), Chinese... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 214 3501 Trousdale Pkwy, 214, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6R: Recent Adavances in Interval Arithmetic based Methods
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Recent Adavances in Interval Arithmetic based Methods
Chair: Frederic Messine
Cluster: Global Optimization

Talk 1: Boosting the performance of rigorous global optimization
Speaker: Victor Reyes Rodriguez
Abstract: In this talk, I will present some boosting methods to improve the performance of rigorous global optimization.

Talk 2: Node selection through upper bounding local search methods in branch & bound solvers for NCOPs
Speaker: Ignacio Araya
Abstract: In this talk, I will present how to improve the convergence of branch & bound solvers for NCOPs by making node selection through upper bounding local search methods.

Talk 3: Interval Branch and Bound Code with PDE-constraints
Speaker: Frédéric Messine
Abstract: In this article, I will show how, in a global optimization solver whose bound calculations are performed by interval arithmetic, we can insert constraints based on numerical simulations. The aim of this extension is to provide exact solutions to concrete complex problems: an example will be given for the design of electric motors.

Speakers
VR

Victor Reyes Rodriguez

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
IA

Ignacio Araya

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
FM

Frédéric Messine

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 154 3518 Trousdale Pkwy, 154, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6S: Multi-agent learning in games and markets
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Multi-agent learning in games and markets
Chair: Manxi Wu
Cluster: Multi-agent Optimization and Games

Talk 1: Learning Optimal Stable Matches in Decentralized Markets with Unknown Preferences
Speaker: Bryce Ferguson
Abstract: Matching algorithms have demonstrated great success in several practical applications, but they often require centralized coordination and plentiful information. In many modern online marketplaces, agents must independently seek out and match with another using little to no information. For these kinds of settings, can we design decentralized, limited-information matching algorithms that preserve the desirable properties of standard centralized techniques? In this work, we constructively answer this question in the affirmative. We model a two-sided matching market as a game consisting of two disjoint sets of agents, referred to as proposers and acceptors, each of whom seeks to match with their most preferable partner on the opposite side of the market. However, each proposer has no knowledge of their own preferences, so they must learn their preferences while forming matches in the market. We present a simple online learning rule that guarantees a strong notion of probabilistic convergence to the welfare-maximizing equilibrium of the game, referred to as the proposer-optimal stable match. To the best of our knowledge, this represents the first completely decoupled, communication-free algorithm that guarantees probabilistic convergence to an optimal stable match, irrespective of the structure of the matching market.

Talk 2: Stochastic Online Fisher Markets: Static Pricing Limits and Adaptive Enhancements
Speaker: Devansh Jalota
Abstract: Fisher markets are one of the most fundamental models for resource allocation. However, the problem of computing equilibrium prices in Fisher markets typically relies on complete knowledge of users' budgets and utility functions and requires transactions to happen in a static market where all users are present simultaneously. Motivated by these practical considerations, we study an online variant of Fisher markets, wherein users with privately known utility and budget parameters, drawn i.i.d. from a distribution, arrive sequentially. In this setting, we first study the limitations of static pricing algorithms, which set uniform prices for all users, along two performance metrics: (i) regret, i.e., the optimality gap in the objective of the Eisenberg-Gale program between an online algorithm and an oracle with complete information, and (ii) capacity violations, i.e., the over-consumption of goods relative to their capacities. Given the limitations of static pricing, we design adaptive posted-pricing algorithms, one with knowledge of the distribution of users' budget and utility parameters and another that adjusts prices solely based on past observations of user consumption, i.e., revealed preference feedback, with improved performance guarantees. Finally, we present numerical experiments to compare our revealed preference algorithm's performance to several benchmarks.

Talk 3: Decentralized learning in Markov potential games and beyond
Speaker: Manxi Wu
Abstract: Infinite-horizon stochastic games provide a versatile framework for studying the repeated interaction among multiple strategic agents in dynamic environments. However, computing equilibria in such games is highly complex, and the long-run outcomes of decentralized learning algorithms in multi-agent settings remain poorly understood. The first part of this talk introduces a multi-agent reinforcement learning dynamics tailored for independent and decentralized settings, where players lack knowledge of the game model and cannot coordinate. The proposed dynamics guarantee convergence to a stationary Nash equilibrium in Markov potential games, demonstrating the effectiveness of simple learning dynamics even with limited information. In the second part of the talk, we extend the learning framework to encompass Markov near potential games, offering flexibility to incorporate a wide range of practically-relevant multi-agent interaction settings. We present efficient algorithms for approximating the stationary Nash equilibrium and substantiate their effectiveness through regret analysis and numerical experiments.

Speakers
BF

Bryce Ferguson

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
DJ

Devansh Jalota

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
MW

Manxi Wu

Assistant Professor, University of California, Berkeley
Name: Dr. Slothington "Slow Convergence" McNapfaceTitle: Distinguished Professor of Continuous Optimization & Energy MinimizationAffiliation: The Lush Canopy Institute of Sluggish AlgorithmsBio:Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 155 3518 Trousdale Pkwy, 155, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6T: Randomized optimization algorithms 1/2
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Randomized optimization algorithms 1/2
Chair: Laurent Condat
Cluster: Nonlinear Optimization

Talk 1: Variance reduction for stochastic proximal algorithms
Speaker: Cheik Traoré
Abstract: In the context of finite sums minimization, variance reduction techniques are widely used to improve the performance of state-of-the-art stochastic gradient methods. Their practical impact is clear, as well as their theoretical properties. Stochastic proximal point algorithms have been studied as an alternative to stochastic gradient algorithms since they are more stable with respect to the choice of the stepsize but their variance reduced versions are not as studied as the gradient ones. In this work, we propose the first unified study of variance reduction techniques for stochastic proximal point algorithms. We introduce a generic stochastic proximal algorithm that can be specified to give the proximal version of SVRG, SAGA, and some of their variants for smooth and convex functions. We provide several convergence results for the iterates and the objective function values. In addition, under the Polyak-Łojasiewicz (PL) condition, we obtain linear convergence rates for the iterates and the function values. Our numerical experiments demonstrate the advantages of the proximal variance reduction methods over their gradient counterparts, especially about the stability with respect to the choice of the stepsize for difficult problems.

Talk 2: Taming Nonconvex Stochastic Mirror Descent with General Bregman Divergence
Speaker: Ilyas Fatkhullin
Abstract: This paper revisits the convergence of Stochastic Mirror Descent (SMD) in the contemporary nonconvex optimization setting. Existing results for batch-free nonconvex SMD restrict the choice of the distance generating function (DGF) to be differentiable with Lipschitz continuous gradients, thereby excluding important setups such as Shannon entropy. In this work, we present a new convergence analysis of nonconvex SMD supporting general DGF, that overcomes the above limitations and relies solely on the standard assumptions. Moreover, our convergence is established with respect to the Bregman Forward-Backward envelope, which is a stronger measure than the commonly used squared norm of gradient mapping. We further extend our results to guarantee high probability convergence under sub-Gaussian noise and global convergence under the generalized Bregman Proximal Polyak-{Ł}ojasiewicz condition. Additionally, we illustrate the advantages of our improved SMD theory in various nonconvex machine learning tasks by harnessing nonsmooth DGFs. Notably, in the context of nonconvex differentially private (DP) learning, our theory yields a simple algorithm with a (nearly) dimension-independent utility bound. For the problem of training linear neural networks, we develop provably convergent stochastic algorithms.

Talk 3: Adaptive Bregman-Kaczmarz: An approach to solve linear inverse problems with independent noise exactly
Speaker: Lionel Tondji
Abstract: We consider the block Bregman–Kaczmarz method for finite dimensional linear inverse problems. The block Bregman–Kaczmarz method uses blocks of the linear system and performs iterative steps with these blocks only. We assume a noise model that we call independent noise, i.e. each time the method performs a step for some block, one obtains a noisy sample of the respective part of the right-hand side which is contaminated with new noise that is independent of all previous steps of the method. One can view these noise models as making a fresh noisy measurement of the respective block each time it is used. In this framework, we are able to show that a well-chosen adaptive stepsize of the block Bregman–Kaczmarz method is able to converge to the exact solution of the linear inverse problem. The plain form of this adaptive stepsize relies on unknown quantities (like the Bregman distance to the solution), but we show a way how these quantities can be estimated purely from given data. We illustrate the finding in numerical experiments and confirm that these heuristic estimates lead to effective stepsizes.

Speakers
avatar for Laurent Condat

Laurent Condat

Senior Research Scientist, King Abdullah University of Science and Technology (KAUST)
Laurent Condat received a PhD in applied mathematics in 2006 from Grenoble Institute of Technology, Grenoble, France. After a postdoc in the Helmholtz Zentrum Muenchen, Munich, Germany, he was hired in 2008 as a permanent researcher by the French National Center for Scientific Research... Read More →
CT

Cheik Traoré

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
IF

Ilyas Fatkhullin

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
LT

Lionel Tondji

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 158 3518 Trousdale Pkwy, 158, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6U: Applications of derivative-free optimization
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Applications of derivative-free optimization
Chair: Juliane Müller
Cluster: Derivative-free Optimization

Talk 1: Surrogate model guided search for optimizing metabolic models
Speaker: Juliane Müller
Abstract: The development and scaleup of novel biofuels require accurate prediction of microbial performance. While there are very sophisticated metabolic and expression (ME) models that simulate cellular metabolism and expression, their accuracy depends on a variety of parameters that must be tuned such that the model agrees with observation data. This yields a continuous optimization problem in which the quality of parameter sets must be assessed by running the compute intensive ME models. In this talk, we will present the challenge associated with this application and propose derivative-free surrogate models guided optimization approaches to tackle it.

Talk 2: A multilevel stochastic regularized first-order method with application to training
Speaker: Filippo Marini
Abstract: We present a new multilevel stochastic framework for the solution of optimization problems. The proposed approach uses random regularized first-order models that exploit an available hierarchical description of the problem, being either in the classical variable space or in the function space, meaning that different levels of accuracy for the objective function are available. We present the converge analysis of the method and show its numerical behavior on the solution of finite-sum minimization problems arising in binary classification problems.

Talk 3: Surrogate-based evolutionary optimization extended to categorical and dependent variables
Speaker: Charlotte Beauthier
Abstract: The objective of this work is the development of methods combining numerical simulation tools and advanced optimization algorithms, which play a crucial role in exploring new conceptual designs. Nowadays the exploitation of multi-disciplinary optimization using high-fidelity simulation models is common to many engineering design problems. A globally effective approach to optimization problems based on computationally expensive analysis lies in the exploitation of surrogate models. They act as cheap-to-evaluate alternatives to the original high-fidelity models reducing the computational cost, while still providing improved designs. Furthermore, in practice, various industrial design problems have continuous, discrete, and categorical variables. From an engineering point of view, the specific case of categorical variables is of great practical interest by their ability to represent the choice of a material, the type of engine architecture, the shape of a cross-section for a beam profile, etc. The contributions of this talk are focused on the management of mixed variables, in particular the categorical ones, in a surrogate-based evolutionary algorithm where dependency can also be defined between variables. More precisely, the dependency considered here is when the definition domain of a variable can be linked to another variable’s value. In order to deal with mixed and dependent variables, different methods are proposed to refine the notion of distance between mixed variables, using the notion of affinity, for instance, and to adapt the genetic operators in the evolutionary algorithm accordingly. These novel developments have been implemented in Minamo, Cenaero’s in-house design space exploration and multi-disciplinary optimization platform. Numerical results will be presented on specific test problems coming from structural and mechanical design frameworks to study the impact of the proposed strategies.

Speakers
JM

Juliane Müller

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 108 3501 Trousdale Pkwy, 108, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6V: Trust-Region Methods
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Trust-Region Methods
Chair: Mohamed Laghdaf Habiboullah
Cluster: nan

Talk 1: Complexity of trust-region methods in the presence of unbounded Hessian approximations
Speaker: Mohamed Laghdaf Habiboullah
Abstract: We extend traditional complexity analyses of trust-region methods for unconstrained, possibly nonconvex, optimization. Whereas most complexity analyses assume uniform boundedness of the model Hessians, we work with potentially unbounded model Hessians. Boundedness is not guaranteed in practical implementations, in particular ones based on quasi-Newton updates such as PSB, BFGS and SR1. Our analysis is conducted for a family of trust-region methods that includes most known methods as special cases. We examine two regimes of Hessian growth: one bounded by a power of the number of successful iterations, and one bounded by a power of the number of iterations. This allows us to formalize and confirm the profound intuition of Powell [IMA J. Numer. Ana. 30(1):289-301,2010], who studied convergence under a special case of our assumptions, but whose proof contained complexity arguments. Specifically, for \(0 \leq p < 1\), we establish sharp \(O(\epsilon^{-2/(1-p)})\) evaluation complexity to find an \(\epsilon\)-stationary point when model Hessians are \(O(k^p)\), where \(k\) is the iteration counter. For \(p = 1\), which is the case studied by Powell, we establish a sharp \(O(\exp(c\epsilon^{-2}))\) evaluation complexity for a certain constant \(c > 0\). This is as Powell suspected and is far worse than other bounds surmised elsewhere in the literature. We establish similar bounds when model Hessians are \(O(|\mathcal{S}_k|^p)\), where \(|\mathcal{S}_k|\) is the number of iterations where the step was accepted, up to iteration \(k\). To the best of our knowledge, ours is the first work to provide complexity bounds when model Hessians grow linearly with \(|\mathcal{S}_k|\) or at most linearly with \(k\), which covers multiple quasi-Newton approximations. Link: https://arxiv.org/abs/2408.06243

Talk 2: Advancing Trust-Region Methods: Gradient-Based Radius Updates and Stochastic Optimization
Speaker: Fabian Bastin
Abstract: We consider the framework of nonlinear, unconstrained, smooth optimization, where the trust-region approach has been established as one of the most prominent methods over the last decades. However, significant investigation has been devoted to updating the trust-region radius in recent years, leading to various schemes that exploit the structure of the problem under consideration. In particular, a more intricate link between the trust-region radius and the norm of the gradient at the current iterate has enabled the derivation of upper bounds on the number of iterations required to obtain an ϵ-optimal solution while preserving the main traditional convergence properties. Such results have also been extended to the context of stochastic programming with adaptive sampling to ensure convergence guarantees, with the key ingredient being the ability to achieve a sufficient decrease with high probability. We revisit the progress made regarding the relationship between the trust-region radius update and the gradient norm to establish that, even under these schemes, the trust region ultimately becomes inactive, even when using an adaptive sampling approach in stochastic optimization. This ensures that, under mild conditions, minimizing the model inside the trust region can ultimately yield an approximate quasi-Newton direction. We illustrate these ideas using a trust-region method with adaptive sampling and truncated conjugate gradient on a set of benchmark problems.

Talk 3: Beyond Newton Interpolation: p-Factor Polynomials for Function Approximation in Optimization
Speaker: Olga Brezhneva
Abstract: We introduce a new interpolation method for nonlinear functions over an interval, focusing on cases where the approximated function is non-regular at a solution. In such cases, classical interpolation methods, such as Newton’s interpolation polynomial, do not necessarily provide the required accuracy for approximating solutions to the equation $f(x)=0$. In contrast, our method, based on p-factor interpolation polynomials, ensures the necessary accuracy, leading to improved function approximations in solving nonlinear equations. The presented results are based on the theory of p-regularity. Our approach provides new insights into function approximation techniques in nonlinear optimization. Specifically, it has potential applications in derivative-free optimization, where function evaluations are costly or noisy, as well as in trust-region methods and polynomial-based approaches for solving nonlinear problems.

Speakers
FB

Fabian Bastin

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
OB

Olga Brezhneva

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 110 3501 Trousdale Pkwy, 110, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6W: Stochastic Optimization
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Session: Stochastic Optimization
Chair: Honghao Zhang
Cluster: nan

Talk 1: Multi-cut stochastic approximation methods for solving stochastic convex composite optimization
Speaker: Honghao Zhang
Abstract: The development of a multi-cut stochastic approximation (SA) method for solving stochastic convex composite optimization (SCCO) problems has remained an open challenge. The difficulty arises from the fact that the stochastic multi-cut model, constructed as the pointwise maximum of individual stochastic linearizations, provides a biased estimate of the objective function, with the error being uncontrollable. This paper introduces multi-cut SA methods for solving SCCO problems, achieving near-optimal convergence rates. The cutting-plane models used in these methods are the pointwise maxima of appropriately chosen one-cut models. To the best of our knowledge, these are the first multi-cut SA methods specifically designed for SCCO problems.

Talk 2: Data-driven Policies for Two-Stage Stochastic Linear Programs
Speaker: Chhavi Sharma
Abstract: A stochastic program typically involves several parameters including deterministic first-stage parameter, and stochastic second-stage elements that serve as input data. These programs are usually re-solved whenever there is a change in any of the input parameters. For example, a stochastic dispatch problem is solved multiple times a day due to fluctuations in electricity prices, demand, and renewable energy availability. However, in practical situations, quick decision-making is crucial, and solving a stochastic program from scratch for every change in input data can be computationally costly. This work addresses this challenge for two-stage stochastic linear programs (2-SLPs) with varying first-stage right-hand sides. We employ data-driven approaches to first construct a novel piecewise linear difference of max-affine policy (PLDC) for deterministic linear programs. This is achieved by leveraging optimal basis matrices from previous solutions and the piecewise linear nature of the optimal solution trajectory. This PLDC policy retains optimal solutions for previously encountered parameters and is expected to provide good-quality solutions for new parameters. Our developed policy applies directly to the extensive form of 2-SLP. When algorithms such as the L-shaped method are applied to solve 2-SLP, we construct the policy using local outer approximations of the recourse function and optimal basis matrices from previous solves. We assess the performance of our policy through both numerical and theoretical analyses. Our numerical experiments show small feasibility and relative optimal gap at solutions returned by the policy.

Talk 3: Statistical Robustness Analysis of In-CVaR Based Regression
Speaker: Yulei You
Abstract: Based on the interval conditional value-at-risk (In-CVaR) proposed in Liu & Pang (2022), this paper investigates the robustness of In-CVaR based regression under data contamination and perturbation. To quantify robustness under data contamination, we introduce the concept of “distributional breakdown point (BP) value”. Our results provide upper and lower bounds for the distributional BP value, which can be widely applied to classic regression tasks, including linear regression, piecewise affine regression, and feedforward neural network regression with homogeneous activation functions. Furthermore, we demonstrate that under data perturbation, the In-CVaR based estimator is qualitatively robust against optimization if and only if the largest portion of the loss is trimmed. Overall, this research complements the robust analysis of In-CVaR and shows that In-CVaR outperforms conditional value-at-risk and sample average approximation in terms of robustness for regression tasks. This talk is based on a joint work with Prof. Junyi Liu at Tsinghua University.

Speakers
HZ

Honghao Zhang

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
CS

Chhavi Sharma

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
YY

Yulei You

Name: Dr. Slothington "Slow Convergence" McNapface Title: Distinguished Professor of Continuous Optimization & Energy Minimization Affiliation: The Lush Canopy Institute of Sluggish Algorithms Bio: Dr. Slothington McNapface is a leading expert in continuous optimization, specializing... Read More →
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 112 3501 Trousdale Pkwy, 112, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6X
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Taper Hall (THH) 215 3501 Trousdale Pkwy, 215, Los Angeles, CA 90089

4:15pm PDT

Parallel Sessions 6Y
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Tuesday July 22, 2025 4:15pm - 5:30pm PDT
Joseph Medicine Crow Center for International and Public Affairs (DMC) 200 3518 Trousdale Pkwy, 200, Los Angeles, CA 90089
 
  • Filter By Date
  • Filter By Venue
  • Filter By Type
  • Timezone


Share Modal

Share this link via

Or copy link

Filter sessions
Apply filters to sessions.
Filtered by Date -