Export Citations
Save this search
Please login to be able to save your searches and receive alerts for new content matching your search criteria.
- research-articleDecember 2024
Accuracy Certificates for Convex Minimization with Inexact Oracle
Journal of Optimization Theory and Applications (JOPT), Volume 204, Issue 1https://doi.org/10.1007/s10957-024-02599-9AbstractAccuracy certificates for convex minimization problems allow for online verification of the accuracy of approximate solutions and provide a theoretically valid online stopping criterion. When solving the Lagrange dual problem, accuracy ...
- research-articleOctober 2024
High-Probability Complexity Bounds for Non-smooth Stochastic Convex Optimization with Heavy-Tailed Noise
Journal of Optimization Theory and Applications (JOPT), Volume 203, Issue 3Pages 2679–2738https://doi.org/10.1007/s10957-024-02533-zAbstractStochastic first-order methods are standard for training large-scale machine learning models. Random behavior may cause a particular run of an algorithm to result in a highly suboptimal objective value, whereas theoretical guarantees are usually ...
- research-articleJuly 2024
High-probability convergence for composite and distributed stochastic minimization and variational inequalities with heavy-tailed noise
- Eduard Gorbunov,
- Abdurakhmon Sadiev,
- Marina Danilova,
- Samuel Horváth,
- Gauthier Gidel,
- Pavel Dvurechensky,
- Alexander Gasnikov,
- Peter Richtárik
ICML'24: Proceedings of the 41st International Conference on Machine LearningArticle No.: 640, Pages 15951–16070High-probability analysis of stochastic first-order optimization methods under mild assumptions on the noise has been gaining a lot of attention in recent years. Typically, gradient clipping is one of the key algorithmic ingredients to derive good high-...
- research-articleJanuary 2025
Barrier algorithms for constrained non-convex optimization
ICML'24: Proceedings of the 41st International Conference on Machine LearningArticle No.: 484, Pages 12190–12214In this paper, we theoretically show that interior-point methods based on self-concordant barriers possess favorable global complexity beyond their standard application area of convex optimization. To do that we propose first- and second-order methods ...
-
- research-articleJuly 2023
High-probability bounds for stochastic optimization and variational inequalities: the case of unbounded variance
- Abdurakhmon Sadiev,
- Marina Danilova,
- Eduard Gorbunov,
- Samuel Horváth,
- Gauthier Gidel,
- Pavel Dvurechensky,
- Alexander Gasnikov,
- Peter Richtárik
ICML'23: Proceedings of the 40th International Conference on Machine LearningArticle No.: 1230, Pages 29563–29648During recent years the interest of optimization and machine learning communities in high-probability convergence of stochastic optimization methods has been growing. One of the main reasons for this is that high-probability complexity bounds are more ...
- research-articleApril 2024
Decentralized local stochastic extra-gradient for variational inequalities
- Aleksandr Beznosikov,
- Pavel Dvurechensky,
- Anastasia Koloskova,
- Valentin Samokhin,
- Sebastian U. Stich,
- Alexander Gasnikov
NIPS '22: Proceedings of the 36th International Conference on Neural Information Processing SystemsArticle No.: 2762, Pages 38116–38133We consider distributed stochastic variational inequalities (VIs) on unbounded domains with the problem data that is heterogeneous (non-IID) and distributed across many devices. We make a very general assumption on the computational network that, in ...
- research-articleApril 2024
Clipped stochastic methods for variational inequalities with heavy-tailed noise
NIPS '22: Proceedings of the 36th International Conference on Neural Information Processing SystemsArticle No.: 2271, Pages 31319–31332Stochastic first-order methods such as Stochastic Extragradient (SEG) or Stochastic Gradient Descent-Ascent (SGDA) for solving smooth minimax problems and, more generally, variational inequality problems (VIP) have been gaining a lot of attention in ...
- research-articleSeptember 2022
Generalized Mirror Prox Algorithm for Monotone Variational Inequalities: Universality and Inexact Oracle
Journal of Optimization Theory and Applications (JOPT), Volume 194, Issue 3Pages 988–1013https://doi.org/10.1007/s10957-022-02062-7AbstractWe introduce an inexact oracle model for variational inequalities with monotone operators, propose a numerical method that solves such variational inequalities, and analyze its convergence rate. As a particular case, we consider variational ...
- research-articleJune 2022
Oracle Complexity Separation in Convex Optimization
- Anastasiya Ivanova,
- Pavel Dvurechensky,
- Evgeniya Vorontsova,
- Dmitry Pasechnyuk,
- Alexander Gasnikov,
- Darina Dvinskikh,
- Alexander Tyurin
Journal of Optimization Theory and Applications (JOPT), Volume 193, Issue 1-3Pages 462–490https://doi.org/10.1007/s10957-022-02038-7AbstractMany convex optimization problems have structured objective functions written as a sum of functions with different oracle types (e.g., full gradient, coordinate derivative, stochastic gradient) and different arithmetic operations complexity of ...
- research-articleJanuary 2022
An Accelerated Method for Derivative-Free Smooth Stochastic Convex Optimization
SIAM Journal on Optimization (SIOPT), Volume 32, Issue 2Pages 1210–1238https://doi.org/10.1137/19M1259225We consider an unconstrained problem of minimizing a smooth convex function which is only available through noisy observations of its values, the noise consisting of two parts. Similar to stochastic optimization problems, the first part is of stochastic ...
- research-articleDecember 2021
An Accelerated Second-Order Method for Distributed Stochastic Optimization
- Artem Agafonov,
- Pavel Dvurechensky,
- Gesualdo Scutari,
- Alexander Gasnikov,
- Dmitry Kamzolov,
- Aleksandr Lukashevich,
- Amir Daneshmand
2021 60th IEEE Conference on Decision and Control (CDC)Pages 2407–2413https://doi.org/10.1109/CDC45484.2021.9683400We consider centralized distributed algorithms for general stochastic convex optimization problems which we approximate by a finite-sum minimization problem with summands distributed among computational nodes. We exploit statistical similarity between the ...
- research-articleDecember 2021
An Accelerated Method For Decentralized Distributed Stochastic Optimization Over Time-Varying Graphs
2021 60th IEEE Conference on Decision and Control (CDC)Pages 3367–3373https://doi.org/10.1109/CDC45484.2021.9683110We consider a distributed stochastic optimization problem that is solved by a decentralized network of agents with only local communication between neighboring agents. The goal of the whole system is to minimize a global objective function given as a sum ...
- research-articleDecember 2020
Multimarginal Optimal Transport by Accelerated Alternating Minimization
2020 59th IEEE Conference on Decision and Control (CDC)Pages 6132–6137https://doi.org/10.1109/CDC42340.2020.9304010We study multimarginal optimal transport (MOT) problems, which include, as a particular case, the Wasserstein barycenter problem. In MOT problems, one has to find an optimal coupling between m probability measures, which amounts to finding a tensor of ...
- ArticleSeptember 2020
Optimal Combination of Tensor Optimization Methods
Optimization and ApplicationsPages 166–183https://doi.org/10.1007/978-3-030-62867-3_13AbstractWe consider the minimization problem of a sum of a number of functions having Lipshitz p-th order derivatives with different Lipschitz constants. In this case, to accelerate optimization, we propose a general framework allowing to obtain near-...
- research-articleJuly 2020
Self-concordant analysis of Frank-Wolfe algorithms
ICML'20: Proceedings of the 37th International Conference on Machine LearningArticle No.: 264, Pages 2814–2824Projection-free optimization via different variants of the Frank-Wolfe (FW), a.k.a. Conditional Gradient method has become one of the cornerstones in optimization for machine learning since in many cases the linear minimization oracle is much cheaper to ...
- research-articleDecember 2019
On Primal and Dual Approaches for Distributed Stochastic Convex Optimization over Networks
2019 IEEE 58th Conference on Decision and Control (CDC)Pages 7435–7440https://doi.org/10.1109/CDC40024.2019.9029798We introduce primal and dual stochastic gradient oracle methods for distributed convex optimization problems over networks. We show that the proposed methods are optimal (in terms of communication steps) for primal and dual oracles. Additionally, for a ...
- research-articleDecember 2018
Distributed Computation of Wasserstein Barycenters Over Networks
2018 IEEE Conference on Decision and Control (CDC)Pages 6544–6549https://doi.org/10.1109/CDC.2018.8619160We propose a new class-optimal algorithm for the distributed computation of Wasserstein Barycenters over networks. Assuming that each node in a graph has a probability distribution, we prove that every node reaches the barycenter of all distributions held ...