Tootfinder

Opt-in global Mastodon full text search. Join the index!

@arXiv_csAI_bot@mastoxiv.page
2025-09-16 10:33:07

MAPGD: Multi-Agent Prompt Gradient Descent for Collaborative Prompt Optimization
Yichen Han, Bojun Liu, Zhengpeng zhou, Guanyu Liu, Zeng Zhang, Yang Yang, Wenli Wang, Isaac N Shi, Yunyan, Lewei He, Tianyu Shi
arxiv.org/abs/2509.11361

@arXiv_statML_bot@mastoxiv.page
2025-10-15 09:49:02

Statistical Guarantees for High-Dimensional Stochastic Gradient Descent
Jiaqi Li, Zhipeng Lou, Johannes Schmidt-Hieber, Wei Biao Wu
arxiv.org/abs/2510.12013

@arXiv_mathOC_bot@mastoxiv.page
2025-07-16 08:41:31

Non-smooth stochastic gradient descent using smoothing functions
Tommaso Giovannelli, Jingfu Tan, Luis Nunes Vicente
arxiv.org/abs/2507.10901

@arXiv_csGT_bot@mastoxiv.page
2025-07-16 08:00:21

A Parallelizable Approach for Characterizing NE in Zero-Sum Games After a Linear Number of Iterations of Gradient Descent
Taemin Kim, James P. Bailey
arxiv.org/abs/2507.11366

@arXiv_csLG_bot@mastoxiv.page
2025-09-15 09:57:11

Understanding Outer Optimizers in Local SGD: Learning Rates, Momentum, and Acceleration
Ahmed Khaled, Satyen Kale, Arthur Douillard, Chi Jin, Rob Fergus, Manzil Zaheer
arxiv.org/abs/2509.10439

@arXiv_mathOC_bot@mastoxiv.page
2025-07-16 09:19:11

Deep Equilibrium models for Poisson Imaging Inverse problems via Mirror Descent
Christian Daniele, Silvia Villa, Samuel Vaiter, Luca Calatroni
arxiv.org/abs/2507.11461

@arXiv_quantph_bot@mastoxiv.page
2025-08-12 11:39:43

Calculating the Projective Norm of higher-order tensors using a gradient descent algorithm
Aaditya Rudra, Maria Anastasia Jivulescu
arxiv.org/abs/2508.07933

@arXiv_eessSP_bot@mastoxiv.page
2025-09-15 08:24:01

Locally Permuted Low Rank Column-wise Sensing
Ahmed Ali Abbasi, Namrata Vaswani
arxiv.org/abs/2509.09820 arxiv.org/pdf/2509.09820

@arXiv_astrophIM_bot@mastoxiv.page
2025-09-15 08:37:41

A Differentiable Surrogate Model for the Generation of Radio Pulses from In-Ice Neutrino Interactions
Philipp Pilar, Martin Ravn, Christian Glaser, Niklas Wahlstr\"om
arxiv.org/abs/2509.10274

@arXiv_mathOC_bot@mastoxiv.page
2025-10-14 11:44:38

Adaptive Conditional Gradient Descent
Abbas Khademi, Antonio Silveti-Falls
arxiv.org/abs/2510.11440 arxiv.org/pdf/2510.11440

@arXiv_csLG_bot@mastoxiv.page
2025-09-11 10:10:33

Modified Loss of Momentum Gradient Descent: Fine-Grained Analysis
Matias D. Cattaneo, Boris Shigida
arxiv.org/abs/2509.08483 arxiv.org/pdf/…

@arXiv_eessAS_bot@mastoxiv.page
2025-09-16 08:55:46

Sound Matching an Analogue Levelling Amplifier Using the Newton-Raphson Method
Chin-Yun Yu, Gy\"orgy Fazekas
arxiv.org/abs/2509.10706

@arXiv_mathNA_bot@mastoxiv.page
2025-10-10 08:55:59

Stochastic Gradient Descent for Incomplete Tensor Linear Systems
Anna Ma, Deanna Needell, Alexander Xue
arxiv.org/abs/2510.07630 arxiv.org/…

@arXiv_mathOC_bot@mastoxiv.page
2025-08-13 08:40:32

Decentralized Relaxed Smooth Optimization with Gradient Descent Methods
Zhanhong Jiang, Aditya Balu, Soumik Sarkar
arxiv.org/abs/2508.08413

@arXiv_csSD_bot@mastoxiv.page
2025-10-13 08:23:10

Audible Networks: Deconstructing and Manipulating Sounds with Deep Non-Negative Autoencoders
Juan Jos\'e Burred, Carmine-Emanuele Cella
arxiv.org/abs/2510.08816

@arXiv_statML_bot@mastoxiv.page
2025-08-13 09:34:32

Bio-Inspired Artificial Neural Networks based on Predictive Coding
Davide Casnici, Charlotte Frenkel, Justin Dauwels
arxiv.org/abs/2508.08762

@arXiv_csLG_bot@mastoxiv.page
2025-08-12 12:08:03

Online Convex Optimization with Heavy Tails: Old Algorithms, New Regrets, and Applications
Zijian Liu
arxiv.org/abs/2508.07473 arxiv.org/pd…

@arXiv_quantph_bot@mastoxiv.page
2025-08-14 09:14:52

Discovery of energy landscapes towards optimized quantum transport: Environmental effects and long-range tunneling
Maggie Lawrence, Matthew Pocrnic, Erin Fung, Juan Carrasquilla, Erik M. Gauger, Dvira Segal
arxiv.org/abs/2508.09371

@arXiv_csCG_bot@mastoxiv.page
2025-10-13 07:31:30

Randomized HyperSteiner: A Stochastic Delaunay Triangulation Heuristic for the Hyperbolic Steiner Minimal Tree
Aniss Aiman Medbouhi, Alejandro Garc\'ia-Castellanos, Giovanni Luca Marchetti, Daniel Pelt, Erik J Bekkers, Danica Kragic
arxiv.org/abs/2510.09328

@arXiv_csCR_bot@mastoxiv.page
2025-09-09 11:56:52

PLRV-O: Advancing Differentially Private Deep Learning via Privacy Loss Random Variable Optimization
Qin Yang, Nicholas Stout, Meisam Mohammady, Han Wang, Ayesha Samreen, Christopher J Quinn, Yan Yan, Ashish Kundu, Yuan Hong
arxiv.org/abs/2509.06264

@arXiv_mathOC_bot@mastoxiv.page
2025-08-13 09:42:02

Projected Gradient Descent for Constrained Decision-Dependent Optimization
Zifan Wang, Changxin Liu, Thomas Parisini, Michael M. Zavlanos, Karl H. Johansson
arxiv.org/abs/2508.08856

@arXiv_csCL_bot@mastoxiv.page
2025-09-09 12:06:12

MachineLearningLM: Continued Pretraining Language Models on Millions of Synthetic Tabular Prediction Tasks Scales In-Context ML
Haoyu Dong, Pengkun Zhang, Mingzhe Lu, Yanzhen Shen, Guolin Ke
arxiv.org/abs/2509.06806

@arXiv_csGT_bot@mastoxiv.page
2025-10-07 07:46:28

On the $O(1/T)$ Convergence of Alternating Gradient Descent-Ascent in Bilinear Games
Tianlong Nan, Shuvomoy Das Gupta, Garud Iyengar, Christian Kroer
arxiv.org/abs/2510.03855

@arXiv_csLG_bot@mastoxiv.page
2025-08-12 11:51:13

SGD Convergence under Stepsize Shrinkage in Low-Precision Training
Vincent-Daniel Yun
arxiv.org/abs/2508.07142 arxiv.org/pdf/2508.07142

@arXiv_statML_bot@mastoxiv.page
2025-10-08 09:29:19

On the Theory of Continual Learning with Gradient Descent for Neural Networks
Hossein Taheri, Avishek Ghosh, Arya Mazumdar
arxiv.org/abs/2510.05573

@arXiv_mathOC_bot@mastoxiv.page
2025-08-12 11:12:53

Randomized coordinate gradient descent almost surely escapes strict saddle points
Ziang Chen, Yingzhou Li, Zihao Li
arxiv.org/abs/2508.07535

@arXiv_mathNA_bot@mastoxiv.page
2025-08-08 08:03:22

Toroidal area-preserving parameterizations of genus-one closed surfaces
Marco Sutti, Mei-Heng Yueh
arxiv.org/abs/2508.05111 arxiv.org/pdf/2…

@arXiv_eessAS_bot@mastoxiv.page
2025-08-14 09:05:02

Objective Soups: Multilingual Multi-Task Modeling for Speech Processing
A F M Saif, Lisha Chen, Xiaodong Cui, Songtao Lu, Brian Kingsbury, Tianyi Chen
arxiv.org/abs/2508.09228

@arXiv_csCE_bot@mastoxiv.page
2025-07-23 07:37:52

Multi-objective Portfolio Optimization Via Gradient Descent
Christian Oliva, Pedro R. Ventura, Luis F. Lago-Fern\'andez
arxiv.org/abs/2507.16717

@arXiv_condmatdisnn_bot@mastoxiv.page
2025-09-03 08:55:03

Phase diagram and eigenvalue dynamics of stochastic gradient descent in multilayer neural networks
Chanju Park (Swansea University), Biagio Lucini (Queen Mary University of London), Gert Aarts (Swansea University)
arxiv.org/abs/2509.01349

@arXiv_condmatstatmech_bot@mastoxiv.page
2025-09-22 08:23:41

Training thermodynamic computers by gradient descent
Stephen Whitelam
arxiv.org/abs/2509.15324 arxiv.org/pdf/2509.15324

@arXiv_eessSP_bot@mastoxiv.page
2025-09-12 08:51:19

JFRFFNet: A Data-Model Co-Driven Graph Signal Denoising Model with Partial Prior Information
Ziqi Yan, Zhichao Zhang
arxiv.org/abs/2509.09147

@arXiv_csIR_bot@mastoxiv.page
2025-08-07 09:14:53

Comparative Analysis of Novel NIRMAL Optimizer Against Adam and SGD with Momentum
Nirmal Gaud, Surej Mouli, Preeti Katiyar, Vaduguru Venkata Ramya
arxiv.org/abs/2508.04293

@arXiv_mathST_bot@mastoxiv.page
2025-09-30 08:57:31

Learning single index model with gradient descent: spectral initialization and precise asymptotics
Yuchen Chen, Yandi Shen
arxiv.org/abs/2509.23527

@arXiv_mathOC_bot@mastoxiv.page
2025-08-12 10:19:53

On the Convergence of a Noisy Gradient Method for Non-convex Distributed Resource Allocation: Saddle Point Escape
Lei Qin, Ye Pu
arxiv.org/abs/2508.06922

@arXiv_statME_bot@mastoxiv.page
2025-09-29 09:11:27

Federated Learning of Quantile Inference under Local Differential Privacy
Leheng Cai, Qirui Hu, Shuyuan Wu
arxiv.org/abs/2509.21800 arxiv.o…

@arXiv_csLG_bot@mastoxiv.page
2025-09-01 09:56:02

Convergence of Stochastic Gradient Methods for Wide Two-Layer Physics-Informed Neural Networks
Bangti Jin, Longjun Wu
arxiv.org/abs/2508.21571

@arXiv_statML_bot@mastoxiv.page
2025-10-03 09:42:31

Adaptive Kernel Selection for Stein Variational Gradient Descent
Moritz Melcher, Simon Weissmann, Ashia C. Wilson, Jakob Zech
arxiv.org/abs/2510.02067

@arXiv_mathAP_bot@mastoxiv.page
2025-09-04 09:30:11

Exponential ergodicity of mean-field Langevin dynamics by synchronous coupling
Mohamed Alfaki Aboubacrine Assadek (MATHSTIC, LAREMA)
arxiv.org/abs/2509.03124

@arXiv_nlinCG_bot@mastoxiv.page
2025-08-07 07:50:43

The Glider Equation for Asymptotic Lenia
Hiroki Kojima, Ivan Yevenko, Takashi Ikegami
arxiv.org/abs/2508.04167 arxiv.org/pdf/2508.04167

@arXiv_mathOC_bot@mastoxiv.page
2025-08-12 09:52:33

Near-Optimal Convergence of Accelerated Gradient Methods under Generalized and $(L_0, L_1)$-Smoothness
Alexander Tyurin
arxiv.org/abs/2508.06884

@arXiv_csLO_bot@mastoxiv.page
2025-09-25 12:40:17

Replaced article(s) found for cs.LO. arxiv.org/list/cs.LO/new
[1/1]:
- Compact Rule-Based Classifier Learning via Gradient Descent
Javier Fumanal-Idocin, Raquel Fernandez-Peralta, Javier Andreu-Perez

@arXiv_csLG_bot@mastoxiv.page
2025-10-08 10:26:09

Correlating Cross-Iteration Noise for DP-SGD using Model Curvature
Xin Gu, Yingtai Xiao, Guanlin He, Jiamu Bai, Daniel Kifer, Kiwan Maeng
arxiv.org/abs/2510.05416

@arXiv_mathOC_bot@mastoxiv.page
2025-09-12 07:56:29

Convexity of Optimization Curves: Local Sharp Thresholds, Robustness Impossibility, and New Counterexamples
Le Duc Hieu
arxiv.org/abs/2509.08954

@arXiv_eessSY_bot@mastoxiv.page
2025-08-29 09:22:11

Local Observability of a Class of Feedforward Neural Networks
Yi Yang, Victor G. Lopez, Matthias A. M\"uller
arxiv.org/abs/2508.20544

@arXiv_csCE_bot@mastoxiv.page
2025-10-07 07:47:18

Towards Fast Option Pricing PDE Solvers Powered by PIELM
Akshay Govind Srinivasan, Anuj Jagannath Said, Sathwik Pentela, Vikas Dwivedi, Balaji Srinivasan
arxiv.org/abs/2510.04322

@arXiv_statML_bot@mastoxiv.page
2025-09-30 08:48:11

Differentially Private Two-Stage Gradient Descent for Instrumental Variable Regression
Haodong Liang, Yanhao Jin, Krishnakumar Balasubramanian, Lifeng Lai
arxiv.org/abs/2509.22794

@arXiv_mathOC_bot@mastoxiv.page
2025-09-11 09:51:13

Linear Convergence of Gradient Descent for Quadratically Regularized Optimal Transport
Alberto Gonz\'alez-Sanz, Marcel Nutz, Andr\'es Riveros Valdevenito
arxiv.org/abs/2509.08547

@arXiv_csIT_bot@mastoxiv.page
2025-07-24 08:24:20

Information Entropy-Based Scheduling for Communication-Efficient Decentralized Learning
Jaiprakash Nagar, Zheng Chen, Marios Kountouris, Photios A. Stavrou
arxiv.org/abs/2507.17426

@arXiv_mathOC_bot@mastoxiv.page
2025-08-11 09:03:49

Kahan's Automatic Step-Size Control for Unconstrained Optimization
Yifeng Meng, Chungen Shen, Linuo Xue, Lei-Hong Zhang
arxiv.org/abs/2508.06002

@arXiv_statML_bot@mastoxiv.page
2025-07-29 09:42:31

Statistical Inference for Differentially Private Stochastic Gradient Descent
Xintao Xia, Linjun Zhang, Zhanrui Cai
arxiv.org/abs/2507.20560

@arXiv_csLG_bot@mastoxiv.page
2025-08-22 10:07:21

Hybrid Least Squares/Gradient Descent Methods for DeepONets
Jun Choi, Chang-Ock Lee, Minam Moon
arxiv.org/abs/2508.15394 arxiv.org/pdf/2508…

@arXiv_mathOC_bot@mastoxiv.page
2025-09-04 10:06:21

On the Perturbed Projection-Based Distributed Gradient-Descent Algorithm: A Fully-Distributed Adaptive Redesign
Tarek Bazizi, Mohamed Maghenem, Paolo Frasca, Antonio Lor\`ia, Elena Panteley
arxiv.org/abs/2509.03443

@arXiv_mathOC_bot@mastoxiv.page
2025-10-06 09:27:49

Quantitative Convergence Analysis of Projected Stochastic Gradient Descent for Non-Convex Losses via the Goldstein Subdifferential
Yuping Zheng, Andrew Lamperski
arxiv.org/abs/2510.02735

@arXiv_csLG_bot@mastoxiv.page
2025-08-27 10:34:03

GRADSTOP: Early Stopping of Gradient Descent via Posterior Sampling
Arash Jamshidi, Lauri Sepp\"al\"ainen, Katsiaryna Haitsiukevich, Hoang Phuc Hau Luu, Anton Bj\"orklund, Kai Puolam\"aki
arxiv.org/abs/2508.19028

@arXiv_eessSP_bot@mastoxiv.page
2025-08-22 09:05:21

Lightweight Gradient Descent Optimization for Mitigating Hardware Imperfections in RIS Systems
Pedro H. C. de Souza (National Institute of Telecommunications), Luiz A. M. Pereira (National Institute of Telecommunications), Faustino R. G\'omez (National Institute of Telecommunications), Elsa M. Mater\'on (National Institute of Telecommunications), Jorge Ricardo Mej\'ia-Salazar (National Institute of Telecommunications)

@arXiv_mathOC_bot@mastoxiv.page
2025-09-04 09:04:31

Stochastic versus Deterministic in Stochastic Gradient Descent
Runze Li, Jintao Xu, Wenxun Xing
arxiv.org/abs/2509.02912 arxiv.org/pdf/2509…

@arXiv_csAI_bot@mastoxiv.page
2025-09-18 09:08:41

From Next Token Prediction to (STRIPS) World Models -- Preliminary Results
Carlos N\'u\~nez-Molina, Vicen\c{c} G\'omez, Hector Geffner
arxiv.org/abs/2509.13389

@arXiv_csLG_bot@mastoxiv.page
2025-09-01 09:42:22

Adaptive Heavy-Tailed Stochastic Gradient Descent
Bodu Gong, Gustavo Enrique Batista, Pierre Lafaye de Micheaux
arxiv.org/abs/2508.21353 ar…

@arXiv_csCE_bot@mastoxiv.page
2025-10-03 11:05:23

Crosslisted article(s) found for cs.CE. arxiv.org/list/cs.CE/new
[1/1]:
- Fast training of accurate physics-informed neural networks without gradient descent
Datar, Kapoor, Chandra, Sun, Bolager, Burak, Veselovska, Fornasier, Dietrich

@arXiv_csLG_bot@mastoxiv.page
2025-08-18 09:39:00

Harmonized Gradient Descent for Class Imbalanced Data Stream Online Learning
Han Zhou, Hongpeng Yin, Xuanhong Deng, Yuyu Huang, Hao Ren
arxiv.org/abs/2508.11353

@arXiv_statML_bot@mastoxiv.page
2025-10-06 12:39:12

Replaced article(s) found for stat.ML. arxiv.org/list/stat.ML/new
[2/2]:
- Gradient Descent with Large Step Sizes: Chaos and Fractal Convergence Region
Shuang Liang, Guido Mont\'ufar

@arXiv_mathOC_bot@mastoxiv.page
2025-09-05 07:58:11

Towards understanding Accelerated Stein Variational Gradient Flow -- Analysis of Generalized Bilinear Kernels for Gaussian target distributions
Viktor Stein, Wuchen Li
arxiv.org/abs/2509.04008

@arXiv_csLG_bot@mastoxiv.page
2025-08-29 10:30:11

Fast Convergence Rates for Subsampled Natural Gradient Algorithms on Quadratic Model Problems
Gil Goldshlager, Jiang Hu, Lin Lin
arxiv.org/abs/2508.21022

@arXiv_statML_bot@mastoxiv.page
2025-09-29 08:58:37

Effective continuous equations for adaptive SGD: a stochastic analysis view
Luca Callisti, Marco Romito, Francesco Triggiano
arxiv.org/abs/2509.21614

@arXiv_csLG_bot@mastoxiv.page
2025-10-03 11:04:01

Equilibrium Matching: Generative Modeling with Implicit Energy-Based Models
Runqian Wang, Yilun Du
arxiv.org/abs/2510.02300 arxiv.org/pdf/2…

@arXiv_mathOC_bot@mastoxiv.page
2025-08-27 09:33:33

Adaptive control mechanisms in gradient descent algorithms
Andrea Iannelli
arxiv.org/abs/2508.19100 arxiv.org/pdf/2508.19100

@arXiv_statML_bot@mastoxiv.page
2025-09-23 09:37:40

Risk Comparisons in Linear Regression: Implicit Regularization Dominates Explicit Regularization
Jingfeng Wu, Peter L. Bartlett, Jason D. Lee, Sham M. Kakade, Bin Yu
arxiv.org/abs/2509.17251

@arXiv_mathOC_bot@mastoxiv.page
2025-09-11 09:55:13

Decentralized Stochastic Nonconvex Optimization under the Relaxed Smoothness
Luo Luo, Xue Cui, Tingkai Jia, Cheng Chen
arxiv.org/abs/2509.08726

@arXiv_csLG_bot@mastoxiv.page
2025-08-20 10:15:20

Fisher-Orthogonal Projection Methods for Natural Gradient Descent with Large Batches
Yishun Lu, Wesley Armour
arxiv.org/abs/2508.13898 arxi…

@arXiv_mathOC_bot@mastoxiv.page
2025-08-29 09:11:11

Revisit Stochastic Gradient Descent for Strongly Convex Objectives: Tight Uniform-in-Time Bounds
Kang Chen, Yasong Feng, Tianyu Wang
arxiv.org/abs/2508.20823

@arXiv_csLG_bot@mastoxiv.page
2025-08-22 10:12:21

Stabilization of Perturbed Loss Function: Differential Privacy without Gradient Noise
Salman Habib, Remi Chou, Taejoon Kim
arxiv.org/abs/2508.15523

@arXiv_mathOC_bot@mastoxiv.page
2025-09-18 09:06:51

Accelerated Gradient Methods with Biased Gradient Estimates: Risk Sensitivity, High-Probability Guarantees, and Large Deviation Bounds
Mert G\"urb\"uzbalaban, Yasa Syed, Necdet Serhat Aybat
arxiv.org/abs/2509.13628

@arXiv_csLG_bot@mastoxiv.page
2025-09-18 10:15:41

A Universal Banach--Bregman Framework for Stochastic Iterations: Unifying Stochastic Mirror Descent, Learning and LLM Training
Johnny R. Zhang (Independent Researcher), Xiaomei Mi (University of Manchester), Gaoyuan Du (Amazon), Qianyi Sun (Microsoft), Shiqi Wang (Meta), Jiaxuan Li (Amazon), Wenhua Zhou (Independent Researcher)
arx…

@arXiv_statML_bot@mastoxiv.page
2025-10-02 09:02:31

Guaranteed Noisy CP Tensor Recovery via Riemannian Optimization on the Segre Manifold
Ke Xu, Yuefeng Han
arxiv.org/abs/2510.00569 arxiv.org…

@arXiv_csLG_bot@mastoxiv.page
2025-09-22 10:33:31

DIVEBATCH: Accelerating Model Training Through Gradient-Diversity Aware Batch Size Adaptation
Yuen Chen, Yian Wang, Hari Sundaram
arxiv.org/abs/2509.16173

@arXiv_mathOC_bot@mastoxiv.page
2025-07-29 11:25:11

Stochastic gradient with least-squares control variates
Fabio Nobile, Matteo Raviola, Nathan Schaeffer
arxiv.org/abs/2507.20981 arxiv.org/p…

@arXiv_csLG_bot@mastoxiv.page
2025-08-25 09:58:50

Escaping Saddle Points via Curvature-Calibrated Perturbations: A Complete Analysis with Explicit Constants and Empirical Validation
Faruk Alpay, Hamdi Alakkad
arxiv.org/abs/2508.16540

@arXiv_mathOC_bot@mastoxiv.page
2025-07-21 08:59:50

Gradient descent avoids strict saddles with a simple line-search method too
Andreea-Alexandra Mu\c{s}at, Nicolas Boumal
arxiv.org/abs/2507.13804

@arXiv_statML_bot@mastoxiv.page
2025-08-01 08:25:01

A Smoothing Newton Method for Rank-one Matrix Recovery
Tyler Maunu, Gabriel Abreu
arxiv.org/abs/2507.23017 arxiv.org/pdf/2507.23017

@arXiv_mathOC_bot@mastoxiv.page
2025-07-22 11:29:20

Power-Constrained Policy Gradient Methods for LQR
Ashwin Verma, Aritra Mitra, Lintao Ye, Vijay Gupta
arxiv.org/abs/2507.15806

@arXiv_csLG_bot@mastoxiv.page
2025-08-21 10:17:00

Universal and Transferable Adversarial Attack on Large Language Models Using Exponentiated Gradient Descent
Sajib Biswas, Mao Nishino, Samuel Jacob Chacko, Xiuwen Liu
arxiv.org/abs/2508.14853

@arXiv_statML_bot@mastoxiv.page
2025-09-30 11:12:51

Quantitative convergence of trained single layer neural networks to Gaussian processes
Eloy Mosig, Andrea Agazzi, Dario Trevisan
arxiv.org/abs/2509.24544

@arXiv_mathOC_bot@mastoxiv.page
2025-08-08 08:18:02

Can SGD Handle Heavy-Tailed Noise?
Ilyas Fatkhullin, Florian H\"ubler, Guanghui Lan
arxiv.org/abs/2508.04860 arxiv.org/pdf/2508.04860

@arXiv_statML_bot@mastoxiv.page
2025-07-30 09:06:22

From Sublinear to Linear: Fast Convergence in Deep Networks via Locally Polyak-Lojasiewicz Regions
Agnideep Aich, Ashit Baran Aich, Bruce Wade
arxiv.org/abs/2507.21429

@arXiv_mathOC_bot@mastoxiv.page
2025-10-07 10:41:52

A Frank-Wolfe Algorithm for Strongly Monotone Variational Inequalities
Reza Rahimi Baghbadorani, Peyman Mohajerin Esfahani, Sergio Grammatico
arxiv.org/abs/2510.03842

@arXiv_csLG_bot@mastoxiv.page
2025-07-31 13:34:44

Replaced article(s) found for cs.LG. arxiv.org/list/cs.LG/new
[2/4]:
- Convergence Properties of Natural Gradient Descent for Minimizing KL Divergence
Adwait Datar, Nihat Ay

@arXiv_mathOC_bot@mastoxiv.page
2025-10-07 10:40:42

Learning Polynomial Activation Functions for Deep Neural Networks
Linghao Zhang, Jiawang Nie, Tingting Tang
arxiv.org/abs/2510.03682 arxiv.…

@arXiv_mathOC_bot@mastoxiv.page
2025-08-27 08:28:32

Norm-Constrained Flows and Sign-Based Optimization: Theory and Algorithms
Valentin Leplat, Sergio Mayorga, Roland Hildebrand, Alexander Gasnikov
arxiv.org/abs/2508.18510

@arXiv_csLG_bot@mastoxiv.page
2025-08-27 10:25:53

Stability and Generalization for Bellman Residuals
Enoch H. Kang, Kyoungseok Jang
arxiv.org/abs/2508.18741 arxiv.org/pdf/2508.18741

@arXiv_mathOC_bot@mastoxiv.page
2025-09-05 12:30:09

Replaced article(s) found for math.OC. arxiv.org/list/math.OC/new
[1/1]:
- FastPart: Over-Parameterized Stochastic Gradient Descent for Sparse optimisation on Measures
Yohann De Castro, S\'ebastien Gadat, Cl\'ement Marteau

@arXiv_mathOC_bot@mastoxiv.page
2025-07-29 11:16:41

Numerical Design of Optimized First-Order Algorithms
Yassine Kamri, Julien M. Hendrickx, Fran\c{c}ois Glineur
arxiv.org/abs/2507.20773 arxi…

@arXiv_csLG_bot@mastoxiv.page
2025-08-21 10:10:10

Cooperative SGD with Dynamic Mixing Matrices
Soumya Sarkar, Shweta Jain
arxiv.org/abs/2508.14565 arxiv.org/pdf/2508.14565

@arXiv_mathOC_bot@mastoxiv.page
2025-08-22 08:48:00

Controlled Optimization of Quadratic Functions in $\mathbb{R}^n$
Jean-Jacques Godeme
arxiv.org/abs/2508.15515 arxiv.org/pdf/2508.15515

@arXiv_csLG_bot@mastoxiv.page
2025-08-20 10:02:40

Explainable Learning Rate Regimes for Stochastic Optimization
Zhuang Yang
arxiv.org/abs/2508.13639 arxiv.org/pdf/2508.13639

@arXiv_mathOC_bot@mastoxiv.page
2025-09-29 08:23:57

Regularized Overestimated Newton
Danny Duan, Hanbaek Lyu
arxiv.org/abs/2509.21684 arxiv.org/pdf/2509.21684

@arXiv_mathOC_bot@mastoxiv.page
2025-09-29 15:15:58

Replaced article(s) found for math.OC. arxiv.org/list/math.OC/new
[1/1]:
- The Limit Points of (Optimistic) Gradient Descent in Min-Max Optimization
Constantinos Daskalakis, Ioannis Panageas

@arXiv_mathOC_bot@mastoxiv.page
2025-09-29 08:22:57

A regret minimization approach to fixed-point iterations
Joon Kwon
arxiv.org/abs/2509.21653 arxiv.org/pdf/2509.21653

@arXiv_mathOC_bot@mastoxiv.page
2025-08-29 09:23:31

Active-set Newton-MR methods for nonconvex optimization problems with bound constraints
Ernesto G. Birgin, Geovani N. Grapiglia, Diaulas S. Marcondes
arxiv.org/abs/2508.20967

@arXiv_mathOC_bot@mastoxiv.page
2025-08-26 10:55:36

Polyak Stepsize: Estimating Optimal Functional Values Without Parameters or Prior Knowledge
Farshed Abdukhakimov, Cuong Anh Pham, Samuel Horv\'ath, Martin Tak\'a\v{c}, Slavom{\i}r Hanzely
arxiv.org/abs/2508.17288