
2025-06-11 08:43:15
Stochastic gradient descent based variational inference for infinite-dimensional inverse problems
Jiaming Sui, Junxiong Jia, Jinglai Li
https://arxiv.org/abs/2506.08380
Stochastic gradient descent based variational inference for infinite-dimensional inverse problems
Jiaming Sui, Junxiong Jia, Jinglai Li
https://arxiv.org/abs/2506.08380
Online Quantum State Tomography via Stochastic Gradient Descent
Jian-Feng Cai, Yuling Jiao, Yinan Li, Xiliang Lu, Jerry Zhijian Yang, Juntao You
https://arxiv.org/abs/2507.07601
Almost Sure Convergence for the Last Iterate of Stochastic Gradient Descent Schemes
Marcel Hudiani
https://arxiv.org/abs/2507.07281 https://
An Adaptive Order Caputo Fractional Gradient Descent Method for Multi-objective Optimization Problems
Barsha Shaw, Md Abu Talhamainuddin Ansary
https://arxiv.org/abs/2507.07674
This https://arxiv.org/abs/2409.08469 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_mat…
Gradient Similarity Surgery in Multi-Task Deep Learning
Thomas Borsani, Andrea Rosani, Giuseppe Nicosia, Giuseppe Di Fatta
https://arxiv.org/abs/2506.06130
Non-Asymptotic Analysis of Online Local Private Learning with SGD
Enze Shi, Jinhan Xie, Bei Jiang, Linglong Kong, Xuming He
https://arxiv.org/abs/2507.07041
Stochastic Gradient-Descent Calibration of Pyragas Delayed-Feedback Control for Chaos Suppression in the Sprott Circuit
Adib Kabir, Onil Morshed, Oishi Kabir
https://arxiv.org/abs/2506.06639
This https://arxiv.org/abs/2504.14730 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csIT_…
Decentralized Optimization with Amplified Privacy via Efficient Communication
Wei Huo, Changxin Liu, Kemi Ding, Karl Henrik Johansson, Ling Shi
https://arxiv.org/abs/2506.07102
Can Biologically Plausible Temporal Credit Assignment Rules Match BPTT for Neural Similarity? E-prop as an Example
Yuhan Helena Liu, Guangyu Robert Yang, Christopher J. Cueva
https://arxiv.org/abs/2506.06904
Simple Convergence Proof of Adam From a Sign-like Descent Perspective
Hanyang Peng, Shuang Qin, Yue Yu, Fangqing Jiang, Hui Wang, Zhouchen Lin
https://arxiv.org/abs/2507.05966
Linear Discriminant Analysis with Gradient Optimization on Covariance Inverse
Cencheng Shen, Yuexiao Dong
https://arxiv.org/abs/2506.06845 https://
This https://arxiv.org/abs/2503.16398 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_mat…
Kahan's Automatic Step-Size Control for Unconstrained Optimization
Yifeng Meng, Chungen Shen, Linuo Xue, Lei-Hong Zhang
https://arxiv.org/abs/2508.06002 https://
Constrained Stein Variational Gradient Descent for Robot Perception, Planning, and Identification
Griffin Tabor, Tucker Hermans
https://arxiv.org/abs/2506.00589
This https://arxiv.org/abs/2505.17282 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csLG_…
This month's newsletter/digest is on it's way to the email box of all the amazing friendly good people who asked for it.
The rest of you bitchy vicious augmentative people in public spats can read it here:
#newsletter #digest
Toroidal area-preserving parameterizations of genus-one closed surfaces
Marco Sutti, Mei-Heng Yueh
https://arxiv.org/abs/2508.05111 https://arxiv.org/pdf/2…
Statistical Inference for Differentially Private Stochastic Gradient Descent
Xintao Xia, Linjun Zhang, Zhanrui Cai
https://arxiv.org/abs/2507.20560 https://
This https://arxiv.org/abs/2503.14353 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_ees…
A Dynamical Systems Perspective on the Analysis of Neural Networks
Dennis Chemnitz, Maximilian Engel, Christian Kuehn, Sara-Viola Kuntz
https://arxiv.org/abs/2507.05164
Comparative Analysis of Novel NIRMAL Optimizer Against Adam and SGD with Momentum
Nirmal Gaud, Surej Mouli, Preeti Katiyar, Vaduguru Venkata Ramya
https://arxiv.org/abs/2508.04293
On the Inherent Privacy of Zeroth Order Projected Gradient Descent
Devansh Gupta, Meisam Razaviyayn, Vatsal Sharan
https://arxiv.org/abs/2507.05610 https:/…
Solving the Gross-Pitaevskii Equation with Quantic Tensor Trains: Ground States and Nonlinear Dynamics
Qian-Can Chen, I-Kang Liu, Jheng-Wei Li, Chia-Min Chung
https://arxiv.org/abs/2507.04279
Riemannian Inexact Gradient Descent for Quadratic Discrimination
Uday Talwar, Meredith K. Kupinski, Afrooz Jalilzadeh
https://arxiv.org/abs/2507.04670 http…
A gradient flow that is none: Heat flow with Wentzell boundary condition
Marie Bormann, L\'eonard Monsaingeon, D. R. Michiel Renger, Max von Renesse
https://arxiv.org/abs/2506.22093
Fast and Provable Hankel Tensor Completion for Multi-measurement Spectral Compressed Sensing
Jinsheng Li, Xu Zhang, Shuang Wu, Wei Cui
https://arxiv.org/abs/2507.04847
Multi-objective Portfolio Optimization Via Gradient Descent
Christian Oliva, Pedro R. Ventura, Luis F. Lago-Fern\'andez
https://arxiv.org/abs/2507.16717 https://
On gradient descent-ascent flows in metric spaces
Noboru Isobe, Sho Shimoyama
https://arxiv.org/abs/2506.20258 https://arxiv.org/pdf/…
On the Benefits of Accelerated Optimization in Robust and Private Estimation
Laurentiu Andrei Marchis, Po-Ling Loh
https://arxiv.org/abs/2506.03044 https:/…
Generalized Gradient Norm Clipping & Non-Euclidean $(L_0,L_1)$-Smoothness
Thomas Pethick, Wanyun Xie, Mete Erdogan, Kimon Antonakopoulos, Tony Silveti-Falls, Volkan Cevher
https://arxiv.org/abs/2506.01913
The Glider Equation for Asymptotic Lenia
Hiroki Kojima, Ivan Yevenko, Takashi Ikegami
https://arxiv.org/abs/2508.04167 https://arxiv.org/pdf/2508.04167
Enhancing Biosecurity in Tamper-Resistant Large Language Models With Quantum Gradient Descent
Fahmida Hai, Saif Nirzhor, Rubayat Khan, Don Roosan
https://arxiv.org/abs/2506.19086 …
Multilevel Stochastic Gradient Descent for Optimal Control Under Uncertainty
Niklas Baumgarten, David Schneiderhan
https://arxiv.org/abs/2506.02647 https:/…
A Proximal Variable Smoothing for Minimization of Nonlinearly Composite Nonsmooth Function -- Maxmin Dispersion and MIMO Applications
Keita Kume, Isao Yamada
https://arxiv.org/abs/2506.05974
Antenna Q-Factor Topology Optimization with Auxiliary Edge Resistivities
Stepan Bosak, Miloslav Capek, Jiri Matas
https://arxiv.org/abs/2506.00595 https://…
A Parallelizable Approach for Characterizing NE in Zero-Sum Games After a Linear Number of Iterations of Gradient Descent
Taemin Kim, James P. Bailey
https://arxiv.org/abs/2507.11366
Experimental Multiport-Network Parameter Estimation and Optimization for Multi-Bit RIS
Philipp del Hougne
https://arxiv.org/abs/2507.02168 https://
Replaced article(s) found for stat.ML. https://arxiv.org/list/stat.ML/new
[1/1]:
- Rank-1 Matrix Completion with Gradient Descent and Small Random Initialization
Daesung Kim, Hye Won Chung
Rethinking LLM Training through Information Geometry and Quantum Metrics
Riccardo Di Sipio
https://arxiv.org/abs/2506.15830 https://a…
Perturbed Gradient Descent Algorithms are Small-Disturbance Input-to-State Stable
Leilei Cui, Zhong-Ping Jiang, Eduardo D. Sontag, Richard D. Braatz
https://arxiv.org/abs/2507.02131
Emergent universal long-range structure in random-organizing systems
Satyam Anand, Guanming Zhang, Stefano Martiniani
https://arxiv.org/abs/2505.22933 http…
Multilevel Bregman Proximal Gradient Descent
Yara Elshiaty, Stefania Petra
https://arxiv.org/abs/2506.03950 https://arxiv.org/pdf/250…
Extreme Learning Machines for Exoplanet Simulations: A Faster, Lightweight Alternative to Deep Learning
Tara P. A. Tahseen, Lu\'is F. Sim\~oes, Kai Hou Yip, Nikolaos Nikolaou, Jo\~ao M. Mendon\c{c}a, Ingo P. Waldmann
https://arxiv.org/abs/2506.19679
This https://arxiv.org/abs/2502.03701 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_mat…
A Smoothing Newton Method for Rank-one Matrix Recovery
Tyler Maunu, Gabriel Abreu
https://arxiv.org/abs/2507.23017 https://arxiv.org/pdf/2507.23017
Can SGD Handle Heavy-Tailed Noise?
Ilyas Fatkhullin, Florian H\"ubler, Guanghui Lan
https://arxiv.org/abs/2508.04860 https://arxiv.org/pdf/2508.04860
Replaced article(s) found for cs.LG. https://arxiv.org/list/cs.LG/new
[4/5]:
- SPGD: Steepest Perturbed Gradient Descent Optimization
Amir M. Vahedi, Horea T. Ilies
This https://arxiv.org/abs/2502.16492 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_mat…
SLASH: Self-Supervised Speech Pitch Estimation Leveraging DSP-derived Absolute Pitch
Ryo Terashima, Yuma Shirahata, Masaya Kawamura
https://arxiv.org/abs/2507.17208 https://
From Sublinear to Linear: Fast Convergence in Deep Networks via Locally Polyak-Lojasiewicz Regions
Agnideep Aich, Ashit Baran Aich, Bruce Wade
https://arxiv.org/abs/2507.21429 h…
Learning to Solve Parametric Mixed-Integer Optimal Control Problems via Differentiable Predictive Control
J\'an Boldock\'y, Shahriar Dadras Javan, Martin Gulan, Martin M\"onnigmann, J\'an Drgo\v{n}a
https://arxiv.org/abs/2506.19646
A Study of Hybrid and Evolutionary Metaheuristics for Single Hidden Layer Feedforward Neural Network Architecture
Gautam Siddharth Kashyap, Md Tabrez Nafis, Samar Wazir
https://arxiv.org/abs/2506.15737
On Universality of Non-Separable Approximate Message Passing Algorithms
Max Lovig, Tianhao Wang, Zhou Fan
https://arxiv.org/abs/2506.23010 https://
Worst-case convergence analysis of relatively inexact gradient descent on smooth convex functions
Pierre Vernimmen, Fran\c{c}ois Glineur
https://arxiv.org/abs/2506.17145
Faithful-Newton Framework: Bridging Inner and Outer Solvers for Enhanced Optimization
Alexander Lim, Fred Roosta
https://arxiv.org/abs/2506.13154 https://
Replaced article(s) found for cs.LG. https://arxiv.org/list/cs.LG/new
[2/4]:
- Convergence Properties of Natural Gradient Descent for Minimizing KL Divergence
Adwait Datar, Nihat Ay
Efficient Feedback Design for Unsourced Random Access with Integrated Sensing and Communication
Mohammad Javad Ahmadi, Mohammad Kazemi, Rafael F. Schaefer
https://arxiv.org/abs/2506.20262
Stable Minima of ReLU Neural Networks Suffer from the Curse of Dimensionality: The Neural Shattering Phenomenon
Tongtong Liang, Dan Qiao, Yu-Xiang Wang, Rahul Parhi
https://arxiv.org/abs/2506.20779
Beyond Rate Coding: Surrogate Gradients Enable Spike Timing Learning in Spiking Neural Networks
Ziqiao Yu, Pengfei Sun, Dan F. M. Goodman
https://arxiv.org/abs/2507.16043 https:…
Information Entropy-Based Scheduling for Communication-Efficient Decentralized Learning
Jaiprakash Nagar, Zheng Chen, Marios Kountouris, Photios A. Stavrou
https://arxiv.org/abs/2507.17426
Stochastic gradient with least-squares control variates
Fabio Nobile, Matteo Raviola, Nathan Schaeffer
https://arxiv.org/abs/2507.20981 https://arxiv.org/p…
Gradient descent avoids strict saddles with a simple line-search method too
Andreea-Alexandra Mu\c{s}at, Nicolas Boumal
https://arxiv.org/abs/2507.13804 ht…
Power-Constrained Policy Gradient Methods for LQR
Ashwin Verma, Aritra Mitra, Lintao Ye, Vijay Gupta
https://arxiv.org/abs/2507.15806 https://
Random feature approximation for general spectral methods
Mike Nguyen, Nicole M\"ucke
https://arxiv.org/abs/2506.16283 https://a…
An Enhanced Privacy-preserving Federated Few-shot Learning Framework for Respiratory Disease Diagnosis
Ming Wang, Zhaoyang Duan, Dong Xue, Fangzhou Liu, Zhongheng Zhang
https://arxiv.org/abs/2507.08050 https://arxiv.org/pdf/2507.08050 https://arxiv.org/html/2507.08050
arXiv:2507.08050v1 Announce Type: new
Abstract: The labor-intensive nature of medical data annotation presents a significant challenge for respiratory disease diagnosis, resulting in a scarcity of high-quality labeled datasets in resource-constrained settings. Moreover, patient privacy concerns complicate the direct sharing of local medical data across institutions, and existing centralized data-driven approaches, which rely on amounts of available data, often compromise data privacy. This study proposes a federated few-shot learning framework with privacy-preserving mechanisms to address the issues of limited labeled data and privacy protection in diagnosing respiratory diseases. In particular, a meta-stochastic gradient descent algorithm is proposed to mitigate the overfitting problem that arises from insufficient data when employing traditional gradient descent methods for neural network training. Furthermore, to ensure data privacy against gradient leakage, differential privacy noise from a standard Gaussian distribution is integrated into the gradients during the training of private models with local data, thereby preventing the reconstruction of medical images. Given the impracticality of centralizing respiratory disease data dispersed across various medical institutions, a weighted average algorithm is employed to aggregate local diagnostic models from different clients, enhancing the adaptability of a model across diverse scenarios. Experimental results show that the proposed method yields compelling results with the implementation of differential privacy, while effectively diagnosing respiratory diseases using data from different structures, categories, and distributions.
toXiv_bot_toot
Keep the beat going: Automatic drum transcription with momentum
Alisha L. Foster, Robert J. Webber
https://arxiv.org/abs/2507.12596 https://
High Probability Convergence of Distributed Clipped Stochastic Gradient Descent with Heavy-tailed Noise
Yuchen Yang, Kaihong Lu, Long Wang
https://arxiv.org/abs/2506.11647
Replaced article(s) found for math.OC. https://arxiv.org/list/math.OC/new
[1/1]:
- Low-rank optimization methods based on projected-projected gradient descent that accumulate at Bo...
Guillaume Olikier, Kyle A. Gallivan, P. -A. Absil
This https://arxiv.org/abs/2505.08408 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_mat…
This https://arxiv.org/abs/2505.08408 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_mat…
Replaced article(s) found for stat.CO. https://arxiv.org/list/stat.CO/new
[1/1]:
- Error bounds for particle gradient descent, and extensions of the log-Sobolev and Talagrand inequ...
Rocco Caprio, Juan Kuntz, Samuel Power, Adam M. Johansen
Primal-Dual Coordinate Descent for Nonconvex-Nonconcave Saddle Point Problems Under the Weak MVI Assumption
Iyad Walwil, Olivier Fercoq
https://arxiv.org/abs/2506.15597
Efficient Online Mirror Descent Stochastic Approximation for Multi-Stage Stochastic Programming
Junhui Zhang, Patrick Jaillet
https://arxiv.org/abs/2506.15392
This https://arxiv.org/abs/2401.06738 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_mat…
Numerical Design of Optimized First-Order Algorithms
Yassine Kamri, Julien M. Hendrickx, Fran\c{c}ois Glineur
https://arxiv.org/abs/2507.20773 https://arxi…
Replaced article(s) found for math.OC. https://arxiv.org/list/math.OC/new
[1/2]:
- Projected gradient descent accumulates at Bouligand stationary points
Guillaume Olikier, Ir\`ene Waldspurger
Non-smooth stochastic gradient descent using smoothing functions
Tommaso Giovannelli, Jingfu Tan, Luis Nunes Vicente
https://arxiv.org/abs/2507.10901 https…
Deep Equilibrium models for Poisson Imaging Inverse problems via Mirror Descent
Christian Daniele, Silvia Villa, Samuel Vaiter, Luca Calatroni
https://arxiv.org/abs/2507.11461
Empirical and computer-aided robustness analysis of long-step and accelerated methods in smooth convex optimization
Pierre Vernimmen, Fran\c{c}ois Glineur
https://arxiv.org/abs/2506.09730
Learning Acceleration Algorithms for Fast Parametric Convex Optimization with Certified Robustness
Rajiv Sambharya, Jinho Bok, Nikolai Matni, George Pappas
https://arxiv.org/abs/2507.16264
Last-Iterate Complexity of SGD for Convex and Smooth Stochastic Problems
Guillaume Garrigos, Daniel Cortild, Lucas Ketels, Juan Peypouquet
https://arxiv.org/abs/2507.14122
An Extended Variational Barzilai-Borwein Method
Xin Xu
https://arxiv.org/abs/2506.12731 https://arxiv.org/pdf/2506.12731