Tootfinder

Opt-in global Mastodon full text search. Join the index!

@arXiv_csLG_bot@mastoxiv.page
2025-09-22 10:33:31

DIVEBATCH: Accelerating Model Training Through Gradient-Diversity Aware Batch Size Adaptation
Yuen Chen, Yian Wang, Hari Sundaram
arxiv.org/abs/2509.16173

@arXiv_eessSP_bot@mastoxiv.page
2025-09-22 09:36:41

Scalable Hessian-free Proximal Conjugate Gradient Method for Nonconvex and Nonsmooth Optimization
Yiming Zhou, Wei Dai
arxiv.org/abs/2509.15973

@arXiv_condmatstatmech_bot@mastoxiv.page
2025-09-22 08:23:41

Training thermodynamic computers by gradient descent
Stephen Whitelam
arxiv.org/abs/2509.15324 arxiv.org/pdf/2509.15324

@arXiv_physicsfludyn_bot@mastoxiv.page
2025-09-22 08:39:31

Assessment of the Gradient Jump Penalisation in Large-Eddy Simulations of Turbulence
Shiyu Du, Manuel M\"unsch, Niclas Jansson, Philipp Schlatter
arxiv.org/abs/2509.16013

@arXiv_mathNA_bot@mastoxiv.page
2025-09-22 08:08:41

Variable-preconditioned transformed primal-dual method for generalized Wasserstein Gradient Flows
Jin Zeng, Dawei Zhan, Ruchi Guo, Chaozhen Wei
arxiv.org/abs/2509.15385

@beyondwatts@beyondwatts.social
2025-12-21 07:46:33

Shortest Day 😃

A serene view of the ocean at dusk, with a gradient sky transitioning from blue to warm hues near the horizon. Subtle waves can be seen on the water, along with distant silhouettes of structures and poles.
@arXiv_csCL_bot@mastoxiv.page
2025-09-22 10:05:21

Sparse-Autoencoder-Guided Internal Representation Unlearning for Large Language Models
Tomoya Yamashita, Akira Ito, Yuuki Yamanaka, Masanori Yamada, Takayuki Miura, Toshiki Shibahara
arxiv.org/abs/2509.15631

@arXiv_quantph_bot@mastoxiv.page
2025-09-22 10:06:01

Training Variational Quantum Circuits Using Particle Swarm Optimization
Marco Mordacci, Michele Amoretti
arxiv.org/abs/2509.15726 arxiv.org…

@arXiv_mathOC_bot@mastoxiv.page
2025-09-22 08:43:21

Escaping saddle points without Lipschitz smoothness: the power of nonlinear preconditioning
Alexander Bodard, Panagiotis Patrinos
arxiv.org/abs/2509.15817

@arXiv_eessIV_bot@mastoxiv.page
2025-09-22 08:18:31

Analysis Plug-and-Play Methods for Imaging Inverse Problems
Edward P. Chandler, Shirin Shoushtari, Brendt Wohlberg, Ulugbek S. Kamilov
arxiv.org/abs/2509.15422

@arXiv_csCE_bot@mastoxiv.page
2025-09-22 12:05:47

Replaced article(s) found for cs.CE. arxiv.org/list/cs.CE/new
[1/1]:
- A comparative analysis for different finite element types in strain-gradient elasticity simulatio...
B. Cagri Sarar, M. Erden Yildizdag, Francesco Fabbrocino, B. Emek Abali

@arXiv_csSD_bot@mastoxiv.page
2025-09-22 09:59:11

Differentiable Acoustic Radiance Transfer
Sungho Lee, Matteo Scerbo, Seungu Han, Min Jun Choi, Kyogu Lee, Enzo De Sena
arxiv.org/abs/2509.15946

@toxi@mastodon.thi.ng
2025-12-16 09:54:11

Recursive polygon subdivision inspired by thin-section mineralogy...
(The area of each polygon is mapped to a color from a gradient. Made with thi.ng/geom, see next message for example & source code...)
1/2

Stop frame animation of a randomized abstract composition of initially thousands of small polygons, slowly converging into only a handful of larger cells/shards. The animation shows the recursive subdivision process in reverse order, i.e. the larger cells at the end are actually some of the first polygons created by randomly slicing the seed polygon (a circular 40-gon). The area/size of each individual poly is mapped to a color from a gradient, with small polys in orange/yellow/pink and large o…
@arXiv_physicsoptics_bot@mastoxiv.page
2025-09-22 09:28:01

The critical role of substrates in mitigating the power-efficiency trade-off in near-field thermophotovoltaics
Kartika N. Nimje, Julien Legendre, Michela F. Picardi, Alejandro W. Rodriguez, Georgia T. Papadakis
arxiv.org/abs/2509.16048

@Techmeme@techhub.social
2025-09-29 14:05:45

Google rolls out its new gradient "G" icon company-wide, saying it "now represents all of Google ... and visually reflects our evolution in the AI era" (Abner Li/9to5Google)
9to5google.com/2025/09/29/goog

@arXiv_csLG_bot@mastoxiv.page
2025-09-22 10:33:51

Inverting Trojans in LLMs
Zhengxing Li, Guangmingmei Yang, Jayaram Raghuram, David J. Miller, George Kesidis
arxiv.org/abs/2509.16203 arxiv…

@arXiv_physicsspaceph_bot@mastoxiv.page
2025-09-22 08:24:31

Particle in cell simulation on mode conversion of Saturn's 20 kHz narrowband radio emission
Zhoufan Mu, Yao Chen, Tangmu Li, Sulan Ni, Zilong Zhang, Hao Ning
arxiv.org/abs/2509.15542

@cosmos4u@scicomm.xyz
2025-12-18 01:24:45

Ionospheric gradient estimation using ground-based GEO observations for monitoring multi-scale ionospheric dynamics: #Ionosphere in motion - a new way to track space weather in real time: eurekalert.org/news-releases/1

@arXiv_statML_bot@mastoxiv.page
2025-10-15 09:49:02

Statistical Guarantees for High-Dimensional Stochastic Gradient Descent
Jiaqi Li, Zhipeng Lou, Johannes Schmidt-Hieber, Wei Biao Wu
arxiv.org/abs/2510.12013

@arXiv_mathOC_bot@mastoxiv.page
2025-09-22 09:11:01

A generalized canonical metric for optimization on the indefinite Stiefel manifold
Dinh Van Tiep, Duong Thi Viet An, Nguyen Thi Ngoc Oanh, Nguyen Thanh Son
arxiv.org/abs/2509.16113

@arXiv_mathSG_bot@mastoxiv.page
2025-10-14 08:14:58

An Invitation to Obstruction Bundle Gluing Through Morse Flow Lines
Ipsita Datta, Yuan Yao
arxiv.org/abs/2510.10393 arxiv.org/pdf/2510.1039…

@arXiv_csLG_bot@mastoxiv.page
2025-09-22 10:32:11

Dynamic Classifier-Free Diffusion Guidance via Online Feedback
Pinelopi Papalampidi, Olivia Wiles, Ira Ktena, Aleksandar Shtedritski, Emanuele Bugliarello, Ivana Kajic, Isabela Albuquerque, Aida Nematzadeh
arxiv.org/abs/2509.16131

@arXiv_hepph_bot@mastoxiv.page
2025-10-15 09:58:12

Gradient-flowed operator product expansion without IR renormalons
Martin Beneke (TU Munich), Hiromasa Takaura (Kyoto University)
arxiv.org/abs/2510.12193

@arXiv_csSD_bot@mastoxiv.page
2025-09-22 10:01:11

Reverse Engineering of Music Mixing Graphs with Differentiable Processors and Iterative Pruning
Sungho Lee, Marco Mart\'inez-Ram\'irez, Wei-Hsiang Liao, Stefan Uhlich, Giorgio Fabbro, Kyogu Lee, Yuki Mitsufuji
arxiv.org/abs/2509.15948

@arXiv_mathAP_bot@mastoxiv.page
2025-10-14 10:33:18

Optimal gradient estimates for conductivity problems with imperfect low-conductivity interfaces
Hongjie Dong, Haigang Li, Yan Zhao
arxiv.org/abs/2510.10615

@arXiv_csRO_bot@mastoxiv.page
2025-10-07 11:43:02

Building Gradient by Gradient: Decentralised Energy Functions for Bimanual Robot Assembly
Alexander L. Mitchell, Joe Watson, Ingmar Posner
arxiv.org/abs/2510.04696

@arXiv_mathDG_bot@mastoxiv.page
2025-10-08 08:33:39

On curvature estimates for four-dimensional gradient Ricci solitons
Huai-Dong Cao
arxiv.org/abs/2510.06059 arxiv.org/pdf/2510.06059

@arXiv_mathOC_bot@mastoxiv.page
2025-09-22 08:25:11

Introducing the method of ellipcenters, a new first order technique for unconstrained optimization
Roger Behling, Ramyro Aquines Correa, Eduarda Ferreira Zanatta, Vincent Guigues
arxiv.org/abs/2509.15471

@arXiv_statCO_bot@mastoxiv.page
2025-09-22 12:32:56

Replaced article(s) found for stat.CO. arxiv.org/list/stat.CO/new
[1/1]:
- Gradient-Free Sequential Bayesian Experimental Design via Interacting Particle Systems
Robert Gruhlke, Matei Hanu, Claudia Schillings, Philipp Wacker

@arXiv_csCL_bot@mastoxiv.page
2025-10-13 10:39:20

SPG: Sandwiched Policy Gradient for Masked Diffusion Language Models
Chengyu Wang, Paria Rashidinejad, DiJia Su, Song Jiang, Sid Wang, Siyan Zhao, Cai Zhou, Shannon Zejiang Shen, Feiyu Chen, Tommi Jaakkola, Yuandong Tian, Bo Liu
arxiv.org/abs/2510.09541

@arXiv_statME_bot@mastoxiv.page
2025-10-13 09:25:10

Reliability Sensitivity with Response Gradient
Siu-Kui Au, Zi-Jun Cao
arxiv.org/abs/2510.09315 arxiv.org/pdf/2510.09315

@arXiv_condmatmeshall_bot@mastoxiv.page
2025-10-09 09:32:21

Thermal gradient-driven skyrmion dynamics with near-zero skyrmion Hall angle
Yogesh Kumar, Hurmal Saren, Pintu Das
arxiv.org/abs/2510.07020

@arXiv_mathNA_bot@mastoxiv.page
2025-10-14 10:54:08

Forward and backward error bounds for a mixed precision preconditioned conjugate gradient algorithm
Thomas Bake, Erin Carson, Yuxin Ma
arxiv.org/abs/2510.11379

@arXiv_statML_bot@mastoxiv.page
2025-10-13 08:27:00

Gradient-Guided Furthest Point Sampling for Robust Training Set Selection
Morris Trestman, Stefan Gugler, Felix A. Faber, O. A. von Lilienfeld
arxiv.org/abs/2510.08906

@arXiv_eessSY_bot@mastoxiv.page
2025-09-30 11:15:01

Small-Covariance Noise-to-State Stability of Stochastic Systems and Its Applications to Stochastic Gradient Dynamics
Leilei Cui, Zhong-Ping Jiang, Eduardo D. Sontag
arxiv.org/abs/2509.24277

@arXiv_csGT_bot@mastoxiv.page
2025-10-07 07:46:28

On the $O(1/T)$ Convergence of Alternating Gradient Descent-Ascent in Bilinear Games
Tianlong Nan, Shuvomoy Das Gupta, Garud Iyengar, Christian Kroer
arxiv.org/abs/2510.03855

@arXiv_csCR_bot@mastoxiv.page
2025-10-09 09:21:21

Reading Between the Lines: Towards Reliable Black-box LLM Fingerprinting via Zeroth-order Gradient Estimation
Shuo Shao, Yiming Li, Hongwei Yao, Yifei Chen, Yuchen Yang, Zhan Qin
arxiv.org/abs/2510.06605

@arXiv_csCV_bot@mastoxiv.page
2025-10-14 22:04:05

Replaced article(s) found for cs.CV. arxiv.org/list/cs.CV/new
[4/8]:
- Boosting Adversarial Transferability via Commonality-Oriented Gradient Optimization
Yanting Gao, Yepeng Liu, Junming Liu, Qi Zhang, Hongyun Zhang, Duoqian Miao, Cairong Zhao

@arXiv_mathAP_bot@mastoxiv.page
2025-10-15 09:54:41

Liouville results for $(p,q)$-Laplacian elliptic equations with source terms involving gradient nonlinearities
Mousomi Bhakta, Anup Biswas, Roberta Filippucci
arxiv.org/abs/2510.12486

@arXiv_physicsmedph_bot@mastoxiv.page
2025-10-07 09:13:32

Human brain high-resolution diffusion MRI with optimized slice-by-slice B0 field shimming in head-only high-performance gradient MRI systems
Patricia Lan, Sherry S. Huang, Chitresh Bhushan, Xinzeng Wang, Seung-Kyun Lee, Raymond Y. Huang, Jerome J. Maller, Jennifer A. McNab, Ante Zhu
arxiv.org/abs/2510.03586

@arXiv_mathOC_bot@mastoxiv.page
2025-10-14 11:35:58

Nonlinearly Preconditioned Gradient Methods: Momentum and Stochastic Analysis
Konstantinos Oikonomidis, Jan Quan, Panagiotis Patrinos
arxiv.org/abs/2510.11312

@arXiv_qbioNC_bot@mastoxiv.page
2025-10-09 09:04:41

Gradient of White Matter Functional Variability via fALFF Differential Identifiability
Xinle Chang, Yang Yang, Yueran Li, Zhengcen Li, Haijin Zeng, Jingyong Su
arxiv.org/abs/2510.06914

@arXiv_astrophGA_bot@mastoxiv.page
2025-09-30 09:39:31

A gradient boosting and broadband approach to finding Lyman-{\alpha} emitting galaxies beyond narrowband surveys
A. Vale, A. Paulino-Afonso, A. Humphrey, P. A. C. Cunha, B. Ribeiro, B. Cerqueira, R. Carvajal, J. Fonseca
arxiv.org/abs/2509.22915

@arXiv_mathDG_bot@mastoxiv.page
2025-10-09 08:19:10

Stability of asymptotically conical gradient K\"ahler-Ricci expanders
Longteng Chen
arxiv.org/abs/2510.06850 arxiv.org/pdf/2510.06850

@arXiv_quantph_bot@mastoxiv.page
2025-10-13 09:24:10

Statistical Benchmarking of Optimization Methods for Variational Quantum Eigensolver under Quantum Noise
Silvie Ill\'esov\'a, Tom\'a\v{s} Bezd\v{e}k, Vojt\v{e}ch Nov\'ak, Bruno Senjean, Martin Beseda
arxiv.org/abs/2510.08727

@arXiv_mathOC_bot@mastoxiv.page
2025-10-15 10:19:31

Temporal Variabilities Limit Convergence Rates in Gradient-Based Online Optimization
Bryan Van Scoy, Gianluca Bianchin
arxiv.org/abs/2510.12512

@arXiv_condmatstatmech_bot@mastoxiv.page
2025-10-10 07:57:48

Thermodynamically Consistent Continuum Theory of Magnetic Particles in High-Gradient Fields
Marko Tesanovic, Daniel M. Markiewitz, Marcus L. Popp, Martin Z. Bazant, Sonja Berensmeier
arxiv.org/abs/2510.07552

@cosmos4u@scicomm.xyz
2025-12-09 17:57:01

ALMA Reveals an Eccentricity Gradient in the #Fomalhaut Debris Disk: iopscience.iop.org/article/10. -> A Planet Carving the Fomalhaut Debris Disk? aasnova.org/2025/12/09/michela

@arXiv_statML_bot@mastoxiv.page
2025-09-30 09:30:11

Statistical Inference for Gradient Boosting Regression
Haimo Fang, Kevin Tan, Giles Hooker
arxiv.org/abs/2509.23127 arxiv.org/pdf/2509.2312…

@arXiv_csLG_bot@mastoxiv.page
2025-10-03 11:00:21

Flatness-Aware Stochastic Gradient Langevin Dynamics
Stefano Bruno, Youngsik Hwang, Jaehyeon An, Sotirios Sabanis, Dong-Young Lim
arxiv.org/abs/2510.02174

@arXiv_mathSG_bot@mastoxiv.page
2025-10-14 09:28:58

From Morse Functions to Lefschetz Fibrations on Cotangent Bundles
Emmanuel Giroux
arxiv.org/abs/2510.10669 arxiv.org/pdf/2510.10669

@arXiv_eessSY_bot@mastoxiv.page
2025-10-06 09:27:19

Global Convergence of Policy Gradient for Entropy Regularized Linear-Quadratic Control with multiplicative noise
Gabriel Diaz, Lucky Li, Wenhao Zhang
arxiv.org/abs/2510.02896

@arXiv_mathOC_bot@mastoxiv.page
2025-10-14 11:44:38

Adaptive Conditional Gradient Descent
Abbas Khademi, Antonio Silveti-Falls
arxiv.org/abs/2510.11440 arxiv.org/pdf/2510.11440

@arXiv_csCV_bot@mastoxiv.page
2025-10-14 16:14:34

Crosslisted article(s) found for cs.CV. arxiv.org/list/cs.CV/new
[1/3]:
- Gradient-Sign Masking for Task Vector Transport Across Pre-Trained Models
Rinaldi, Panariello, Salici, Liu, Ciccone, Porrello, Calderara

@arXiv_mathDG_bot@mastoxiv.page
2025-10-07 10:16:22

Curvature pinching of asymptotically conical gradient expanding Ricci solitons
Huai-Dong Cao, Junming Xie
arxiv.org/abs/2510.05075 arxiv.or…

@arXiv_csCR_bot@mastoxiv.page
2025-10-06 09:58:19

Untargeted Jailbreak Attack
Xinzhe Huang, Wenjing Hu, Tianhang Zheng, Kedong Xiu, Xiaojun Jia, Di Wang, Zhan Qin, Kui Ren
arxiv.org/abs/2510.02999

@arXiv_mathNA_bot@mastoxiv.page
2025-10-10 08:55:59

Stochastic Gradient Descent for Incomplete Tensor Linear Systems
Anna Ma, Deanna Needell, Alexander Xue
arxiv.org/abs/2510.07630 arxiv.org/…

@arXiv_csLG_bot@mastoxiv.page
2025-10-10 11:18:19

On the optimization dynamics of RLVR: Gradient gap and step size thresholds
Joe Suk, Yaqi Duan
arxiv.org/abs/2510.08539 arxiv.org/pdf/2510.…

@arXiv_mathOC_bot@mastoxiv.page
2025-10-15 09:38:41

A Gradient Guided Diffusion Framework for Chance Constrained Programming
Boyang Zhang, Zhiguo Wang, Ya-Feng Liu
arxiv.org/abs/2510.12238 ar…

@arXiv_statML_bot@mastoxiv.page
2025-10-08 09:29:19

On the Theory of Continual Learning with Gradient Descent for Neural Networks
Hossein Taheri, Avishek Ghosh, Arya Mazumdar
arxiv.org/abs/2510.05573

@arXiv_quantph_bot@mastoxiv.page
2025-10-08 10:24:49

Hybrid Quantum-Classical Policy Gradient for Adaptive Control of Cyber-Physical Systems: A Comparative Study of VQC vs. MLP
Aueaphum Aueawatthanaphisut, Nyi Wunna Tun
arxiv.org/abs/2510.06010

@arXiv_csCL_bot@mastoxiv.page
2025-10-06 10:16:39

PGMEL: Policy Gradient-based Generative Adversarial Network for Multimodal Entity Linking
KM Pooja, Cheng Long, Aixin Sun
arxiv.org/abs/2510.02726

@arXiv_mathAP_bot@mastoxiv.page
2025-10-07 11:29:32

Riesz fractional gradient functionals defined on partitions: nonlocal-to-local variational limits
Almi Stefano, Maicol Caponi, Manuel Friedrich, Francesco Solombrino
arxiv.org/abs/2510.04881

@arXiv_mathOC_bot@mastoxiv.page
2025-10-15 09:08:11

New Classes of Non-monotone Variational Inequality Problems Solvable via Proximal Gradient on Smooth Gap Functions
Lei Zhao, Daoli Zhu, Shuzhong Zhang
arxiv.org/abs/2510.12105

@arXiv_mathDG_bot@mastoxiv.page
2025-09-26 07:44:41

Four-dimensional Gradient Ricci Solitons Gradient shrinking Ricci Solitons and Modified Sectional Curvature
Xiaodong Cao, Ernani Ribeiro Jr, Hosea Wondo
arxiv.org/abs/2509.20669

@arXiv_csLG_bot@mastoxiv.page
2025-10-08 10:38:29

NeST-BO: Fast Local Bayesian Optimization via Newton-Step Targeting of Gradient and Hessian Information
Wei-Ting Tang, Akshay Kudva, Joel A. Paulson
arxiv.org/abs/2510.05516

@arXiv_statML_bot@mastoxiv.page
2025-10-07 10:11:52

Computing Wasserstein Barycenters through Gradient Flows
Eduardo Fernandes Montesuma, Yassir Bendou, Mike Gartrell
arxiv.org/abs/2510.04602

@arXiv_eessSY_bot@mastoxiv.page
2025-10-03 09:00:41

Off-Policy Reinforcement Learning with Anytime Safety Guarantees via Robust Safe Gradient Flow
Pol Mestres, Arnau Marzabal, Jorge Cort\'es
arxiv.org/abs/2510.01492

@arXiv_csCL_bot@mastoxiv.page
2025-10-15 10:38:41

SMEC: Rethinking Matryoshka Representation Learning for Retrieval Embedding Compression
Biao Zhang, Lixin Chen, Tong Liu, Bo Zheng
arxiv.org/abs/2510.12474

@arXiv_csLG_bot@mastoxiv.page
2025-10-06 10:25:09

AdaBet: Gradient-free Layer Selection for Efficient Training of Deep Neural Networks
Irene Tenison, Soumyajit Chatterjee, Fahim Kawsar, Mohammad Malekzadeh
arxiv.org/abs/2510.03101

@arXiv_mathAP_bot@mastoxiv.page
2025-10-10 09:28:29

Gradient regularity for widely degenerate parabolic equations
Michael Strunk
arxiv.org/abs/2510.07999 arxiv.org/pdf/2510.07999

@arXiv_statML_bot@mastoxiv.page
2025-10-03 09:42:31

Adaptive Kernel Selection for Stein Variational Gradient Descent
Moritz Melcher, Simon Weissmann, Ashia C. Wilson, Jakob Zech
arxiv.org/abs/2510.02067

@arXiv_mathOC_bot@mastoxiv.page
2025-11-14 09:50:00

(Adaptive) Scaled gradient methods beyond locally Holder smoothness: Lyapunov analysis, convergence rate and complexity
Susan Ghaderi, Morteza Rahimi, Yves Moreau, Masoud Ahookhosh
arxiv.org/abs/2511.10425 arxiv.org/pdf/2511.10425 arxiv.org/html/2511.10425
arXiv:2511.10425v1 Announce Type: new
Abstract: This paper addresses the unconstrained minimization of smooth convex functions whose gradients are locally Holder continuous. Building on these results, we analyze the Scaled Gradient Algorithm (SGA) under local smoothness assumptions, proving its global convergence and iteration complexity. Furthermore, under local strong convexity and the Kurdyka-Lojasiewicz (KL) inequality, we establish linear convergence rates and provide explicit complexity bounds. In particular, we show that when the gradient is locally Lipschitz continuous, SGA attains linear convergence for any KL exponent. We then introduce and analyze an adaptive variant of SGA (AdaSGA), which automatically adjusts the scaling and step-size parameters. For this method, we show global convergence, and derive local linear rates under strong convexity.
toXiv_bot_toot

@arXiv_mathDG_bot@mastoxiv.page
2025-10-10 08:47:49

Asymptotic behaviour of the weak inverse anisotropic mean curvature flow
Chaoqun Gao, Yong Wei, Rong Zhou
arxiv.org/abs/2510.08168 arxiv.or…

@arXiv_csLG_bot@mastoxiv.page
2025-10-03 11:04:31

Robust Tangent Space Estimation via Laplacian Eigenvector Gradient Orthogonalization
Dhruv Kohli, Sawyer J. Robertson, Gal Mishne, Alexander Cloninger
arxiv.org/abs/2510.02308

@arXiv_mathOC_bot@mastoxiv.page
2025-10-06 09:27:49

Quantitative Convergence Analysis of Projected Stochastic Gradient Descent for Non-Convex Losses via the Goldstein Subdifferential
Yuping Zheng, Andrew Lamperski
arxiv.org/abs/2510.02735

@arXiv_statML_bot@mastoxiv.page
2025-10-15 09:29:01

Active Subspaces in Infinite Dimension
Poorbita Kundu, Nathan Wycoff
arxiv.org/abs/2510.11871 arxiv.org/pdf/2510.11871

@arXiv_csLG_bot@mastoxiv.page
2025-10-08 10:46:59

Inductive inference of gradient-boosted decision trees on graphs for insurance fraud detection
F\'elix Vandervorst, Bruno Deprez, Wouter Verbeke, Tim Verdonck
arxiv.org/abs/2510.05676

@arXiv_mathOC_bot@mastoxiv.page
2025-10-09 09:24:51

Approximate Bregman proximal gradient algorithm with variable metric Armijo--Wolfe line search
Kiwamu Fujiki, Shota Takahashi, Akiko Takeda
arxiv.org/abs/2510.06615

@arXiv_csLG_bot@mastoxiv.page
2025-10-02 11:07:51

CurES: From Gradient Analysis to Efficient Curriculum Learning for Reasoning LLMs
Yongcheng Zeng, Zexu Sun, Bokai Ji, Erxue Min, Hengyi Cai, Shuaiqiang Wang, Dawei Yin, Haifeng Zhang, Xu Chen, Jun Wang
arxiv.org/abs/2510.01037

@arXiv_statML_bot@mastoxiv.page
2025-10-01 09:43:28

When Langevin Monte Carlo Meets Randomization: Non-asymptotic Error Bounds beyond Log-Concavity and Gradient Lipschitzness
Xiaojie Wang, Bin Yang
arxiv.org/abs/2509.25630

@arXiv_mathOC_bot@mastoxiv.page
2025-11-14 09:19:00

Global Convergence of Four-Layer Matrix Factorization under Random Initialization
Minrui Luo, Weihang Xu, Xiang Gao, Maryam Fazel, Simon Shaolei Du
arxiv.org/abs/2511.09925 arxiv.org/pdf/2511.09925 arxiv.org/html/2511.09925
arXiv:2511.09925v1 Announce Type: new
Abstract: Gradient descent dynamics on the deep matrix factorization problem is extensively studied as a simplified theoretical model for deep neural networks. Although the convergence theory for two-layer matrix factorization is well-established, no global convergence guarantee for general deep matrix factorization under random initialization has been established to date. To address this gap, we provide a polynomial-time global convergence guarantee for randomly initialized gradient descent on four-layer matrix factorization, given certain conditions on the target matrix and a standard balanced regularization term. Our analysis employs new techniques to show saddle-avoidance properties of gradient decent dynamics, and extends previous theories to characterize the change in eigenvalues of layer weights.
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2025-10-02 11:10:31

Sample-Efficient Differentially Private Fine-Tuning via Gradient Matrix Denoising
Ali Dadsetan, Frank Rudzicz
arxiv.org/abs/2510.01137 arxi…

@arXiv_statML_bot@mastoxiv.page
2025-09-30 08:48:11

Differentially Private Two-Stage Gradient Descent for Instrumental Variable Regression
Haodong Liang, Yanhao Jin, Krishnakumar Balasubramanian, Lifeng Lai
arxiv.org/abs/2509.22794

@arXiv_csLG_bot@mastoxiv.page
2025-10-08 10:26:09

Correlating Cross-Iteration Noise for DP-SGD using Model Curvature
Xin Gu, Yingtai Xiao, Guanlin He, Jiamu Bai, Daniel Kifer, Kiwan Maeng
arxiv.org/abs/2510.05416

@arXiv_mathOC_bot@mastoxiv.page
2025-11-14 10:01:50

Low-Discrepancy Set Post-Processing via Gradient Descent
Fran\c{c}ois Cl\'ement, Linhang Huang, Woorim Lee, Cole Smidt, Braeden Sodt, Xuan Zhang
arxiv.org/abs/2511.10496 arxiv.org/pdf/2511.10496 arxiv.org/html/2511.10496
arXiv:2511.10496v1 Announce Type: new
Abstract: The construction of low-discrepancy sets, used for uniform sampling and numerical integration, has recently seen great improvements based on optimization and machine learning techniques. However, these methods are computationally expensive, often requiring days of computation or access to GPU clusters. We show that simple gradient descent-based techniques allow for comparable results when starting with a reasonably uniform point set. Not only is this method much more efficient and accessible, but it can be applied as post-processing to any low-discrepancy set generation method for a variety of standard discrepancy measures.
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2025-10-07 13:05:42

Reinforce-Ada: An Adaptive Sampling Framework for Reinforce-Style LLM Training
Wei Xiong, Chenlu Ye, Baohao Liao, Hanze Dong, Xinxing Xu, Christof Monz, Jiang Bian, Nan Jiang, Tong Zhang
arxiv.org/abs/2510.04996

@arXiv_mathOC_bot@mastoxiv.page
2025-09-29 09:16:07

A Riemannian Accelerated Proximal Gradient Method
Shuailing Feng, Yuhang Jiang, Wen Huang, Shihui Ying
arxiv.org/abs/2509.21897 arxiv.org/p…

@arXiv_csLG_bot@mastoxiv.page
2025-10-07 13:05:22

Adaptive Memory Momentum via a Model-Based Framework for Deep Learning Optimization
Kristi Topollai, Anna Choromanska
arxiv.org/abs/2510.04988

@arXiv_mathOC_bot@mastoxiv.page
2025-11-14 09:37:10

S-D-RSM: Stochastic Distributed Regularized Splitting Method for Large-Scale Convex Optimization Problems
Maoran Wang, Xingju Cai, Yongxin Chen
arxiv.org/abs/2511.10133 arxiv.org/pdf/2511.10133 arxiv.org/html/2511.10133
arXiv:2511.10133v1 Announce Type: new
Abstract: This paper investigates the problems large-scale distributed composite convex optimization, with motivations from a broad range of applications, including multi-agent systems, federated learning, smart grids, wireless sensor networks, compressed sensing, and so on. Stochastic gradient descent (SGD) and its variants are commonly employed to solve such problems. However, existing algorithms often rely on vanishing step sizes, strong convexity assumptions, or entail substantial computational overhead to ensure convergence or obtain favorable complexity. To bridge the gap between theory and practice, we integrate consensus optimization and operator splitting techniques (see Problem Reformulation) to develop a novel stochastic splitting algorithm, termed the \emph{stochastic distributed regularized splitting method} (S-D-RSM). In practice, S-D-RSM performs parallel updates of proximal mappings and gradient information for only a randomly selected subset of agents at each iteration. By introducing regularization terms, it effectively mitigates consensus discrepancies among distributed nodes. In contrast to conventional stochastic methods, our theoretical analysis establishes that S-D-RSM achieves global convergence without requiring diminishing step sizes or strong convexity assumptions. Furthermore, it achieves an iteration complexity of $\mathcal{O}(1/\epsilon)$ with respect to both the objective function value and the consensus error. Numerical experiments show that S-D-RSM achieves up to 2--3$\times$ speedup compared to state-of-the-art baselines, while maintaining comparable or better accuracy. These results not only validate the algorithm's theoretical guarantees but also demonstrate its effectiveness in practical tasks such as compressed sensing and empirical risk minimization.
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2025-10-13 10:44:10

Weight Initialization and Variance Dynamics in Deep Neural Networks and Large Language Models
Yankun Han
arxiv.org/abs/2510.09423 arxiv.org…

@arXiv_csLG_bot@mastoxiv.page
2025-10-14 13:41:28

Boundary-Guided Policy Optimization for Memory-efficient RL of Diffusion Large Language Models
Nianyi Lin, Jiajie Zhang, Lei Hou, Juanzi Li
arxiv.org/abs/2510.11683

@arXiv_mathOC_bot@mastoxiv.page
2025-09-25 08:48:02

Inexact and Stochastic Gradient Optimization Algorithms with Inertia and Hessian Driven Damping
Harsh Choudhary, Jalal Fadili, Vyachelav Kungurtsev
arxiv.org/abs/2509.19561

@arXiv_mathOC_bot@mastoxiv.page
2025-09-30 12:01:41

Proximal gradient methods in Banach spaces
Gerd Wachsmuth, Daniel Walter
arxiv.org/abs/2509.24685 arxiv.org/pdf/2509.24685

@arXiv_csLG_bot@mastoxiv.page
2025-10-02 11:08:01

Gated X-TFC: Soft Domain Decomposition for Forward and Inverse Problems in Sharp-Gradient PDEs
Vikas Dwivedi, Enrico Schiassi, Monica Sigovan, Bruno Sixou
arxiv.org/abs/2510.01039

@arXiv_mathOC_bot@mastoxiv.page
2025-10-03 09:48:41

On the (almost) Global Exponential Convergence of the Overparameterized Policy Optimization for the LQR Problem
Moh Kamalul Wafi, Arthur Castello B. de Oliveira, Eduardo D. Sontag
arxiv.org/abs/2510.02140

@arXiv_csLG_bot@mastoxiv.page
2025-09-23 12:49:50

Unveiling m-Sharpness Through the Structure of Stochastic Gradient Noise
Haocheng Luo, Mehrtash Harandi, Dinh Phung, Trung Le
arxiv.org/abs/2509.18001

@arXiv_mathOC_bot@mastoxiv.page
2025-10-01 08:57:27

A Single-Loop Gradient Algorithm for Pessimistic Bilevel Optimization via Smooth Approximation
Cao Qichao, Zeng Shangzhi, Zhang jin
arxiv.org/abs/2509.26240

@arXiv_mathOC_bot@mastoxiv.page
2025-09-23 08:49:20

Hessian-guided Perturbed Wasserstein Gradient Flows for Escaping Saddle Points
Naoya Yamamoto, Juno Kim, Taiji Suzuki
arxiv.org/abs/2509.16974

@arXiv_mathOC_bot@mastoxiv.page
2025-11-14 13:23:10

Replaced article(s) found for math.OC. arxiv.org/list/math.OC/new
[1/1]:
- A robust BFGS algorithm for unconstrained nonlinear optimization problems
Yaguang Yang
arxiv.org/abs/1212.5929
- Quantum computing and the stable set problem
Alja\v{z} Krpan, Janez Povh, Dunja Pucher
arxiv.org/abs/2405.12845 mastoxiv.page/@arXiv_mathOC_bo
- Mean Field Game with Reflected Jump Diffusion Dynamics: A Linear Programming Approach
Zongxia Liang, Xiang Yu, Keyu Zhang
arxiv.org/abs/2508.20388 mastoxiv.page/@arXiv_mathOC_bo
- Differential Dynamic Programming for the Optimal Control Problem with an Ellipsoidal Target Set a...
Sungjun Eom, Gyunghoon Park
arxiv.org/abs/2509.07546 mastoxiv.page/@arXiv_mathOC_bo
- On the Moreau envelope properties of weakly convex functions
Marien Renaud, Arthur Leclaire, Nicolas Papadakis
arxiv.org/abs/2509.13960 mastoxiv.page/@arXiv_mathOC_bo
- Automated algorithm design via Nevanlinna-Pick interpolation
Ibrahim K. Ozaslan, Tryphon T. Georgiou, Mihailo R. Jovanovic
arxiv.org/abs/2509.21416 mastoxiv.page/@arXiv_mathOC_bo
- Optimal Control of a Bioeconomic Crop-Energy System with Energy Reinvestment
Othman Cherkaoui Dekkaki
arxiv.org/abs/2510.11381 mastoxiv.page/@arXiv_mathOC_bo
- Point Convergence Analysis of the Accelerated Gradient Method for Multiobjective Optimization: Co...
Yingdong Yin
arxiv.org/abs/2510.26382 mastoxiv.page/@arXiv_mathOC_bo
- History-Aware Adaptive High-Order Tensor Regularization
Chang He, Bo Jiang, Yuntian Jiang, Chuwen Zhang, Shuzhong Zhang
arxiv.org/abs/2511.05788
- Equivalence of entropy solutions and gradient flows for pressureless 1D Euler systems
Jos\'e Antonio Carrillo, Sondre Tesdal Galtung
arxiv.org/abs/2312.04932 mastoxiv.page/@arXiv_mathAP_bo
- Kernel Modelling of Fading Memory Systems
Yongkang Huo, Thomas Chaffey, Rodolphe Sepulchre
arxiv.org/abs/2403.11945 mastoxiv.page/@arXiv_eessSY_bo
- The Maximum Theoretical Ground Speed of the Wheeled Vehicle
Altay Zhakatayev, Mukatai Nemerebayev
arxiv.org/abs/2502.15341 mastoxiv.page/@arXiv_physicscl
- Hessian stability and convergence rates for entropic and Sinkhorn potentials via semiconcavity
Giacomo Greco, Luca Tamanini
arxiv.org/abs/2504.11133 mastoxiv.page/@arXiv_mathPR_bo
- Optimizing the ground state energy of the three-dimensional magnetic Dirichlet Laplacian with con...
Matthias Baur
arxiv.org/abs/2504.21597 mastoxiv.page/@arXiv_mathph_bo
- A localized consensus-based sampling algorithm
Arne Bouillon, Alexander Bodard, Panagiotis Patrinos, Dirk Nuyens, Giovanni Samaey
arxiv.org/abs/2505.24861 mastoxiv.page/@arXiv_mathNA_bo
- A Novel Sliced Fused Gromov-Wasserstein Distance
Moritz Piening, Robert Beinert
arxiv.org/abs/2508.02364 mastoxiv.page/@arXiv_csLG_bot/
- Minimal Regret Walras Equilibria for Combinatorial Markets via Duality, Integrality, and Sensitiv...
Alo\"is Duguet, Tobias Harks, Martin Schmidt, Julian Schwarz
arxiv.org/abs/2511.09021 mastoxiv.page/@arXiv_csGT_bot/
toXiv_bot_toot