2025-09-22 10:33:31
DIVEBATCH: Accelerating Model Training Through Gradient-Diversity Aware Batch Size Adaptation
Yuen Chen, Yian Wang, Hari Sundaram
https://arxiv.org/abs/2509.16173 https://
DIVEBATCH: Accelerating Model Training Through Gradient-Diversity Aware Batch Size Adaptation
Yuen Chen, Yian Wang, Hari Sundaram
https://arxiv.org/abs/2509.16173 https://
Targeted Fine-Tuning of DNN-Based Receivers via Influence Functions
Marko Tuononen, Heikki Penttinen, Ville Hautam\"aki
https://arxiv.org/abs/2509.15950 https://
State-of-the-Art Dysarthric Speech Recognition with MetaICL for on-the-fly Personalization
Dhruuv Agarwal, Harry Zhang, Yang Yu, Quan Wang
https://arxiv.org/abs/2509.15516 https…
Dimension-Free Minimax Rates for Learning Pairwise Interactions in Attention-Style Models
Shai Zucker, Xiong Wang, Fei Lu, Inbar Seroussi
https://arxiv.org/abs/2510.11789 https:…
Im still "e-learning" how to generate aura and hype moments at a market-competitive rate so bear with me
Operand Quant: A Single-Agent Architecture for Autonomous Machine Learning Engineering
Arjun Sahney, Ram Gorthi, Cezary {\L}astowski, Javier Vega
https://arxiv.org/abs/2510.11694
Based on Deep Neural Networks: A Machine Learning-Assisted Channel Estimation Method for MIMO Systems
Haoran He
https://arxiv.org/abs/2510.11891 https://ar…
Multi-agent learning under uncertainty: Recurrence vs. concentration
Kyriakos Lotidis, Panayotis Mertikopoulos, Nicholas Bambos, Jose Blanchet
https://arxiv.org/abs/2512.08132 https://arxiv.org/pdf/2512.08132 https://arxiv.org/html/2512.08132
arXiv:2512.08132v1 Announce Type: new
Abstract: In this paper, we examine the convergence landscape of multi-agent learning under uncertainty. Specifically, we analyze two stochastic models of regularized learning in continuous games -- one in continuous and one in discrete time with the aim of characterizing the long-run behavior of the induced sequence of play. In stark contrast to deterministic, full-information models of learning (or models with a vanishing learning rate), we show that the resulting dynamics do not converge in general. In lieu of this, we ask instead which actions are played more often in the long run, and by how much. We show that, in strongly monotone games, the dynamics of regularized learning may wander away from equilibrium infinitely often, but they always return to its vicinity in finite time (which we estimate), and their long-run distribution is sharply concentrated around a neighborhood thereof. We quantify the degree of this concentration, and we show that these favorable properties may all break down if the underlying game is not strongly monotone -- underscoring in this way the limits of regularized learning in the presence of persistent randomness and uncertainty.
toXiv_bot_toot
Rate optimal learning of equilibria from data
Till Freihaut, Luca Viano, Emanuele Nevali, Volkan Cevher, Matthieu Geist, Giorgia Ramponi
https://arxiv.org/abs/2510.09325 https:/…
Statistical Guarantees for High-Dimensional Stochastic Gradient Descent
Jiaqi Li, Zhipeng Lou, Johannes Schmidt-Hieber, Wei Biao Wu
https://arxiv.org/abs/2510.12013 https://
Energy distance and evolution problems: a promising tool for kinetic equations
Gennaro Auricchio, Giuseppe Toscani
https://arxiv.org/abs/2510.09123 https://
Finetune Once: Decoupling General & Domain Learning with Dynamic Boosted Annealing
Yang Tang, Ruijie Liu, Yifan Wang, Shiyu Li, Xi Chen
https://arxiv.org/abs/2509.26242 http…
Adaptive Deception Framework with Behavioral Analysis for Enhanced Cybersecurity Defense
Basil Abdullah AL-Zahrani
https://arxiv.org/abs/2510.02424 https://
Temperature derivative divergence of the electric conductivity and thermal photon emission rate at the critical end point from holography
Yi-Ping Si, Danning Li, Mei Huang
https://arxiv.org/abs/2509.25636
False Discovery Rate Control via Bayesian Mirror Statistic
Marco Molinari, Magne Thoresen
https://arxiv.org/abs/2510.00875 https://arxiv.org/pdf/2510.00875…
Tight Regret Upper and Lower Bounds for Optimistic Hedge in Two-Player Zero-Sum Games
Taira Tsuchiya
https://arxiv.org/abs/2510.11691 https://arxiv.org/pdf…
Ultra-Reliable Risk-Aggregated Sum Rate Maximization via Model-Aided Deep Learning
Hassaan Hashmi, Spyridon Pougkakiotis, Dionysis Kalogerias
https://arxiv.org/abs/2509.26311 ht…
Efficient Breast and Ovarian Cancer Classification via ViT-Based Preprocessing and Transfer Learning
Richa Rawat, Faisal Ahmed
https://arxiv.org/abs/2509.18553 https://
Physics-informed learning under mixing: How physical knowledge speeds up learning
Anna Scampicchio, Leonardo F. Toso, Rahel Rickenbach, James Anderson, Melanie N. Zeilinger
https://arxiv.org/abs/2509.24801
Scale-Invariant Regret Matching and Online Learning with Optimal Convergence: Bridging Theory and Practice in Zero-Sum Games
Brian Hu Zhang, Ioannis Anagnostides, Tuomas Sandholm
https://arxiv.org/abs/2510.04407
On the Adversarial Robustness of Learning-based Conformal Novelty Detection
Daofu Zhang, Mehrdad Pournaderi, Hanne M. Clifford, Yu Xiang, Pramod K. Varshney
https://arxiv.org/abs/2510.00463
Why Do We Need Warm-up? A Theoretical Perspective
Foivos Alimisis, Rustem Islamov, Aurelien Lucchi
https://arxiv.org/abs/2510.03164 https://arxiv.org/pdf/2…
Bayesian Self-Calibration and Parametric Channel Estimation for 6G Antenna Arrays
Patrick H\"odl, Jakob M\"oderl, Erik Leitinger, Klaus Witrisal
https://arxiv.org/abs/2510.11628
A Bio-Inspired Minimal Model for Non-Stationary K-Armed Bandits
Krubeal Danieli, Mikkel Elle Lepper{\o}d
https://arxiv.org/abs/2509.22209 https://arxiv.org…
MOCLIP: A Foundation Model for Large-Scale Nanophotonic Inverse Design
S. Rodionov, A. Burguete-Lopez, M. Makarenko, Q. Wang, F. Getman, A. Fratalocchi
https://arxiv.org/abs/2511.18980 https://arxiv.org/pdf/2511.18980 https://arxiv.org/html/2511.18980
arXiv:2511.18980v1 Announce Type: new
Abstract: Foundation models (FM) are transforming artificial intelligence by enabling generalizable, data-efficient solutions across different domains for a broad range of applications. However, the lack of large and diverse datasets limits the development of FM in nanophotonics. This work presents MOCLIP (Metasurface Optics Contrastive Learning Pretrained), a nanophotonic foundation model that integrates metasurface geometry and spectra within a shared latent space. MOCLIP employs contrastive learning to align geometry and spectral representations using an experimentally acquired dataset with a sample density comparable to ImageNet-1K. The study demonstrates MOCLIP inverse design capabilities for high-throughput zero-shot prediction at a rate of 0.2 million samples per second, enabling the design of a full 4-inch wafer populated with high-density metasurfaces in minutes. It also shows generative latent-space optimization reaching 97 percent accuracy. Finally, we introduce an optical information storage concept that uses MOCLIP to achieve a density of 0.1 Gbit per square millimeter at the resolution limit, exceeding commercial optical media by a factor of six. These results position MOCLIP as a scalable and versatile platform for next-generation photonic design and data-driven applications.
toXiv_bot_toot
Generalization Analysis for Classification on Korobov Space
Yuqing Liu
https://arxiv.org/abs/2509.22748 https://arxiv.org/pdf/2509.22748
When Scores Learn Geometry: Rate Separations under the Manifold Hypothesis
Xiang Li, Zebang Shen, Ya-Ping Hsieh, Niao He
https://arxiv.org/abs/2509.24912 https://
Score the Steps, Not Just the Goal: VLM-Based Subgoal Evaluation for Robotic Manipulation
Ramy ElMallah, Krish Chhajer, Chi-Guhn Lee
https://arxiv.org/abs/2509.19524 https://
Robust equilibria in continuous games: From strategic to dynamic robustness
Kyriakos Lotidis, Panayotis Mertikopoulos, Nicholas Bambos, Jose Blanchet
https://arxiv.org/abs/2512.08138 https://arxiv.org/pdf/2512.08138 https://arxiv.org/html/2512.08138
arXiv:2512.08138v1 Announce Type: new
Abstract: In this paper, we examine the robustness of Nash equilibria in continuous games, under both strategic and dynamic uncertainty. Starting with the former, we introduce the notion of a robust equilibrium as those equilibria that remain invariant to small -- but otherwise arbitrary -- perturbations to the game's payoff structure, and we provide a crisp geometric characterization thereof. Subsequently, we turn to the question of dynamic robustness, and we examine which equilibria may arise as stable limit points of the dynamics of "follow the regularized leader" (FTRL) in the presence of randomness and uncertainty. Despite their very distinct origins, we establish a structural correspondence between these two notions of robustness: strategic robustness implies dynamic robustness, and, conversely, the requirement of strategic robustness cannot be relaxed if dynamic robustness is to be maintained. Finally, we examine the rate of convergence to robust equilibria as a function of the underlying regularizer, and we show that entropically regularized learning converges at a geometric rate in games with affinely constrained action spaces.
toXiv_bot_toot
Federated Computation of ROC and PR Curves
Xuefeng Xu, Graham Cormode
https://arxiv.org/abs/2510.04979 https://arxiv.org/pdf/2510.04979
Score the Steps, Not Just the Goal: VLM-Based Subgoal Evaluation for Robotic Manipulation
Ramy ElMallah, Krish Chhajer, Chi-Guhn Lee
https://arxiv.org/abs/2509.19524 https://
Improving accuracy in short mortality rate series: Exploring Multi-step Forecasting Approaches in Hybrid Systems
Filipe C. L. Duarte, Paulo S. G. de Mattos Neto, Paulo R. A. Firmino
https://arxiv.org/abs/2509.22395
Data-Driven Reconstruction of Significant Wave Heights from Sparse Observations
Hongyuan Shi, Yilin Zhai, Ping Dong, Zaijin You, Chao Zhan, Qing Wang
https://arxiv.org/abs/2509.19384
Effective continuous equations for adaptive SGD: a stochastic analysis view
Luca Callisti, Marco Romito, Francesco Triggiano
https://arxiv.org/abs/2509.21614 https://
Group Relative Policy Optimization for Text-to-Speech with Large Language Models
Chang Liu, Ya-Jun Hu, Ying-Ying Gao, Shi-Lei Zhang, Zhen-Hua Ling
https://arxiv.org/abs/2509.18798
Efficient Hyperparameter Tuning via Trajectory Invariance Principle
Bingrui Li, Jiaxin Wen, Zhanpeng Zhou, Jun Zhu, Jianfei Chen
https://arxiv.org/abs/2509.25049 https://…
Scaling with Collapse: Efficient and Predictable Training of LLM Families
Shane Bergsma, Bin Claire Zhang, Nolan Dey, Shaheer Muhammad, Gurpreet Gosal, Joel Hestness
https://arxiv.org/abs/2509.25087