
2025-08-29 10:18:31
Unbiased Stochastic Optimization for Gaussian Processes on Finite Dimensional RKHS
Neta Shoham, Haim Avron
https://arxiv.org/abs/2508.20588 https://arxiv.o…
Unbiased Stochastic Optimization for Gaussian Processes on Finite Dimensional RKHS
Neta Shoham, Haim Avron
https://arxiv.org/abs/2508.20588 https://arxiv.o…
Towards Generalized Parameter Tuning in Coherent Ising Machines: A Portfolio-Based Approach
Tatsuro Hanyu, Takahiro Katagiri, Daichi Mukunoki, Tetsuya Hoshino
https://arxiv.org/abs/2507.20295
Investigating the Role of LLMs Hyperparameter Tuning and Prompt Engineering to Support Domain Modeling
Vladyslav Bulhakov, Giordano d'Aloisio, Claudio Di Sipio, Antinisca Di Marco, Davide Di Ruscio
https://arxiv.org/abs/2507.14735
Private Hyperparameter Tuning with Ex-Post Guarantee
Badih Ghazi, Pritish Kamath, Alexander Knop, Ravi Kumar, Pasin Manurangsi, Chiyuan Zhang
https://arxiv.org/abs/2508.15183 ht…
Bridging Offline and Online Reinforcement Learning for LLMs
Jack Lanchantin, Angelica Chen, Janice Lan, Xian Li, Swarnadeep Saha, Tianlu Wang, Jing Xu, Ping Yu, Weizhe Yuan, Jason E Weston, Sainbayar Sukhbaatar, Ilia Kulikov
https://arxiv.org/abs/2506.21495 https://arxiv.org/pdf/2506.21495 https://arxiv.org/html/2506.21495
arXiv:2506.21495v1 Announce Type: new
Abstract: We investigate the effectiveness of reinforcement learning methods for finetuning large language models when transitioning from offline to semi-online to fully online regimes for both verifiable and non-verifiable tasks. Our experiments cover training on verifiable math as well as non-verifiable instruction following with a set of benchmark evaluations for both. Across these settings, we extensively compare online and semi-online Direct Preference Optimization and Group Reward Policy Optimization objectives, and surprisingly find similar performance and convergence between these variants, which all strongly outperform offline methods. We provide a detailed analysis of the training dynamics and hyperparameter selection strategies to achieve optimal results. Finally, we show that multi-tasking with verifiable and non-verifiable rewards jointly yields improved performance across both task types.
toXiv_bot_toot
Hyperparameter-free minimum-lengthscale constraints for topology optimization
Rodrigo Arrieta, Giuseppe Romano, Steven G. Johnson
https://arxiv.org/abs/2507.16108
Quantum Annealing Hyperparameter Analysis for Optimal Sensor Placement in Production Environments
Nico Kraus, Marvin Erdmann, Alexander Kuzmany, Daniel Porawski, Jonas Stein
https://arxiv.org/abs/2507.16584
Black-box optimization using factorization and Ising machines
Ryo Tamura, Yuya Seki, Yuki Minamoto, Koki Kitai, Yoshiki Matsuda, Shu Tanaka, Koji Tsuda
https://arxiv.org/abs/2507.18003
From Linearity to Non-Linearity: How Masked Autoencoders Capture Spatial Correlations
Anthony Bisulco, Rahul Ramesh, Randall Balestriero, Pratik Chaudhari
https://arxiv.org/abs/2508.15404
Sparse but Wrong: Incorrect L0 Leads to Incorrect Features in Sparse Autoencoders
David Chanin, Adri\`a Garriga-Alonso
https://arxiv.org/abs/2508.16560 https://
Learning Acceleration Algorithms for Fast Parametric Convex Optimization with Certified Robustness
Rajiv Sambharya, Jinho Bok, Nikolai Matni, George Pappas
https://arxiv.org/abs/2507.16264
Evaluation and Optimization of Leave-one-out Cross-validation for the Lasso
Ryan Burn
https://arxiv.org/abs/2508.14368 https://arxiv.org/pdf/2508.14368
RLGS: Reinforcement Learning-Based Adaptive Hyperparameter Tuning for Gaussian Splatting
Zhan Li, Huangying Zhan, Changyang Li, Qingan Yan, Yi Xu
https://arxiv.org/abs/2508.04078
Analyzing Internal Activity and Robustness of SNNs Across Neuron Parameter Space
Szymon Mazurek, Jakub Caputa, Maciej Wielgosz
https://arxiv.org/abs/2507.14757
Dual-Attention U-Net with Class-Specific Ensembles and Bayesian Hyperparameter Optimization for Precise Wound and Scale Marker Segmentation
Daniel Cie\'slak, Miriam Reca, Olena Onyshchenko, Jacek Rumi\'nski
https://arxiv.org/abs/2507.05314
In-Context Decision Making for Optimizing Complex AutoML Pipelines
Amir Rezaei Balef, Katharina Eggensperger
https://arxiv.org/abs/2508.13657 https://arxiv…
Great Restraining Wall in Multidimentional Collective Variable Space
Zhijun Pan, Maodong Li, Dechin Chen, Yi Isaac Yang
https://arxiv.org/abs/2506.17043 ht…
Dataset-Adaptive Dimensionality Reduction
Hyeon Jeon, Jeongin Park, Soohyun Lee, Dae Hyun Kim, Sungbok Shin, Jinwook Seo
https://arxiv.org/abs/2507.11984 h…
Harnessing data-driven methods for precise model independent event shape estimation in relativistic heavy-ion collisions
Dipankar Basak, H. Hushnud, Kalyan Dey
https://arxiv.org/abs/2508.13349
This https://arxiv.org/abs/2412.06481 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_mat…
Towards Effective Offensive Security LLM Agents: Hyperparameter Tuning, LLM as a Judge, and a Lightweight CTF Benchmark
Minghao Shao, Nanda Rani, Kimberly Milner, Haoran Xi, Meet Udeshi, Saksham Aggarwal, Venkata Sai Charan Putrevu, Sandeep Kumar Shukla, Prashanth Krishnamurthy, Farshad Khorrami, Ramesh Karri, Muhammad Shafique
https://arx…
CrossDenoise: Denoising Implicit Feedback via a Lightweight Entity-Aware Synergistic Framework
Ze Liu, Xianquan Wang, Shuochen Liu, Jie Ma, Huibo Xu, Yupeng Han, Zhe Yang, Kai Zhang, Longfei Li, Jun Zhou
https://arxiv.org/abs/2508.10851
carps: A Framework for Comparing N Hyperparameter Optimizers on M Benchmarks
Carolin Benjamins, Helena Graf, Sarah Segel, Difan Deng, Tim Ruhkopf, Leona Hennig, Soham Basu, Neeratyoy Mallik, Edward Bergman, Deyao Chen, Fran\c{c}ois Cl\'ement, Matthias Feurer, Katharina Eggensperger, Frank Hutter, Carola Doerr, Marius Lindauer
https://
This https://arxiv.org/abs/2412.06481 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_mat…
Uniform convergence for Gaussian kernel ridge regression
Paul Dommel, Rajmadan Lakshmanan
https://arxiv.org/abs/2508.11274 https://arxiv.org/pdf/2508.11274…
Predictive posteriors under hidden confounding
Carlos Garc\'ia Meixide, David R\'ios Insua
https://arxiv.org/abs/2507.05170 https://
SO-PIFRNN: Self-optimization physics-informed Fourier-features randomized neural network for solving partial differential equations
Jiale Linghu, Weifeng Gao, Hao Dong, Yufeng Nie
https://arxiv.org/abs/2508.10921
Successive Halving with Learning Curve Prediction via Latent Kronecker Gaussian Processes
Jihao Andreas Lin, Nicolas Mayoraz, Steffen Rendle, Dima Kuzmin, Emil Praun, Berivan Isik
https://arxiv.org/abs/2508.14818
The Role of Review Process Failures in Affective State Estimation: An Empirical Investigation of DEAP Dataset
Nazmun N Khan, Taylor Sweet, Chase A Harvey, Calder Knapp, Dean J. Krusienski, David E Thompson
https://arxiv.org/abs/2508.02417
Posterior Transition Modeling for Unsupervised Diffusion-Based Speech Enhancement
Mostafa Sadeghi (MULTISPEECH), Jean-Eudes Ayilo (MULTISPEECH), Romain Serizel (MULTISPEECH), Xavier Alameda-Pineda (ROBOTLEARN)
https://arxiv.org/abs/2507.02391
ADMIRE-BayesOpt: Accelerated Data MIxture RE-weighting for Language Models with Bayesian Optimization
Shengzhuang Chen, Xu Ouyang, Michael Arthur Leopold Pearce, Thomas Hartvigsen, Jonathan Richard Schwarz
https://arxiv.org/abs/2508.11551
Enhancing Power Flow Estimation with Topology-Aware Gated Graph Neural Networks
Shrenik Jadhav, Birva Sevak, Srijita Das, Wencong Su, Van-Hai Bui
https://arxiv.org/abs/2507.02078 …
A statistical physics framework for optimal learning
Francesca Mignacco, Francesco Mori
https://arxiv.org/abs/2507.07907 https://arxi…
Error Breakdown and Sensitivity Analysis of Dynamical Quantities in Markov State Models
Yehor Tuchkov, Luke Evans, Sonya M. Hanson, Erik H. Thiede
https://arxiv.org/abs/2508.06735
Improving Neural Network Training using Dynamic Learning Rate Schedule for PINNs and Image Classification
D. Veerababu, Ashwin A. Raikar, Prasanta K. Ghosh
https://arxiv.org/abs/2507.21749
AutoScale: Linear Scalarization Guided by Multi-Task Optimization Metrics
Yi Yang, Kei Ikemura, Qingwen Zhang, Xiaomeng Zhu, Ci Li, Nazre Batool, Sina Sharif Mansouri, John Folkesson
https://arxiv.org/abs/2508.13979
SPEAR: Subset-sampled Performance Evaluation via Automated Ground Truth Generation for RAG
Zou Yuheng, Wang Yiran, Tian Yuzhu, Zhu Min, Huang Yanhua
https://arxiv.org/abs/2507.06554
On the Effectiveness of Classical Regression Methods for Optimal Switching Problems
Martin Andersson, Benny Avelin, Marcus Olofsson
https://arxiv.org/abs/2506.15436
Machine Learning-Driven High-Precision Model for $\alpha$-Decay Energy and Half-Life Prediction of superheavy nuclei
Qingning Yuan, Panpan Qi, Xuanpen Xiao, Xue Wang, Juan He, Guimei Long, Zhengwei Duan, Yangyan Dai, Runchao Yan, Gongming Yu, Haitao Yang, Qiang Hu
https://arxiv.org/abs/2508.03155
Vid2Sim: Generalizable, Video-based Reconstruction of Appearance, Geometry and Physics for Mesh-free Simulation
Chuhao Chen, Zhiyang Dou, Chen Wang, Yiming Huang, Anjun Chen, Qiao Feng, Jiatao Gu, Lingjie Liu
https://arxiv.org/abs/2506.06440
trAIce3D: A Prompt-Driven Transformer Based U-Net for Semantic Segmentation of Microglial Cells from Large-Scale 3D Microscopy Images
MohammadAmin Alamalhoda, Arsalan Firoozi, Alessandro Venturino, Sandra Siegert
https://arxiv.org/abs/2507.22635
A Practical Guide to Tuning Spiking Neuronal Dynamics
William Gebhardt, Alexander G. Ororbia, Nathan McDonald, Clare Thiem, Jack Lombardi
https://arxiv.org/abs/2506.08138
Minimal Deterministic Echo State Networks Outperform Random Reservoirs in Learning Chaotic Dynamics
Francesco Martinuzzi
https://arxiv.org/abs/2507.06050 h…
Improving Robustness of Foundation Models in Domain Adaptation with Soup-Adapters
Marco Roschkowski
https://arxiv.org/abs/2507.05807 https://
This https://arxiv.org/abs/2502.06044 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_sta…
Efficient inference of dynamic gene regulatory networks using discrete penalty
Visweswaran Ravikumar, Aaresh Bhathena, Wajd N Al-Holou, Salar Fattahi, Arvind Rao
https://arxiv.org/abs/2507.23106
This https://arxiv.org/abs/2506.05673 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csLG_…
Fine-tuning for Data-enabled Predictive Control of Noisy Systems by Reinforcement Learning
Jinbao Wang, Shiliang Zhang, Jun Liu, Xuehui Ma, Haolin Liu
https://arxiv.org/abs/2505.24572
This https://arxiv.org/abs/2503.22733 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csLG_…
Replaced article(s) found for math.OC. https://arxiv.org/list/math.OC/new
[1/1]:
- Global relaxation-based LP-Newton method for multiple hyperparameter selection in support vector ...
Yaru Qian, Qingna Li, Alain Zemkoho
This https://arxiv.org/abs/2505.00812 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csLG_…
Tree-Structured Parzen Estimator Can Solve Black-Box Combinatorial Optimization More Efficiently
Kenshin Abe, Yunzhuo Wang, Shuhei Watanabe
https://arxiv.org/abs/2507.08053 https://arxiv.org/pdf/2507.08053 https://arxiv.org/html/2507.08053
arXiv:2507.08053v1 Announce Type: new
Abstract: Tree-structured Parzen estimator (TPE) is a versatile hyperparameter optimization (HPO) method supported by popular HPO tools. Since these HPO tools have been developed in line with the trend of deep learning (DL), the problem setups often used in the DL domain have been discussed for TPE such as multi-objective optimization and multi-fidelity optimization. However, the practical applications of HPO are not limited to DL, and black-box combinatorial optimization is actively utilized in some domains, e.g., chemistry and biology. As combinatorial optimization has been an untouched, yet very important, topic in TPE, we propose an efficient combinatorial optimization algorithm for TPE. In this paper, we first generalize the categorical kernel with the numerical kernel in TPE, enabling us to introduce a distance structure to the categorical kernel. Then we discuss modifications for the newly developed kernel to handle a large combinatorial search space. These modifications reduce the time complexity of the kernel calculation with respect to the size of a combinatorial search space. In the experiments using synthetic problems, we verified that our proposed method identifies better solutions with fewer evaluations than the original TPE. Our algorithm is available in Optuna, an open-source framework for HPO.
toXiv_bot_toot