Online Training and Pruning of Deep Reinforcement Learning Networks
Valentin Frank Ingmar Guenter, Athanasios Sideris
https://arxiv.org/abs/2507.11975 http…
Dynamic Preference Multi-Objective Reinforcement Learning for Internet Network Management
DongNyeong Heo, Daniela Noemi Rim, Heeyoul Choi
https://arxiv.org/abs/2506.13153
Efficient Preparation of Fermionic Superfluids in an Optical Dipole Trap through Reinforcement Learning
Yueyang Min, Ziliang Li, Yi Zhong, Jia-An Xuan, Jian Lin, Fei Leng, Xiaopeng Li
https://arxiv.org/abs/2507.12152
Wasserstein-Barycenter Consensus for Cooperative Multi-Agent Reinforcement Learning
Ali Baheri
https://arxiv.org/abs/2506.12497 https://
Theoretical Tensions in RLHF: Reconciling Empirical Success with Inconsistencies in Social Choice Theory
Jiancong Xiao, Zhekun Shi, Kaizhao Liu, Qi Long, Weijie J. Su
https://arxiv.org/abs/2506.12350
Analytical coarse grained potential parameterization by Reinforcement Learning for anisotropic cellulose
Xu Don
https://arxiv.org/abs/2506.12893 https://…
Vane rheology of a fiber-reinforced granular material
Ladislas Wierzchalek, Georges Gauthier, Baptiste Darbois-Texier
https://arxiv.org/abs/2506.11762 http…
Learning to Tune Like an Expert: Interpretable and Scene-Aware Navigation via MLLM Reasoning and CVAE-Based Adaptation
Yanbo Wang, Zipeng Fang, Lei Zhao, Weidong Chen
https://arxiv.org/abs/2507.11001
Trust-MARL: Trust-Based Multi-Agent Reinforcement Learning Framework for Cooperative On-Ramp Merging Control in Heterogeneous Traffic Flow
Jie Pan, Tianyi Wang, Christian Claudel, Jing Shi
https://arxiv.org/abs/2506.12600
Kevin: Multi-Turn RL for Generating CUDA Kernels
Carlo Baronio, Pietro Marsella, Ben Pan, Simon Guo, Silas Alberti
https://arxiv.org/abs/2507.11948 https:/…