Tootfinder

Opt-in global Mastodon full text search. Join the index!

No exact results. Similar results found.
@arXiv_astrophGA_bot@mastoxiv.page
2025-06-16 09:55:09

The Karl G. Jansky Very Large Array Local Group L-band Survey (LGLBS)
Eric W. Koch, Adam K. Leroy, Erik W. Rosolowsky, Laura Chomiuk, Julianne J. Dalcanton, Nickolas M. Pingel, Sumit K. Sarbadhicary, Sne\v{z}ana Stanimirovi\'c, Fabian Walter, Haylee N. Archer, Alberto D. Bolatto, Michael P. Busch, Hongxing Chen, Ryan Chown, Harrisen Corbould, Serena A. Cronin, Jeremy Darling, Thomas Do, Jennifer Donovan Meyer, Cosima Eibensteiner, Deidre Hunter, R\'emy Indebetouw, Preshanth Jag…

@arXiv_mathRA_bot@mastoxiv.page
2025-07-16 08:31:11

Structure of Galois rings and the Gelfand-Kirillov Conjecture
Vyacheslav Futorny, Jonas T. Hartwig, Erich C. Jauch, Jo\~ao Schwarz
arxiv.org/abs/2507.10782

@arXiv_csCL_bot@mastoxiv.page
2025-06-27 09:58:19

Bridging Offline and Online Reinforcement Learning for LLMs
Jack Lanchantin, Angelica Chen, Janice Lan, Xian Li, Swarnadeep Saha, Tianlu Wang, Jing Xu, Ping Yu, Weizhe Yuan, Jason E Weston, Sainbayar Sukhbaatar, Ilia Kulikov
arxiv.org/abs/2506.21495 arxiv.org/pdf/2506.21495 arxiv.org/html/2506.21495
arXiv:2506.21495v1 Announce Type: new
Abstract: We investigate the effectiveness of reinforcement learning methods for finetuning large language models when transitioning from offline to semi-online to fully online regimes for both verifiable and non-verifiable tasks. Our experiments cover training on verifiable math as well as non-verifiable instruction following with a set of benchmark evaluations for both. Across these settings, we extensively compare online and semi-online Direct Preference Optimization and Group Reward Policy Optimization objectives, and surprisingly find similar performance and convergence between these variants, which all strongly outperform offline methods. We provide a detailed analysis of the training dynamics and hyperparameter selection strategies to achieve optimal results. Finally, we show that multi-tasking with verifiable and non-verifiable rewards jointly yields improved performance across both task types.
toXiv_bot_toot