Parallels Between VLA Model Post-Training and Human Motor Learning: Progress, Challenges, and Trends
Tian-Yu Xiang, Ao-Qun Jin, Xiao-Hu Zhou, Mei-Jiang Gui, Xiao-Liang Xie, Shi-Qi Liu, Shuang-Yi Wang, Sheng-Bin Duan, Fu-Chao Xie, Wen-Kai Wang, Si-Cheng Wang, Ling-Yun Li, Tian Tu, Zeng-Guang Hou
https://arxiv.org/abs/2506.20966…
Data Efficacy for Language Model Training
Yalun Dai, Yangyu Huang, Xin Zhang, Wenshan Wu, Chong Li, Wenhui Lu, Shijie Cao, Li Dong, Scarlett Li
https://arxiv.org/abs/2506.21545 https://arxiv.org/pdf/2506.21545 https://arxiv.org/html/2506.21545
arXiv:2506.21545v1 Announce Type: new
Abstract: Data is fundamental to the training of language models (LM). Recent research has been dedicated to data efficiency, which aims to maximize performance by selecting a minimal or optimal subset of training data. Techniques such as data filtering, sampling, and selection play a crucial role in this area. To complement it, we define Data Efficacy, which focuses on maximizing performance by optimizing the organization of training data and remains relatively underexplored. This work introduces a general paradigm, DELT, for considering data efficacy in LM training, which highlights the significance of training data organization. DELT comprises three components: Data Scoring, Data Selection, and Data Ordering. Among these components, we design Learnability-Quality Scoring (LQS), as a new instance of Data Scoring, which considers both the learnability and quality of each data sample from the gradient consistency perspective. We also devise Folding Ordering (FO), as a novel instance of Data Ordering, which addresses issues such as model forgetting and data distribution bias. Comprehensive experiments validate the data efficacy in LM training, which demonstrates the following: Firstly, various instances of the proposed DELT enhance LM performance to varying degrees without increasing the data scale and model size. Secondly, among these instances, the combination of our proposed LQS for data scoring and Folding for data ordering achieves the most significant improvement. Lastly, data efficacy can be achieved together with data efficiency by applying data selection. Therefore, we believe that data efficacy is a promising foundational area in LM training.
toXiv_bot_toot
InceptionMamba: Efficient Multi-Stage Feature Enhancement with Selective State Space Model for Microscopic Medical Image Segmentation
Daniya Najiha Abdul Kareem, Abdul Hannan, Mubashir Noman, Jean Lahoud, Mustansar Fiaz, Hisham Cholakkal
https://arxiv.org/abs/2506.12208
Coriolis force acting on near-surface horizontal flows during simulations of flux emergence produces a tilt angle consistent with Joy's law on the Sun
William Roland-Batty, Hannah Schunker, Robert H. Cameron, Damien Przybylski, Laurent Gizon, David I. Pontin
https://arxiv.org/abs/2506.15935
The role of small-angle electron-electron scattering in transverse magnetic focusing experiment
Dmitry A. Egorov, Dmitriy A. Pokhabov, Evgeny Yu. Zhdanov, Andrey A. Shevyrin, Askhat K. Bakarov, Alexander A. Shklyaev, Arthur G. Pogosov
https://arxiv.org/abs/2506.09432
Role of topotactic hydrogen in Superconductivity of Infinite-layer Nickelate NdNiO$_{2}$: A first-principles and variational Monte Carlo study
Manoj Gupta, Arun Kumar Maurya, Amal Medhi, Tanusri Saha Dasgupta
https://arxiv.org/abs/2506.13399
Quantifying Flow State Dynamics: A Prefrontal Cortex EEG-Based Model Validation Study. Unveiling the Prefrontal Cortex's Role in Flow State Experience: An Empirical EEG Analysis
Gianluca Rosso, Raffaella Ricci, Lorenzo Pia, Giovanni Rebaudo, Michele Guindani, Alberto Marocchino, Giorgio De Pieri, Andrea Filippo Rosso
https://
Towards Efficient Speech-Text Jointly Decoding within One Speech Language Model
Haibin Wu, Yuxuan Hu, Ruchao Fan, Xiaofei Wang, Kenichi Kumatani, Bo Ren, Jianwei Yu, Heng Lu, Lijuan Wang, Yao Qian, Jinyu Li
https://arxiv.org/abs/2506.04518
The Category Implementation Role Model Award goes to #SchleswigHolstein
The award is handed over by @… from #DNIP
#nextcloudsummit25