Tootfinder

Opt-in global Mastodon full text search. Join the index!

@hex@kolektiva.social
2026-02-20 10:37:46

In my head I'm just replacing "counter insurgency" with "horse cavalry."
"We're going to keep learning how to leverage horse cavalry against machine guns and tanks until we get it right."
No. No you will not. You will keep trying until you learn the hard way that it can't be done.

@compfu@mograph.social
2026-02-18 19:39:19

RE: mograph.social/@thevfxfeed/116
To be honest, a machine-learning model for Nuke where the training data is clearly sourced and free of legal baggage would be a really good thing. There are so many image segmentation and inpainti…

@Mediagazer@mstdn.social
2026-04-13 18:35:35

Internal memo: CNN hires Chris Wiggins, who was chief data scientist at the NYT for more than a decade, for a newly created role as head of ML and AI science (Todd Spangler/Variety)
variety.com/2026/tv/news/cnn-c

@Techmeme@techhub.social
2026-02-06 13:40:52

How an appeal changed the way the USPTO assesses AI patents under the US Patent Act, signaling a shift towards more favorable treatment of AI and ML inventions (Matthew Carey/Bloomberg Law)
news.bloomberglaw.com/tech-and

@UP8@mastodon.social
2026-03-16 21:39:44

🎶 TweetyBERT parses canary songs to better understand how brains learn language
#birds

@inthehands@hachyderm.io
2026-02-14 20:01:58

RE: mastodon.social/@airspeedswift
Reflections on Trusting Trust and the Trust You’re Trusting is an Opaque, Nondeterministic Machine Learning Model

CETI is a nonprofit organization
applying advanced machine learning and state-of-the-art robotics
to listen to and translate the communication of sperm whales.
Our research focus is in Dominica
in the Eastern Caribbean
projectceti.org/

@ruth_mottram@fediscience.org
2026-03-09 11:37:16

Machine Learning techniques are upending multiple scientific fields. Operational 5-day forecasting of air quality in 1 minute in this paper from Chinese researchers.
This is awesome work with very clear public health implications.
EDIT for clarity: I am.not suggesting LLMs have anything to do with this work, but many people hear AI and imagine LLMs. And many of them.are perhaps rightly sceptical of AI as a result.
But AI or ML techniques can be useful for lots of things, not just chatbots. And we should probably invest more in those.

nature.com/articles/s41586-026

@UP8@mastodon.social
2026-03-03 15:14:58

⚰️ Fentanyl or phony? Machine learning algorithm learns to pick out opioid signatures
#sensors

@cosmos4u@scicomm.xyz
2026-02-10 23:49:42

Possible identification of the Luna 9 Moon landing site using a novel machine learning algorithm: #Luna9 Spacecraft, 60 Years After It Vanished: iflscience.com/nasas-lunar-orb

@chpietsch@fedifreu.de
2026-04-07 14:52:29

Der VCD hat mich heute in seinem Newsletter daran erinnert, dass es bahnvorhersage.de/ gibt. Das ist der Bahnroutenplaner, der dir für jede Verbindung verrät, wie wahrscheinlich du alle Anschlüsse bekommst. Das klappt wunderbar. Diese Funktion sollte in bahn.de und alle Bahn-Apps eingebaut werden!
Wie…

@drgeraint@glasgow.social
2026-02-09 21:58:38
Content warning: Injury details

The dangers of relying on machine learning uncritically.
reuters.com/investigations/ai-

@ruth_mottram@fediscience.org
2026-03-03 16:55:44

Proud PhD supervisor moment: @… have nice write up to Elke Schlager's brilliant work, (now a preprint in @…) on how we can use #MachineLearning methods to emulate #Greenland ice sheet melt via European Weather Cloud computing
europeanweather.cloud/use-case

@Carwil@mastodon.online
2026-03-04 01:40:03

Confirmation in the press that the US DoD and Palantir are using AI and machine learning software to produce and speed up target listing. thetimes.com/world/middle-east

@datascience@genomic.social
2026-02-01 11:00:00

Tidy Modeling with R: #rstats #machinelearning

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 16:07:47

Replaced article(s) found for cs.LG. arxiv.org/list/cs.LG/new
[2/6]:
- Performance Asymmetry in Model-Based Reinforcement Learning
Jing Yu Lim, Rushi Shah, Zarif Ikram, Samson Yu, Haozhe Ma, Tze-Yun Leong, Dianbo Liu
arxiv.org/abs/2505.19698 mastoxiv.page/@arXiv_csLG_bot/
- Towards Robust Real-World Multivariate Time Series Forecasting: A Unified Framework for Dependenc...
Jinkwan Jang, Hyungjin Park, Jinmyeong Choi, Taesup Kim
arxiv.org/abs/2506.08660 mastoxiv.page/@arXiv_csLG_bot/
- Wasserstein Barycenter Soft Actor-Critic
Zahra Shahrooei, Ali Baheri
arxiv.org/abs/2506.10167 mastoxiv.page/@arXiv_csLG_bot/
- Foundation Models for Causal Inference via Prior-Data Fitted Networks
Yuchen Ma, Dennis Frauen, Emil Javurek, Stefan Feuerriegel
arxiv.org/abs/2506.10914 mastoxiv.page/@arXiv_csLG_bot/
- FREQuency ATTribution: benchmarking frequency-based occlusion for time series data
Dominique Mercier, Andreas Dengel, Sheraz Ahmed
arxiv.org/abs/2506.18481 mastoxiv.page/@arXiv_csLG_bot/
- Complexity-aware fine-tuning
Andrey Goncharov, Daniil Vyazhev, Petr Sychev, Edvard Khalafyan, Alexey Zaytsev
arxiv.org/abs/2506.21220 mastoxiv.page/@arXiv_csLG_bot/
- Transfer Learning in Infinite Width Feature Learning Networks
Clarissa Lauditi, Blake Bordelon, Cengiz Pehlevan
arxiv.org/abs/2507.04448 mastoxiv.page/@arXiv_csLG_bot/
- A hierarchy tree data structure for behavior-based user segment representation
Liu, Kang, Iyer, Malik, Li, Wang, Lu, Zhao, Wang, Liu, Liu, Liang, Yu
arxiv.org/abs/2508.01115 mastoxiv.page/@arXiv_csLG_bot/
- One-Step Flow Q-Learning: Addressing the Diffusion Policy Bottleneck in Offline Reinforcement Lea...
Thanh Nguyen, Chang D. Yoo
arxiv.org/abs/2508.13904 mastoxiv.page/@arXiv_csLG_bot/
- Uncertainty Propagation Networks for Neural Ordinary Differential Equations
Hadi Jahanshahi, Zheng H. Zhu
arxiv.org/abs/2508.16815 mastoxiv.page/@arXiv_csLG_bot/
- Learning Unified Representations from Heterogeneous Data for Robust Heart Rate Modeling
Zhengdong Huang, Zicheng Xie, Wentao Tian, Jingyu Liu, Lunhong Dong, Peng Yang
arxiv.org/abs/2508.21785 mastoxiv.page/@arXiv_csLG_bot/
- Monte Carlo Tree Diffusion with Multiple Experts for Protein Design
Liu, Cao, Jiang, Luo, Duan, Wang, Sosnick, Xu, Stevens
arxiv.org/abs/2509.15796 mastoxiv.page/@arXiv_csLG_bot/
- From Samples to Scenarios: A New Paradigm for Probabilistic Forecasting
Xilin Dai, Zhijian Xu, Wanxu Cai, Qiang Xu
arxiv.org/abs/2509.19975 mastoxiv.page/@arXiv_csLG_bot/
- Why High-rank Neural Networks Generalize?: An Algebraic Framework with RKHSs
Yuka Hashimoto, Sho Sonoda, Isao Ishikawa, Masahiro Ikeda
arxiv.org/abs/2509.21895 mastoxiv.page/@arXiv_csLG_bot/
- From Parameters to Behaviors: Unsupervised Compression of the Policy Space
Davide Tenedini, Riccardo Zamboni, Mirco Mutti, Marcello Restelli
arxiv.org/abs/2509.22566 mastoxiv.page/@arXiv_csLG_bot/
- RHYTHM: Reasoning with Hierarchical Temporal Tokenization for Human Mobility
Haoyu He, Haozheng Luo, Yan Chen, Qi R. Wang
arxiv.org/abs/2509.23115 mastoxiv.page/@arXiv_csLG_bot/
- Polychromic Objectives for Reinforcement Learning
Jubayer Ibn Hamid, Ifdita Hasan Orney, Ellen Xu, Chelsea Finn, Dorsa Sadigh
arxiv.org/abs/2509.25424 mastoxiv.page/@arXiv_csLG_bot/
- Recursive Self-Aggregation Unlocks Deep Thinking in Large Language Models
Siddarth Venkatraman, et al.
arxiv.org/abs/2509.26626 mastoxiv.page/@arXiv_csLG_bot/
- Cautious Weight Decay
Chen, Li, Liang, Su, Xie, Pierse, Liang, Lao, Liu
arxiv.org/abs/2510.12402 mastoxiv.page/@arXiv_csLG_bot/
- TeamFormer: Shallow Parallel Transformers with Progressive Approximation
Wei Wang, Xiao-Yong Wei, Qing Li
arxiv.org/abs/2510.15425 mastoxiv.page/@arXiv_csLG_bot/
- Latent-Augmented Discrete Diffusion Models
Dario Shariatian, Alain Durmus, Umut Simsekli, Stefano Peluchetti
arxiv.org/abs/2510.18114 mastoxiv.page/@arXiv_csLG_bot/
- Predicting Metabolic Dysfunction-Associated Steatotic Liver Disease using Machine Learning Method...
Mary E. An, Paul Griffin, Jonathan G. Stine, Ramakrishna Balakrishnan, Soundar Kumara
arxiv.org/abs/2510.22293 mastoxiv.page/@arXiv_csLG_bot/
toXiv_bot_toot

@shriramk@mastodon.social
2026-03-28 14:28:58

An article about using "AI" for drug discovery (that fails to distinguish ML from LLMs from…), but nicely profiles Matt Might's work!
nytimes.com/2025/03/20/well/ai

@seeingwithsound@mas.to
2026-02-27 14:24:49

A machine learning-based decoder framework for the cortical voltage-sensitive dye responses to retinal neuromorphic microstimulation: A proof-of-concept simulation study mdpi.com/2306-5354/13/2/231 Seizure risks not accounted for (e.g. edge-only vision), nor receptive field sizes, etc;

Decoding of images from simulated VSD signals
@jorgecandeias@mastodon.social
2026-04-01 15:08:01

RE: mastodon.social/@nikclayton/11
I very much doubt copyright law will survive the era of generative machine learning.
In code, in music, in literature, in anything.
The absolute abuse and disregard for copyright the…

@Techmeme@techhub.social
2026-04-06 03:45:37

Microsoft is updating devices from Windows 11 24H2 to version 25H2 with no way to fully opt out, and says an "intelligent" ML-based system handles the rollout (Kunal Khullar/Tom's Hardware)
tomshardware.com/software/wind

@ruth_mottram@fediscience.org
2026-03-04 06:39:44

The preprint of the paper is here btw: #MachineLearning methods to emulate #Greenland ice sheet melt via European Weather Cloud computing
europeanweather.cloud/use-case

@underdarkGIS@fosstodon.org
2026-03-20 21:56:04

It's going to be a busy #EGU poster session for our team this year:
🌲 Explainable Machine Learning for diagnosing Data Quality Issues in Dendrometer-Based Tree Growth Time Series meetingorganizer.coperni…

@v_i_o_l_a@openbiblio.social
2026-01-25 15:29:38

"Author Name Disambiguation in Scholarly Research: A Bibliometric Perspective"
doi.org/10.1515/opis-2025-0035
"The rapid expansion of scholarly publishing has amplified the long-standing challenge of author name ambiguity in academic databases. This issue, manifesting a…

@arXiv_csDS_bot@mastoxiv.page
2026-02-10 10:45:35

Incremental (k, z)-Clustering on Graphs
Emilio Cruciani, Sebastian Forster, Antonis Skarlatos
arxiv.org/abs/2602.08542 arxiv.org/pdf/2602.08542 arxiv.org/html/2602.08542
arXiv:2602.08542v1 Announce Type: new
Abstract: Given a weighted undirected graph, a number of clusters $k$, and an exponent $z$, the goal in the $(k, z)$-clustering problem on graphs is to select $k$ vertices as centers that minimize the sum of the distances raised to the power $z$ of each vertex to its closest center. In the dynamic setting, the graph is subject to adversarial edge updates, and the goal is to maintain explicitly an exact $(k, z)$-clustering solution in the induced shortest-path metric.
While efficient dynamic $k$-center approximation algorithms on graphs exist [Cruciani et al. SODA 2024], to the best of our knowledge, no prior work provides similar results for the dynamic $(k,z)$-clustering problem. As the main result of this paper, we develop a randomized incremental $(k, z)$-clustering algorithm that maintains with high probability a constant-factor approximation in a graph undergoing edge insertions with a total update time of $\tilde O(k m^{1 o(1)} k^{1 \frac{1}{\lambda}} m)$, where $\lambda \geq 1$ is an arbitrary fixed constant. Our incremental algorithm consists of two stages. In the first stage, we maintain a constant-factor bicriteria approximate solution of size $\tilde{O}(k)$ with a total update time of $m^{1 o(1)}$ over all adversarial edge insertions. This first stage is an intricate adaptation of the bicriteria approximation algorithm by Mettu and Plaxton [Machine Learning 2004] to incremental graphs. One of our key technical results is that the radii in their algorithm can be assumed to be non-decreasing while the approximation ratio remains constant, a property that may be of independent interest.
In the second stage, we maintain a constant-factor approximate $(k,z)$-clustering solution on a dynamic weighted instance induced by the bicriteria approximate solution. For this subproblem, we employ a dynamic spanner algorithm together with a static $(k,z)$-clustering algorithm.
toXiv_bot_toot

@primonatura@mstdn.social
2026-02-20 15:00:14

"Satellite images indicate that the Doñana Marshland will disappear within 60 years"
#Environment
phys.org/news/2026-02-satellit

@cellfourteen@social.petertoushkov.eu
2026-01-21 15:05:17

I like AI. I like robots. I love machine learning automation. I don't like it when their use cases are replacing people, spying on or profiling people, prosecuting people, submitting people into subscription, creating "art", "videos", "pictures" and scumbag "memes", warfare, propaganda, disinformation, deepfakes of any kind, ads, trolling, pumping up stocks, just plain wrong search results that force you to waste twice as much time to confirm they …

@hw@fediscience.org
2026-02-23 07:19:51

Audrey Watters writes about how the #AI 'tsunami' in #edtech follows the same trajectory as all the previous technological hype cycles:
"There will be no “AI” tutor revolution just as there was no MOOC revolution just as there was no personalized learning revolution just as there was no computer-assisted instruction revolution just as there was no teaching machine revolution."
2ndbreakfast.audreywatters.com

@arXiv_physicschemph_bot@mastoxiv.page
2026-03-27 08:19:37

Autotuning T-PaiNN: Enabling Data-Efficient GNN Interatomic Potential Development via Classical-to-Quantum Transfer Learning
Vivienne Pelletier, Vedant Bhat, Daniel J. Rivera, Steven A. Wilson, Christopher L. Muhich
arxiv.org/abs/2603.24752 arxiv.org/pdf/2603.24752 arxiv.org/html/2603.24752
arXiv:2603.24752v1 Announce Type: new
Abstract: Machine-learned interatomic potentials (MLIPs), particularly graph neural network (GNN)-based models, offer a promising route to achieving near-density functional theory (DFT) accuracy at significantly reduced computational cost. However, their practical deployment is often limited by the large volumes of expensive quantum mechanical training data required. In this work, we introduce a transfer learning framework, Transfer-PaiNN (T-PaiNN), that substantially improves the data efficiency of GNN-MLIPs by leveraging inexpensive classical force field data. The approach consists of pretraining a PaiNN MLIP architecture on large-scale datasets generated from classical molecular simulations, followed by fine-tuning (dubbed autotuning) using a comparatively small DFT dataset. We demonstrate the effectiveness of autotuning T-PaiNN on both gas-phase molecular systems (QM9 dataset) and condensed-phase liquid water. Across all cases, T-PaiNN significantly outperforms models trained solely on DFT data, achieving order-of-magnitude reductions in mean absolute error while accelerating training convergence. For example, using the QM9 data set, error reductions of up to 25 times are observed in low-data regimes, while liquid water simulations show improved predictions of energies, forces, and experimentally relevant properties such as density and diffusion. These gains arise from the model's ability to learn general features of the potential energy surface from extensive classical sampling, which are subsequently refined to quantum accuracy. Overall, this work establishes transfer learning from classical force fields as a practical and computationally efficient strategy for developing high-accuracy, data-efficient GNN interatomic potentials, enabling broader application of MLIPs to complex chemical systems.
toXiv_bot_toot

@NFL@darktundra.xyz
2026-01-24 19:26:37

NFL player props, 2026 AFC, NFC Championship picks, odds, AI predictions: Puka Nacua Over 92.5 receiving yards

cbssports.com/nfl/news/nfl-pla

@arXiv_qbioGN_bot@mastoxiv.page
2026-03-10 08:59:39

Identifying genes associated with phenotypes using machine and deep learning
Muhammad Muneeb, David B. Ascher, YooChan Myung
arxiv.org/abs/2603.06804

@arXiv_condmatstrel_bot@mastoxiv.page
2026-02-03 08:18:44

Machine Learning to Predict Spectral Anisotropy in Valence-to-Core X-ray Emission Spectroscopy
Charles A. Cardot, John Tichenor, Seth M. Shjandemaar, Josh J. Kas, Fernando D. Vila, Gerald T. Seidler, John J. Rehr
arxiv.org/abs/2602.00242

@arXiv_csCL_bot@mastoxiv.page
2026-03-31 10:40:59

Crosslisted article(s) found for cs.CL. arxiv.org/list/cs.CL/new
[2/2]:
- The Geometry of Harmful Intent: Training-Free Anomaly Detection via Angular Deviation in LLM Resi...
Isaac Llorente-Saguer
arxiv.org/abs/2603.27412 mastoxiv.page/@arXiv_csLG_bot/
- LongCat-Next: Lexicalizing Modalities as Discrete Tokens
Meituan LongCat Team, et al.
arxiv.org/abs/2603.27538 mastoxiv.page/@arXiv_csCV_bot/
- Emergent Social Intelligence Risks in Generative Multi-Agent Systems
Huang, Jiang, Wang, Zhuang, Luo, Ma, Xu, Chen, Moniz, Lin, Chen, Chawla, Dziri, Sun, Zhang
arxiv.org/abs/2603.27771 mastoxiv.page/@arXiv_csMA_bot/
- KVSculpt: KV Cache Compression as Distillation
Bo Jiang, Sian Jin
arxiv.org/abs/2603.27819 mastoxiv.page/@arXiv_csLG_bot/
- Q-Bridge: Code Translation for Quantum Machine Learning via LLMs
Runjia Zeng, Priyabrata Senapati, Ruixiang Tang, Dongfang Liu, Qiang Guan
arxiv.org/abs/2603.27836 mastoxiv.page/@arXiv_quantph_b
- EffiSkill: Agent Skill Based Automated Code Efficiency Optimization
Zimu Wang, Yuling Shi, Mengfan Li, Zijun Liu, Jie M. Zhang, Chengcheng Wan, Xiaodong Gu
arxiv.org/abs/2603.27850 mastoxiv.page/@arXiv_csSE_bot/
- Efficient Inference of Large Vision Language Models
Surendra Pathak
arxiv.org/abs/2603.27960 mastoxiv.page/@arXiv_csLG_bot/
- CDH-Bench: A Commonsense-Driven Hallucination Benchmark for Evaluating Visual Fidelity in Vision-...
Kesheng Chen, Yamin Hu, Qi Zhou, Zhenqian Zhu, Wenjian Luo
arxiv.org/abs/2603.27982 mastoxiv.page/@arXiv_csCV_bot/
- MOSS-VoiceGenerator: Create Realistic Voices with Natural Language Descriptions
Huang, Fan, Jiang, Jiang, Tu, Zhu, Zhang, Zhao, Yang, Fei, Li, Yang, Cheng, Qiu
arxiv.org/abs/2603.28086 mastoxiv.page/@arXiv_csSD_bot/
- Does Claude's Constitution Have a Culture?
Parham Pourdavood
arxiv.org/abs/2603.28123 mastoxiv.page/@arXiv_csCY_bot/
- MiroEval: Benchmarking Multimodal Deep Research Agents in Process and Outcome
Fangda Ye, et al.
arxiv.org/abs/2603.28407 mastoxiv.page/@arXiv_csAI_bot/
- IsoQuant: Hardware-Aligned SO(4) Isoclinic Rotations for LLM KV Cache Compression
Zhongping Ji
arxiv.org/abs/2603.28430 mastoxiv.page/@arXiv_csLG_bot/
- Entropic Claim Resolution: Uncertainty-Driven Evidence Selection for RAG
Davide Di Gioia
arxiv.org/abs/2603.28444 mastoxiv.page/@arXiv_csAI_bot/
- Moving Beyond Review: Applying Language Models to Planning and Translation in Reflection
Seyed Parsa Neshaei, Richard Lee Davis, Tanja K\"aser
arxiv.org/abs/2603.28596 mastoxiv.page/@arXiv_csHC_bot/
- ResAdapt: Adaptive Resolution for Efficient Multimodal Reasoning
Huanxuan Liao, Zhongtao Jiang, Yupu Hao, Yuqiao Tan, Shizhu He, Jun Zhao, Kun Xu, Kang Liu
arxiv.org/abs/2603.28610 mastoxiv.page/@arXiv_csCV_bot/
- The Ultimate Tutorial for AI-driven Scale Development in Generative Psychometrics: Releasing AIGE...
Lara Russell-Lasalandra, Hudson Golino, Luis Eduardo Garrido, Alexander P. Christensen
arxiv.org/abs/2603.28643 mastoxiv.page/@arXiv_csAI_bot/
- SOLE-R1: Video-Language Reasoning as the Sole Reward for On-Robot Reinforcement Learning
Philip Schroeder, Thomas Weng, Karl Schmeckpeper, Eric Rosen, Stephen Hart, Ondrej Biza
arxiv.org/abs/2603.28730 mastoxiv.page/@arXiv_csRO_bot/
- ParaSpeechCLAP: A Dual-Encoder Speech-Text Model for Rich Stylistic Language-Audio Pretraining
Anuj Diwan, Eunsol Choi, David Harwath
arxiv.org/abs/2603.28737 mastoxiv.page/@arXiv_eessAS_bo
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 10:36:21

On Electric Vehicle Energy Demand Forecasting and the Effect of Federated Learning
Andreas Tritsarolis, Gil Sampaio, Nikos Pelekis, Yannis Theodoridis
arxiv.org/abs/2602.20782 arxiv.org/pdf/2602.20782 arxiv.org/html/2602.20782
arXiv:2602.20782v1 Announce Type: new
Abstract: The wide spread of new energy resources, smart devices, and demand side management strategies has motivated several analytics operations, from infrastructure load modeling to user behavior profiling. Energy Demand Forecasting (EDF) of Electric Vehicle Supply Equipments (EVSEs) is one of the most critical operations for ensuring efficient energy management and sustainability, since it enables utility providers to anticipate energy/power demand, optimize resource allocation, and implement proactive measures to improve grid reliability. However, accurate EDF is a challenging problem due to external factors, such as the varying user routines, weather conditions, driving behaviors, unknown state of charge, etc. Furthermore, as concerns and restrictions about privacy and sustainability have grown, training data has become increasingly fragmented, resulting in distributed datasets scattered across different data silos and/or edge devices, calling for federated learning solutions. In this paper, we investigate different well-established time series forecasting methodologies to address the EDF problem, from statistical methods (the ARIMA family) to traditional machine learning models (such as XGBoost) and deep neural networks (GRU and LSTM). We provide an overview of these methods through a performance comparison over four real-world EVSE datasets, evaluated under both centralized and federated learning paradigms, focusing on the trade-offs between forecasting fidelity, privacy preservation, and energy overheads. Our experimental results demonstrate, on the one hand, the superiority of gradient boosted trees (XGBoost) over statistical and NN-based models in both prediction accuracy and energy efficiency and, on the other hand, an insight that Federated Learning-enabled models balance these factors, offering a promising direction for decentralized energy demand forecasting.
toXiv_bot_toot

Short-range kamikaze drones are one of the fastest moving facets of the defense sector today —
The Marine Corps "Organic Precision Fires-Light" (OPF-L) program, is designed to provide dismounted Marine infantry rifle squads with a man-packable, easy-to-operate precision strike drone to engage adversaries beyond line of sight.
A recent announcement of a $23.9-million contract to provide the U.S. Marine Corps with more than 600 "Bolt-M" drones is the next phas…

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 16:07:58

Replaced article(s) found for cs.LG. arxiv.org/list/cs.LG/new
[3/6]:
- Towards Scalable Oversight via Partitioned Human Supervision
Ren Yin, Takashi Ishida, Masashi Sugiyama
arxiv.org/abs/2510.22500 mastoxiv.page/@arXiv_csLG_bot/
- ContextPilot: Fast Long-Context Inference via Context Reuse
Yinsicheng Jiang, Yeqi Huang, Liang Cheng, Cheng Deng, Xuan Sun, Luo Mai
arxiv.org/abs/2511.03475 mastoxiv.page/@arXiv_csLG_bot/
- Metabolomic Biomarker Discovery for ADHD Diagnosis Using Interpretable Machine Learning
Nabil Belacel, Mohamed Rachid Boulassel
arxiv.org/abs/2601.11283 mastoxiv.page/@arXiv_csLG_bot/
- PhysE-Inv: A Physics-Encoded Inverse Modeling approach for Arctic Snow Depth Prediction
Akila Sampath, Vandana Janeja, Jianwu Wang
arxiv.org/abs/2601.17074
- SAGE-5GC: Security-Aware Guidelines for Evaluating Anomaly Detection in the 5G Core Network
Cristian Manca, Christian Scano, Giorgio Piras, Fabio Brau, Maura Pintor, Battista Biggio
arxiv.org/abs/2602.03596
- LORE: Jointly Learning the Intrinsic Dimensionality and Relative Similarity Structure From Ordina...
Anand, Helbling, Davenport, Berman, Alagapan, Rozell
arxiv.org/abs/2602.04192
- Towards Robust Scaling Laws for Optimizers
Alexandra Volkova, Mher Safaryan, Christoph H. Lampert, Dan Alistarh
arxiv.org/abs/2602.07712 mastoxiv.page/@arXiv_csLG_bot/
- Do We Need Adam? Surprisingly Strong and Sparse Reinforcement Learning with SGD in LLMs
Sagnik Mukherjee, Lifan Yuan, Pavan Jayasinha, Dilek Hakkani-T\"ur, Hao Peng
arxiv.org/abs/2602.07729 mastoxiv.page/@arXiv_csLG_bot/
- AceGRPO: Adaptive Curriculum Enhanced Group Relative Policy Optimization for Autonomous Machine L...
Yuzhu Cai, Zexi Liu, Xinyu Zhu, Cheng Wang, Siheng Chen
arxiv.org/abs/2602.07906 mastoxiv.page/@arXiv_csLG_bot/
- VESPO: Variational Sequence-Level Soft Policy Optimization for Stable Off-Policy LLM Training
Guobin Shen, Chenxiao Zhao, Xiang Cheng, Lei Huang, Xing Yu
arxiv.org/abs/2602.10693 mastoxiv.page/@arXiv_csLG_bot/
- KBVQ-MoE: KLT-guided SVD with Bias-Corrected Vector Quantization for MoE Large Language Models
Zukang Xu, Zhixiong Zhao, Xing Hu, Zhixuan Chen, Dawei Yang
arxiv.org/abs/2602.11184 mastoxiv.page/@arXiv_csLG_bot/
- MUSE: Multi-Tenant Model Serving With Seamless Model Updates
Correia, Ferreira, Martins, Bento, Guerreiro, Pereira, Gomes, Bono, Ferreira, Bizarro
arxiv.org/abs/2602.11776 mastoxiv.page/@arXiv_csLG_bot/
- Pawsterior: Variational Flow Matching for Structured Simulation-Based Inference
Jorge Carrasco-Pollo, Floor Eijkelboom, Jan-Willem van de Meent
arxiv.org/abs/2602.13813 mastoxiv.page/@arXiv_csLG_bot/
- Silent Inconsistency in Data-Parallel Full Fine-Tuning: Diagnosing Worker-Level Optimization Misa...
Hong Li, Zhen Zhou, Honggang Zhang, Yuping Luo, Xinyue Wang, Han Gong, Zhiyuan Liu
arxiv.org/abs/2602.14462 mastoxiv.page/@arXiv_csLG_bot/
- Divine Benevolence is an $x^2$: GLUs scale asymptotically faster than MLPs
Alejandro Francisco Queiruga
arxiv.org/abs/2602.14495 mastoxiv.page/@arXiv_csLG_bot/
- \"UberWeb: Insights from Multilingual Curation for a 20-Trillion-Token Dataset
DatologyAI, et al.
arxiv.org/abs/2602.15210 mastoxiv.page/@arXiv_csLG_bot/
- GLM-5: from Vibe Coding to Agentic Engineering
GLM-5-Team, et al.
arxiv.org/abs/2602.15763 mastoxiv.page/@arXiv_csLG_bot/
- Anatomy of Capability Emergence: Scale-Invariant Representation Collapse and Top-Down Reorganizat...
Jayadev Billa
arxiv.org/abs/2602.15997 mastoxiv.page/@arXiv_csLG_bot/
- AI-CARE: Carbon-Aware Reporting Evaluation Metric for AI Models
KC Santosh, Srikanth Baride, Rodrigue Rizk
arxiv.org/abs/2602.16042 mastoxiv.page/@arXiv_csLG_bot/
- Beyond Message Passing: A Symbolic Alternative for Expressive and Interpretable Graph Learning
Chuqin Geng, Li Zhang, Haolin Ye, Ziyu Zhao, Yuhe Jiang, Tara Saba, Xinyu Wang, Xujie Si
arxiv.org/abs/2602.16947 mastoxiv.page/@arXiv_csLG_bot/
toXiv_bot_toot

@berlinbuzzwords@floss.social
2026-02-12 15:25:25

Our Call for Papers for Berlin Buzzwords closes this Sunday, February 15!
We encourage everyone in modern data infrastructure, search and machine learning and focused on open source software projects to submit their talk proposals, especially first-timers and people from underrepresented groups! #bbuzz #OpenSource #Berlin #Conference #MachineLearning #Search #DataInfrastructure #DataScience

@arXiv_qbioGN_bot@mastoxiv.page
2026-03-09 08:08:21

Machine Learning for analysis of Multiple Sclerosis cross-tissue bulk and single-cell transcriptomics data
Francesco Massafra, Samuele Punzo, Silvia Giulia Galfr\'e, Alessandro Maglione, Simone Pernice, Stefano Forti, Simona Rolla, Marco Beccuti, Marinella Clerico, Corrado Priami, Alina S\^irbu
arxiv.org/abs/2603.05572

@arXiv_nlinAO_bot@mastoxiv.page
2026-03-23 09:42:03

Replaced article(s) found for nlin.AO. arxiv.org/list/nlin.AO/new
[1/1]:
- Collective dynamics on higher-order networks
Battiston, Bick, Lucas, Mill\'an, Skardal, Zhang
arxiv.org/abs/2510.05253 mastoxiv.page/@arXiv_nlinAO_bo
- Interpretable Early Warnings using Machine Learning in an Online Game-experiment
Guillaume Falmagne, Anna B. Stephenson, Simon A. Levin
arxiv.org/abs/2502.09880 mastoxiv.page/@arXiv_physicsso
toXiv_bot_toot

@mgorny@social.treehouse.systems
2026-04-05 13:14:07

I'm sorry to say that I actually wrote it:
"The pinnacle of enshittification, or Large Language Models"
#AI #LLM #NoAI #NoLLM

@arXiv_physicschemph_bot@mastoxiv.page
2026-03-27 09:36:37

Crosslisted article(s) found for physics.chem-ph. arxiv.org/list/physics.chem-ph
[1/1]:
- How unconstrained machine-learning models learn physical symmetries
Michelangelo Domina, Joseph William Abbott, Paolo Pegolo, Filippo Bigi, Michele Ceriotti
arxiv.org/abs/2603.24638 mastoxiv.page/@arXiv_csLG_bot/
- Concerted Electron-Ion Transport by Polyacrylonitrile Elucidated with Reactive Deep Learning Pote...
Chahal-Crockett, Toomey, Kearney, Gao, Damron, Naskar, Roy
arxiv.org/abs/2603.24798 mastoxiv.page/@arXiv_condmatmt
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 16:08:08

Replaced article(s) found for cs.LG. arxiv.org/list/cs.LG/new
[4/6]:
- Neural Proposals, Symbolic Guarantees: Neuro-Symbolic Graph Generation with Hard Constraints
Chuqin Geng, Li Zhang, Mark Zhang, Haolin Ye, Ziyu Zhao, Xujie Si
arxiv.org/abs/2602.16954 mastoxiv.page/@arXiv_csLG_bot/
- Multi-Probe Zero Collision Hash (MPZCH): Mitigating Embedding Collisions and Enhancing Model Fres...
Ziliang Zhao, et al.
arxiv.org/abs/2602.17050 mastoxiv.page/@arXiv_csLG_bot/
- MASPO: Unifying Gradient Utilization, Probability Mass, and Signal Reliability for Robust and Sam...
Fu, Lin, Fang, Zheng, Hu, Shao, Qin, Pan, Zeng, Cai
arxiv.org/abs/2602.17550 mastoxiv.page/@arXiv_csLG_bot/
- A Theoretical Framework for Modular Learning of Robust Generative Models
Corinna Cortes, Mehryar Mohri, Yutao Zhong
arxiv.org/abs/2602.17554 mastoxiv.page/@arXiv_csLG_bot/
- Multi-Round Human-AI Collaboration with User-Specified Requirements
Sima Noorani, Shayan Kiyani, Hamed Hassani, George Pappas
arxiv.org/abs/2602.17646 mastoxiv.page/@arXiv_csLG_bot/
- NEXUS: A compact neural architecture for high-resolution spatiotemporal air quality forecasting i...
Rampunit Kumar, Aditya Maheshwari
arxiv.org/abs/2602.19654 mastoxiv.page/@arXiv_csLG_bot/
- Augmenting Lateral Thinking in Language Models with Humor and Riddle Data for the BRAINTEASER Task
Mina Ghashami, Soumya Smruti Mishra
arxiv.org/abs/2405.10385 mastoxiv.page/@arXiv_csCL_bot/
- Watermarking Language Models with Error Correcting Codes
Patrick Chao, Yan Sun, Edgar Dobriban, Hamed Hassani
arxiv.org/abs/2406.10281 mastoxiv.page/@arXiv_csCR_bot/
- Learning to Control Unknown Strongly Monotone Games
Siddharth Chandak, Ilai Bistritz, Nicholas Bambos
arxiv.org/abs/2407.00575 mastoxiv.page/@arXiv_csMA_bot/
- Classification and reconstruction for single-pixel imaging with classical and quantum neural netw...
Sofya Manko, Dmitry Frolovtsev
arxiv.org/abs/2407.12506 mastoxiv.page/@arXiv_quantph_b
- Statistical Inference for Temporal Difference Learning with Linear Function Approximation
Weichen Wu, Gen Li, Yuting Wei, Alessandro Rinaldo
arxiv.org/abs/2410.16106 mastoxiv.page/@arXiv_statML_bo
- Big data approach to Kazhdan-Lusztig polynomials
Abel Lacabanne, Daniel Tubbenhauer, Pedro Vaz
arxiv.org/abs/2412.01283 mastoxiv.page/@arXiv_mathRT_bo
- MoEMba: A Mamba-based Mixture of Experts for High-Density EMG-based Hand Gesture Recognition
Mehran Shabanpour, Kasra Rad, Sadaf Khademi, Arash Mohammadi
arxiv.org/abs/2502.17457 mastoxiv.page/@arXiv_eessSP_bo
- Tightening Optimality gap with confidence through conformal prediction
Miao Li, Michael Klamkin, Russell Bent, Pascal Van Hentenryck
arxiv.org/abs/2503.04071 mastoxiv.page/@arXiv_statML_bo
- SEED: Towards More Accurate Semantic Evaluation for Visual Brain Decoding
Juhyeon Park, Peter Yongho Kim, Jiook Cha, Shinjae Yoo, Taesup Moon
arxiv.org/abs/2503.06437 mastoxiv.page/@arXiv_csCV_bot/
- How much does context affect the accuracy of AI health advice?
Prashant Garg, Thiemo Fetzer
arxiv.org/abs/2504.18310 mastoxiv.page/@arXiv_econGN_bo
- Reproducing and Improving CheXNet: Deep Learning for Chest X-ray Disease Classification
Daniel J. Strick, Carlos Garcia, Anthony Huang, Thomas Gardos
arxiv.org/abs/2505.06646 mastoxiv.page/@arXiv_eessIV_bo
- Sharp Gaussian approximations for Decentralized Federated Learning
Soham Bonnerjee, Sayar Karmakar, Wei Biao Wu
arxiv.org/abs/2505.08125 mastoxiv.page/@arXiv_statML_bo
- HoloLLM: Multisensory Foundation Model for Language-Grounded Human Sensing and Reasoning
Chuhao Zhou, Jianfei Yang
arxiv.org/abs/2505.17645 mastoxiv.page/@arXiv_csCV_bot/
- A Copula Based Supervised Filter for Feature Selection in Diabetes Risk Prediction Using Machine ...
Agnideep Aich, Md Monzur Murshed, Sameera Hewage, Amanda Mayeaux
arxiv.org/abs/2505.22554 mastoxiv.page/@arXiv_statML_bo
- Synthesis of discrete-continuous quantum circuits with multimodal diffusion models
Florian F\"urrutter, Zohim Chandani, Ikko Hamamura, Hans J. Briegel, Gorka Mu\~noz-Gil
arxiv.org/abs/2506.01666 mastoxiv.page/@arXiv_quantph_b
toXiv_bot_toot

@arXiv_csCL_bot@mastoxiv.page
2026-03-31 10:12:22

Training data generation for context-dependent rubric-based short answer grading
Pavel \v{S}indel\'a\v{r}, D\'avid Slivka, Christopher Bouma, Filip Pr\'a\v{s}il, Ond\v{r}ej Bojar
arxiv.org/abs/2603.28537 arxiv.org/pdf/2603.28537 arxiv.org/html/2603.28537
arXiv:2603.28537v1 Announce Type: new
Abstract: Every 4 years, the PISA test is administered by the OECD to test the knowledge of teenage students worldwide and allow for comparisons of educational systems. However, having to avoid language differences and annotator bias makes the grading of student answers challenging. For these reasons, it would be interesting to compare methods of automatic student answer grading. To train some of these methods, which require machine learning, or to compute parameters or select hyperparameters for those that do not, a large amount of domain-specific data is needed. In this work, we explore a small number of methods for creating a large-scale training dataset using only a relatively small confidential dataset as a reference, leveraging a set of very simple derived text formats to preserve confidentiality. Using these methods, we successfully created three surrogate datasets that are, at the very least, superficially more similar to the reference dataset than purely the result of prompt-based generation. Early experiments suggest one of these approaches might also lead to improved model training.
toXiv_bot_toot

@arXiv_condmatdisnn_bot@mastoxiv.page
2026-01-21 22:50:45

Replaced article(s) found for cond-mat.dis-nn. arxiv.org/list/cond-mat.dis-nn
[1/1]:
- Machine Learning Symmetry Discovery for Integrable Hamiltonian Dynamics
Wanda Hou, Molan Li, Yi-Zhuang You

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 16:08:29

Replaced article(s) found for cs.LG. arxiv.org/list/cs.LG/new
[6/6]:
- Fast-ThinkAct: Efficient Vision-Language-Action Reasoning via Verbalizable Latent Planning
Chi-Pin Huang, Yunze Man, Zhiding Yu, Min-Hung Chen, Jan Kautz, Yu-Chiang Frank Wang, Fu-En Yang
arxiv.org/abs/2601.09708 mastoxiv.page/@arXiv_csCV_bot/
- Universality of Many-body Projected Ensemble for Learning Quantum Data Distribution
Quoc Hoan Tran, Koki Chinzei, Yasuhiro Endo, Hirotaka Oshima
arxiv.org/abs/2601.18637 mastoxiv.page/@arXiv_quantph_b
- FROST: Filtering Reasoning Outliers with Attention for Efficient Reasoning
Haozheng Luo, Zhuolin Jiang, Md Zahid Hasan, Yan Chen, Soumalya Sarkar
arxiv.org/abs/2601.19001 mastoxiv.page/@arXiv_csCL_bot/
- Analysis of Shuffling Beyond Pure Local Differential Privacy
Shun Takagi, Seng Pei Liew
arxiv.org/abs/2601.19154 mastoxiv.page/@arXiv_csDS_bot/
- CryoLVM: Self-supervised Learning from Cryo-EM Density Maps with Large Vision Models
Weining Fu, Kai Shu, Kui Xu, Qiangfeng Cliff Zhang
arxiv.org/abs/2602.02620
- XtraLight-MedMamba for Classification of Neoplastic Tubular Adenomas
Sultana, Afsar, Rahu, Singh, Shula, Combs, Forchetti, Asari
arxiv.org/abs/2602.04819
- Flow-Based Conformal Predictive Distributions
Trevor Harris
arxiv.org/abs/2602.07633 mastoxiv.page/@arXiv_statML_bo
- GOT-Edit: Geometry-Aware Generic Object Tracking via Online Model Editing
Shih-Fang Chen, Jun-Cheng Chen, I-Hong Jhuo, Yen-Yu Lin
arxiv.org/abs/2602.08550 mastoxiv.page/@arXiv_csCV_bot/
- UI-Venus-1.5 Technical Report
Venus Team, et al.
arxiv.org/abs/2602.09082 mastoxiv.page/@arXiv_csCV_bot/
- The Wisdom of Many Queries: Complexity-Diversity Principle for Dense Retriever Training
Xincan Feng, Noriki Nishida, Yusuke Sakai, Yuji Matsumoto
arxiv.org/abs/2602.09448 mastoxiv.page/@arXiv_csIR_bot/
- Intent Laundering: AI Safety Datasets Are Not What They Seem
Shahriar Golchin, Marc Wetter
arxiv.org/abs/2602.16729 mastoxiv.page/@arXiv_csCR_bot/
- The Metaphysics We Train: A Heideggerian Reading of Machine Learning
Heman Shakeri
arxiv.org/abs/2602.19028 mastoxiv.page/@arXiv_csCY_bot/
- Skill-Inject: Measuring Agent Vulnerability to Skill File Attacks
David Schmotz, Luca Beurer-Kellner, Sahar Abdelnabi, Maksym Andriushchenko
arxiv.org/abs/2602.20156 mastoxiv.page/@arXiv_csCR_bot/
- A Very Big Video Reasoning Suite
Maijunxian Wang, et al.
arxiv.org/abs/2602.20159 mastoxiv.page/@arXiv_csCV_bot/
toXiv_bot_toot

@arXiv_physicschemph_bot@mastoxiv.page
2026-03-27 09:36:37

Crosslisted article(s) found for physics.chem-ph. arxiv.org/list/physics.chem-ph
[1/1]:
- How unconstrained machine-learning models learn physical symmetries
Michelangelo Domina, Joseph William Abbott, Paolo Pegolo, Filippo Bigi, Michele Ceriotti

@arXiv_csCL_bot@mastoxiv.page
2026-03-31 11:12:53

Replaced article(s) found for cs.CL. arxiv.org/list/cs.CL/new
[3/5]:
- Can Small Language Models Handle Context-Summarized Multi-Turn Customer-Service QA? A Synthetic D...
Lakshan Cooray, Deshan Sumanathilaka, Pattigadapa Venkatesh Raju
arxiv.org/abs/2602.00665 mastoxiv.page/@arXiv_csCL_bot/
- SEAD: Self-Evolving Agent for Multi-Turn Service Dialogue
Dai, Gao, Zhang, Wang, Luo, Wang, Wang, Wu, Wang
arxiv.org/abs/2602.03548
- OmniRAG-Agent: Agentic Omnimodal Reasoning for Low-Resource Long Audio-Video Question Answering
Yifan Zhu, Xinyu Mu, Tao Feng, Zhonghong Ou, Yuning Gong, Haoran Luo
arxiv.org/abs/2602.03707
- GreekMMLU: A Native-Sourced Multitask Benchmark for Evaluating Language Models in Greek
Zhang, Konomi, Xypolopoulos, Divriotis, Skianis, Nikolentzos, Stamou, Shang, Vazirgiannis
arxiv.org/abs/2602.05150
- Using LLMs for Knowledge Component-level Correctness Labeling in Open-ended Coding Problems
Zhangqi Duan, Arnav Kankaria, Dhruv Kartik, Andrew Lan
arxiv.org/abs/2602.17542 mastoxiv.page/@arXiv_csCL_bot/
- MetaState: Persistent Working Memory Enhances Reasoning in Discrete Diffusion Language Models
Kejing Xia, Mingzhe Li, Lixuan Wei, Zhenbang Du, Xiangchi Yuan, Dachuan Shi, Qirui Jin, Wenke Lee
arxiv.org/abs/2603.01331 mastoxiv.page/@arXiv_csCL_bot/
- A Browser-based Open Source Assistant for Multimodal Content Verification
Milner, Foster, Karmakharm, Razuvayevskaya, Roberts, Porcellini, Teyssou, Bontcheva
arxiv.org/abs/2603.02842 mastoxiv.page/@arXiv_csCL_bot/
- Nw\=ach\=a Mun\=a: A Devanagari Speech Corpus and Proximal Transfer Benchmark for Nepal Bhasha ASR
Sharma, Shrestha, Poudel, Tiwari, Shrestha, Ghimire, Bal
arxiv.org/abs/2603.07554 mastoxiv.page/@arXiv_csCL_bot/
- Model Merging in the Era of Large Language Models: Methods, Applications, and Future Directions
Mingyang Song, Mao Zheng
arxiv.org/abs/2603.09938 mastoxiv.page/@arXiv_csCL_bot/
- AgentDrift: Unsafe Recommendation Drift Under Tool Corruption Hidden by Ranking Metrics in LLM Ag...
Zekun Wu, Adriano Koshiyama, Sahan Bulathwela, Maria Perez-Ortiz
arxiv.org/abs/2603.12564 mastoxiv.page/@arXiv_csCL_bot/
- GhanaNLP Parallel Corpora: Comprehensive Multilingual Resources for Low-Resource Ghanaian Languages
Gyamfi, Azunre, Moore, Budu, Asare, Owusu, Asiamah
arxiv.org/abs/2603.13793 mastoxiv.page/@arXiv_csCL_bot/
- sebis at ArchEHR-QA 2026: How Much Can You Do Locally? Evaluating Grounded EHR QA on a Single Not...
Ibrahim Ebrar Yurt, Fabian Karl, Tejaswi Choppa, Florian Matthes
arxiv.org/abs/2603.13962 mastoxiv.page/@arXiv_csCL_bot/
- ExPosST: Explicit Positioning with Adaptive Masking for LLM-Based Simultaneous Machine Translation
Yuzhe Shang, Pengzhi Gao, Yazheng Yang, Jiayao Ma, Wei Liu, Jian Luan, Jinsong Su
arxiv.org/abs/2603.14903 mastoxiv.page/@arXiv_csCL_bot/
- BanglaSocialBench: A Benchmark for Evaluating Sociopragmatic and Cultural Alignment of LLMs in Ba...
Tanvir Ahmed Sijan, S. M Golam Rifat, Pankaj Chowdhury Partha, Md. Tanjeed Islam, Md. Musfique Anwar
arxiv.org/abs/2603.15949 mastoxiv.page/@arXiv_csCL_bot/
- EngGPT2: Sovereign, Efficient and Open Intelligence
G. Ciarfaglia, et al.
arxiv.org/abs/2603.16430 mastoxiv.page/@arXiv_csCL_bot/
- HypeLoRA: Hyper-Network-Generated LoRA Adapters for Calibrated Language Model Fine-Tuning
Bartosz Trojan, Filip G\k{e}bala
arxiv.org/abs/2603.19278 mastoxiv.page/@arXiv_csCL_bot/
- Automatic Analysis of Collaboration Through Human Conversational Data Resources: A Review
Yi Yu, Maria Boritchev, Chlo\'e Clavel
arxiv.org/abs/2603.19292 mastoxiv.page/@arXiv_csCL_bot/
- Alignment Whack-a-Mole : Finetuning Activates Verbatim Recall of Copyrighted Books in Large Langu...
Xinyue Liu, Niloofar Mireshghallah, Jane C. Ginsburg, Tuhin Chakrabarty
arxiv.org/abs/2603.20957 mastoxiv.page/@arXiv_csCL_bot/
- KG-Hopper: Empowering Compact Open LLMs with Knowledge Graph Reasoning via Reinforcement Learning
Shuai Wang, Yinan Yu
arxiv.org/abs/2603.21440 mastoxiv.page/@arXiv_csCL_bot/
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 10:38:41

On the Generalization Behavior of Deep Residual Networks From a Dynamical System Perspective
Jinshu Huang, Mingfei Sun, Chunlin Wu
arxiv.org/abs/2602.20921 arxiv.org/pdf/2602.20921 arxiv.org/html/2602.20921
arXiv:2602.20921v1 Announce Type: new
Abstract: Deep neural networks (DNNs) have significantly advanced machine learning, with model depth playing a central role in their successes. The dynamical system modeling approach has recently emerged as a powerful framework, offering new mathematical insights into the structure and learning behavior of DNNs. In this work, we establish generalization error bounds for both discrete- and continuous-time residual networks (ResNets) by combining Rademacher complexity, flow maps of dynamical systems, and the convergence behavior of ResNets in the deep-layer limit. The resulting bounds are of order $O(1/\sqrt{S})$ with respect to the number of training samples $S$, and include a structure-dependent negative term, yielding depth-uniform and asymptotic generalization bounds under milder assumptions. These findings provide a unified understanding of generalization across both discrete- and continuous-time ResNets, helping to close the gap in both the order of sample complexity and assumptions between the discrete- and continuous-time settings.
toXiv_bot_toot

@arXiv_physicschemph_bot@mastoxiv.page
2026-03-26 09:57:22

Replaced article(s) found for physics.chem-ph. arxiv.org/list/physics.chem-ph
[1/1]:
- Proposal on the Calculation of the Ionisation-Cluster Size Distribution (I). The Model and Its Si...
Bernd Heide
arxiv.org/abs/2404.03961 mastoxiv.page/@arXiv_physicsco
- Bridging chemistry and Gaussian boson sampling: A photonic hierarchy of approximations for molecu...
Jan-Lucas Eickmann, et al.
arxiv.org/abs/2507.19442 mastoxiv.page/@arXiv_quantph_b
- Benchmarking Universal Machine Learning Interatomic Potentials for Supported Nanoparticles: Decou...
Jiayan Xu, Abhirup Patra, Amar Deep Pathak, Sharan Shetty, Detlef Hohl, Roberto Car
arxiv.org/abs/2512.05221 mastoxiv.page/@arXiv_condmatmt
- Knowledge Distillation of a Protein Language Model Yields a Foundational Implicit Solvent Model
Justin Airas, Bin Zhang
arxiv.org/abs/2601.05388 mastoxiv.page/@arXiv_physicsbi
- Universal Foundations of Thermodynamics: Entropy and Energy Beyond Equilibrium and Without Extens...
Gian Paolo Beretta
arxiv.org/abs/2602.09986 mastoxiv.page/@arXiv_quantph_b
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 10:34:01

Bikelution: Federated Gradient-Boosting for Scalable Shared Micro-Mobility Demand Forecasting
Antonios Tziorvas, Andreas Tritsarolis, Yannis Theodoridis
arxiv.org/abs/2602.20671 arxiv.org/pdf/2602.20671 arxiv.org/html/2602.20671
arXiv:2602.20671v1 Announce Type: new
Abstract: The rapid growth of dockless bike-sharing systems has generated massive spatio-temporal datasets useful for fleet allocation, congestion reduction, and sustainable mobility. Bike demand, however, depends on several external factors, making traditional time-series models insufficient. Centralized Machine Learning (CML) yields high-accuracy forecasts but raises privacy and bandwidth issues when data are distributed across edge devices. To overcome these limitations, we propose Bikelution, an efficient Federated Learning (FL) solution based on gradient-boosted trees that preserves privacy while delivering accurate mid-term demand forecasts up to six hours ahead. Experiments on three real-world BSS datasets show that Bikelution is comparable to its CML-based variant and outperforms the current state-of-the-art. The results highlight the feasibility of privacy-aware demand forecasting and outline the trade-offs between FL and CML approaches.
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 10:38:51

Hierarchic-EEG2Text: Assessing EEG-To-Text Decoding across Hierarchical Abstraction Levels
Anupam Sharma, Harish Katti, Prajwal Singh, Shanmuganathan Raman, Krishna Miyapuram
arxiv.org/abs/2602.20932 arxiv.org/pdf/2602.20932 arxiv.org/html/2602.20932
arXiv:2602.20932v1 Announce Type: new
Abstract: An electroencephalogram (EEG) records the spatially averaged electrical activity of neurons in the brain, measured from the human scalp. Prior studies have explored EEG-based classification of objects or concepts, often for passive viewing of briefly presented image or video stimuli, with limited classes. Because EEG exhibits a low signal-to-noise ratio, recognizing fine-grained representations across a large number of classes remains challenging; however, abstract-level object representations may exist. In this work, we investigate whether EEG captures object representations across multiple hierarchical levels, and propose episodic analysis, in which a Machine Learning (ML) model is evaluated across various, yet related, classification tasks (episodes). Unlike prior episodic EEG studies that rely on fixed or randomly sampled classes of equal cardinality, we adopt hierarchy-aware episode sampling using WordNet to generate episodes with variable classes of diverse hierarchy. We also present the largest episodic framework in the EEG domain for detecting observed text from EEG signals in the PEERS dataset, comprising $931538$ EEG samples under $1610$ object labels, acquired from $264$ human participants (subjects) performing controlled cognitive tasks, enabling the study of neural dynamics underlying perception, decision-making, and performance monitoring.
We examine how the semantic abstraction level affects classification performance across multiple learning techniques and architectures, providing a comprehensive analysis. The models tend to improve performance when the classification categories are drawn from higher levels of the hierarchy, suggesting sensitivity to abstraction. Our work highlights abstraction depth as an underexplored dimension of EEG decoding and motivates future research in this direction.
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 12:33:36

Crosslisted article(s) found for cs.LG. arxiv.org/list/cs.LG/new
[2/3]:
- Diffusion Modulation via Environment Mechanism Modeling for Planning
Hanping Zhang, Yuhong Guo
arxiv.org/abs/2602.20422 mastoxiv.page/@arXiv_csAI_bot/
- Heterogeneity-Aware Client Selection Methodology For Efficient Federated Learning
Nihal Balivada, Shrey Gupta, Shashank Shreedhar Bhatt, Suyash Gupta
arxiv.org/abs/2602.20450 mastoxiv.page/@arXiv_csDC_bot/
- Prior-Agnostic Incentive-Compatible Exploration
Ramya Ramalingam, Osbert Bastani, Aaron Roth
arxiv.org/abs/2602.20465 mastoxiv.page/@arXiv_csGT_bot/
- PhyGHT: Physics-Guided HyperGraph Transformer for Signal Purification at the HL-LHC
Mohammed Rakib, Luke Vaughan, Shivang Patel, Flera Rizatdinova, Alexander Khanov, Atriya Sen
arxiv.org/abs/2602.20475 mastoxiv.page/@arXiv_hepex_bot
- ActionEngine: From Reactive to Programmatic GUI Agents via State Machine Memory
Zhong, Faisal, Fran\c{c}a, Leesatapornwongsa, Szekeres, Rong, Nath
arxiv.org/abs/2602.20502 mastoxiv.page/@arXiv_csAI_bot/
- Inner Speech as Behavior Guides: Steerable Imitation of Diverse Behaviors for Human-AI coordination
Rakshit Trivedi, Kartik Sharma, David C Parkes
arxiv.org/abs/2602.20517 mastoxiv.page/@arXiv_csAI_bot/
- Stop-Think-AutoRegress: Language Modeling with Latent Diffusion Planning
Lovelace, Belardi, Zalouk, Polavaram, Kundurthy, Weinberger
arxiv.org/abs/2602.20528 mastoxiv.page/@arXiv_csCL_bot/
- Standard Transformers Achieve the Minimax Rate in Nonparametric Regression with $C^{s,\lambda}$ T...
Yanming Lai, Defeng Sun
arxiv.org/abs/2602.20555 mastoxiv.page/@arXiv_statML_bo
- Personal Information Parroting in Language Models
Nishant Subramani, Kshitish Ghate, Mona Diab
arxiv.org/abs/2602.20580 mastoxiv.page/@arXiv_csCL_bot/
- Characterizing Online and Private Learnability under Distributional Constraints via Generalized S...
Mo\"ise Blanchard, Abhishek Shetty, Alexander Rakhlin
arxiv.org/abs/2602.20585 mastoxiv.page/@arXiv_statML_bo
- Amortized Bayesian inference for actigraph time sheet data from mobile devices
Daniel Zhou, Sudipto Banerjee
arxiv.org/abs/2602.20611 mastoxiv.page/@arXiv_statML_bo
- Knowing the Unknown: Interpretable Open-World Object Detection via Concept Decomposition Model
Xueqiang Lv, Shizhou Zhang, Yinghui Xing, Di Xu, Peng Wang, Yanning Zhang
arxiv.org/abs/2602.20616 mastoxiv.page/@arXiv_csCV_bot/
- On the Convergence of Stochastic Gradient Descent with Perturbed Forward-Backward Passes
Boao Kong, Hengrui Zhang, Kun Yuan
arxiv.org/abs/2602.20646 mastoxiv.page/@arXiv_mathOC_bo
- DANCE: Doubly Adaptive Neighborhood Conformal Estimation
Feng, Reich, Beaglehole, Luo, Park, Yoo, Huang, Mao, Boz, Kim
arxiv.org/abs/2602.20652 mastoxiv.page/@arXiv_statML_bo
- Vision-Language Models for Ergonomic Assessment of Manual Lifting Tasks: Estimating Horizontal an...
Mohammad Sadra Rajabi, Aanuoluwapo Ojelade, Sunwook Kim, Maury A. Nussbaum
arxiv.org/abs/2602.20658 mastoxiv.page/@arXiv_csCV_bot/
- F10.7 Index Prediction: A Multiscale Decomposition Strategy with Wavelet Transform for Performanc...
Xuran Ma, et al.
arxiv.org/abs/2602.20712 mastoxiv.page/@arXiv_astrophIM
- Communication-Inspired Tokenization for Structured Image Representations
Davtyan, Sahin, Haghighi, Stapf, Acuaviva, Alahi, Favaro
arxiv.org/abs/2602.20731 mastoxiv.page/@arXiv_csCV_bot/
- SibylSense: Adaptive Rubric Learning via Memory Tuning and Adversarial Probing
Yifei Xu, et al.
arxiv.org/abs/2602.20751 mastoxiv.page/@arXiv_csCL_bot/
- Assessing the Impact of Speaker Identity in Speech Spoofing Detection
Anh-Tuan Dao, Driss Matrouf, Nicholas Evans
arxiv.org/abs/2602.20805 mastoxiv.page/@arXiv_csSD_bot/
- Don't Ignore the Tail: Decoupling top-K Probabilities for Efficient Language Model Distillation
Sayantan Dasgupta, Trevor Cohn, Timothy Baldwin
arxiv.org/abs/2602.20816 mastoxiv.page/@arXiv_csCL_bot/
- DRESS: A Continuous Framework for Structural Graph Refinement
Eduar Castrillo Velilla
arxiv.org/abs/2602.20833 mastoxiv.page/@arXiv_csDS_bot/
toXiv_bot_toot