Tootfinder

Opt-in global Mastodon full text search. Join the index!

@nemobis@mamot.fr
2026-01-30 10:10:28

Presented without comment
«Economic convergence, demography, labour markets: what progress have EU candidate countries made?
Candidate countries remain far from European Union averages in terms of GDP per capita and their demographic and labour market characteristics»

@seeingwithsound@mas.to
2026-01-10 09:24:46

(LinkedIn) #Neuroscience in 2026: Realistic milestones the scientific community can achieve linkedin.com/pulse/neuroscienc

@arXiv_mathAP_bot@mastoxiv.page
2026-02-10 14:58:34

Crosslisted article(s) found for math.AP. arxiv.org/list/math.AP/new
[1/1]:
- Stability and Convergence of Modal Approximations in Coupled Thermoelastic Systems: Theory and Si...
I. Essadeq, S. Nafiri, S. Benjelloun, A. E. Fettouh

@arXiv_csLG_bot@mastoxiv.page
2025-12-22 10:32:40

Convergence Guarantees for Federated SARSA with Local Training and Heterogeneous Agents
Paul Mangold, Elo\"ise Berthier, Eric Moulines
arxiv.org/abs/2512.17688 arxiv.org/pdf/2512.17688 arxiv.org/html/2512.17688
arXiv:2512.17688v1 Announce Type: new
Abstract: We present a novel theoretical analysis of Federated SARSA (FedSARSA) with linear function approximation and local training. We establish convergence guarantees for FedSARSA in the presence of heterogeneity, both in local transitions and rewards, providing the first sample and communication complexity bounds in this setting. At the core of our analysis is a new, exact multi-step error expansion for single-agent SARSA, which is of independent interest. Our analysis precisely quantifies the impact of heterogeneity, demonstrating the convergence of FedSARSA with multiple local updates. Crucially, we show that FedSARSA achieves linear speed-up with respect to the number of agents, up to higher-order terms due to Markovian sampling. Numerical experiments support our theoretical findings.
toXiv_bot_toot

@cosmos4u@scicomm.xyz
2026-01-22 04:59:09

Dripping to Destruction - Exploring Salt-driven Viscous Surface Convergence in Europa’s Icy Shell: #Europa

@@arXiv_physicsatomph_bot@mastoxiv.page@mastoxiv.page
2026-01-06 14:20:02

Crosslisted article(s) found for physics.atom-ph. arxiv.org/list/physics.atom-ph
[1/1]:
- A quadratic-scaling algorithm with guaranteed convergence for quantum coupled-channel calculations
Hubert J. J\'o\'zwiak, Md Muktadir Rahman, Timur V. Tscherbul

@tiotasram@kolektiva.social
2026-01-18 23:17:10
Content warning: ICE & resistance

In case anyone was wondering about the relevance of #LandBack in the current moment, via CrimeThinc an article on the Minneapolis resistance states:
"""
The Whipple, a federal building in Fort Snelling on the outskirts of the Minneapolis and St. Paul, has long been a regional headquarters for ICE, having previously housed other federal agencies. The complex is located across the street from a National Guard base, down the road from a military base, and next to the preserved fort itself. The fort sits on the sacred site of the convergence of two rivers. It was one of the earliest sites of colonization in the area; at one time, it was a concentration camp holding native Dakota people.
"""
If at any point in the past you ever felt that maybe Native soverignty was a niche issue, or so far from being realized that other causes were more important or relevant, things like this are a good reminder that that cause: overturning the colonial order, is the *same* cause as any meaningful change from the fascist status quo. Things like a "return to democracy" aren't necessarily bad, but the rot runs to the root of this nation, and any intervention that doesn't go that deep is going to leave us right back in this situation again later on.
The fact that ICE is detaining Native Americans is not at all a mistake given their white supremacist aims.
Article link: #ICE #LandBack

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 10:38:41

On the Generalization Behavior of Deep Residual Networks From a Dynamical System Perspective
Jinshu Huang, Mingfei Sun, Chunlin Wu
arxiv.org/abs/2602.20921 arxiv.org/pdf/2602.20921 arxiv.org/html/2602.20921
arXiv:2602.20921v1 Announce Type: new
Abstract: Deep neural networks (DNNs) have significantly advanced machine learning, with model depth playing a central role in their successes. The dynamical system modeling approach has recently emerged as a powerful framework, offering new mathematical insights into the structure and learning behavior of DNNs. In this work, we establish generalization error bounds for both discrete- and continuous-time residual networks (ResNets) by combining Rademacher complexity, flow maps of dynamical systems, and the convergence behavior of ResNets in the deep-layer limit. The resulting bounds are of order $O(1/\sqrt{S})$ with respect to the number of training samples $S$, and include a structure-dependent negative term, yielding depth-uniform and asymptotic generalization bounds under milder assumptions. These findings provide a unified understanding of generalization across both discrete- and continuous-time ResNets, helping to close the gap in both the order of sample complexity and assumptions between the discrete- and continuous-time settings.
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 10:43:41

SOM-VQ: Topology-Aware Tokenization for Interactive Generative Models
Alessandro Londei, Denise Lanzieri, Matteo Benati
arxiv.org/abs/2602.21133 arxiv.org/pdf/2602.21133 arxiv.org/html/2602.21133
arXiv:2602.21133v1 Announce Type: new
Abstract: Vector-quantized representations enable powerful discrete generative models but lack semantic structure in token space, limiting interpretable human control. We introduce SOM-VQ, a tokenization method that combines vector quantization with Self-Organizing Maps to learn discrete codebooks with explicit low-dimensional topology. Unlike standard VQ-VAE, SOM-VQ uses topology-aware updates that preserve neighborhood structure: nearby tokens on a learned grid correspond to semantically similar states, enabling direct geometric manipulation of the latent space. We demonstrate that SOM-VQ produces more learnable token sequences in the evaluated domains while providing an explicit navigable geometry in code space. Critically, the topological organization enables intuitive human-in-the-loop control: users can steer generation by manipulating distances in token space, achieving semantic alignment without frame-level constraints. We focus on human motion generation - a domain where kinematic structure, smooth temporal continuity, and interactive use cases (choreography, rehabilitation, HCI) make topology-aware control especially natural - demonstrating controlled divergence and convergence from reference sequences through simple grid-based sampling. SOM-VQ provides a general framework for interpretable discrete representations applicable to music, gesture, and other interactive generative domains.
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 10:42:31

ProxyFL: A Proxy-Guided Framework for Federated Semi-Supervised Learning
Duowen Chen, Yan Wang
arxiv.org/abs/2602.21078 arxiv.org/pdf/2602.21078 arxiv.org/html/2602.21078
arXiv:2602.21078v1 Announce Type: new
Abstract: Federated Semi-Supervised Learning (FSSL) aims to collaboratively train a global model across clients by leveraging partially-annotated local data in a privacy-preserving manner. In FSSL, data heterogeneity is a challenging issue, which exists both across clients and within clients. External heterogeneity refers to the data distribution discrepancy across different clients, while internal heterogeneity represents the mismatch between labeled and unlabeled data within clients. Most FSSL methods typically design fixed or dynamic parameter aggregation strategies to collect client knowledge on the server (external) and / or filter out low-confidence unlabeled samples to reduce mistakes in local client (internal). But, the former is hard to precisely fit the ideal global distribution via direct weights, and the latter results in fewer data participation into FL training. To this end, we propose a proxy-guided framework called ProxyFL that focuses on simultaneously mitigating external and internal heterogeneity via a unified proxy. I.e., we consider the learnable weights of classifier as proxy to simulate the category distribution both locally and globally. For external, we explicitly optimize global proxy against outliers instead of direct weights; for internal, we re-include the discarded samples into training by a positive-negative proxy pool to mitigate the impact of potentially-incorrect pseudo-labels. Insight experiments & theoretical analysis show our significant performance and convergence in FSSL.
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 16:07:37

Replaced article(s) found for cs.LG. arxiv.org/list/cs.LG/new
[1/6]:
- Towards Attributions of Input Variables in a Coalition
Xinhao Zheng, Huiqi Deng, Quanshi Zhang
arxiv.org/abs/2309.13411
- Knee or ROC
Veronica Wendt, Jacob Steiner, Byunggu Yu, Caleb Kelly, Justin Kim
arxiv.org/abs/2401.07390
- Rethinking Disentanglement under Dependent Factors of Variation
Antonio Almud\'evar, Alfonso Ortega
arxiv.org/abs/2408.07016 mastoxiv.page/@arXiv_csLG_bot/
- Minibatch Optimal Transport and Perplexity Bound Estimation in Discrete Flow Matching
Etrit Haxholli, Yeti Z. Gurbuz, Ogul Can, Eli Waxman
arxiv.org/abs/2411.00759 mastoxiv.page/@arXiv_csLG_bot/
- Predicting Subway Passenger Flows under Incident Situation with Causality
Xiannan Huang, Shuhan Qiu, Quan Yuan, Chao Yang
arxiv.org/abs/2412.06871 mastoxiv.page/@arXiv_csLG_bot/
- Characterizing LLM Inference Energy-Performance Tradeoffs across Workloads and GPU Scaling
Paul Joe Maliakel, Shashikant Ilager, Ivona Brandic
arxiv.org/abs/2501.08219 mastoxiv.page/@arXiv_csLG_bot/
- Universality of Benign Overfitting in Binary Linear Classification
Ichiro Hashimoto, Stanislav Volgushev, Piotr Zwiernik
arxiv.org/abs/2501.10538 mastoxiv.page/@arXiv_csLG_bot/
- Safe Reinforcement Learning for Real-World Engine Control
Julian Bedei, Lucas Koch, Kevin Badalian, Alexander Winkler, Patrick Schaber, Jakob Andert
arxiv.org/abs/2501.16613 mastoxiv.page/@arXiv_csLG_bot/
- A Statistical Learning Perspective on Semi-dual Adversarial Neural Optimal Transport Solvers
Roman Tarasov, Petr Mokrov, Milena Gazdieva, Evgeny Burnaev, Alexander Korotin
arxiv.org/abs/2502.01310
- Improving the Convergence of Private Shuffled Gradient Methods with Public Data
Shuli Jiang, Pranay Sharma, Zhiwei Steven Wu, Gauri Joshi
arxiv.org/abs/2502.03652 mastoxiv.page/@arXiv_csLG_bot/
- Using the Path of Least Resistance to Explain Deep Networks
Sina Salek, Joseph Enguehard
arxiv.org/abs/2502.12108 mastoxiv.page/@arXiv_csLG_bot/
- Distributional Vision-Language Alignment by Cauchy-Schwarz Divergence
Wenzhe Yin, Zehao Xiao, Pan Zhou, Shujian Yu, Jiayi Shen, Jan-Jakob Sonke, Efstratios Gavves
arxiv.org/abs/2502.17028 mastoxiv.page/@arXiv_csLG_bot/
- Armijo Line-search Can Make (Stochastic) Gradient Descent Provably Faster
Sharan Vaswani, Reza Babanezhad
arxiv.org/abs/2503.00229 mastoxiv.page/@arXiv_csLG_bot/
- Semantic Parallelism: Redefining Efficient MoE Inference via Model-Data Co-Scheduling
Yan Li, Zhenyu Zhang, Zhengang Wang, Pengfei Chen, Pengfei Zheng
arxiv.org/abs/2503.04398 mastoxiv.page/@arXiv_csLG_bot/
- A Survey on Federated Fine-tuning of Large Language Models
Wu, Tian, Li, Sun, Tam, Zhou, Liao, Xiong, Guo, Li, Xu
arxiv.org/abs/2503.12016 mastoxiv.page/@arXiv_csLG_bot/
- Towards Trustworthy GUI Agents: A Survey
Yucheng Shi, Wenhao Yu, Jingyuan Huang, Wenlin Yao, Wenhu Chen, Ninghao Liu
arxiv.org/abs/2503.23434 mastoxiv.page/@arXiv_csLG_bot/
- CONTINA: Confidence Interval for Traffic Demand Prediction with Coverage Guarantee
Chao Yang, Xiannan Huang, Shuhan Qiu, Yan Cheng
arxiv.org/abs/2504.13961 mastoxiv.page/@arXiv_csLG_bot/
- Regularity and Stability Properties of Selective SSMs with Discontinuous Gating
Nikola Zubi\'c, Davide Scaramuzza
arxiv.org/abs/2505.11602 mastoxiv.page/@arXiv_csLG_bot/
- RECON: Robust symmetry discovery via Explicit Canonical Orientation Normalization
Alonso Urbano, David W. Romero, Max Zimmer, Sebastian Pokutta
arxiv.org/abs/2505.13289 mastoxiv.page/@arXiv_csLG_bot/
- RefLoRA: Refactored Low-Rank Adaptation for Efficient Fine-Tuning of Large Models
Yilang Zhang, Bingcong Li, Georgios B. Giannakis
arxiv.org/abs/2505.18877 mastoxiv.page/@arXiv_csLG_bot/
- SuperMAN: Interpretable and Expressive Networks over Temporally Sparse Heterogeneous Data
Bechler-Speicher, Zerio, Huri, Vestergaard, Gilad-Bachrach, Jess, Bhatt, Sazonovs
arxiv.org/abs/2505.19193 mastoxiv.page/@arXiv_csLG_bot/
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 12:33:36

Crosslisted article(s) found for cs.LG. arxiv.org/list/cs.LG/new
[2/3]:
- Diffusion Modulation via Environment Mechanism Modeling for Planning
Hanping Zhang, Yuhong Guo
arxiv.org/abs/2602.20422 mastoxiv.page/@arXiv_csAI_bot/
- Heterogeneity-Aware Client Selection Methodology For Efficient Federated Learning
Nihal Balivada, Shrey Gupta, Shashank Shreedhar Bhatt, Suyash Gupta
arxiv.org/abs/2602.20450 mastoxiv.page/@arXiv_csDC_bot/
- Prior-Agnostic Incentive-Compatible Exploration
Ramya Ramalingam, Osbert Bastani, Aaron Roth
arxiv.org/abs/2602.20465 mastoxiv.page/@arXiv_csGT_bot/
- PhyGHT: Physics-Guided HyperGraph Transformer for Signal Purification at the HL-LHC
Mohammed Rakib, Luke Vaughan, Shivang Patel, Flera Rizatdinova, Alexander Khanov, Atriya Sen
arxiv.org/abs/2602.20475 mastoxiv.page/@arXiv_hepex_bot
- ActionEngine: From Reactive to Programmatic GUI Agents via State Machine Memory
Zhong, Faisal, Fran\c{c}a, Leesatapornwongsa, Szekeres, Rong, Nath
arxiv.org/abs/2602.20502 mastoxiv.page/@arXiv_csAI_bot/
- Inner Speech as Behavior Guides: Steerable Imitation of Diverse Behaviors for Human-AI coordination
Rakshit Trivedi, Kartik Sharma, David C Parkes
arxiv.org/abs/2602.20517 mastoxiv.page/@arXiv_csAI_bot/
- Stop-Think-AutoRegress: Language Modeling with Latent Diffusion Planning
Lovelace, Belardi, Zalouk, Polavaram, Kundurthy, Weinberger
arxiv.org/abs/2602.20528 mastoxiv.page/@arXiv_csCL_bot/
- Standard Transformers Achieve the Minimax Rate in Nonparametric Regression with $C^{s,\lambda}$ T...
Yanming Lai, Defeng Sun
arxiv.org/abs/2602.20555 mastoxiv.page/@arXiv_statML_bo
- Personal Information Parroting in Language Models
Nishant Subramani, Kshitish Ghate, Mona Diab
arxiv.org/abs/2602.20580 mastoxiv.page/@arXiv_csCL_bot/
- Characterizing Online and Private Learnability under Distributional Constraints via Generalized S...
Mo\"ise Blanchard, Abhishek Shetty, Alexander Rakhlin
arxiv.org/abs/2602.20585 mastoxiv.page/@arXiv_statML_bo
- Amortized Bayesian inference for actigraph time sheet data from mobile devices
Daniel Zhou, Sudipto Banerjee
arxiv.org/abs/2602.20611 mastoxiv.page/@arXiv_statML_bo
- Knowing the Unknown: Interpretable Open-World Object Detection via Concept Decomposition Model
Xueqiang Lv, Shizhou Zhang, Yinghui Xing, Di Xu, Peng Wang, Yanning Zhang
arxiv.org/abs/2602.20616 mastoxiv.page/@arXiv_csCV_bot/
- On the Convergence of Stochastic Gradient Descent with Perturbed Forward-Backward Passes
Boao Kong, Hengrui Zhang, Kun Yuan
arxiv.org/abs/2602.20646 mastoxiv.page/@arXiv_mathOC_bo
- DANCE: Doubly Adaptive Neighborhood Conformal Estimation
Feng, Reich, Beaglehole, Luo, Park, Yoo, Huang, Mao, Boz, Kim
arxiv.org/abs/2602.20652 mastoxiv.page/@arXiv_statML_bo
- Vision-Language Models for Ergonomic Assessment of Manual Lifting Tasks: Estimating Horizontal an...
Mohammad Sadra Rajabi, Aanuoluwapo Ojelade, Sunwook Kim, Maury A. Nussbaum
arxiv.org/abs/2602.20658 mastoxiv.page/@arXiv_csCV_bot/
- F10.7 Index Prediction: A Multiscale Decomposition Strategy with Wavelet Transform for Performanc...
Xuran Ma, et al.
arxiv.org/abs/2602.20712 mastoxiv.page/@arXiv_astrophIM
- Communication-Inspired Tokenization for Structured Image Representations
Davtyan, Sahin, Haghighi, Stapf, Acuaviva, Alahi, Favaro
arxiv.org/abs/2602.20731 mastoxiv.page/@arXiv_csCV_bot/
- SibylSense: Adaptive Rubric Learning via Memory Tuning and Adversarial Probing
Yifei Xu, et al.
arxiv.org/abs/2602.20751 mastoxiv.page/@arXiv_csCL_bot/
- Assessing the Impact of Speaker Identity in Speech Spoofing Detection
Anh-Tuan Dao, Driss Matrouf, Nicholas Evans
arxiv.org/abs/2602.20805 mastoxiv.page/@arXiv_csSD_bot/
- Don't Ignore the Tail: Decoupling top-K Probabilities for Efficient Language Model Distillation
Sayantan Dasgupta, Trevor Cohn, Timothy Baldwin
arxiv.org/abs/2602.20816 mastoxiv.page/@arXiv_csCL_bot/
- DRESS: A Continuous Framework for Structural Graph Refinement
Eduar Castrillo Velilla
arxiv.org/abs/2602.20833 mastoxiv.page/@arXiv_csDS_bot/
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2026-02-25 16:08:18

Replaced article(s) found for cs.LG. arxiv.org/list/cs.LG/new
[5/6]:
- Watermarking Degrades Alignment in Language Models: Analysis and Mitigation
Apurv Verma, NhatHai Phan, Shubhendu Trivedi
arxiv.org/abs/2506.04462 mastoxiv.page/@arXiv_csCL_bot/
- Sensory-Motor Control with Large Language Models via Iterative Policy Refinement
J\^onata Tyska Carvalho, Stefano Nolfi
arxiv.org/abs/2506.04867 mastoxiv.page/@arXiv_csAI_bot/
- ICE-ID: A Novel Historical Census Dataset for Longitudinal Identity Resolution
de Carvalho, Popov, Kaatee, Correia, Th\'orisson, Li, Bj\"ornsson, Sigur{\dh}arson, Dibangoye
arxiv.org/abs/2506.13792 mastoxiv.page/@arXiv_csAI_bot/
- Feedback-driven recurrent quantum neural network universality
Lukas Gonon, Rodrigo Mart\'inez-Pe\~na, Juan-Pablo Ortega
arxiv.org/abs/2506.16332 mastoxiv.page/@arXiv_quantph_b
- Programming by Backprop: An Instruction is Worth 100 Examples When Finetuning LLMs
Cook, Sapora, Ahmadian, Khan, Rocktaschel, Foerster, Ruis
arxiv.org/abs/2506.18777 mastoxiv.page/@arXiv_csAI_bot/
- Stochastic Quantum Spiking Neural Networks with Quantum Memory and Local Learning
Jiechen Chen, Bipin Rajendran, Osvaldo Simeone
arxiv.org/abs/2506.21324 mastoxiv.page/@arXiv_csNE_bot/
- Enjoying Non-linearity in Multinomial Logistic Bandits: A Minimax-Optimal Algorithm
Pierre Boudart (SIERRA), Pierre Gaillard (Thoth), Alessandro Rudi (PSL, DI-ENS, Inria)
arxiv.org/abs/2507.05306 mastoxiv.page/@arXiv_statML_bo
- Characterizing State Space Model and Hybrid Language Model Performance with Long Context
Saptarshi Mitra, Rachid Karami, Haocheng Xu, Sitao Huang, Hyoukjun Kwon
arxiv.org/abs/2507.12442 mastoxiv.page/@arXiv_csAR_bot/
- Is Exchangeability better than I.I.D to handle Data Distribution Shifts while Pooling Data for Da...
Ayush Roy, Samin Enam, Jun Xia, Won Hwa Kim, Vishnu Suresh Lokhande
arxiv.org/abs/2507.19575 mastoxiv.page/@arXiv_csCV_bot/
- TASER: Table Agents for Schema-guided Extraction and Recommendation
Nicole Cho, Kirsty Fielding, William Watson, Sumitra Ganesh, Manuela Veloso
arxiv.org/abs/2508.13404 mastoxiv.page/@arXiv_csAI_bot/
- Morphology-Aware Peptide Discovery via Masked Conditional Generative Modeling
Nuno Costa, Julija Zavadlav
arxiv.org/abs/2509.02060 mastoxiv.page/@arXiv_qbioBM_bo
- PCPO: Proportionate Credit Policy Optimization for Aligning Image Generation Models
Jeongjae Lee, Jong Chul Ye
arxiv.org/abs/2509.25774 mastoxiv.page/@arXiv_csCV_bot/
- Multi-hop Deep Joint Source-Channel Coding with Deep Hash Distillation for Semantically Aligned I...
Didrik Bergstr\"om, Deniz G\"und\"uz, Onur G\"unl\"u
arxiv.org/abs/2510.06868 mastoxiv.page/@arXiv_csIT_bot/
- MoMaGen: Generating Demonstrations under Soft and Hard Constraints for Multi-Step Bimanual Mobile...
Chengshu Li, et al.
arxiv.org/abs/2510.18316 mastoxiv.page/@arXiv_csRO_bot/
- A Spectral Framework for Graph Neural Operators: Convergence Guarantees and Tradeoffs
Roxanne Holden, Luana Ruiz
arxiv.org/abs/2510.20954 mastoxiv.page/@arXiv_statML_bo
- Breaking Agent Backbones: Evaluating the Security of Backbone LLMs in AI Agents
Bazinska, Mathys, Casucci, Rojas-Carulla, Davies, Souly, Pfister
arxiv.org/abs/2510.22620 mastoxiv.page/@arXiv_csCR_bot/
- Uncertainty Calibration of Multi-Label Bird Sound Classifiers
Raphael Schwinger, Ben McEwen, Vincent S. Kather, Ren\'e Heinrich, Lukas Rauch, Sven Tomforde
arxiv.org/abs/2511.08261 mastoxiv.page/@arXiv_csSD_bot/
- Two-dimensional RMSD projections for reaction path visualization and validation
Rohit Goswami (Institute IMX and Lab-COSMO, \'Ecole polytechnique f\'ed\'erale de Lausanne)
arxiv.org/abs/2512.07329 mastoxiv.page/@arXiv_physicsch
- Distribution-informed Online Conformal Prediction
Dongjian Hu, Junxi Wu, Shu-Tao Xia, Changliang Zou
arxiv.org/abs/2512.07770 mastoxiv.page/@arXiv_statML_bo
- Coupling Experts and Routers in Mixture-of-Experts via an Auxiliary Loss
Ang Lv, Jin Ma, Yiyuan Ma, Siyuan Qiao
arxiv.org/abs/2512.23447 mastoxiv.page/@arXiv_csCL_bot/
toXiv_bot_toot

@arXiv_csLG_bot@mastoxiv.page
2025-12-22 10:34:40

Weighted Stochastic Differential Equation to Implement Wasserstein-Fisher-Rao Gradient Flow
Herlock Rahimi
arxiv.org/abs/2512.17878 arxiv.org/pdf/2512.17878 arxiv.org/html/2512.17878
arXiv:2512.17878v1 Announce Type: new
Abstract: Score-based diffusion models currently constitute the state of the art in continuous generative modeling. These methods are typically formulated via overdamped or underdamped Ornstein--Uhlenbeck-type stochastic differential equations, in which sampling is driven by a combination of deterministic drift and Brownian diffusion, resulting in continuous particle trajectories in the ambient space. While such dynamics enjoy exponential convergence guarantees for strongly log-concave target distributions, it is well known that their mixing rates deteriorate exponentially in the presence of nonconvex or multimodal landscapes, such as double-well potentials. Since many practical generative modeling tasks involve highly non-log-concave target distributions, considerable recent effort has been devoted to developing sampling schemes that improve exploration beyond classical diffusion dynamics.
A promising line of work leverages tools from information geometry to augment diffusion-based samplers with controlled mass reweighting mechanisms. This perspective leads naturally to Wasserstein--Fisher--Rao (WFR) geometries, which couple transport in the sample space with vertical (reaction) dynamics on the space of probability measures. In this work, we formulate such reweighting mechanisms through the introduction of explicit correction terms and show how they can be implemented via weighted stochastic differential equations using the Feynman--Kac representation. Our study provides a preliminary but rigorous investigation of WFR-based sampling dynamics, and aims to clarify their geometric and operator-theoretic structure as a foundation for future theoretical and algorithmic developments.
toXiv_bot_toot