Tootfinder

Opt-in global Mastodon full text search. Join the index!

@arXiv_quantph_bot@mastoxiv.page
2025-10-02 10:06:01

Quantum Probabilistic Label Refining: Enhancing Label Quality for Robust Image Classification
Fang Qi, Lu Peng, Zhengming Ding
arxiv.org/abs/2510.00528

@arXiv_mathDS_bot@mastoxiv.page
2025-09-01 10:16:21

Crosslisted article(s) found for math.DS. arxiv.org/list/math.DS/new
[1/1]:
- Manifold Trajectories in Next-Token Prediction: From Replicator Dynamics to Softmax Equilibrium
Christopher R. Lee-Jenkins

@arXiv_csCL_bot@mastoxiv.page
2025-09-30 14:12:02

An empirical study on the limitation of Transformers in program trace generation
Simeng Sun
arxiv.org/abs/2509.25073 arxiv.org/pdf/2509.250…

@arXiv_statML_bot@mastoxiv.page
2025-10-15 10:21:31

Dendrograms of Mixing Measures for Softmax-Gated Gaussian Mixture of Experts: Consistency without Model Sweeps
Do Tien Hai, Trung Nguyen Mai, TrungTin Nguyen, Nhat Ho, Binh T. Nguyen, Christopher Drovandi
arxiv.org/abs/2510.12744

@arXiv_csLG_bot@mastoxiv.page
2025-09-15 09:56:11

Multipole Semantic Attention: A Fast Approximation of Softmax Attention for Pretraining
Rupert Mitchell, Kristian Kersting
arxiv.org/abs/2509.10406

@arXiv_csCR_bot@mastoxiv.page
2025-08-28 09:43:21

Breaking the Layer Barrier: Remodeling Private Transformer Inference with Hybrid CKKS and MPC
Tianshi Xu, Wen-jie Lu, Jiangrui Yu, Chen Yi, Chenqi Lin, Runsheng Wang, Meng Li
arxiv.org/abs/2508.19525

@arXiv_csCV_bot@mastoxiv.page
2025-09-25 10:16:02

ExpFace: Exponential Angular Margin Loss for Deep Face Recognition
Jinhui Zheng, Xueyuan Gong
arxiv.org/abs/2509.19753 arxiv.org/pdf/2509.1…

@arXiv_eessSY_bot@mastoxiv.page
2025-10-03 10:03:41

Computing Control Lyapunov-Barrier Functions: Softmax Relaxation and Smooth Patching with Formal Guarantees
Jun Liu, Maxwell Fitzsimmons
arxiv.org/abs/2510.02223

@arXiv_csLG_bot@mastoxiv.page
2025-09-10 10:44:41

Customizing the Inductive Biases of Softmax Attention using Structured Matrices
Yilun Kuang, Noah Amsel, Sanae Lotfi, Shikai Qiu, Andres Potapczynski, Andrew Gordon Wilson
arxiv.org/abs/2509.07963

@arXiv_csCL_bot@mastoxiv.page
2025-09-22 10:23:01

Localmax dynamics for attention in transformers and its asymptotic behavior
Henri Cimeti\`ere, Maria Teresa Chiri, Bahman Gharesifard
arxiv.org/abs/2509.15958

@arXiv_csLG_bot@mastoxiv.page
2025-10-13 10:42:30

Task-Level Insights from Eigenvalues across Sequence Models
Rahel Rickenbach, Jelena Trisovic, Alexandre Didier, Jerome Sieber, Melanie N. Zeilinger
arxiv.org/abs/2510.09379

@arXiv_csCL_bot@mastoxiv.page
2025-08-21 09:57:50

Improving in-context learning with a better scoring function
Omar Naim, Swarnadeep Bhar, J\'er\^ome Bolte, Nicholas Asher
arxiv.org/abs/2508.14685

@arXiv_csCR_bot@mastoxiv.page
2025-08-14 07:39:02

Can AI Keep a Secret? Contextual Integrity Verification: A Provable Security Architecture for LLMs
Aayush Gupta
arxiv.org/abs/2508.09288 ar…

@arXiv_econTH_bot@mastoxiv.page
2025-10-07 13:35:57

Crosslisted article(s) found for econ.TH. arxiv.org/list/econ.TH/new
[1/1]:
- Beyond Softmax: A New Perspective on Gradient Bandits
Emerson Melo, David M\"uller

@arXiv_eessAS_bot@mastoxiv.page
2025-10-08 09:42:39

TokenChain: A Discrete Speech Chain via Semantic Token Modeling
Mingxuan Wang, Satoshi Nakamura
arxiv.org/abs/2510.06201 arxiv.org/pdf/2510…

@arXiv_csCL_bot@mastoxiv.page
2025-10-15 10:27:41

Credal Transformer: A Principled Approach for Quantifying and Mitigating Hallucinations in Large Language Models
Shihao Ji, Zihui Song, Jiajie Huang
arxiv.org/abs/2510.12137

@arXiv_csCL_bot@mastoxiv.page
2025-10-14 13:16:18

Deconstructing Attention: Investigating Design Principles for Effective Language Modeling
Huiyin Xue, Nafise Sadat Moosavi, Nikolaos Aletras
arxiv.org/abs/2510.11602

@arXiv_eessAS_bot@mastoxiv.page
2025-09-04 08:49:21

Speech Intelligibility Assessment with Uncertainty-Aware Whisper Embeddings and sLSTM
Ryandhimas E. Zezario, Dyah A. M. G. Wisnu, Hsin-Min Wang, Yu Tsao
arxiv.org/abs/2509.03013