Tootfinder

Opt-in global Mastodon full text search. Join the index!

@arXiv_csCL_bot@mastoxiv.page
2025-08-19 11:44:50

Word Meanings in Transformer Language Models
Jumbly Grindrod, Peter Grindrod
arxiv.org/abs/2508.12863 arxiv.org/pdf/2508.12863

@arXiv_csLG_bot@mastoxiv.page
2025-09-19 10:26:51

Attention Beyond Neighborhoods: Reviving Transformer for Graph Clustering
Xuanting Xie, Bingheng Li, Erlin Pan, Rui Hou, Wenyu Chen, Zhao Kang
arxiv.org/abs/2509.15024

@arXiv_csCV_bot@mastoxiv.page
2025-08-18 09:54:00

Handwritten Text Recognition of Historical Manuscripts Using Transformer-Based Models
Erez Meoded
arxiv.org/abs/2508.11499 arxiv.org/pdf/25…

@arXiv_csRO_bot@mastoxiv.page
2025-08-18 09:24:30

OVSegDT: Segmenting Transformer for Open-Vocabulary Object Goal Navigation
Tatiana Zemskova, Aleksei Staroverov, Dmitry Yudin, Aleksandr Panov
arxiv.org/abs/2508.11479

@arXiv_csSE_bot@mastoxiv.page
2025-07-18 09:20:02

ROSE: Transformer-Based Refactoring Recommendation for Architectural Smells
Samal Nursapa, Anastassiya Samuilova, Alessio Bucaioni. Phuong T. Nguyen
arxiv.org/abs/2507.12561

@peterhoneyman@a2mi.social
2025-08-18 20:00:51

i am determined to read the attention/transformer paper
i even printed it out

Attention Is All You Need
Ashish Vaswani
Noam Shazeer
Niki Parmar
Jakob Uszkoreit
Llion Jones
Aidan N. Gomez
Łukasz Kaiser
Illia Polosukhin

The dominant sequence transduction models are based on complex recurrent or convolutional neural networks that include an encoder and a decoder. The best performing models also connect the encoder and decoder through an attention mechanism. We propose a new simple network architecture, the Transformer, based solely on attention mechanisms, dispensing with …
@arXiv_eessIV_bot@mastoxiv.page
2025-08-19 08:22:30

FractMorph: A Fractional Fourier-Based Multi-Domain Transformer for Deformable Image Registration
Shayan Kebriti, Shahabedin Nabavi, Ali Gooya
arxiv.org/abs/2508.12445

@arXiv_csGR_bot@mastoxiv.page
2025-07-18 08:44:42

HairFormer: Transformer-Based Dynamic Neural Hair Simulation
Joy Xiaoji Zhang, Jingsen Zhu, Hanyu Chen, Steve Marschner
arxiv.org/abs/2507.12600

@Techmeme@techhub.social
2025-08-18 23:35:45

Nvidia debuts the Nemotron-Nano-9B-v2, a hybrid Mamba-transformer model, saying it achieves scores comparable to or better than Qwen3-8B on reasoning benchmarks (Carl Franzen/VentureBeat)
venturebeat.com/ai/nvidia-rele

@arXiv_eessAS_bot@mastoxiv.page
2025-09-19 09:40:01

MELA-TTS: Joint transformer-diffusion model with representation alignment for speech synthesis
Keyu An, Zhiyu Zhang, Changfeng Gao, Yabin Li, Zhendong Peng, Haoxu Wang, Zhihao Du, Han Zhao, Zhifu Gao, Xiangang Li
arxiv.org/abs/2509.14784

@arXiv_csCL_bot@mastoxiv.page
2025-08-18 09:40:50

Rationalizing Transformer Predictions via End-To-End Differentiable Self-Training
Marc Brinner, Sina Zarrie{\ss}
arxiv.org/abs/2508.11393 a…

@arXiv_physicsplasmph_bot@mastoxiv.page
2025-07-18 08:19:52

Early Prediction of Current Quench Events in the ADITYA Tokamak using Transformer based Data Driven Models
Jyoti Agarwal, Bhaskar Chaudhury, Jaykumar Navadiya, Shrichand Jakhar, Manika Sharma
arxiv.org/abs/2507.12797

@arXiv_csSD_bot@mastoxiv.page
2025-07-18 09:35:52

Multi-Class-Token Transformer for Multitask Self-supervised Music Information Retrieval
Yuexuan Kong, Vincent Lostanlen, Romain Hennequin, Mathieu Lagrange, Gabriel Meseguer-Brocal
arxiv.org/abs/2507.12996

@arXiv_eessSY_bot@mastoxiv.page
2025-09-18 09:51:51

Large Language Model-Empowered Decision Transformer for UAV-Enabled Data Collection
Zhixion Chen, Jiangzhou Wang, and Hyundong Shin, Arumugam Nallanathan
arxiv.org/abs/2509.13934

@arXiv_csRO_bot@mastoxiv.page
2025-09-18 10:03:51

UltraHiT: A Hierarchical Transformer Architecture for Generalizable Internal Carotid Artery Robotic Ultrasonography
Teng Wang, Haojun Jiang, Yuxuan Wang, Zhenguo Sun, Xiangjie Yan, Xiang Li, Gao Huang
arxiv.org/abs/2509.13832

@arXiv_qbioQM_bot@mastoxiv.page
2025-08-18 08:24:30

Brain Tumor Segmentation in Sub-Sahara Africa with Advanced Transformer and ConvNet Methods: Fine-Tuning, Data Mixing and Ensembling
Toufiq Musah, Chantelle Amoako-Atta, John Amankwaah Otu, Lukman E. Ismaila, Swallah Alhaji Suraka, Oladimeji Williams, Isaac Tigbee, Kato Hussein Wabbi, Samantha Katsande, Kanyiri Ahmed Yakubu, Adedayo Kehinde Lawal, Anita Nsiah Donkor, Naeem Mwinlanaah Adamu, Adebowale Akande, John Othieno, Prince Ebenezer Adjei, Zhang Dong, Confidence Raymond, Udunna C.…

@pbloem@sigmoid.social
2025-08-18 11:58:02

The HRM paper has been mostly debunked by the ARC-AGI people.
arcprize.org/blog/hrm-analysis
The results are legit but most of them are not down to the architecture (swapping it out for a transformer doesn't change that much).
Also, the model is purely transductive. It onl…

@arXiv_astrophIM_bot@mastoxiv.page
2025-07-17 08:48:00

Image-Based Multi-Survey Classification of Light Curves with a Pre-Trained Vision Transformer
Daniel Moreno-Cartagena, Guillermo Cabrera-Vives, Alejandra M. Mu\~noz Arancibia, Pavlos Protopapas, Francisco F\"orster, M\'arcio Catelan, A. Bayo, Pablo A. Est\'evez, P. S\'anchez-S\'aez, Franz E. Bauer, M. Pavez-Herrera, L. Hern\'andez-Garc\'ia, Gonzalo Rojas

@arXiv_csCV_bot@mastoxiv.page
2025-07-18 10:21:22

Taming Diffusion Transformer for Real-Time Mobile Video Generation
Yushu Wu, Yanyu Li, Anil Kag, Ivan Skorokhodov, Willi Menapace, Ke Ma, Arpit Sahni, Ju Hu, Aliaksandr Siarohin, Dhritiman Sagar, Yanzhi Wang, Sergey Tulyakov
arxiv.org/abs/2507.13343

@arXiv_csAR_bot@mastoxiv.page
2025-07-17 08:18:40

Characterizing State Space Model (SSM) and SSM-Transformer Hybrid Language Model Performance with Long Context Length
Saptarshi Mitra, Rachid Karami, Haocheng Xu, Sitao Huang, Hyoukjun Kwon
arxiv.org/abs/2507.12442

@arXiv_csAI_bot@mastoxiv.page
2025-09-18 08:13:11

Semantic Fusion with Fuzzy-Membership Features for Controllable Language Modelling
Yongchao Huang, Hassan Raza
arxiv.org/abs/2509.13357 arx…

@arXiv_csGR_bot@mastoxiv.page
2025-08-19 08:22:39

MixCache: Mixture-of-Cache for Video Diffusion Transformer Acceleration
Yuanxin Wei, Lansong Diao, Bujiao Chen, Shenggan Cheng, Zhengping Qian, Wenyuan Yu, Nong Xiao, Wei Lin, Jiangsu Du
arxiv.org/abs/2508.12691

@arXiv_statML_bot@mastoxiv.page
2025-09-19 08:40:31

Asymptotic Study of In-context Learning with Random Transformers through Equivalent Models
Samet Demir, Zafer Dogan
arxiv.org/abs/2509.15152

@arXiv_csCR_bot@mastoxiv.page
2025-07-18 12:17:34

Replaced article(s) found for cs.CR. arxiv.org/list/cs.CR/new
[1/1]:
- TBDetector:Transformer-Based Detector for Advanced Persistent Threats with Provenance Graph
Wang, Wen, Zhang, Zhao, Ma, Luo, Xu, Nie, Wu, Liu

@arXiv_eessAS_bot@mastoxiv.page
2025-09-18 09:39:11

SV-Mixer: Replacing the Transformer Encoder with Lightweight MLPs for Self-Supervised Model Compression in Speaker Verification
Jungwoo Heo, Hyun-seo Shin, Chan-yeong Lim, Kyo-won Koo, Seung-bin Kim, Jisoo Son, Ha-Jin Yu
arxiv.org/abs/2509.14136

@arXiv_eessSP_bot@mastoxiv.page
2025-09-17 08:56:49

NEFT: A Unified Transformer Framework for Efficient Near-Field CSI Feedback in XL-MIMO Systems
Haiyang Li, Tianqi Mao, Pengyu Wang, Ruiqi Liu, Shunyu Li, Zhaocheng Wang
arxiv.org/abs/2509.12748

@arXiv_csCV_bot@mastoxiv.page
2025-07-18 10:20:32

Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy
Yiting Yang, Hao Luo, Yuan Sun, Qingsen Yan, Haokui Zhang, Wei Dong, Guoqing Wang, Peng Wang, Yang Yang, Hengtao Shen
arxiv.org/abs/2507.13260

@arXiv_grqc_bot@mastoxiv.page
2025-09-16 10:44:37

Transformer Networks for Continuous Gravitational-wave Searches
Prasanna. M. Joshi, Reinhard Prix
arxiv.org/abs/2509.10912 arxiv.org/pdf/25…

@arXiv_astrophHE_bot@mastoxiv.page
2025-07-17 09:10:10

Enhancements to the IceCube Extremely High Energy Neutrino Selection using Graph & Transformer Based Neural Networks
Maxwell Nakos (for the IceCube Collaboration), Aske Rosted (for the IceCube Collaboration), Lu Lu (for the IceCube Collaboration)
arxiv.org/abs/2507.11774

@arXiv_csRO_bot@mastoxiv.page
2025-08-18 09:30:00

MultiPark: Multimodal Parking Transformer with Next-Segment Prediction
Han Zheng, Zikang Zhou, Guli Zhang, Zhepei Wang, Kaixuan Wang, Peiliang Li, Shaojie Shen, Ming Yang, Tong Qin
arxiv.org/abs/2508.11537

@arXiv_csDL_bot@mastoxiv.page
2025-08-18 10:02:57

Crosslisted article(s) found for cs.DL. arxiv.org/list/cs.DL/new
[1/1]:
- Handwritten Text Recognition of Historical Manuscripts Using Transformer-Based Models
Erez Meoded

@arXiv_csAI_bot@mastoxiv.page
2025-08-18 10:56:31

Crosslisted article(s) found for cs.AI. arxiv.org/list/cs.AI/new
[5/5]:
- Handwritten Text Recognition of Historical Manuscripts Using Transformer-Based Models
Erez Meoded

@arXiv_csHC_bot@mastoxiv.page
2025-08-15 07:47:02

Pre-trained Transformer-models using chronic invasive electrophysiology for symptom decoding without patient-individual training
Timon Merk, Saeed Salehi, Richard M. Koehler, Qiming Cui, Maria Olaru, Amelia Hahn, Nicole R. Provenza, Simon Little, Reza Abbasi-Asl, Phil A. Starr, Wolf-Julian Neumann
arxiv.org/abs/2508.10160

@arXiv_csRO_bot@mastoxiv.page
2025-09-18 10:04:51

Pre-Manipulation Alignment Prediction with Parallel Deep State-Space and Transformer Models
Motonari Kambara, Komei Sugiura
arxiv.org/abs/2509.13839

@arXiv_csPL_bot@mastoxiv.page
2025-07-17 08:17:00

Universal Synthesis of Differentiably Tunable Numerical Abstract Transformers
Shaurya Gomber, Debangshu Banerjee, Gagandeep Singh
arxiv.org/abs/2507.11827

@arXiv_qfinMF_bot@mastoxiv.page
2025-08-19 14:44:06

Replaced article(s) found for q-fin.MF. arxiv.org/list/q-fin.MF/new
[1/1]:
- Quantformer: from attention to profit with a quantitative transformer trading strategy
Zhaofeng Zhang, Banghao Chen, Shengxin Zhu, Nicolas Langren\'e

@arXiv_csCR_bot@mastoxiv.page
2025-08-19 11:21:20

A Robust Cross-Domain IDS using BiGRU-LSTM-Attention for Medical and Industrial IoT Security
Afrah Gueriani, Hamza Kheddar, Ahmed Cherif Mazari, Mohamed Chahine Ghanem
arxiv.org/abs/2508.12470

@arXiv_eessIV_bot@mastoxiv.page
2025-08-18 08:31:50

LKFMixer: Exploring Large Kernel Feature For Efficient Image Super-Resolution
Yinggan Tang, Quanwei Hu
arxiv.org/abs/2508.11391 arxiv.org/p…

@arXiv_csIT_bot@mastoxiv.page
2025-08-15 09:13:22

Predictive Position Control for Movable Antenna Arrays in UAV Communications: A Spatio-Temporal Transformer-LSTM Framework
Kan Yu, Kaixuan Li, Xiaowu Liu, Qixun Zhang, Zhiyong Feng
arxiv.org/abs/2508.10720

@arXiv_csIR_bot@mastoxiv.page
2025-08-11 09:55:39

eSASRec: Enhancing Transformer-based Recommendations in a Modular Fashion
Daria Tikhonovich, Nikita Zelinskiy, Aleksandr V. Petrov, Mayya Spirina, Andrei Semenov, Andrey V. Savchenko, Sergei Kuliev
arxiv.org/abs/2508.06450

@arXiv_qbioNC_bot@mastoxiv.page
2025-08-14 08:06:52

Perceptual Reality Transformer: Neural Architectures for Simulating Neurological Perception Conditions
Baihan Lin
arxiv.org/abs/2508.09852

@arXiv_csLG_bot@mastoxiv.page
2025-09-16 12:46:37

Dynamic Relational Priming Improves Transformer in Multivariate Time Series
Hunjae Lee, Corey Clark
arxiv.org/abs/2509.12196 arxiv.org/pdf/…

@arXiv_csCL_bot@mastoxiv.page
2025-09-19 10:33:21

Patent Language Model Pretraining with ModernBERT
Amirhossein Yousefiramandi, Ciaran Cooney
arxiv.org/abs/2509.14926 arxiv.org/pdf/2509.149…

@arXiv_csCV_bot@mastoxiv.page
2025-07-16 10:38:21

Streaming 4D Visual Geometry Transformer
Dong Zhuo, Wenzhao Zheng, Jiahe Guo, Yuqi Wu, Jie Zhou, Jiwen Lu
arxiv.org/abs/2507.11539

@arXiv_csAR_bot@mastoxiv.page
2025-07-18 08:08:42

An ultra-low-power CGRA for accelerating Transformers at the edge
Rohit Prasad
arxiv.org/abs/2507.12904 arxiv.org/pdf…

@arXiv_csAI_bot@mastoxiv.page
2025-09-18 09:08:41

From Next Token Prediction to (STRIPS) World Models -- Preliminary Results
Carlos N\'u\~nez-Molina, Vicen\c{c} G\'omez, Hector Geffner
arxiv.org/abs/2509.13389

@arXiv_csRO_bot@mastoxiv.page
2025-09-16 12:05:07

Tenma: Robust Cross-Embodiment Robot Manipulation with Diffusion Transformer
Travis Davies, Yiqi Huang, Yunxin Liu, Xiang Chen, Huxian Liu, Luhui Hu
arxiv.org/abs/2509.11865

@arXiv_eessAS_bot@mastoxiv.page
2025-08-19 08:45:09

Arabic ASR on the SADA Large-Scale Arabic Speech Corpus with Transformer-Based Models
Branislav Gerazov, Marcello Politi, S\'ebastien Brati\`eres
arxiv.org/abs/2508.12968

@arXiv_astrophIM_bot@mastoxiv.page
2025-07-18 08:04:22

Astro-MoE: Mixture of Experts for Multiband Astronomical Time Series
Martina C\'adiz-Leyton, Guillermo Cabrera-Vives, Pavlos Protopapas, Daniel Moreno-Cartagena, Ignacio Becker
arxiv.org/abs/2507.12611

@arXiv_csCL_bot@mastoxiv.page
2025-07-16 10:31:11

Addressing Data Imbalance in Transformer-Based Multi-Label Emotion Detection with Weighted Loss
Xia Cui
arxiv.org/abs/2507.11384

@arXiv_csRO_bot@mastoxiv.page
2025-09-17 10:43:40

An Uncertainty-Weighted Decision Transformer for Navigation in Dense, Complex Driving Scenarios
Zhihao Zhang, Chengyang Peng, Minghao Zhu, Ekim Yurtsever, Keith A. Redmill
arxiv.org/abs/2509.13132

@arXiv_eessIV_bot@mastoxiv.page
2025-07-18 08:56:52

Unleashing Vision Foundation Models for Coronary Artery Segmentation: Parallel ViT-CNN Encoding and Variational Fusion
Caixia Dong, Duwei Dai, Xinyi Han, Fan Liu, Xu Yang, Zongfang Li, Songhua Xu
arxiv.org/abs/2507.12938

@arXiv_csSD_bot@mastoxiv.page
2025-09-12 08:30:29

Efficient Transformer-Based Piano Transcription With Sparse Attention Mechanisms
Weixing Wei, Kazuyoshi Yoshii
arxiv.org/abs/2509.09318 arx…

@arXiv_csCV_bot@mastoxiv.page
2025-08-19 12:05:40

Compact Attention: Exploiting Structured Spatio-Temporal Sparsity for Fast Video Generation
Qirui Li, Guangcong Zheng, Qi Zhao, Jie Li, Bin Dong, Yiwu Yao, Xi Li
arxiv.org/abs/2508.12969

@arXiv_csCL_bot@mastoxiv.page
2025-09-19 10:30:31

Llama-Mimi: Speech Language Models with Interleaved Semantic and Acoustic Tokens
Issa Sugiura, Shuhei Kurita, Yusuke Oda, Ryuichiro Higashinaka
arxiv.org/abs/2509.14882

@arXiv_eessSP_bot@mastoxiv.page
2025-09-11 08:39:53

RTR: A Transformer-Based Lossless Crossover with Perfect Phase Alignment
Xiangying Li, Jiankuan Li, Yong Tang
arxiv.org/abs/2509.08272 arxi…

@arXiv_csIR_bot@mastoxiv.page
2025-07-17 08:06:40

Sparse Autoencoders for Sequential Recommendation Models: Interpretation and Flexible Control
Anton Klenitskiy, Konstantin Polev, Daria Denisova, Alexey Vasilev, Dmitry Simakov, Gleb Gusev
arxiv.org/abs/2507.12202

@arXiv_csLG_bot@mastoxiv.page
2025-08-15 10:10:12

Self-Supervised Temporal Super-Resolution of Energy Data using Generative Adversarial Transformer
Xuanhao Mu, G\"okhan Demirel, Yuzhe Zhang, Jianlei Liu, Thorsten Schlachter, Veit Hagenmeyer
arxiv.org/abs/2508.10587

@arXiv_csCR_bot@mastoxiv.page
2025-08-15 09:41:42

A Transformer-Based Approach for DDoS Attack Detection in IoT Networks
Sandipan Dey, Payal Santosh Kate, Vatsala Upadhyay, Abhishek Vaish
arxiv.org/abs/2508.10636

@arXiv_astrophIM_bot@mastoxiv.page
2025-07-18 08:17:42

A Semi-Supervised Learning Method for the Identification of Bad Exposures in Large Imaging Surveys
Yufeng Luo, Adam D. Myers, Alex Drlica-Wagner, Dario Dematties, Salma Borchani, Frank Valdes, Arjun Dey, David Schlegel, Rongpu Zhou, DESI Legacy Imaging Surveys Team
arxiv.org/abs/2507.12784

@arXiv_csCV_bot@mastoxiv.page
2025-07-17 10:29:00

CytoSAE: Interpretable Cell Embeddings for Hematology
Muhammed Furkan Dasdelen, Hyesu Lim, Michele Buck, Katharina S. G\"otze, Carsten Marr, Steffen Schneider
arxiv.org/abs/2507.12464

@arXiv_csCL_bot@mastoxiv.page
2025-07-18 09:57:02

The Imitation Game: Turing Machine Imitator is Length Generalizable Reasoner
Zhouqi Hua, Wenwei Zhang, Chengqi Lyu, Yuzhe Gu, Songyang Gao, Kuikun Liu, Kai Chen
arxiv.org/abs/2507.13332

@arXiv_csRO_bot@mastoxiv.page
2025-09-17 10:31:10

GRATE: a Graph transformer-based deep Reinforcement learning Approach for Time-efficient autonomous robot Exploration
Haozhan Ni, Jingsong Liang, Chenyu He, Yuhong Cao, Guillaume Sartoretti
arxiv.org/abs/2509.12863

@arXiv_csCV_bot@mastoxiv.page
2025-08-19 12:08:00

Real-Time Beach Litter Detection and Counting: A Comparative Analysis of RT-DETR Model Variants
Miftahul Huda, Arsyiah Azahra, Putri Maulida Chairani, Dimas Rizky Ramadhani, Nabila Azhari, Ade Lailani
arxiv.org/abs/2508.13101

@arXiv_csCL_bot@mastoxiv.page
2025-09-18 10:16:51

Canary-1B-v2 & Parakeet-TDT-0.6B-v3: Efficient and High-Performance Models for Multilingual ASR and AST
Monica Sekoyan, Nithin Rao Koluguri, Nune Tadevosyan, Piotr Zelasko, Travis Bartley, Nick Karpov, Jagadeesh Balam, Boris Ginsburg
arxiv.org/abs/2509.14128

@arXiv_csAR_bot@mastoxiv.page
2025-07-15 09:17:41

Pimba: A Processing-in-Memory Acceleration for Post-Transformer Large Language Model Serving
Wonung Kim, Yubin Lee, Yoonsung Kim, Jinwoo Hwang, Seongryong Oh, Jiyong Jung, Aziz Huseynov, Woong Gyu Park, Chang Hyun Park, Divya Mahajan, Jongse Park
arxiv.org/abs/2507.10178

@arXiv_csCR_bot@mastoxiv.page
2025-09-15 07:37:11

Differential Robustness in Transformer Language Models: Empirical Evaluation Under Adversarial Text Attacks
Taniya Gidatkar, Oluwaseun Ajao, Matthew Shardlow
arxiv.org/abs/2509.09706

@arXiv_csCV_bot@mastoxiv.page
2025-07-17 10:26:40

DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition
Hayat Ullah, Muhammad Ali Shafique, Abbas Khan, Arslan Munir
arxiv.org/abs/2507.12426

@arXiv_csIR_bot@mastoxiv.page
2025-09-15 07:38:41

Faster and Memory-Efficient Training of Sequential Recommendation Models for Large Catalogs
Maxim Zhelnin, Dmitry Redko, Volkov Daniil, Anna Volodkevich, Petr Sokerin, Valeriy Shevchenko, Egor Shvetsov, Alexey Vasilev, Darya Denisova, Ruslan Izmailov, Alexey Zaytsev
arxiv.org/abs/2509.09682

@arXiv_csLG_bot@mastoxiv.page
2025-09-12 09:53:19

Adaptive Pareto-Optimal Token Merging for Edge Transformer Models in Semantic Communication
Omar Erak, Omar Alhussein, Hatem Abou-Zeid, Mehdi Bennis
arxiv.org/abs/2509.09168

@arXiv_csRO_bot@mastoxiv.page
2025-07-16 09:47:01

Whom to Respond To? A Transformer-Based Model for Multi-Party Social Robot Interaction
He Zhu, Ryo Miyoshi, Yuki Okafuji
arxiv.org/abs/2507.10960

@arXiv_csCL_bot@mastoxiv.page
2025-08-18 09:37:20

LLM Compression: How Far Can We Go in Balancing Size and Performance?
Sahil Sk, Debasish Dhal, Sonal Khosla, Sk Shahid, Sambit Shekhar, Akash Dhaka, Shantipriya Parida, Dilip K. Prasad, Ond\v{r}ej Bojar
arxiv.org/abs/2508.11318

@arXiv_csSD_bot@mastoxiv.page
2025-07-17 09:03:20

RUMAA: Repeat-Aware Unified Music Audio Analysis for Score-Performance Alignment, Transcription, and Mistake Detection
Sungkyun Chang, Simon Dixon, Emmanouil Benetos
arxiv.org/abs/2507.12175

@arXiv_csCV_bot@mastoxiv.page
2025-08-18 09:51:30

RMFAT: Recurrent Multi-scale Feature Atmospheric Turbulence Mitigator
Zhiming Liu, Nantheera Anantrasirichai
arxiv.org/abs/2508.11409 arxiv…

@arXiv_astrophIM_bot@mastoxiv.page
2025-07-16 08:45:41

Mapping Diffuse Radio Sources Using TUNA: A Transformer-Based Deep Learning Approach
Nicoletta Sanvitale, Claudio Gheller, Franco Vazza, Annalisa Bonafede, Virginia Cuciti, Emanuele De Rubeis, Federica Govoni, Matteo Murgia, Valentina Vacca
arxiv.org/abs/2507.11320

@arXiv_csCL_bot@mastoxiv.page
2025-08-18 09:08:00

Beyond the Rosetta Stone: Unification Forces in Generalization Dynamics
Carter Blum, Katja Filipova, Ann Yuan, Asma Ghandeharioun, Julian Zimmert, Fred Zhang, Jessica Hoffmann, Tal Linzen, Martin Wattenberg, Lucas Dixon, Mor Geva
arxiv.org/abs/2508.11017

@arXiv_csAI_bot@mastoxiv.page
2025-08-11 09:42:39

AntiCheatPT: A Transformer-Based Approach to Cheat Detection in Competitive Computer Games
Mille Mei Zhen Loo, Gert Luzkov, Paolo Burelli
arxiv.org/abs/2508.06348

@arXiv_csLG_bot@mastoxiv.page
2025-07-17 10:21:20

PRISM: Distributed Inference for Foundation Models at Edge
Muhammad Azlan Qazi, Alexandros Iosifidis, Qi Zhang
arxiv.org/abs/2507.12145

@arXiv_csCV_bot@mastoxiv.page
2025-08-18 09:52:40

Data-Driven Deepfake Image Detection Method -- The 2024 Global Deepfake Image Detection Challenge
Xiaoya Zhu, Yibing Nan, Shiguo Lian
arxiv.org/abs/2508.11464

@arXiv_eessAS_bot@mastoxiv.page
2025-09-19 09:29:31

SpeechMLC: Speech Multi-label Classification
Miseul Kim, Seyun Um, Hyeonjin Cha, Hong-goo Kang
arxiv.org/abs/2509.14677 arxiv.org/pdf/2509.…

@arXiv_eessIV_bot@mastoxiv.page
2025-07-17 09:24:30

Unit-Based Histopathology Tissue Segmentation via Multi-Level Feature Representation
Ashkan Shakarami, Azade Farshad, Yousef Yeganeh, Lorenzo Nicole, Peter Schuffler, Stefano Ghidoni, Nassir Navab
arxiv.org/abs/2507.12427

@arXiv_csCR_bot@mastoxiv.page
2025-07-17 08:12:10

Effective Fine-Tuning of Vision Transformers with Low-Rank Adaptation for Privacy-Preserving Image Classification
Haiwei Lin, Shoko Imaizumi, Hitoshi Kiya
arxiv.org/abs/2507.11943

@arXiv_csSD_bot@mastoxiv.page
2025-09-17 12:32:20

Replaced article(s) found for cs.SD. arxiv.org/list/cs.SD/new
[1/1]:
- SwinSRGAN: Swin Transformer-based Generative Adversarial Network for High-Fidelity Speech Super-R...
Jiajun Yuan, Xiaochen Wang, Yuhang Xiao, Yulin Wu, Chenhao Hu, Xueyang Lv

@arXiv_csCV_bot@mastoxiv.page
2025-08-15 10:26:02

STream3R: Scalable Sequential 3D Reconstruction with Causal Transformer
Yushi Lan, Yihang Luo, Fangzhou Hong, Shangchen Zhou, Honghua Chen, Zhaoyang Lyu, Shuai Yang, Bo Dai, Chen Change Loy, Xingang Pan
arxiv.org/abs/2508.10893

@arXiv_csLG_bot@mastoxiv.page
2025-07-09 10:13:12

The Fourier Spectral Transformer Networks For Efficient and Generalizable Nonlinear PDEs Prediction
Beibei Li
arxiv.org/abs/2507.05584

@arXiv_eessIV_bot@mastoxiv.page
2025-07-14 08:37:32

Depth-Sequence Transformer (DST) for Segment-Specific ICA Calcification Mapping on Non-Contrast CT
Xiangjian Hou, Ebru Yaman Akcicek, Xin Wang, Kazem Hashemizadeh, Scott Mcnally, Chun Yuan, Xiaodong Ma
arxiv.org/abs/2507.08214

@arXiv_csAI_bot@mastoxiv.page
2025-07-16 10:09:51

Modeling Code: Is Text All You Need?
Daniel Nichols, Konstantinos Parasyris, Harshitha Menon, Brian R. Bartoldson, Giorgis Georgakoudis, Tal Ben-Nun, Abhinav Bhatele
arxiv.org/abs/2507.11467

@arXiv_csCV_bot@mastoxiv.page
2025-09-16 12:44:07

3DViT-GAT: A Unified Atlas-Based 3D Vision Transformer and Graph Learning Framework for Major Depressive Disorder Detection Using Structural MRI Data
Nojod M. Alotaibi, Areej M. Alhothali, Manar S. Ali
arxiv.org/abs/2509.12143

@arXiv_csLG_bot@mastoxiv.page
2025-09-08 09:53:10

VARMA-Enhanced Transformer for Time Series Forecasting
Jiajun Song, Xiaoou Liu
arxiv.org/abs/2509.04782 arxiv.org/pdf/2509.04782

@arXiv_csCV_bot@mastoxiv.page
2025-07-14 10:03:02

Generalizable 7T T1-map Synthesis from 1.5T and 3T T1 MRI with an Efficient Transformer Model
Zach Eidex, Mojtaba Safari, Tonghe Wang, Vanessa Wildman, David S. Yu, Hui Mao, Erik Middlebrooks, Aparna Kesewala, Xiaofeng Yang
arxiv.org/abs/2507.08655

@arXiv_csLG_bot@mastoxiv.page
2025-09-10 10:37:41

Transformer-Based Approach to Optimal Sensor Placement for Structural Health Monitoring of Probe Cards
Mehdi Bejani, Marco Mauri, Daniele Acconcia, Simone Todaro, Stefano Mariani
arxiv.org/abs/2509.07603

@arXiv_csRO_bot@mastoxiv.page
2025-09-17 10:34:50

Out of Distribution Detection in Self-adaptive Robots with AI-powered Digital Twins
Erblin Isaku, Hassan Sartaj, Shaukat Ali, Beatriz Sanguino, Tongtong Wang, Guoyuan Li, Houxiang Zhang, Thomas Peyrucain
arxiv.org/abs/2509.12982

@arXiv_csCL_bot@mastoxiv.page
2025-09-12 09:23:49

ViRanker: A BGE-M3 & Blockwise Parallel Transformer Cross-Encoder for Vietnamese Reranking
Phuong-Nam Dang, Kieu-Linh Nguyen, Thanh-Hieu Pham
arxiv.org/abs/2509.09131

@arXiv_csCV_bot@mastoxiv.page
2025-09-15 10:02:31

I-Segmenter: Integer-Only Vision Transformer for Efficient Semantic Segmentation
Jordan Sassoon, Michal Szczepanski, Martyna Poreba
arxiv.org/abs/2509.10334

@arXiv_csCL_bot@mastoxiv.page
2025-08-14 09:48:12

Speed Always Wins: A Survey on Efficient Architectures for Large Language Models
Weigao Sun, Jiaxi Hu, Yucheng Zhou, Jusen Du, Disen Lan, Kexin Wang, Tong Zhu, Xiaoye Qu, Yu Zhang, Xiaoyu Mo, Daizong Liu, Yuxuan Liang, Wenliang Chen, Guoqi Li, Yu Cheng
arxiv.org/abs/2508.09834

@arXiv_csCV_bot@mastoxiv.page
2025-09-08 09:53:50

Scale-interaction transformer: a hybrid cnn-transformer model for facial beauty prediction
Djamel Eddine Boukhari
arxiv.org/abs/2509.05078

@arXiv_csCV_bot@mastoxiv.page
2025-08-12 12:46:43

THAT: Token-wise High-frequency Augmentation Transformer for Hyperspectral Pansharpening
Hongkun Jin, Hongcheng Jiang, Zejun Zhang, Yuan Zhang, Jia Fu, Tingfeng Li, Kai Luo
arxiv.org/abs/2508.08183

@arXiv_csCV_bot@mastoxiv.page
2025-09-17 10:55:20

TexTAR : Textual Attribute Recognition in Multi-domain and Multi-lingual Document Images
Rohan Kumar, Jyothi Swaroopa Jinka, Ravi Kiran Sarvadevabhatla
arxiv.org/abs/2509.13151

@arXiv_csCV_bot@mastoxiv.page
2025-07-10 08:30:21

Hierarchical Multi-Stage Transformer Architecture for Context-Aware Temporal Action Localization
Hayat Ullah, Arslan Munir, Oliver Nina
arxiv.org/abs/2507.06411

@arXiv_csCV_bot@mastoxiv.page
2025-09-11 09:13:53

Symmetry Interactive Transformer with CNN Framework for Diagnosis of Alzheimer's Disease Using Structural MRI
Zheng Yang, Yanteng Zhang, Xupeng Kou, Yang Liu, Chao Ren
arxiv.org/abs/2509.08243