
2025-08-19 11:44:50
Word Meanings in Transformer Language Models
Jumbly Grindrod, Peter Grindrod
https://arxiv.org/abs/2508.12863 https://arxiv.org/pdf/2508.12863
Word Meanings in Transformer Language Models
Jumbly Grindrod, Peter Grindrod
https://arxiv.org/abs/2508.12863 https://arxiv.org/pdf/2508.12863
Attention Beyond Neighborhoods: Reviving Transformer for Graph Clustering
Xuanting Xie, Bingheng Li, Erlin Pan, Rui Hou, Wenyu Chen, Zhao Kang
https://arxiv.org/abs/2509.15024 h…
Handwritten Text Recognition of Historical Manuscripts Using Transformer-Based Models
Erez Meoded
https://arxiv.org/abs/2508.11499 https://arxiv.org/pdf/25…
OVSegDT: Segmenting Transformer for Open-Vocabulary Object Goal Navigation
Tatiana Zemskova, Aleksei Staroverov, Dmitry Yudin, Aleksandr Panov
https://arxiv.org/abs/2508.11479 h…
ROSE: Transformer-Based Refactoring Recommendation for Architectural Smells
Samal Nursapa, Anastassiya Samuilova, Alessio Bucaioni. Phuong T. Nguyen
https://arxiv.org/abs/2507.12561
FractMorph: A Fractional Fourier-Based Multi-Domain Transformer for Deformable Image Registration
Shayan Kebriti, Shahabedin Nabavi, Ali Gooya
https://arxiv.org/abs/2508.12445 h…
HairFormer: Transformer-Based Dynamic Neural Hair Simulation
Joy Xiaoji Zhang, Jingsen Zhu, Hanyu Chen, Steve Marschner
https://arxiv.org/abs/2507.12600 ht…
Nvidia debuts the Nemotron-Nano-9B-v2, a hybrid Mamba-transformer model, saying it achieves scores comparable to or better than Qwen3-8B on reasoning benchmarks (Carl Franzen/VentureBeat)
https://venturebeat.com/ai/nvidia-rele
MELA-TTS: Joint transformer-diffusion model with representation alignment for speech synthesis
Keyu An, Zhiyu Zhang, Changfeng Gao, Yabin Li, Zhendong Peng, Haoxu Wang, Zhihao Du, Han Zhao, Zhifu Gao, Xiangang Li
https://arxiv.org/abs/2509.14784
Rationalizing Transformer Predictions via End-To-End Differentiable Self-Training
Marc Brinner, Sina Zarrie{\ss}
https://arxiv.org/abs/2508.11393 https://a…
Early Prediction of Current Quench Events in the ADITYA Tokamak using Transformer based Data Driven Models
Jyoti Agarwal, Bhaskar Chaudhury, Jaykumar Navadiya, Shrichand Jakhar, Manika Sharma
https://arxiv.org/abs/2507.12797
Multi-Class-Token Transformer for Multitask Self-supervised Music Information Retrieval
Yuexuan Kong, Vincent Lostanlen, Romain Hennequin, Mathieu Lagrange, Gabriel Meseguer-Brocal
https://arxiv.org/abs/2507.12996
Large Language Model-Empowered Decision Transformer for UAV-Enabled Data Collection
Zhixion Chen, Jiangzhou Wang, and Hyundong Shin, Arumugam Nallanathan
https://arxiv.org/abs/2509.13934
UltraHiT: A Hierarchical Transformer Architecture for Generalizable Internal Carotid Artery Robotic Ultrasonography
Teng Wang, Haojun Jiang, Yuxuan Wang, Zhenguo Sun, Xiangjie Yan, Xiang Li, Gao Huang
https://arxiv.org/abs/2509.13832
Brain Tumor Segmentation in Sub-Sahara Africa with Advanced Transformer and ConvNet Methods: Fine-Tuning, Data Mixing and Ensembling
Toufiq Musah, Chantelle Amoako-Atta, John Amankwaah Otu, Lukman E. Ismaila, Swallah Alhaji Suraka, Oladimeji Williams, Isaac Tigbee, Kato Hussein Wabbi, Samantha Katsande, Kanyiri Ahmed Yakubu, Adedayo Kehinde Lawal, Anita Nsiah Donkor, Naeem Mwinlanaah Adamu, Adebowale Akande, John Othieno, Prince Ebenezer Adjei, Zhang Dong, Confidence Raymond, Udunna C.…
The HRM paper has been mostly debunked by the ARC-AGI people.
https://arcprize.org/blog/hrm-analysis
The results are legit but most of them are not down to the architecture (swapping it out for a transformer doesn't change that much).
Also, the model is purely transductive. It onl…
Image-Based Multi-Survey Classification of Light Curves with a Pre-Trained Vision Transformer
Daniel Moreno-Cartagena, Guillermo Cabrera-Vives, Alejandra M. Mu\~noz Arancibia, Pavlos Protopapas, Francisco F\"orster, M\'arcio Catelan, A. Bayo, Pablo A. Est\'evez, P. S\'anchez-S\'aez, Franz E. Bauer, M. Pavez-Herrera, L. Hern\'andez-Garc\'ia, Gonzalo Rojas
Taming Diffusion Transformer for Real-Time Mobile Video Generation
Yushu Wu, Yanyu Li, Anil Kag, Ivan Skorokhodov, Willi Menapace, Ke Ma, Arpit Sahni, Ju Hu, Aliaksandr Siarohin, Dhritiman Sagar, Yanzhi Wang, Sergey Tulyakov
https://arxiv.org/abs/2507.13343
Characterizing State Space Model (SSM) and SSM-Transformer Hybrid Language Model Performance with Long Context Length
Saptarshi Mitra, Rachid Karami, Haocheng Xu, Sitao Huang, Hyoukjun Kwon
https://arxiv.org/abs/2507.12442
Semantic Fusion with Fuzzy-Membership Features for Controllable Language Modelling
Yongchao Huang, Hassan Raza
https://arxiv.org/abs/2509.13357 https://arx…
MixCache: Mixture-of-Cache for Video Diffusion Transformer Acceleration
Yuanxin Wei, Lansong Diao, Bujiao Chen, Shenggan Cheng, Zhengping Qian, Wenyuan Yu, Nong Xiao, Wei Lin, Jiangsu Du
https://arxiv.org/abs/2508.12691
Asymptotic Study of In-context Learning with Random Transformers through Equivalent Models
Samet Demir, Zafer Dogan
https://arxiv.org/abs/2509.15152 https://
Replaced article(s) found for cs.CR. https://arxiv.org/list/cs.CR/new
[1/1]:
- TBDetector:Transformer-Based Detector for Advanced Persistent Threats with Provenance Graph
Wang, Wen, Zhang, Zhao, Ma, Luo, Xu, Nie, Wu, Liu
SV-Mixer: Replacing the Transformer Encoder with Lightweight MLPs for Self-Supervised Model Compression in Speaker Verification
Jungwoo Heo, Hyun-seo Shin, Chan-yeong Lim, Kyo-won Koo, Seung-bin Kim, Jisoo Son, Ha-Jin Yu
https://arxiv.org/abs/2509.14136
NEFT: A Unified Transformer Framework for Efficient Near-Field CSI Feedback in XL-MIMO Systems
Haiyang Li, Tianqi Mao, Pengyu Wang, Ruiqi Liu, Shunyu Li, Zhaocheng Wang
https://arxiv.org/abs/2509.12748
Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy
Yiting Yang, Hao Luo, Yuan Sun, Qingsen Yan, Haokui Zhang, Wei Dong, Guoqing Wang, Peng Wang, Yang Yang, Hengtao Shen
https://arxiv.org/abs/2507.13260
Transformer Networks for Continuous Gravitational-wave Searches
Prasanna. M. Joshi, Reinhard Prix
https://arxiv.org/abs/2509.10912 https://arxiv.org/pdf/25…
Enhancements to the IceCube Extremely High Energy Neutrino Selection using Graph & Transformer Based Neural Networks
Maxwell Nakos (for the IceCube Collaboration), Aske Rosted (for the IceCube Collaboration), Lu Lu (for the IceCube Collaboration)
https://arxiv.org/abs/2507.11774
MultiPark: Multimodal Parking Transformer with Next-Segment Prediction
Han Zheng, Zikang Zhou, Guli Zhang, Zhepei Wang, Kaixuan Wang, Peiliang Li, Shaojie Shen, Ming Yang, Tong Qin
https://arxiv.org/abs/2508.11537
Crosslisted article(s) found for cs.DL. https://arxiv.org/list/cs.DL/new
[1/1]:
- Handwritten Text Recognition of Historical Manuscripts Using Transformer-Based Models
Erez Meoded
Crosslisted article(s) found for cs.AI. https://arxiv.org/list/cs.AI/new
[5/5]:
- Handwritten Text Recognition of Historical Manuscripts Using Transformer-Based Models
Erez Meoded
Pre-trained Transformer-models using chronic invasive electrophysiology for symptom decoding without patient-individual training
Timon Merk, Saeed Salehi, Richard M. Koehler, Qiming Cui, Maria Olaru, Amelia Hahn, Nicole R. Provenza, Simon Little, Reza Abbasi-Asl, Phil A. Starr, Wolf-Julian Neumann
https://arxiv.org/abs/2508.10160
Pre-Manipulation Alignment Prediction with Parallel Deep State-Space and Transformer Models
Motonari Kambara, Komei Sugiura
https://arxiv.org/abs/2509.13839 https://
Universal Synthesis of Differentiably Tunable Numerical Abstract Transformers
Shaurya Gomber, Debangshu Banerjee, Gagandeep Singh
https://arxiv.org/abs/2507.11827
Replaced article(s) found for q-fin.MF. https://arxiv.org/list/q-fin.MF/new
[1/1]:
- Quantformer: from attention to profit with a quantitative transformer trading strategy
Zhaofeng Zhang, Banghao Chen, Shengxin Zhu, Nicolas Langren\'e
A Robust Cross-Domain IDS using BiGRU-LSTM-Attention for Medical and Industrial IoT Security
Afrah Gueriani, Hamza Kheddar, Ahmed Cherif Mazari, Mohamed Chahine Ghanem
https://arxiv.org/abs/2508.12470 …
LKFMixer: Exploring Large Kernel Feature For Efficient Image Super-Resolution
Yinggan Tang, Quanwei Hu
https://arxiv.org/abs/2508.11391 https://arxiv.org/p…
Predictive Position Control for Movable Antenna Arrays in UAV Communications: A Spatio-Temporal Transformer-LSTM Framework
Kan Yu, Kaixuan Li, Xiaowu Liu, Qixun Zhang, Zhiyong Feng
https://arxiv.org/abs/2508.10720
eSASRec: Enhancing Transformer-based Recommendations in a Modular Fashion
Daria Tikhonovich, Nikita Zelinskiy, Aleksandr V. Petrov, Mayya Spirina, Andrei Semenov, Andrey V. Savchenko, Sergei Kuliev
https://arxiv.org/abs/2508.06450
Perceptual Reality Transformer: Neural Architectures for Simulating Neurological Perception Conditions
Baihan Lin
https://arxiv.org/abs/2508.09852 https://…
Dynamic Relational Priming Improves Transformer in Multivariate Time Series
Hunjae Lee, Corey Clark
https://arxiv.org/abs/2509.12196 https://arxiv.org/pdf/…
Patent Language Model Pretraining with ModernBERT
Amirhossein Yousefiramandi, Ciaran Cooney
https://arxiv.org/abs/2509.14926 https://arxiv.org/pdf/2509.149…
Streaming 4D Visual Geometry Transformer
Dong Zhuo, Wenzhao Zheng, Jiahe Guo, Yuqi Wu, Jie Zhou, Jiwen Lu
https://arxiv.org/abs/2507.11539 https://
An ultra-low-power CGRA for accelerating Transformers at the edge
Rohit Prasad
https://arxiv.org/abs/2507.12904 https://arxiv.org/pdf…
From Next Token Prediction to (STRIPS) World Models -- Preliminary Results
Carlos N\'u\~nez-Molina, Vicen\c{c} G\'omez, Hector Geffner
https://arxiv.org/abs/2509.13389 h…
Tenma: Robust Cross-Embodiment Robot Manipulation with Diffusion Transformer
Travis Davies, Yiqi Huang, Yunxin Liu, Xiang Chen, Huxian Liu, Luhui Hu
https://arxiv.org/abs/2509.11865
Arabic ASR on the SADA Large-Scale Arabic Speech Corpus with Transformer-Based Models
Branislav Gerazov, Marcello Politi, S\'ebastien Brati\`eres
https://arxiv.org/abs/2508.12968
Astro-MoE: Mixture of Experts for Multiband Astronomical Time Series
Martina C\'adiz-Leyton, Guillermo Cabrera-Vives, Pavlos Protopapas, Daniel Moreno-Cartagena, Ignacio Becker
https://arxiv.org/abs/2507.12611
Addressing Data Imbalance in Transformer-Based Multi-Label Emotion Detection with Weighted Loss
Xia Cui
https://arxiv.org/abs/2507.11384 https://
An Uncertainty-Weighted Decision Transformer for Navigation in Dense, Complex Driving Scenarios
Zhihao Zhang, Chengyang Peng, Minghao Zhu, Ekim Yurtsever, Keith A. Redmill
https://arxiv.org/abs/2509.13132
Unleashing Vision Foundation Models for Coronary Artery Segmentation: Parallel ViT-CNN Encoding and Variational Fusion
Caixia Dong, Duwei Dai, Xinyi Han, Fan Liu, Xu Yang, Zongfang Li, Songhua Xu
https://arxiv.org/abs/2507.12938
Efficient Transformer-Based Piano Transcription With Sparse Attention Mechanisms
Weixing Wei, Kazuyoshi Yoshii
https://arxiv.org/abs/2509.09318 https://arx…
Compact Attention: Exploiting Structured Spatio-Temporal Sparsity for Fast Video Generation
Qirui Li, Guangcong Zheng, Qi Zhao, Jie Li, Bin Dong, Yiwu Yao, Xi Li
https://arxiv.org/abs/2508.12969
Llama-Mimi: Speech Language Models with Interleaved Semantic and Acoustic Tokens
Issa Sugiura, Shuhei Kurita, Yusuke Oda, Ryuichiro Higashinaka
https://arxiv.org/abs/2509.14882 …
RTR: A Transformer-Based Lossless Crossover with Perfect Phase Alignment
Xiangying Li, Jiankuan Li, Yong Tang
https://arxiv.org/abs/2509.08272 https://arxi…
Sparse Autoencoders for Sequential Recommendation Models: Interpretation and Flexible Control
Anton Klenitskiy, Konstantin Polev, Daria Denisova, Alexey Vasilev, Dmitry Simakov, Gleb Gusev
https://arxiv.org/abs/2507.12202
Self-Supervised Temporal Super-Resolution of Energy Data using Generative Adversarial Transformer
Xuanhao Mu, G\"okhan Demirel, Yuzhe Zhang, Jianlei Liu, Thorsten Schlachter, Veit Hagenmeyer
https://arxiv.org/abs/2508.10587
A Transformer-Based Approach for DDoS Attack Detection in IoT Networks
Sandipan Dey, Payal Santosh Kate, Vatsala Upadhyay, Abhishek Vaish
https://arxiv.org/abs/2508.10636 https:…
A Semi-Supervised Learning Method for the Identification of Bad Exposures in Large Imaging Surveys
Yufeng Luo, Adam D. Myers, Alex Drlica-Wagner, Dario Dematties, Salma Borchani, Frank Valdes, Arjun Dey, David Schlegel, Rongpu Zhou, DESI Legacy Imaging Surveys Team
https://arxiv.org/abs/2507.12784
CytoSAE: Interpretable Cell Embeddings for Hematology
Muhammed Furkan Dasdelen, Hyesu Lim, Michele Buck, Katharina S. G\"otze, Carsten Marr, Steffen Schneider
https://arxiv.org/abs/2507.12464
The Imitation Game: Turing Machine Imitator is Length Generalizable Reasoner
Zhouqi Hua, Wenwei Zhang, Chengqi Lyu, Yuzhe Gu, Songyang Gao, Kuikun Liu, Kai Chen
https://arxiv.org/abs/2507.13332
GRATE: a Graph transformer-based deep Reinforcement learning Approach for Time-efficient autonomous robot Exploration
Haozhan Ni, Jingsong Liang, Chenyu He, Yuhong Cao, Guillaume Sartoretti
https://arxiv.org/abs/2509.12863
Real-Time Beach Litter Detection and Counting: A Comparative Analysis of RT-DETR Model Variants
Miftahul Huda, Arsyiah Azahra, Putri Maulida Chairani, Dimas Rizky Ramadhani, Nabila Azhari, Ade Lailani
https://arxiv.org/abs/2508.13101
Canary-1B-v2 & Parakeet-TDT-0.6B-v3: Efficient and High-Performance Models for Multilingual ASR and AST
Monica Sekoyan, Nithin Rao Koluguri, Nune Tadevosyan, Piotr Zelasko, Travis Bartley, Nick Karpov, Jagadeesh Balam, Boris Ginsburg
https://arxiv.org/abs/2509.14128
Pimba: A Processing-in-Memory Acceleration for Post-Transformer Large Language Model Serving
Wonung Kim, Yubin Lee, Yoonsung Kim, Jinwoo Hwang, Seongryong Oh, Jiyong Jung, Aziz Huseynov, Woong Gyu Park, Chang Hyun Park, Divya Mahajan, Jongse Park
https://arxiv.org/abs/2507.10178
Differential Robustness in Transformer Language Models: Empirical Evaluation Under Adversarial Text Attacks
Taniya Gidatkar, Oluwaseun Ajao, Matthew Shardlow
https://arxiv.org/abs/2509.09706
DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition
Hayat Ullah, Muhammad Ali Shafique, Abbas Khan, Arslan Munir
https://arxiv.org/abs/2507.12426
Faster and Memory-Efficient Training of Sequential Recommendation Models for Large Catalogs
Maxim Zhelnin, Dmitry Redko, Volkov Daniil, Anna Volodkevich, Petr Sokerin, Valeriy Shevchenko, Egor Shvetsov, Alexey Vasilev, Darya Denisova, Ruslan Izmailov, Alexey Zaytsev
https://arxiv.org/abs/2509.09682…
Adaptive Pareto-Optimal Token Merging for Edge Transformer Models in Semantic Communication
Omar Erak, Omar Alhussein, Hatem Abou-Zeid, Mehdi Bennis
https://arxiv.org/abs/2509.09168
Whom to Respond To? A Transformer-Based Model for Multi-Party Social Robot Interaction
He Zhu, Ryo Miyoshi, Yuki Okafuji
https://arxiv.org/abs/2507.10960 h…
LLM Compression: How Far Can We Go in Balancing Size and Performance?
Sahil Sk, Debasish Dhal, Sonal Khosla, Sk Shahid, Sambit Shekhar, Akash Dhaka, Shantipriya Parida, Dilip K. Prasad, Ond\v{r}ej Bojar
https://arxiv.org/abs/2508.11318
RUMAA: Repeat-Aware Unified Music Audio Analysis for Score-Performance Alignment, Transcription, and Mistake Detection
Sungkyun Chang, Simon Dixon, Emmanouil Benetos
https://arxiv.org/abs/2507.12175
RMFAT: Recurrent Multi-scale Feature Atmospheric Turbulence Mitigator
Zhiming Liu, Nantheera Anantrasirichai
https://arxiv.org/abs/2508.11409 https://arxiv…
Mapping Diffuse Radio Sources Using TUNA: A Transformer-Based Deep Learning Approach
Nicoletta Sanvitale, Claudio Gheller, Franco Vazza, Annalisa Bonafede, Virginia Cuciti, Emanuele De Rubeis, Federica Govoni, Matteo Murgia, Valentina Vacca
https://arxiv.org/abs/2507.11320
Beyond the Rosetta Stone: Unification Forces in Generalization Dynamics
Carter Blum, Katja Filipova, Ann Yuan, Asma Ghandeharioun, Julian Zimmert, Fred Zhang, Jessica Hoffmann, Tal Linzen, Martin Wattenberg, Lucas Dixon, Mor Geva
https://arxiv.org/abs/2508.11017
AntiCheatPT: A Transformer-Based Approach to Cheat Detection in Competitive Computer Games
Mille Mei Zhen Loo, Gert Luzkov, Paolo Burelli
https://arxiv.org/abs/2508.06348 https:…
PRISM: Distributed Inference for Foundation Models at Edge
Muhammad Azlan Qazi, Alexandros Iosifidis, Qi Zhang
https://arxiv.org/abs/2507.12145 https://
Data-Driven Deepfake Image Detection Method -- The 2024 Global Deepfake Image Detection Challenge
Xiaoya Zhu, Yibing Nan, Shiguo Lian
https://arxiv.org/abs/2508.11464 https://…
SpeechMLC: Speech Multi-label Classification
Miseul Kim, Seyun Um, Hyeonjin Cha, Hong-goo Kang
https://arxiv.org/abs/2509.14677 https://arxiv.org/pdf/2509.…
Unit-Based Histopathology Tissue Segmentation via Multi-Level Feature Representation
Ashkan Shakarami, Azade Farshad, Yousef Yeganeh, Lorenzo Nicole, Peter Schuffler, Stefano Ghidoni, Nassir Navab
https://arxiv.org/abs/2507.12427
Effective Fine-Tuning of Vision Transformers with Low-Rank Adaptation for Privacy-Preserving Image Classification
Haiwei Lin, Shoko Imaizumi, Hitoshi Kiya
https://arxiv.org/abs/2507.11943
Replaced article(s) found for cs.SD. https://arxiv.org/list/cs.SD/new
[1/1]:
- SwinSRGAN: Swin Transformer-based Generative Adversarial Network for High-Fidelity Speech Super-R...
Jiajun Yuan, Xiaochen Wang, Yuhang Xiao, Yulin Wu, Chenhao Hu, Xueyang Lv
STream3R: Scalable Sequential 3D Reconstruction with Causal Transformer
Yushi Lan, Yihang Luo, Fangzhou Hong, Shangchen Zhou, Honghua Chen, Zhaoyang Lyu, Shuai Yang, Bo Dai, Chen Change Loy, Xingang Pan
https://arxiv.org/abs/2508.10893
The Fourier Spectral Transformer Networks For Efficient and Generalizable Nonlinear PDEs Prediction
Beibei Li
https://arxiv.org/abs/2507.05584 https://
Depth-Sequence Transformer (DST) for Segment-Specific ICA Calcification Mapping on Non-Contrast CT
Xiangjian Hou, Ebru Yaman Akcicek, Xin Wang, Kazem Hashemizadeh, Scott Mcnally, Chun Yuan, Xiaodong Ma
https://arxiv.org/abs/2507.08214
Modeling Code: Is Text All You Need?
Daniel Nichols, Konstantinos Parasyris, Harshitha Menon, Brian R. Bartoldson, Giorgis Georgakoudis, Tal Ben-Nun, Abhinav Bhatele
https://arxiv.org/abs/2507.11467
3DViT-GAT: A Unified Atlas-Based 3D Vision Transformer and Graph Learning Framework for Major Depressive Disorder Detection Using Structural MRI Data
Nojod M. Alotaibi, Areej M. Alhothali, Manar S. Ali
https://arxiv.org/abs/2509.12143
VARMA-Enhanced Transformer for Time Series Forecasting
Jiajun Song, Xiaoou Liu
https://arxiv.org/abs/2509.04782 https://arxiv.org/pdf/2509.04782
Generalizable 7T T1-map Synthesis from 1.5T and 3T T1 MRI with an Efficient Transformer Model
Zach Eidex, Mojtaba Safari, Tonghe Wang, Vanessa Wildman, David S. Yu, Hui Mao, Erik Middlebrooks, Aparna Kesewala, Xiaofeng Yang
https://arxiv.org/abs/2507.08655
Transformer-Based Approach to Optimal Sensor Placement for Structural Health Monitoring of Probe Cards
Mehdi Bejani, Marco Mauri, Daniele Acconcia, Simone Todaro, Stefano Mariani
https://arxiv.org/abs/2509.07603
Out of Distribution Detection in Self-adaptive Robots with AI-powered Digital Twins
Erblin Isaku, Hassan Sartaj, Shaukat Ali, Beatriz Sanguino, Tongtong Wang, Guoyuan Li, Houxiang Zhang, Thomas Peyrucain
https://arxiv.org/abs/2509.12982
ViRanker: A BGE-M3 & Blockwise Parallel Transformer Cross-Encoder for Vietnamese Reranking
Phuong-Nam Dang, Kieu-Linh Nguyen, Thanh-Hieu Pham
https://arxiv.org/abs/2509.09131
I-Segmenter: Integer-Only Vision Transformer for Efficient Semantic Segmentation
Jordan Sassoon, Michal Szczepanski, Martyna Poreba
https://arxiv.org/abs/2509.10334 https://
Speed Always Wins: A Survey on Efficient Architectures for Large Language Models
Weigao Sun, Jiaxi Hu, Yucheng Zhou, Jusen Du, Disen Lan, Kexin Wang, Tong Zhu, Xiaoye Qu, Yu Zhang, Xiaoyu Mo, Daizong Liu, Yuxuan Liang, Wenliang Chen, Guoqi Li, Yu Cheng
https://arxiv.org/abs/2508.09834
Scale-interaction transformer: a hybrid cnn-transformer model for facial beauty prediction
Djamel Eddine Boukhari
https://arxiv.org/abs/2509.05078 https://…
THAT: Token-wise High-frequency Augmentation Transformer for Hyperspectral Pansharpening
Hongkun Jin, Hongcheng Jiang, Zejun Zhang, Yuan Zhang, Jia Fu, Tingfeng Li, Kai Luo
https://arxiv.org/abs/2508.08183
TexTAR : Textual Attribute Recognition in Multi-domain and Multi-lingual Document Images
Rohan Kumar, Jyothi Swaroopa Jinka, Ravi Kiran Sarvadevabhatla
https://arxiv.org/abs/2509.13151
Hierarchical Multi-Stage Transformer Architecture for Context-Aware Temporal Action Localization
Hayat Ullah, Arslan Munir, Oliver Nina
https://arxiv.org/abs/2507.06411
Symmetry Interactive Transformer with CNN Framework for Diagnosis of Alzheimer's Disease Using Structural MRI
Zheng Yang, Yanteng Zhang, Xupeng Kou, Yang Liu, Chao Ren
https://arxiv.org/abs/2509.08243