
2025-06-26 09:08:50
Enhancing Large Language Models through Structured Reasoning
Yubo Dong, Hehe Fan
https://arxiv.org/abs/2506.20241 https://arxiv.org/p…
Enhancing Large Language Models through Structured Reasoning
Yubo Dong, Hehe Fan
https://arxiv.org/abs/2506.20241 https://arxiv.org/p…
An advanced AI driven database system
M. Tedeschi, S. Rizwan, C. Shringi, V. Devram Chandgir, S. Belich
https://arxiv.org/abs/2507.17778 https://arxiv.org/…
AraTable: Benchmarking LLMs' Reasoning and Understanding of Arabic Tabular Data
Rana Alshaikh, Israa Alghanmi, Shelan Jeawak
https://arxiv.org/abs/2507.18442 https://…
DistrAttention: An Efficient and Flexible Self-Attention Mechanism on Modern GPUs
Haolin Jin, Mengbai Xiao, Yuan Yuan, Xiao Zhang, Dongxiao Yu, Guanghui Zhang, Haoliang Wang
https://arxiv.org/abs/2507.17245
Characterizing Communication Patterns in Distributed Large Language Model Inference
Lang Xu, Kaushik Kandadi Suresh, Quentin Anthony, Nawras Alnaasan, Dhabaleswar K. Panda
https://arxiv.org/abs/2507.14392
On the Effectiveness of LLM-as-a-judge for Code Generation and Summarization
Giuseppe Crupi, Rosalia Tufano, Alejandro Velasco, Antonio Mastropaolo, Denys Poshyvanyk, Gabriele Bavota
https://arxiv.org/abs/2507.16587
Perspectives in Play: A Multi-Perspective Approach for More Inclusive NLP Systems
Benedetta Muscato, Lucia Passaro, Gizem Gezici, Fosca Giannotti
https://arxiv.org/abs/2506.20209 …
Scaling Recommender Transformers to One Billion Parameters
Kirill Khrylchenko, Artem Matveev, Sergei Makeev, Vladimir Baikalov
https://arxiv.org/abs/2507.15994
Malware Classification Leveraging NLP & Machine Learning for Enhanced Accuracy
Bishwajit Prasad Gond, Rajneekant, Pushkar Kishore, Durga Prasad Mohapatra
https://arxiv.org/abs/2506.16224
SocioXplorer: An Interactive Tool for Topic and Network Analysis in Social Data
Sandrine Chausson, Youssef Al Hariri, Walid Magdy, Bj\"orn Ross
https://arxiv.org/abs/2506.18845
NLP Meets the World: Toward Improving Conversations With the Public About Natural Language Processing Research
Shomir Wilson
https://arxiv.org/abs/2507.10559
Open problems in ageing science: A roadmap for biogerontology
Angelo Talay, Aleksey V. Belikov, Paul Ka Po To, Hamid H. Alfatemi, Uri Alon, Joris Deelen, Collin Y. Ewald, David Gems, Vera Gorbunova, Jan Gruber, Sara H\"agg, John Hemming, Steve Horvath, Alaattin Kaya, Caitlin J. Lewis, Andrea Maier, Maria B Marinova, Graham Pawelec, Shahaf Peleg, Suresh Rattan, Morten Scheibye-Knudsen, Tomas Schmauck-Medina, Vardan Saroyan, Andrei Seluanov, Alexandra Stolzing, Emma Teeling, Robert …
"AI Can Help Limit the Spread of Misinformation During Natural Disaster, Study Finds"
#AI #ArtificialIntelligence
Integrating Quantized LLMs into Robotics Systems as Edge AI to Leverage their Natural Language Processing Capabilities
Miguel \'A. Gonz\'alez-Santamarta, Francisco J. Rodr\'iguez-Lera, David Sobr\'in-Hidalgo, \'Angel Manuel Guerrero-Higueras, Vicente Matell\'An-Olivera
https://arxiv.org/abs/2506.09581
So this guy threw Natural Language Processing at the Voynich Manuscript and concluded that it probably is written in some kind of language and is not just total gibberish. Cool bit of ML research! https://github.com/brianmg/voynich-nlp-analysis
PathWeaver: A High-Throughput Multi-GPU System for Graph-Based Approximate Nearest Neighbor Search
Sukjin Kim, Seongyeon Park, Si Ung Noh, Junguk Hong, Taehee Kwon, Hunseong Lim, Jinho Lee
https://arxiv.org/abs/2507.17094
Finance Language Model Evaluation (FLaME)
Glenn Matlin, Mika Okamoto, Huzaifa Pardawala, Yang Yang, Sudheer Chava
https://arxiv.org/abs/2506.15846 https://…
eSapiens: A Real-World NLP Framework for Multimodal Document Understanding and Enterprise Knowledge Processing
Isaac Shi, Zeyuan Li, Wenli Wang, Lewei He, Yang Yang, Tianyu Shi
https://arxiv.org/abs/2506.16768
ElliottAgents: A Natural Language-Driven Multi-Agent System for Stock Market Analysis and Prediction
Jaros{\l}aw A. Chudziak, Micha{\l} Wawer
https://arxiv.org/abs/2507.03435
Attacking interpretable NLP systems
Eldor Abdukhamidov, Tamer Abuhmed, Joanna C. S. Santos, Mohammed Abuhamad
https://arxiv.org/abs/2507.16164 https://
DSSD: Efficient Edge-Device Deployment and Collaborative Inference via Distributed Split Speculative Decoding
Jiahong Ning, Ce Zheng, Tingting Yang
https://arxiv.org/abs/2507.12000
Reservoir Computing as a Language Model
Felix K\"oster, Atsushi Uchida
https://arxiv.org/abs/2507.15779 https://arxiv.org/pdf/25…
Next stop on our NLP timeline (as part of the #ISE2025 lecture) was Terry Winograd's SHRDLU, an early natural language understanding system developed in 1968-70 that could manipulate blocks in a virtual world.
Winograd, T. Procedures as a Representation for Data in a Computer Program for Understanding Natural Language. MIT AI Technical Report 235.
Human-Centred AI in FinTech: Developing a User Experience (UX) Research Point of View (PoV) Playbook
Festus Adedoyin, Huseyin Dogan
https://arxiv.org/abs/2506.15325
Identifying economic narratives in large text corpora -- An integrated approach using Large Language Models
Tobias Schmidt, Kai-Robin Lange, Matthias Reccius, Henrik M\"uller, Michael Roos, Carsten Jentsch
https://arxiv.org/abs/2506.15041
An ultra-low-power CGRA for accelerating Transformers at the edge
Rohit Prasad
https://arxiv.org/abs/2507.12904 https://arxiv.org/pdf…
Evolutionary Feature-wise Thresholding for Binary Representation of NLP Embeddings
Soumen Sinha, Shahryar Rahnamayan, Azam Asilian Bidgoli
https://arxiv.org/abs/2507.17025
This week, we were discussing the central question Can we "predict" a word? as the basis for statistical language models in our #ISE2025 lecture. Of course, I wasx trying Shakespeare quotes to motivate the (international) students to complement the quotes with "predicted" missing words ;-)
"All the world's a stage, and all the men and women merely...."
Evaluating Large Language Models for Phishing Detection, Self-Consistency, Faithfulness, and Explainability
Shova Kuikel, Aritran Piplai, Palvi Aggarwal
https://arxiv.org/abs/2506.13746
Replaced article(s) found for cs.CL. https://arxiv.org/list/cs.CL/new
[1/3]:
- Modeling the Sacred: Considerations when Using Religious Texts in Natural Language Processing
Ben Hutchinson
Quantum Adiabatic Generation of Human-Like Passwords
Sascha M\"ucke, Raoul Heese, Thore Gerlach, David Biesner, Loong Kuan Lee, Nico Piatkowski
https://arxiv.org/abs/2506.08917
This https://arxiv.org/abs/2505.16978 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csAI_…
Theories of "Sexuality" in Natural Language Processing Bias Research
Jacob Hobbs
https://arxiv.org/abs/2506.22481 https://a…
BMFM-DNA: A SNP-aware DNA foundation model to capture variant effects
Hongyang Li, Sanjoy Dey, Bum Chul Kwon, Michael Danziger, Michal Rosen-Tzvi, Jianying Hu, James Kozloski, Ching-Huei Tsou, Bharath Dandala, Pablo Meyer
https://arxiv.org/abs/2507.05265
Exploiting Primacy Effect To Improve Large Language Models
Bianca Raimondi, Maurizio Gabbrielli
https://arxiv.org/abs/2507.13949 https://
This https://arxiv.org/abs/2412.18407 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_sta…
With the advent of ELIZA, Joseph Weizenbaum's first psychotherapist chatbot, NLP took another major step with pattern-based substitution algorithms based on simple regular expressions.
Weizenbaum, Joseph (1966). ELIZA—a computer program for the study of natural language communication between man and machine. Com. of the ACM. 9: 36–45.
Evaluating Large Language Models for Phishing Detection, Self-Consistency, Faithfulness, and Explainability
Shova Kuikel, Aritran Piplai, Palvi Aggarwal
https://arxiv.org/abs/2506.13746
From Sentences to Sequences: Rethinking Languages in Biological System
Ke Liu, Shuanke Shen, Hao Chen
https://arxiv.org/abs/2507.00953 https://
FerroAI: A Deep Learning Model for Predicting Phase Diagrams of Ferroelectric Materials
Chenbo Zhang, Xian Chen
https://arxiv.org/abs/2506.10970 https://…
FlexiSAGA: A Flexible Systolic Array GEMM Accelerator for Sparse and Dense Processing
Mika Markus M\"uller, Konstantin L\"ubeck, Alexander Louis-Ferdinand Jung, Jannik Steinmetz, Oliver Bringmann
https://arxiv.org/abs/2506.01566
Language Surgery in Multilingual Large Language Models
Joanito Agili Lopo, Muhammad Ravi Shulthan Habibi, Tack Hwa Wong, Muhammad Ilham Ghozali, Fajri Koto, Genta Indra Winata, Peerat Limkonchotiwat, Alham Fikri Aji, Samuel Cahyawijaya
https://arxiv.org/abs/2506.12450
Bookmarked: Talking About Muslims in Middle French: The Potential of Word-to-Vector Models for Studying Semantic Relationships in Medieval Languages – DH Lab #Digital_Humanities
Estimation of An Infinite Dimensional Transition Probability Matrix Using a Generalized Hierarchical Stick-Breaking Process
Agamani Saha, Souvik Roy
https://arxiv.org/abs/2507.07433
Past, Present and Future: Exploring Adaptive AI in Software Development Bots
Omar Elsisi, Glaucia Melo
https://arxiv.org/abs/2507.10822 https://
CSI2Vec: Towards a Universal CSI Feature Representation for Positioning and Channel Charting
Victoria Palhares, Sueda Taner, Christoph Studer
https://arxiv.org/abs/2506.05237
FinBERT2: A Specialized Bidirectional Encoder for Bridging the Gap in Finance-Specific Deployment of Large Language Models
Xuan Xu, Fufang Wen, Beilin Chu, Zhibing Fu, Qinhong Lin, Jiaqi Liu, Binjie Fei, Zhongliang Yang, Linna Zhou, Yu Li
https://arxiv.org/abs/2506.06335
A Topic Modeling Analysis of Stigma Dimensions, Social, and Related Behavioral Circumstances in Clinical Notes Among Patients with HIV
Ziyi Chen, Yiyang Liu, Mattia Prosperi, Krishna Vaddiparti, Robert L Cook, Jiang Bian, Yi Guo, Yonghui Wu
https://arxiv.org/abs/2506.09279
Last leg on our brief history of NLP (so far) is the advent of large language models with GPT-3 in 2020 and the introduction of learning from the prompt (aka few-shot learning).
T. B. Brown et al. (2020). Language models are few-shot learners. NIPS'20
https://…
VRAgent-R1: Boosting Video Recommendation with MLLM-based Agents via Reinforcement Learning
Siran Chen, Boyu Chen, Chenyun Yu, Yuxiao Luo, Ouyang Yi, Lei Cheng, Chengxiang Zhuo, Zang Li, Yali Wang
https://arxiv.org/abs/2507.02626
LLM vs. SAST: A Technical Analysis on Detecting Coding Bugs of GPT4-Advanced Data Analysis
Madjid G. Tehrani, Eldar Sultanow, William J. Buchanan, Mahkame Houmani, Christel H. Djaha Fodja
https://arxiv.org/abs/2506.15212
Interactive Text-to-SQL via Expected Information Gain for Disambiguation
Luyu Qiu, Jianing Li, Chi Su, Lei Chen
https://arxiv.org/abs/2507.06467 https://…
This https://arxiv.org/abs/2412.01753 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csRO_…
A Language-Driven Framework for Improving Personalized Recommendations: Merging LLMs with Traditional Algorithms
Aaron Goldstein, Ayan Dutta
https://arxiv.org/abs/2507.07251
This https://arxiv.org/abs/2312.17294 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csSE_…
Beyond Architectures: Evaluating the Role of Contextual Embeddings in Detecting Bipolar Disorder on Social Media
Khalid Hasan, Jamil Saquer
https://arxiv.org/abs/2507.14231
Multivariate Long-term Time Series Forecasting with Fourier Neural Filter
Chenheng Xu, Dan Wu, Yixin Zhu, Ying Nian Wu
https://arxiv.org/abs/2506.09174 htt…
HyperSumm-RL: A Dialogue Summarization Framework for Modeling Leadership Perception in Social Robots
Subasish Das
https://arxiv.org/abs/2507.04160 https://…
Utilizing AI for Aviation Post-Accident Analysis Classification
Aziida Nanyonga, Graham Wild
https://arxiv.org/abs/2506.00169 https://
Let's Measure the Elephant in the Room: Facilitating Personalized Automated Analysis of Privacy Policies at Scale
Rui Zhao, Vladyslav Melnychuk, Jun Zhao, Jesse Wright, Nigel Shadbolt
https://arxiv.org/abs/2507.14214
VEDA: Efficient LLM Generation Through Voting-based KV Cache Eviction and Dataflow-flexible Accelerator
Zhican Wang, Hongxiang Fan, Haroon Waris, Gang Wang, Zhenyu Li, Jianfei Jiang, Yanan Sun, Guanghui He
https://arxiv.org/abs/2507.00797
Unveiling Privacy Policy Complexity: An Exploratory Study Using Graph Mining, Machine Learning, and Natural Language Processing
Vijayalakshmi Ramasamy, Seth Barrett, Gokila Dorai, Jessica Zumbach
https://arxiv.org/abs/2507.02968
Building on the 90s, statistical n-gram language models, trained on vast text collections, became the backbone of NLP research. They fueled advancements in nearly all NLP techniques of the era, laying the groundwork for today's AI.
F. Jelinek (1997), Statistical Methods for Speech Recognition, MIT Press, Cambridge, MA
#NLP
Is Diversity All You Need for Scalable Robotic Manipulation?
Modi Shi, Li Chen, Jin Chen, Yuxiang Lu, Chiming Liu, Guanghui Ren, Ping Luo, Di Huang, Maoqing Yao, Hongyang Li
https://arxiv.org/abs/2507.06219
Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management
Luis Gasco, Hermenegildo Fabregat, Laura Garc\'ia-Sardi\~na, Paula Estrella, Daniel Deniz, Alvaro Rodrigo, Rabih Zbib
https://arxiv.org/abs/2507.13275
Investigating Vulnerabilities and Defenses Against Audio-Visual Attacks: A Comprehensive Survey Emphasizing Multimodal Models
Jinming Wen, Xinyi Wu, Shuai Zhao, Yanhao Jia, Yuwen Li
https://arxiv.org/abs/2506.11521
Last week, we continued our #ISE2025 lecture on distributional semantics with the introduction of neural language models (NLMs) and compared them to traditional statistical n-gram models.
Benefits of NLMs:
- Capturing Long-Range Dependencies
- Computational and Statistical Tractability
- Improved Generalisation
- Higher Accuracy
@…
Improving Drug Identification in Overdose Death Surveillance using Large Language Models
Arthur J. Funnell, Panayiotis Petousis, Fabrice Harel-Canada, Ruby Romero, Alex A. T. Bui, Adam Koncsol, Hritika Chaturvedi, Chelsea Shover, David Goodman-Meza
https://arxiv.org/abs/2507.12679
INTERPOS: Interaction Rhythm Guided Positional Morphing for Mobile App Recommender Systems
M. H. Maqbool, Moghis Fereidouni, Umar Farooq, A. B. Siddique, Hassan Foroosh
https://arxiv.org/abs/2506.12661
Visualization for interactively adjusting the de-bias effect of word embedding
Arisa Sugino, Takayuki Itoh
https://arxiv.org/abs/2506.02447 https://…
An Evaluation of Large Language Models on Text Summarization Tasks Using Prompt Engineering Techniques
Walid Mohamed Aly, Taysir Hassan A. Soliman, Amr Mohamed AbdelAziz
https://arxiv.org/abs/2507.05123
Next stop in our NLP timeline is 2013, the introduction of low dimensional dense word vectors - so-called "word embeddings" - based on distributed semantics, as e.g. word2vec by Mikolov et al. from Google, which enabled representation learning on text.
T. Mikolov et al. (2013). Efficient Estimation of Word Representations in Vector Space.
…
A Computational Framework to Identify Self-Aspects in Text
Jaya Caporusso, Matthew Purver, Senja Pollak
https://arxiv.org/abs/2507.13115 https://
Verified Language Processing with Hybrid Explainability: A Technical Report
Oliver Robert Fox, Giacomo Bergami, Graham Morgan
https://arxiv.org/abs/2507.05017
The Dark Side of LLMs Agent-based Attacks for Complete Computer Takeover
Matteo Lupinacci, Francesco Aurelio Pironti, Francesco Blefari, Francesco Romeo, Luigi Arena, Angelo Furfaro
https://arxiv.org/abs/2507.06850
Training-free LLM Merging for Multi-task Learning
Zichuan Fu, Xian Wu, Yejing Wang, Wanyu Wang, Shanshan Ye, Hongzhi Yin, Yi Chang, Yefeng Zheng, Xiangyu Zhao
https://arxiv.org/abs/2506.12379
Text-ADBench: Text Anomaly Detection Benchmark based on LLMs Embedding
Feng Xiao, Jicong Fan
https://arxiv.org/abs/2507.12295 https://
SoK: Are Watermarks in LLMs Ready for Deployment?
Kieu Dang, Phung Lai, NhatHai Phan, Yelong Shen, Ruoming Jin, Abdallah Khreishah, My Thai
https://arxiv.org/abs/2506.05594
PLACE: Prompt Learning for Attributed Community Search
Shuheng Fang, Kangfei Zhao, Rener Zhang, Yu Rong, Jeffrey Xu Yu
https://arxiv.org/abs/2507.05311 htt…
False Alarms, Real Damage: Adversarial Attacks Using LLM-based Models on Text-based Cyber Threat Intelligence Systems
Samaneh Shafee, Alysson Bessani, Pedro M. Ferreira
https://arxiv.org/abs/2507.06252
Propaganda and Information Dissemination in the Russo-Ukrainian War: Natural Language Processing of Russian and Western Twitter Narratives
Zaur Gouliev
https://arxiv.org/abs/2506.01807
Adversarial Text Generation with Dynamic Contextual Perturbation
Hetvi Waghela, Jaydip Sen, Sneha Rakshit, Subhasis Dasgupta
https://arxiv.org/abs/2506.09148
Towards Fair Rankings: Leveraging LLMs for Gender Bias Detection and Measurement
Maryam Mousavian, Zahra Abbasiantaeb, Mohammad Aliannejadi, Fabio Crestani
https://arxiv.org/abs/2506.22372
Iterative Augmentation with Summarization Refinement (IASR) Evaluation for Unstructured Survey data Modeling and Analysis
Payal Bhattad, Sai Manoj Pudukotai Dinakarrao, Anju Gupta
https://arxiv.org/abs/2507.12126
This https://arxiv.org/abs/2505.18889 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csCR_…
Natural language processing for African languages
David Ifeoluwa Adelani
https://arxiv.org/abs/2507.00297 https://arxiv.org/pdf/2507.…
From Ambiguity to Accuracy: The Transformative Effect of Coreference Resolution on Retrieval-Augmented Generation systems
Youngjoon Jang, Seongtae Hong, Junyoung Son, Sungjin Park, Chanjun Park, Heuiseok Lim
https://arxiv.org/abs/2507.07847
PROVSYN: Synthesizing Provenance Graphs for Data Augmentation in Intrusion Detection Systems
Yi Huang, Wajih UI Hassan, Yao Guo, Xiangqun Chen, Ding Li
https://arxiv.org/abs/2506.06226
Checklist Engineering Empowers Multilingual LLM Judges
Mohammad Ghiasvand Mohammadkhani, Hamid Beigy
https://arxiv.org/abs/2507.06774 https://
Dialogue-Based Multi-Dimensional Relationship Extraction from Novels
Yuchen Yan, Hanjie Zhao, Senbin Zhu, Hongde Liu, Zhihong Zhang, Yuxiang Jia
https://arxiv.org/abs/2507.04852
MaXIFE: Multilingual and Cross-lingual Instruction Following Evaluation
Yile Liu, Ziwei Ma, Xiu Jiang, Jinglu Hu, Jing Chang, Liang Li
https://arxiv.org/abs/2506.01776
Rethinking the Privacy of Text Embeddings: A Reproducibility Study of "Text Embeddings Reveal (Almost) As Much As Text"
Dominykas Seputis, Yongkang Li, Karsten Langerak, Serghei Mihailov
https://arxiv.org/abs/2507.07700
LLMs as Architects and Critics for Multi-Source Opinion Summarization
Anuj Attri, Arnav Attri, Pushpak Bhattacharyya, Suman Banerjee, Amey Patil, Muthusamy Chelliah, Nikesh Garera
https://arxiv.org/abs/2507.04751
iQUEST: An Iterative Question-Guided Framework for Knowledge Base Question Answering
Shuai Wang, Yinan Yu
https://arxiv.org/abs/2506.01784 https://
R1-RE: Cross-Domain Relationship Extraction with RLVR
Runpeng Dai, Tong Zheng, Run Yang, Hongtu Zhu
https://arxiv.org/abs/2507.04642 https://
Put Teacher in Student's Shoes: Cross-Distillation for Ultra-compact Model Compression Framework
Maolin Wang, Jun Chu, Sicong Xie, Xiaoling Zang, Yao Zhao, Wenliang Zhong, Xiangyu Zhao
https://arxiv.org/abs/2507.04636
Evaluating Scoring Bias in LLM-as-a-Judge
Qingquan Li, Shaoyu Dou, Kailai Shao, Chao Chen, Haixiang Hu
https://arxiv.org/abs/2506.22316 https://
Revisiting Active Learning under (Human) Label Variation
Cornelia Gruber, Helen Alber, Bernd Bischl, G\"oran Kauermann, Barbara Plank, Matthias A{\ss}enmacher
https://arxiv.org/abs/2507.02593