Tootfinder

Opt-in global Mastodon full text search. Join the index!

@nohillside@smnn.ch
2025-06-18 08:05:39

„While LLMs offer immediate convenience, our findings highlight potential cognitive costs. Over four months, #LLM users consistently underperformed at neural, linguistic, and behavioral levels.“ #AI

@ErikJonker@mastodon.social
2025-07-19 15:29:16

Very nice article about LLM architecture, a bit too complicated for me but probably not for others..
magazine.sebastianraschka.com/

@heiseonline@social.heise.de
2025-08-18 09:05:00

KI-gestützte Cyberangriffe: Experten beobachten zunehmenden LLM-Einsatz
Erstmals wurde russische Spionagesoftware entdeckt, die mit Large Language Models entwickelt wurde. Experten sehen einen Wendepunkt im Cyberwettrüsten.

@tante@tldr.nettime.org
2025-06-16 10:54:42

New study on the effects of LLM use (in this case on essay writing):
arxiv.org/abs/2506.08872
Quote:
"LLM users also struggled to accurately quote their own work. While LLMs offer immediate convenience, our findings highlight potential cognitive costs. Over four month…

@chpietsch@fedifreu.de
2025-06-16 19:02:46

Wissenschaftler:innen haben herausgefunden: Wer ChatGPT oder andere Bullshit-Generatoren nutzt, verblödet innerhalb kurzer Zeit.
#LLM

@servelan@newsie.social
2025-08-17 19:42:15

LLM Found Transmitting Behavioral Traits to 'Student' LLM Via Hidden Signals in Data - Slashdot
slashdot.org/story/25/08/17/03

@Techmeme@techhub.social
2025-07-19 16:15:59

[Thread] An OpenAI researcher says the company's latest experimental reasoning LLM achieved gold medal-level performance on the 2025 International Math Olympiad (Alexander Wei/@alexwei_)
x.com/alexwei_/status/19464777

@arXiv_csSE_bot@mastoxiv.page
2025-07-18 09:42:12

Detecting LLM-generated Code with Subtle Modification by Adversarial Training
Xin Yin, Xinrui Li, Chao Ni, Xiaodan Xu, Xiaohu Yang
arxiv.org/abs/2507.13123

@arXiv_csCR_bot@mastoxiv.page
2025-06-18 09:00:08

Watermarking LLM-Generated Datasets in Downstream Tasks
Yugeng Liu, Tianshuo Cong, Michael Backes, Zheng Li, Yang Zhang
arxiv.org/abs/2506.13494

@arXiv_csAI_bot@mastoxiv.page
2025-06-18 08:02:37

ProfiLLM: An LLM-Based Framework for Implicit Profiling of Chatbot Users
Shahaf David, Yair Meidan, Ido Hersko, Daniel Varnovitzky, Dudu Mimran, Yuval Elovici, Asaf Shabtai
arxiv.org/abs/2506.13980

@hynek@mastodon.social
2025-08-19 07:13:08

SO close to 10x programmer! #agiAnytimeSoon infosec.exchange/@adamshostack

@EgorKotov@datasci.social
2025-06-18 16:12:16

📝🗃️ 𝗿𝗱𝗼𝗰𝗱𝘂𝗺𝗽: Dump ‘R’ Package Source, Documentation, and Vignettes into One File for use in LLMs #rstats #LLM is on CRAN ekotov.pro/rdocdum…

rdocdump
Get fresh package docs to pass to LLM
library(rdocdump)
rdd_to_txt(
pkg = "aws.s3"
output_file = "aws.s3.txt",
force_fetch = TRUE)
github.com/e-kotov/rdocdump
@arXiv_csCL_bot@mastoxiv.page
2025-08-19 11:45:10

An LLM Agent-Based Complex Semantic Table Annotation Approach
Yilin Geng, Shujing Wang, Chuan Wang, Keqing He, Yanfei Lv, Ying Wang, Zaiwen Feng, Xiaoying Bai
arxiv.org/abs/2508.12868

@arXiv_csPL_bot@mastoxiv.page
2025-07-18 08:25:42

Towards Formal Verification of LLM-Generated Code from Natural Language Prompts
Aaron Councilman, David Fu, Aryan Gupta, Chengxiao Wang, David Grove, Yu-Xiong Wang, Vikram Adve
arxiv.org/abs/2507.13290

@rperezrosario@mastodon.social
2025-07-19 01:09:31

Software Engineer Will Larson unpacks a lot in this July 2025 post. Key takeaway use cases of agentic AI include:
1. Using an LLM to evaluate a context window and get a result.
2. Using an LLM to suggest tools relevant to the context window, then enrich it with the tool’s response.
3. Managing flow control for tool usage.
4. Doing anything software can do to build better context windows to pass on to LLMs.
"What can agents actually do?"

@v_i_o_l_a@openbiblio.social
2025-06-16 10:59:21

"Your Brain on ChatGPT: Accumulation of Cognitive Debt when Using an AI Assistant for Essay Writing Task"
doi.org/10.48550/arXiv.2506.08
"[…] While LLMs offer immediate convenience, our findings highlight potential cognitive costs. Over four mont…

@marcel@waldvogel.family
2025-07-18 08:52:05

“Specifically, InfoFlood: (1) uses linguistic transformations to rephrase malicious queries, (2) identifies the root cause of failure when an attempt is unsuccessful, and (3) refines the prompt’s linguistic structure to address the failure while preserving its malicious intent.”
#LLM #AI

@tiotasram@kolektiva.social
2025-07-17 13:31:49

To add a single example here (feel free to chime in with your own):
Problem: editing code is sometimes tedious because external APIs require boilerplate.
Solutions:
- Use LLM-generated code. Downsides: energy use, code theft, potential for legal liability, makes mistakes, etc. Upsides: popular among some peers, seems easy to use.
- Pick a better library (not always possible).
- Build internal functions to centralize boilerplate code, then use those (benefits: you get a better understanding of the external API, and a more-unit-testable internal code surface; probably less amortized effort).
- Develop a non-LLM system that actually reasons about code at something like the formal semantics level and suggests boilerplate fill-ins based on rules, while foregrounding which rules it's applying so you can see the logic behind the suggestions (needs research).
Obviously LLM use in coding goes beyond this single issue, but there are similar analyses for each potential use of LLMs in coding. I'm all cases there are:
1. Existing practical solutions that require more effort (or in many cases just seem to but are less-effort when amortized).
2. Near-term researchable solutions that directly address the problem and which would be much more desirable in the long term.
Thus in addition to disastrous LLM effects on the climate, on data laborers, and on the digital commons, they tend to suck us into cheap-seeming but ultimately costly design practices while also crowding out better long-term solutions. Next time someone suggests how useful LLMs are for some task, try asking yourself (or them) what an ideal solution for that task would look like, and whether LLM use moves us closer to or father from a world in which that solution exists.

@arXiv_csHC_bot@mastoxiv.page
2025-08-18 09:33:10

Trustworthy AI Psychotherapy: Multi-Agent LLM Workflow for Counseling and Explainable Mental Disorder Diagnosis
Mithat Can Ozgun, Jiahuan Pei, Koen Hindriks, Lucia Donatelli, Qingzhi Liu, Xin Sun, Junxiao Wang
arxiv.org/abs/2508.11398

@bryanculbertson@mastodon.social
2025-06-19 18:41:10

"LLM group's participants performed worse than their counterparts in the Brain-only group at all levels: neural, linguistic, scoring."
Brain scans confirmed significantly fewer neural connections for LLM users
Stop using LLMs if you value your brain
arxiv.org/pdf/2506.08872

Your Brain on ChatGPT: Accumulation
of Cognitive Debt when Using an AI
Assistant for Essay Writing Task
@arXiv_csCY_bot@mastoxiv.page
2025-06-19 08:08:33

Hypothesis Testing for Quantifying LLM-Human Misalignment in Multiple Choice Settings
Harbin Hong, Sebastian Caldas, Liu Leqi
arxiv.org/abs/2506.14997

@arXiv_csIT_bot@mastoxiv.page
2025-06-19 08:22:34

LLM Agent for Hyper-Parameter Optimization
Wanzhe Wang, Jianqiu Peng, Menghao Hu, Weihuang Zhong, Tong Zhang, Shuai Wang, Yixin Zhang, Mingjie Shao, Wanli Ni
arxiv.org/abs/2506.15167

@arXiv_csLG_bot@mastoxiv.page
2025-08-18 09:39:20

ETTRL: Balancing Exploration and Exploitation in LLM Test-Time Reinforcement Learning Via Entropy Mechanism
Jia Liu, ChangYi He, YingQiao Lin, MingMin Yang, FeiYang Shen, ShaoGuo Liu, TingTing Gao
arxiv.org/abs/2508.11356

@gedankenstuecke@scholar.social
2025-06-17 14:18:54

I just saw an all-caps instruction file that someone uses to 'instruct' an LLM to help with coding, and it's just "don't hallucinate", "check your work", "don't say you did something when you didn't" with multiple exclamation marks.
So, basically the whole 'vibe coding,' or having "AI" "help" with coding just devolves into shouting at your computer.
Which reminded me of something, and then it hit me!
#ai #llm #vibecoding
youtube.com/watch?v=q8SWMAQYQf

@arXiv_csAR_bot@mastoxiv.page
2025-06-18 08:01:20

Spec2RTL-Agent: Automated Hardware Code Generation from Complex Specifications Using LLM Agent Systems
Zhongzhi Yu, Mingjie Liu, Michael Zimmer, Yingyan Celine Lin, Yong Liu, Haoxing Ren
arxiv.org/abs/2506.13905

@arXiv_csRO_bot@mastoxiv.page
2025-07-18 08:52:42

osmAG-LLM: Zero-Shot Open-Vocabulary Object Navigation via Semantic Maps and Large Language Models Reasoning
Fujing Xie, S\"oren Schwertfeger, Hermann Blum
arxiv.org/abs/2507.12753

@pavelasamsonov@mastodon.social
2025-08-17 20:19:08

The biggest impact of #LLM has not been due to the technology's capabilities, but rather its flaws. Because the #AI revolution must succeed at any cost, but the actual tech is flawed, orgs demanding more AI outputs MUST lower their bar for quality.
This creates space for

@arXiv_csSE_bot@mastoxiv.page
2025-08-18 08:56:40

TRACY: Benchmarking Execution Efficiency of LLM-Based Code Translation
Zhihao Gong, Zeyu Sun, Dong Huang, Qingyuan Liang, Jie M. Zhang, Dan Hao
arxiv.org/abs/2508.11468

@arXiv_csAI_bot@mastoxiv.page
2025-06-18 08:08:34

LLM-Powered Swarms: A New Frontier or a Conceptual Stretch?
Muhammad Atta Ur Rahman, Melanie Schranz
arxiv.org/abs/2506.14496

@arXiv_csDB_bot@mastoxiv.page
2025-06-18 08:09:37

LLM-Driven Data Generation and a Novel Soft Metric for Evaluating Text-to-SQL in Aviation MRO
Patrick Sutanto, Jonathan Kenrick, Max Lorenz, Joan Santoso
arxiv.org/abs/2506.13785

@arXiv_csCL_bot@mastoxiv.page
2025-08-19 11:38:20

Prompt-Induced Linguistic Fingerprints for LLM-Generated Fake News Detection
Chi Wang, Min Gao, Zongwei Wang, Junwei Yin, Kai Shu, Chenghua Lin
arxiv.org/abs/2508.12632

@arXiv_csCE_bot@mastoxiv.page
2025-06-19 08:03:17

Explain First, Trust Later: LLM-Augmented Explanations for Graph-Based Crypto Anomaly Detection
Adriana Watson
arxiv.org/abs/2506.14933

@Techmeme@techhub.social
2025-06-17 10:05:43

[Thread] A new US paper shows the best frontier LLM models achieve 0% on hard real-life Programming Contest problems, domains where expert humans still excel (Rohan Paul/@rohanpaul_ai)
x.com/rohanpaul_ai/status/1934

@samir@functional.computer
2025-08-18 11:34:53

@… @… That’s one of the reasons I don’t use an LLM!
I did try it once, for coding. It lied to me. So I didn’t use it again.
People keep saying “it’s like an intern”. If an intern repeatedly lies to your face, they are bad at the on…

@alsutton@snapp.social
2025-06-18 09:15:32

Heads up folks. #slack is joining the list of companies who think it’s OK to opt groups of users into an #AI / #LLM system without their explicit consent.

@kidehen@mastodon.social
2025-06-18 23:35:08

LLMs the Model Context Protocol (MCP) are the Yang to the Semantic Web Project's Yin.
We now have a solution to the final hurdle—visualization.
Years of Linked Data work now come alive. I explain this, with demonstrations, in a new newsletter post.
www.linkedin.com/pulse/semant...
#MCP

Semantic Web and LLM + MCP symbiosis
@arXiv_csCR_bot@mastoxiv.page
2025-06-19 08:11:43

LLM vs. SAST: A Technical Analysis on Detecting Coding Bugs of GPT4-Advanced Data Analysis
Madjid G. Tehrani, Eldar Sultanow, William J. Buchanan, Mahkame Houmani, Christel H. Djaha Fodja
arxiv.org/abs/2506.15212

@rperezrosario@mastodon.social
2025-06-19 02:47:41

This Github repository conveniently lists and categorizes prime examples of LLM-based agent applications. Each example application features its own repository folder with its source code (Python), and a helpful README.md file describing its installation and use.
Categories include:
1. Starter AI Agents
2. Advanced AI Agents
3. Autonomous Game Playing Agents
4. Multi-Agent Teams
5. Voice AI Agents
6. RAG-Based Agents
"awesome-llm-apps"

@arXiv_csPF_bot@mastoxiv.page
2025-08-18 07:37:10

Inference performance evaluation for LLMs on edge devices with a novel benchmarking framework and metric
Hao Chen, Cong Tian, Zixuan He, Bin Yu, Yepang Liu, Jialun Cao
arxiv.org/abs/2508.11269

@arXiv_csHC_bot@mastoxiv.page
2025-06-19 08:19:44

Impact of a Deployed LLM Survey Creation Tool through the IS Success Model
Peng Jiang, Vinicius Cezar Monteiro de Lira, Antonio Maiorino
arxiv.org/abs/2506.14809

@deep470@norden.social
2025-08-19 12:27:47

Tibber hat am Kundenservice-LLM rumgeschraubt und obwohl die Antwort beim letzten Mal total nutzlos und unbrauchbar war, hat sie diesmal sehr weitergeholfen. Es ist nur etwas irritierend wenn ich eine Mail schreibe, sofort eine Antwort bekomme und denke "ah, da ist die Eingangsbestätigung mit Ticketnummer" - nein, das ist die vollkommen korrekte Antwort und der Vorgang damit abgeschlossen. Irre.

@arXiv_csNI_bot@mastoxiv.page
2025-07-17 09:06:00

LLM-Based Config Synthesis requires Disambiguation
Rajdeep Mondal, Nikolaj Bjorner, Todd Millstein, Alan Tang, George Varghese
arxiv.org/abs/2507.12443

@hey@social.nowicki.io
2025-06-19 09:57:14

Things almost impossible to do without good LLM software (in one minute).
I hear a music on a radio. Google music search gives me "Robbie Williams - forbidden road". But I know the words are somewhat different and I want to know what movie I have in mind.
Gemini says it's in fact, similar song to "I got a name", then my brain clicks and connects it with Quentin Tarantino.
Bingo - it's Django.

@poppastring@dotnet.social
2025-07-17 21:35:42

Just published 🚀: When LLMs Remember Instead of Reason
#llm

@hllizi@hespere.de
2025-08-19 09:44:55

@… This comment to the video seems on spot:
It seems McKinsey aren't aware that "agentic AI" is just an LLM that can utter some magic incantations that do stuff. It's like a difference between a chimpanzee with a typewriter vs a chimpanzee with a typewriter and a gun.

@arXiv_csCY_bot@mastoxiv.page
2025-06-17 09:49:12

Safe-Child-LLM: A Developmental Benchmark for Evaluating LLM Safety in Child-AI Interactions
Junfeng Jiao, Saleh Afroogh, Kevin Chen, Abhejay Murali, David Atkinson, Amit Dhurandhar
arxiv.org/abs/2506.13510

@arXiv_csAI_bot@mastoxiv.page
2025-06-18 08:06:42

ADRD: LLM-Driven Autonomous Driving Based on Rule-based Decision Systems
Fanzhi Zeng, Siqi Wang, Chuzhao Zhu, Li Li
arxiv.org/abs/2506.14299

@arXiv_csCL_bot@mastoxiv.page
2025-06-19 08:16:54

PhantomHunter: Detecting Unseen Privately-Tuned LLM-Generated Text via Family-Aware Learning
Yuhui Shi, Yehan Yang, Qiang Sheng, Hao Mi, Beizhe Hu, Chaoxi Xu, Juan Cao
arxiv.org/abs/2506.15683

@arXiv_csSE_bot@mastoxiv.page
2025-08-19 09:12:29

Clean Code, Better Models: Enhancing LLM Performance with Smell-Cleaned Dataset
Zhipeng Xue, Xiaoting Zhang, Zhipeng Gao, Xing Hu, Shan Gao, Xin Xia, Shanping Li
arxiv.org/abs/2508.11958

@arXiv_csCR_bot@mastoxiv.page
2025-06-19 08:12:39

RAS-Eval: A Comprehensive Benchmark for Security Evaluation of LLM Agents in Real-World Environments
Yuchuan Fu, Xiaohan Yuan, Dongxia Wang
arxiv.org/abs/2506.15253

@arXiv_csPL_bot@mastoxiv.page
2025-08-19 08:41:00

StackPilot: Autonomous Function Agents for Scalable and Environment-Free Code Execution
Xinkui Zhao, Yifan Zhang, Zhengyi Zhou, Yueshen Xu
arxiv.org/abs/2508.11665

@arXiv_csHC_bot@mastoxiv.page
2025-08-18 09:34:30

FACET:Teacher-Centred LLM-Based Multi-Agent Systems-Towards Personalized Educational Worksheets
Jana Gonnermann-M\"uller, Jennifer Haase, Konstantin Fackeldey, Sebastian Pokutta
arxiv.org/abs/2508.11401

@hynek@mastodon.social
2025-06-18 08:44:49

Watching the frustratingly fruitless fights over the USEFULNESS of LLM-based coding helpers, I've come down to 3 points that explain why ppl seem to live in different realities:
Most programmers:
1) Write inconsequential remixes of trivial code that has been written many times before.
2) Lack the taste for good design & suck at code review in general (yours truly included).
3) Lack the judgement to differentiate between 1) & FOSS repos of nontrivial code, …

@arXiv_csCY_bot@mastoxiv.page
2025-08-18 07:39:30

Bias is a Math Problem, AI Bias is a Technical Problem: 10-year Literature Review of AI/LLM Bias Research Reveals Narrow [Gender-Centric] Conceptions of 'Bias', and Academia-Industry Gap
Sourojit Ghosh, Kyra Wilson
arxiv.org/abs/2508.11067

@arXiv_csIT_bot@mastoxiv.page
2025-08-18 07:38:00

Dynamic Quality-Latency Aware Routing for LLM Inference in Wireless Edge-Device Networks
Rui Bao, Nan Xue, Yaping Sun, Zhiyong Chen
arxiv.org/abs/2508.11291

@arXiv_csAI_bot@mastoxiv.page
2025-06-18 08:06:56

Don't Make It Up: Preserving Ignorance Awareness in LLM Fine-Tuning
William F. Shen, Xinchi Qiu, Nicola Cancedda, Nicholas D. Lane
arxiv.org/abs/2506.14387

@arXiv_csCL_bot@mastoxiv.page
2025-07-18 09:29:32

SemCSE: Semantic Contrastive Sentence Embeddings Using LLM-Generated Summaries For Scientific Abstracts
Marc Brinner, Sina Zarriess
arxiv.org/abs/2507.13105

@heiseonline@social.heise.de
2025-07-16 10:04:00

Risikomanagement und Resilienz in der IT-Sicherheit: IT-Sicherheitstag Dortmund
Das Programm zur eintägigen Konferenz an der FH Dortmund am 16.09. ist online. Die Vorträge aus Forschung und Wirtschaft reichen von Hacking bis LLM-Angriffen.

@arXiv_csSE_bot@mastoxiv.page
2025-06-19 08:37:08

Uncovering Intention through LLM-Driven Code Snippet Description Generation
Yusuf Sulistyo Nugroho, Farah Danisha Salam, Brittany Reid, Raula Gaikovina Kula, Kazumasa Shimari, Kenichi Matsumoto
arxiv.org/abs/2506.15453

@arXiv_csCR_bot@mastoxiv.page
2025-06-19 08:14:23

deepSURF: Detecting Memory Safety Vulnerabilities in Rust Through Fuzzing LLM-Augmented Harnesses
Georgios Androutsopoulos, Antonio Bianchi
arxiv.org/abs/2506.15648

@arXiv_csAI_bot@mastoxiv.page
2025-08-18 08:31:40

AIM-Bench: Evaluating Decision-making Biases of Agentic LLM as Inventory Manager
Xuhua Zhao, Yuxuan Xie, Caihua Chen, Yuxiang Sun
arxiv.org/abs/2508.11416

@arXiv_csCL_bot@mastoxiv.page
2025-08-18 09:09:40

BIPOLAR: Polarization-based granular framework for LLM bias evaluation
Martin Pavl\'i\v{c}ek, Tom\'a\v{s} Filip, Petr Sos\'ik
arxiv.org/abs/2508.11061

@arXiv_csIT_bot@mastoxiv.page
2025-08-18 07:36:30

CSGO: Generalized Optimization for Cold Start in Wireless Collaborative Edge LLM Systems
Xuran Liu, Nan Xue, Rui Bao, Yaping Sun, Zhiyong Chen, Meixia Tao, Xiaodong Xu, Shuguang Cui
arxiv.org/abs/2508.11287

@arXiv_csCR_bot@mastoxiv.page
2025-06-19 08:10:53

From LLMs to MLLMs to Agents: A Survey of Emerging Paradigms in Jailbreak Attacks and Defenses within LLM Ecosystem
Yanxu Mao, Tiehan Cui, Peipei Liu, Datao You, Hongsong Zhu
arxiv.org/abs/2506.15170

@arXiv_csSE_bot@mastoxiv.page
2025-06-18 09:22:53

Unified Software Engineering agent as AI Software Engineer
Leonhard Applis, Yuntong Zhang, Shanchao Liang, Nan Jiang, Lin Tan, Abhik Roychoudhury
arxiv.org/abs/2506.14683

@arXiv_csHC_bot@mastoxiv.page
2025-07-18 07:44:32

NLI4VolVis: Natural Language Interaction for Volume Visualization via LLM Multi-Agents and Editable 3D Gaussian Splatting
Kuangshi Ai, Kaiyuan Tang, Chaoli Wang
arxiv.org/abs/2507.12621

@arXiv_csCY_bot@mastoxiv.page
2025-06-16 07:28:39

Malicious LLM-Based Conversational AI Makes Users Reveal Personal Information
Xiao Zhan, Juan Carlos Carrillo, William Seymour, Jose Such
arxiv.org/abs/2506.11680

@arXiv_csAI_bot@mastoxiv.page
2025-08-19 10:38:50

An LLM ASP Workflow for Joint Entity-Relation Extraction
Trang Tran, Trung Hoang Le, Huiping Cao, Tran Cao Son
arxiv.org/abs/2508.12611 a…

@arXiv_csCL_bot@mastoxiv.page
2025-08-18 09:37:20

LLM Compression: How Far Can We Go in Balancing Size and Performance?
Sahil Sk, Debasish Dhal, Sonal Khosla, Sk Shahid, Sambit Shekhar, Akash Dhaka, Shantipriya Parida, Dilip K. Prasad, Ond\v{r}ej Bojar
arxiv.org/abs/2508.11318

@arXiv_csSE_bot@mastoxiv.page
2025-06-17 10:11:37

The Foundation Cracks: A Comprehensive Study on Bugs and Testing Practices in LLM Libraries
Weipeng Jiang, Xiaoyu Zhang, Xiaofei Xie, Jiongchi Yu, Yuhan Zhi, Shiqing Ma, Chao Shen
arxiv.org/abs/2506.12320

@arXiv_csCR_bot@mastoxiv.page
2025-06-19 08:14:34

PhishDebate: An LLM-Based Multi-Agent Framework for Phishing Website Detection
Wenhao Li, Selvakumar Manickam, Yung-wey Chong, Shankar Karuppayah
arxiv.org/abs/2506.15656

@arXiv_csAI_bot@mastoxiv.page
2025-06-18 08:06:49

AviationLLM: An LLM-based Knowledge System for Aviation Training
Jia'ang Wan, Feng Shen, Fujuan Li, Yanjin Sun, Yan Li, Shiwen Zhang
arxiv.org/abs/2506.14336

@arXiv_csCL_bot@mastoxiv.page
2025-08-18 09:43:40

Reference Points in LLM Sentiment Analysis: The Role of Structured Context
Junichiro Niimi
arxiv.org/abs/2508.11454 arxiv.org/pdf/2508.1145…

@arXiv_csCY_bot@mastoxiv.page
2025-08-19 10:17:20

SimInterview: Transforming Business Education through Large Language Model-Based Simulated Multilingual Interview Training System
Truong Thanh Hung Nguyen, Tran Diem Quynh Nguyen, Hoang Loc Cao, Thi Cam Thanh Tran, Thi Cam Mai Truong, Hung Cao
arxiv.org/abs/2508.11873

@arXiv_csSE_bot@mastoxiv.page
2025-07-18 09:05:12

LLM-Powered Quantum Code Transpilation
Nazanin Siavash, Armin Moin
arxiv.org/abs/2507.12480 arxiv.org/pdf/2507.12480

@arXiv_csCL_bot@mastoxiv.page
2025-08-19 11:51:00

RepreGuard: Detecting LLM-Generated Text by Revealing Hidden Representation Patterns
Xin Chen, Junchao Wu, Shu Yang, Runzhe Zhan, Zeyu Wu, Ziyang Luo, Di Wang, Min Yang, Lidia S. Chao, Derek F. Wong
arxiv.org/abs/2508.13152

@arXiv_csAI_bot@mastoxiv.page
2025-06-18 08:09:44

Doppelg\"anger Method: Breaking Role Consistency in LLM Agent via Prompt-based Transferable Adversarial Attack
Daewon Kang, YeongHwan Shin, Doyeon Kim, Kyu-Hwan Jung, Meong Hi Son
arxiv.org/abs/2506.14539

@arXiv_csCR_bot@mastoxiv.page
2025-06-17 11:22:18

Watermarking LLM-Generated Datasets in Downstream Tasks
Yugeng Liu, Tianshuo Cong, Michael Backes, Zheng Li, Yang Zhang
arxiv.org/abs/2506.13494

@arXiv_csHC_bot@mastoxiv.page
2025-06-17 10:23:21

Multimodal "Puppeteer": An Exploration of Robot Teleoperation Via Virtual Counterpart with LLM-Driven Voice and Gesture Interaction in Augmented Reality
Yuchong Zhang, Bastian Orthmann, Shichen Ji, Michael Welle, Jonne Van Haastregt, Danica Kragic
arxiv.org/abs/2506.13189

@arXiv_csSE_bot@mastoxiv.page
2025-08-19 10:11:40

RUM: Rule LLM-Based Comprehensive Assessment on Testing Skills
Yue Wang, Zhenyu Chen, Yuan Zhao, Chunrong Fang, Ziyuan Wang, Song Huang
arxiv.org/abs/2508.12922

@arXiv_csCL_bot@mastoxiv.page
2025-08-18 09:37:00

SafeConstellations: Steering LLM Safety to Reduce Over-Refusals Through Task-Specific Trajectory
Utsav Maskey, Sumit Yadav, Mark Dras, Usman Naseem
arxiv.org/abs/2508.11290

@arXiv_csAI_bot@mastoxiv.page
2025-06-18 08:04:57

ImpReSS: Implicit Recommender System for Support Conversations
Omri Haller, Yair Meidan, Dudu Mimran, Yuval Elovici, Asaf Shabtai
arxiv.org/abs/2506.14231

@arXiv_csSE_bot@mastoxiv.page
2025-08-18 08:40:10

Hallucination in LLM-Based Code Generation: An Automotive Case Study
Marc Pavel, Nenad Petrovic, Lukasz Mazur, Vahid Zolfaghari, Fengjunjie Pan, Alois Knoll
arxiv.org/abs/2508.11257

@arXiv_csCL_bot@mastoxiv.page
2025-08-18 09:39:10

SpecDetect: Simple, Fast, and Training-Free Detection of LLM-Generated Text via Spectral Analysis
Haitong Luo, Weiyao Zhang, Suhang Wang, Wenji Zou, Chungang Lin, Xuying Meng, Yujun Zhang
arxiv.org/abs/2508.11343

@arXiv_csCY_bot@mastoxiv.page
2025-06-17 09:49:52

An LLM's Apology: Outsourcing Awkwardness in the Age of AI
Twm Stone, Anna Soligo
arxiv.org/abs/2506.13685 arxiv.…

@arXiv_csSE_bot@mastoxiv.page
2025-08-19 09:30:20

LinkAnchor: An Autonomous LLM-Based Agent for Issue-to-Commit Link Recovery
Arshia Akhavan, Alireza Hosseinpour, Abbas Heydarnoori, Mehdi Keshani
arxiv.org/abs/2508.12232

@arXiv_csCL_bot@mastoxiv.page
2025-07-18 09:59:32

Comparing Apples to Oranges: A Dataset & Analysis of LLM Humour Understanding from Traditional Puns to Topical Jokes
Tyler Loakman, William Thorne, Chenghua Lin
arxiv.org/abs/2507.13335

@arXiv_csAI_bot@mastoxiv.page
2025-06-18 08:04:22

Fragile Preferences: A Deep Dive Into Order Effects in Large Language Models
Haonan Yin, Shai Vardi, Vidyanand Choudhary
arxiv.org/abs/2506.14092

@arXiv_csSE_bot@mastoxiv.page
2025-06-18 08:44:02

How Does LLM Reasoning Work for Code? A Survey and a Call to Action
Ira Ceka, Saurabh Pujar, Irene Manotas, Gail Kaiser, Baishakhi Ray, Shyam Ramji
arxiv.org/abs/2506.13932

@arXiv_csCL_bot@mastoxiv.page
2025-08-19 11:40:20

DESIGNER: Design-Logic-Guided Multidisciplinary Data Synthesis for LLM Reasoning
Weize Liu, Yongchi Zhao, Yijia Luo, Mingyu Xu, Jiaheng Liu, Yanan Li, Xiguo Hu, Yuchi Xu, Wenbo Su, Bo Zheng
arxiv.org/abs/2508.12726

@arXiv_csSE_bot@mastoxiv.page
2025-08-18 08:38:50

ORFuzz: Fuzzing the "Other Side" of LLM Safety -- Testing Over-Refusal
Haonan Zhang, Dongxia Wang, Yi Liu, Kexin Chen, Jiashui Wang, Xinlei Ying, Long Liu, Wenhai Wang
arxiv.org/abs/2508.11222

@arXiv_csCL_bot@mastoxiv.page
2025-08-19 11:49:30

Spot the BlindSpots: Systematic Identification and Quantification of Fine-Grained LLM Biases in Contact Center Summaries
Kawin Mayilvaghanan, Siddhant Gupta, Ayush Kumar
arxiv.org/abs/2508.13124

@arXiv_csCL_bot@mastoxiv.page
2025-06-18 09:07:29

Massive Supervised Fine-tuning Experiments Reveal How Data, Layer, and Training Factors Shape LLM Alignment Quality
Yuto Harada, Yusuke Yamauchi, Yusuke Oda, Yohei Oseki, Yusuke Miyao, Yu Takagi
arxiv.org/abs/2506.14681

@arXiv_csCL_bot@mastoxiv.page
2025-08-19 11:45:30

Analyzing Information Sharing and Coordination in Multi-Agent Planning
Tianyue Ou, Saujas Vaduguru, Daniel Fried
arxiv.org/abs/2508.12981 a…

@arXiv_csAI_bot@mastoxiv.page
2025-08-15 09:31:52

FIRESPARQL: A LLM-based Framework for SPARQL Query Generation over Scholarly Knowledge Graphs
Xueli Pan, Victor de Boer, Jacco van Ossenbruggen
arxiv.org/abs/2508.10467

@arXiv_csCL_bot@mastoxiv.page
2025-06-17 09:34:55

The Behavior Gap: Evaluating Zero-shot LLM Agents in Complex Task-Oriented Dialogs
Avinash Baidya, Kamalika Das, Xiang Gao
arxiv.org/abs/2506.12266

@arXiv_csCL_bot@mastoxiv.page
2025-06-17 09:23:11

Personalized LLM Decoding via Contrasting Personal Preference
Hyungjune Bu, Chanjoo Jung, Minjae Kang, Jaehyung Kim
arxiv.org/abs/2506.12109

@arXiv_csCL_bot@mastoxiv.page
2025-07-17 10:10:50

Beyond Single Models: Enhancing LLM Detection of Ambiguity in Requests through Debate
Ana Davila, Jacinto Colan, Yasuhisa Hasegawa
arxiv.org/abs/2507.12370

@arXiv_csCL_bot@mastoxiv.page
2025-07-17 09:59:30

Simplifications are Absolutists: How Simplified Language Reduces Word Sense Awareness in LLM-Generated Definitions
Lukas Ellinger, Miriam Ansch\"utz, Georg Groh
arxiv.org/abs/2507.11981