Tootfinder

Opt-in global Mastodon full text search. Join the index!

@arXiv_csCR_bot@mastoxiv.page
2025-07-31 09:14:21

Invisible Injections: Exploiting Vision-Language Models Through Steganographic Prompt Embedding
Chetan Pathade
arxiv.org/abs/2507.22304 arx…

@arXiv_csCL_bot@mastoxiv.page
2025-07-29 11:46:21

Soft Injection of Task Embeddings Outperforms Prompt-Based In-Context Learning
Jungwon Park, Wonjong Rhee
arxiv.org/abs/2507.20906 arxiv.or…

@gedankenstuecke@scholar.social
2025-05-25 03:16:56

I'm not surprised that Gitlab decided to run off a cliff to follow GitHub:
«AI coding bot allows prompt injection with a pull request»
Everyday I'm more grateful for @… and @…!
pivot-to-ai.com/2025/05/24/ai-

@publicvoit@graz.social
2025-07-09 07:31:58

"Zero-Click Prompt Injection":
calypsoai.com/insights/prompt-
So instead of trying to trick an employee via phishing

@gedankenstuecke@scholar.social
2025-05-25 03:16:56

I'm not surprised that Gitlab decided to run off a cliff to follow GitHub:
«AI coding bot allows prompt injection with a pull request»
Everyday I'm more grateful for @… and @…!
pivot-to-ai.com/2025/05/24/ai-

@arXiv_csCR_bot@mastoxiv.page
2025-05-30 07:16:44

Securing AI Agents with Information-Flow Control
Manuel Costa, Boris K\"opf, Aashish Kolluri, Andrew Paverd, Mark Russinovich, Ahmed Salem, Shruti Tople, Lukas Wutschitz, Santiago Zanella-B\'eguelin
arxiv.org/abs/2505.23643

@arXiv_csCR_bot@mastoxiv.page
2025-06-25 07:46:20

Enhancing Security in LLM Applications: A Performance Evaluation of Early Detection Systems
Valerii Gakh, Hayretdin Bahsi
arxiv.org/abs/2506.19109

@arXiv_csCR_bot@mastoxiv.page
2025-07-21 08:38:30

TopicAttack: An Indirect Prompt Injection Attack via Topic Transition
Yulin Chen, Haoran Li, Yuexin Li, Yue Liu, Yangqiu Song, Bryan Hooi
arxiv.org/abs/2507.13686

@arXiv_csCR_bot@mastoxiv.page
2025-06-09 07:53:32

To Protect the LLM Agent Against the Prompt Injection Attack with Polymorphic Prompt
Zhilong Wang, Neha Nagaraja, Lan Zhang, Hayretdin Bahsi, Pawan Patil, Peng Liu
arxiv.org/abs/2506.05739

@arXiv_csCR_bot@mastoxiv.page
2025-07-18 08:57:42

Prompt Injection 2.0: Hybrid AI Threats
Jeremy McHugh, Kristina \v{S}ekrst, Jon Cefalu
arxiv.org/abs/2507.13169 arxiv…

@arXiv_csCR_bot@mastoxiv.page
2025-07-18 08:41:32

MAD-Spear: A Conformity-Driven Prompt Injection Attack on Multi-Agent Debate Systems
Yu Cui, Hongyang Du
arxiv.org/abs/2507.13038

@arXiv_csCR_bot@mastoxiv.page
2025-07-11 09:11:11

May I have your Attention? Breaking Fine-Tuning based Prompt Injection Defenses using Architecture-Aware Attacks
Nishit V. Pandya, Andrey Labunets, Sicun Gao, Earlence Fernandes
arxiv.org/abs/2507.07417

@arXiv_csCR_bot@mastoxiv.page
2025-06-12 07:32:21

LLMail-Inject: A Dataset from a Realistic Adaptive Prompt Injection Challenge
Sahar Abdelnabi, Aideen Fay, Ahmed Salem, Egor Zverev, Kai-Chieh Liao, Chi-Huang Liu, Chun-Chih Kuo, Jannis Weigend, Danyael Manlangit, Alex Apostolov, Haris Umair, Jo\~ao Donato, Masayuki Kawakita, Athar Mahboob, Tran Huu Bach, Tsun-Han Chiang, Myeongjin Cho, Hajin Choi, Byeonghyeon Kim, Hyeonjin Lee, Benjamin Pannell, Conor McCauley, Mark Russinovich, Andrew Paverd, Giovanni Cherubin

@arXiv_csCR_bot@mastoxiv.page
2025-07-11 09:39:51

Defending Against Prompt Injection With a Few DefensiveTokens
Sizhe Chen, Yizhu Wang, Nicholas Carlini, Chawin Sitawarin, David Wagner
arxiv.org/abs/2507.07974

@arXiv_csCR_bot@mastoxiv.page
2025-07-04 09:56:21

Meta SecAlign: A Secure Foundation LLM Against Prompt Injection Attacks
Sizhe Chen, Arman Zharmagambetov, David Wagner, Chuan Guo
arxiv.org/abs/2507.02735

@arXiv_csCR_bot@mastoxiv.page
2025-06-09 07:37:32

Sentinel: SOTA model to protect against prompt injections
Dror Ivry, Oran Nahum
arxiv.org/abs/2506.05446 arxiv.org/pd…

@arXiv_csCR_bot@mastoxiv.page
2025-07-23 13:05:34

Replaced article(s) found for cs.CR. arxiv.org/list/cs.CR/new
[1/1]:
- Defense Against Prompt Injection Attack by Leveraging Attack Techniques
Yulin Chen, Haoran Li, Zihao Zheng, Yangqiu Song, Dekai Wu, Bryan Hooi

@arXiv_csCR_bot@mastoxiv.page
2025-07-09 09:23:32

How Not to Detect Prompt Injections with an LLM
Sarthak Choudhary, Divyam Anshumaan, Nils Palumbo, Somesh Jha
arxiv.org/abs/2507.05630

@arXiv_csCR_bot@mastoxiv.page
2025-07-21 11:52:57

Replaced article(s) found for cs.CR. arxiv.org/list/cs.CR/new
[1/1]:
- Defense Against Prompt Injection Attack by Leveraging Attack Techniques
Yulin Chen, Haoran Li, Zihao Zheng, Yangqiu Song, Dekai Wu, Bryan Hooi

@arXiv_csCR_bot@mastoxiv.page
2025-06-06 09:37:16

This arxiv.org/abs/2505.05849 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCR_…

@arXiv_csCR_bot@mastoxiv.page
2025-06-13 07:24:20

Evaluation empirique de la s\'ecurisation et de l'alignement de ChatGPT et Gemini: analyse comparative des vuln\'erabilit\'es par exp\'erimentations de jailbreaks
Rafa\"el Nouailles (GdR)
arxiv.org/abs/2506.10029

@arXiv_csCR_bot@mastoxiv.page
2025-07-10 09:49:31

The Dark Side of LLMs Agent-based Attacks for Complete Computer Takeover
Matteo Lupinacci, Francesco Aurelio Pironti, Francesco Blefari, Francesco Romeo, Luigi Arena, Angelo Furfaro
arxiv.org/abs/2507.06850

@arXiv_csCR_bot@mastoxiv.page
2025-06-02 07:17:42

LLM Agents Should Employ Security Principles
Kaiyuan Zhang, Zian Su, Pin-Yu Chen, Elisa Bertino, Xiangyu Zhang, Ninghui Li
arxiv.org/abs/2505.24019

@arXiv_csCR_bot@mastoxiv.page
2025-06-03 17:52:02

This arxiv.org/abs/2505.18889 has been replaced.
initial toot: mastoxiv.page/@arXiv_csCR_…