❓Wie können Musikwissenschaftler:innen mit einem Knowledge Graph interagieren und relevante Informationen abrufen, ohne SPARQL-Expert:innen zu sein?
🤩 Wir freuen uns diese Frage mit euch im Rahmen einer Data Challenge zu diskutieren!✨Und in Kooperation mit dem @…
✨ From Notes to Nodes – Develop an AI-driven explor…
Incremental (k, z)-Clustering on Graphs
Emilio Cruciani, Sebastian Forster, Antonis Skarlatos
https://arxiv.org/abs/2602.08542 https://arxiv.org/pdf/2602.08542 https://arxiv.org/html/2602.08542
arXiv:2602.08542v1 Announce Type: new
Abstract: Given a weighted undirected graph, a number of clusters $k$, and an exponent $z$, the goal in the $(k, z)$-clustering problem on graphs is to select $k$ vertices as centers that minimize the sum of the distances raised to the power $z$ of each vertex to its closest center. In the dynamic setting, the graph is subject to adversarial edge updates, and the goal is to maintain explicitly an exact $(k, z)$-clustering solution in the induced shortest-path metric.
While efficient dynamic $k$-center approximation algorithms on graphs exist [Cruciani et al. SODA 2024], to the best of our knowledge, no prior work provides similar results for the dynamic $(k,z)$-clustering problem. As the main result of this paper, we develop a randomized incremental $(k, z)$-clustering algorithm that maintains with high probability a constant-factor approximation in a graph undergoing edge insertions with a total update time of $\tilde O(k m^{1 o(1)} k^{1 \frac{1}{\lambda}} m)$, where $\lambda \geq 1$ is an arbitrary fixed constant. Our incremental algorithm consists of two stages. In the first stage, we maintain a constant-factor bicriteria approximate solution of size $\tilde{O}(k)$ with a total update time of $m^{1 o(1)}$ over all adversarial edge insertions. This first stage is an intricate adaptation of the bicriteria approximation algorithm by Mettu and Plaxton [Machine Learning 2004] to incremental graphs. One of our key technical results is that the radii in their algorithm can be assumed to be non-decreasing while the approximation ratio remains constant, a property that may be of independent interest.
In the second stage, we maintain a constant-factor approximate $(k,z)$-clustering solution on a dynamic weighted instance induced by the bicriteria approximate solution. For this subproblem, we employ a dynamic spanner algorithm together with a static $(k,z)$-clustering algorithm.
toXiv_bot_toot
GraphWalker: Agentic Knowledge Graph Question Answering via Synthetic Trajectory Curriculum
Shuwen Xu, Yao Xu, Jiaxiang Liu, Chenhao Yuan, Wenshuo Peng, Jun Zhao, Kang Liu
https://arxiv.org/abs/2603.28533 https://arxiv.org/pdf/2603.28533 https://arxiv.org/html/2603.28533
arXiv:2603.28533v1 Announce Type: new
Abstract: Agentic knowledge graph question answering (KGQA) requires an agent to iteratively interact with knowledge graphs (KGs), posing challenges in both training data scarcity and reasoning generalization. Specifically, existing approaches often restrict agent exploration: prompting-based methods lack autonomous navigation training, while current training pipelines usually confine reasoning to predefined trajectories. To this end, this paper proposes \textit{GraphWalker}, a novel agentic KGQA framework that addresses these challenges through \textit{Automated Trajectory Synthesis} and \textit{Stage-wise Fine-tuning}. GraphWalker adopts a two-stage SFT training paradigm: First, the agent is trained on structurally diverse trajectories synthesized from constrained random-walk paths, establishing a broad exploration prior over the KG; Second, the agent is further fine-tuned on a small set of expert trajectories to develop reflection and error recovery capabilities. Extensive experiments demonstrate that our stage-wise SFT paradigm unlocks a higher performance ceiling for a lightweight reinforcement learning (RL) stage, enabling GraphWalker to achieve state-of-the-art performance on CWQ and WebQSP. Additional results on GrailQA and our constructed GraphWalkerBench confirm that GraphWalker enhances generalization to out-of-distribution reasoning paths. The code is publicly available at https://github.com/XuShuwenn/GraphWalker
toXiv_bot_toot
The Leader's Handbook Study Series
https://www.in2in.org/shop/p/the-leadership-handbook-study-series
In his forward to the Leader’s Handbook Ackoff offers a format to study the book which we will follow in these sessions:
“I suggest a small group …
Crosslisted article(s) found for cs.CL. https://arxiv.org/list/cs.CL/new
[1/2]:
- Bridge-RAG: An Abstract Bridge Tree Based Retrieval Augmented Generation Algorithm With Cuckoo Fi...
Li, Liu, Zong, Tao, Dai, Ren, Liu, Jiang, Yang
https://arxiv.org/abs/2603.26668 https://mastoxiv.page/@arXiv_csIR_bot/116322781593134028
- SRAG: RAG with Structured Data Improves Vector Retrieval
Shalin Shah, Srikanth Ryali, Ramasubbu Venkatesh
https://arxiv.org/abs/2603.26670 https://mastoxiv.page/@arXiv_csIR_bot/116322784870180864
- LITTA: Late-Interaction and Test-Time Alignment for Visually-Grounded Multimodal Retrieval
Seonok Kim
https://arxiv.org/abs/2603.26683 https://mastoxiv.page/@arXiv_csIR_bot/116322841916406330
- Agentic AI for Human Resources: LLM-Driven Candidate Assessment
Yuksel, Anees, Elneima, Hewavitharana, Al-Badrashiny, Sawaf
https://arxiv.org/abs/2603.26710 https://mastoxiv.page/@arXiv_csIR_bot/116322937601675587
- SEAR: Schema-Based Evaluation and Routing for LLM Gateways
Zecheng Zhang, Han Zheng, Yue Xu
https://arxiv.org/abs/2603.26728 https://mastoxiv.page/@arXiv_csDB_bot/116322627580095245
- SleepVLM: Explainable and Rule-Grounded Sleep Staging via a Vision-Language Model
Guifeng Deng, Pan Wang, Jiquan Wang, Shuying Rao, Junyi Xie, Wanjun Guo, Tao Li, Haiteng Jiang
https://arxiv.org/abs/2603.26738 https://mastoxiv.page/@arXiv_csCV_bot/116322739676378309
- Aesthetic Assessment of Chinese Handwritings Based on Vision Language Models
Chen Zheng, Yuxuan Lai, Haoyang Lu, Wentao Ma, Jitao Yang, Jian Wang
https://arxiv.org/abs/2603.26768 https://mastoxiv.page/@arXiv_csCV_bot/116323078149576728
- Learning to Select Visual In-Context Demonstrations
Eugene Lee, Yu-Chi Lin, Jiajie Diao
https://arxiv.org/abs/2603.26775 https://mastoxiv.page/@arXiv_csLG_bot/116322648878995047
- CRISP: Characterizing Relative Impact of Scholarly Publications
Hannah Collison, Benjamin Van Durme, Daniel Khashabi
https://arxiv.org/abs/2603.26791 https://mastoxiv.page/@arXiv_csDL_bot/116322621679820997
- GroupRAG: Cognitively Inspired Group-Aware Retrieval and Reasoning via Knowledge-Driven Problem S...
Xinyi Duan, Yuanrong Tang, Jiangtao Gong
https://arxiv.org/abs/2603.26807 https://mastoxiv.page/@arXiv_csIR_bot/116322959557860848
- In your own words: computationally identifying interpretable themes in free-text survey data
Jenny S Wang, Aliya Saperstein, Emma Pierson
https://arxiv.org/abs/2603.26930 https://mastoxiv.page/@arXiv_csCY_bot/116322780637316287
- Multilingual Stutter Event Detection for English, German, and Mandarin Speech
Felix Haas, Sebastian P. Bayerl
https://arxiv.org/abs/2603.26939 https://mastoxiv.page/@arXiv_csSD_bot/116322704289189130
- FormalProofBench: Can Models Write Graduate Level Math Proofs That Are Formally Verified?
Ravi, Ying, Nesterov, Krishnan, Uskuplu, Xia, Aswedige, Nashold
https://arxiv.org/abs/2603.26996 https://mastoxiv.page/@arXiv_csAI_bot/116322625941412681
- PHONOS: PHOnetic Neutralization for Online Streaming Applications
Waris Quamer, Mu-Ruei Tseng, Ghady Nasrallah, Ricardo Gutierrez-Osuna
https://arxiv.org/abs/2603.27001 https://mastoxiv.page/@arXiv_eessAS_bot/116322763598554193
- ChartNet: A Million-Scale, High-Quality Multimodal Dataset for Robust Chart Understanding
Jovana Kondic, et al.
https://arxiv.org/abs/2603.27064 https://mastoxiv.page/@arXiv_csCV_bot/116323214468792735
- daVinci-LLM:Towards the Science of Pretraining
Qin, Liu, Mi, Xie, Huang, Si, Lu, Feng, Wu, Liu, Luo, Hou, Guo, Qiao, Liu
https://arxiv.org/abs/2603.27164 https://mastoxiv.page/@arXiv_csAI_bot/116322653467105951
- LightMover: Generative Light Movement with Color and Intensity Controls
Zhou, Wang, Kim, Shu, Yu, Hold-Geoffroy, Chaturvedi, Wu, Lin, Cohen
https://arxiv.org/abs/2603.27209 https://mastoxiv.page/@arXiv_csCV_bot/116323263295656104
- Self-evolving AI agents for protein discovery and directed evolution
Tan, Zhang, Li, Yu, Zhong, Zhou, Dong, Hong
https://arxiv.org/abs/2603.27303 https://mastoxiv.page/@arXiv_csAI_bot/116322838641595927
- Inference-Time Structural Reasoning for Compositional Vision-Language Understanding
Amartya Bhattacharya
https://arxiv.org/abs/2603.27349 https://mastoxiv.page/@arXiv_csCV_bot/116323280006044500
- LLM Readiness Harness: Evaluation, Observability, and CI Gates for LLM/RAG Applications
Alexandre Cristov\~ao Maiorano
https://arxiv.org/abs/2603.27355 https://mastoxiv.page/@arXiv_csAI_bot/116322987708962414
- Heterogeneous Debate Engine: Identity-Grounded Cognitive Architecture for Resilient LLM-Based Eth...
Jakub Mas{\l}owski, Jaros{\l}aw A. Chudziak
https://arxiv.org/abs/2603.27404 https://mastoxiv.page/@arXiv_csAI_bot/116322999177460352
toXiv_bot_toot
Replaced article(s) found for cs.CL. https://arxiv.org/list/cs.CL/new
[1/5]:
- Beyond In-Distribution Success: Scaling Curves of CoT Granularity for Language Model Generalization
Ru Wang, Wei Huang, Selena Song, Haoyu Zhang, Qian Niu, Yusuke Iwasawa, Yutaka Matsuo, Jiaxian Guo
https://arxiv.org/abs/2502.18273 https://mastoxiv.page/@arXiv_csCL_bot/114069031700102129
- Benchmarking NLP-supported Language Sample Analysis for Swiss Children's Speech
Anja Ryser, Yingqiang Gao, Sarah Ebling
https://arxiv.org/abs/2504.00780 https://mastoxiv.page/@arXiv_csCL_bot/114267149909002069
- Cultural Biases of Large Language Models and Humans in Historical Interpretation
Fabio Celli, Georgios Spathulas
https://arxiv.org/abs/2504.02572 https://mastoxiv.page/@arXiv_csCL_bot/114278467094094490
- BRIDGE: Benchmarking Large Language Models for Understanding Real-world Clinical Practice Text
Jiageng Wu, et al.
https://arxiv.org/abs/2504.19467 https://mastoxiv.page/@arXiv_csCL_bot/114420036189999973
- Understanding the Anchoring Effect of LLM with Synthetic Data: Existence, Mechanism, and Potentia...
Yiming Huang, Biquan Bie, Zuqiu Na, Weilin Ruan, Songxin Lei, Yutao Yue, Xinlei He
https://arxiv.org/abs/2505.15392 https://mastoxiv.page/@arXiv_csCL_bot/114550277171100272
- Just as Humans Need Vaccines, So Do Models: Model Immunization to Combat Falsehoods
Raza, Qureshi, Farooq, Lotif, Chadha, Pandya, Emmanouilidis
https://arxiv.org/abs/2505.17870 https://mastoxiv.page/@arXiv_csCL_bot/114572956853819813
- LingoLoop Attack: Trapping MLLMs via Linguistic Context and State Entrapment into Endless Loops
Fu, Jiang, Hong, Li, Guo, Yang, Chen, Zhang
https://arxiv.org/abs/2506.14493 https://mastoxiv.page/@arXiv_csCL_bot/114703502552989170
- GHTM: A Graph-based Hybrid Topic Modeling Approach with a Benchmark Dataset for the Low-Resource ...
Farhana Haque, Md. Abdur Rahman, Sumon Ahmed
https://arxiv.org/abs/2508.00605 https://mastoxiv.page/@arXiv_csCL_bot/114969875643478303
- Link Prediction for Event Logs in the Process Industry
Anastasia Zhukova, Thomas Walton, Christian E. Lobm\"uller, Bela Gipp
https://arxiv.org/abs/2508.09096 https://mastoxiv.page/@arXiv_csCL_bot/115020938764936882
- AirQA: A Comprehensive QA Dataset for AI Research with Instance-Level Evaluation
Huang, Cao, Zhang, Kang, Wang, Wang, Luo, Zheng, Qian, Chen, Yu
https://arxiv.org/abs/2509.16952 https://mastoxiv.page/@arXiv_csCL_bot/115253526588472475
- Multi-View Attention Multiple-Instance Learning Enhanced by LLM Reasoning for Cognitive Distortio...
Jun Seo Kim, Hyemi Kim, Woo Joo Oh, Hongjin Cho, Hochul Lee, Hye Hyeon Kim
https://arxiv.org/abs/2509.17292 https://mastoxiv.page/@arXiv_csCL_bot/115253586227941157
- Dual-Space Smoothness for Robust and Balanced LLM Unlearning
Han Yan, Zheyuan Liu, Meng Jiang
https://arxiv.org/abs/2509.23362 https://mastoxiv.page/@arXiv_csCL_bot/115293308293558024
- The Rise of AfricaNLP: Contributions, Contributors, Community Impact, and Bibliometric Analysis
Tadesse Destaw Belay, et al.
https://arxiv.org/abs/2509.25477 https://mastoxiv.page/@arXiv_csCL_bot/115298213432594791
- Open ASR Leaderboard: Towards Reproducible and Transparent Multilingual and Long-Form Speech Reco...
Srivastav, Zheng, Bezzam, Le Bihan, Koluguri, \.Zelasko, Majumdar, Moumen, Gandhi
https://arxiv.org/abs/2510.06961 https://mastoxiv.page/@arXiv_csCL_bot/115343748052193267
- Neuron-Level Analysis of Cultural Understanding in Large Language Models
Taisei Yamamoto, Ryoma Kumon, Danushka Bollegala, Hitomi Yanaka
https://arxiv.org/abs/2510.08284 https://mastoxiv.page/@arXiv_csCL_bot/115349533441895984
- CLMN: Concept based Language Models via Neural Symbolic Reasoning
Yibo Yang
https://arxiv.org/abs/2510.10063 https://mastoxiv.page/@arXiv_csCL_bot/115372392366793754
- Schema for In-Context Learning
Chen, Chen, Wang, Leong, Fung, Bernales, Aspuru-Guzik
https://arxiv.org/abs/2510.13905 https://mastoxiv.page/@arXiv_csCL_bot/115389057899856601
- Evaluating Latent Knowledge of Public Tabular Datasets in Large Language Models
Matteo Silvestri, Fabiano Veglianti, Flavio Giorgi, Fabrizio Silvestri, Gabriele Tolomei
https://arxiv.org/abs/2510.20351 https://mastoxiv.page/@arXiv_csCL_bot/115428615784704418
- LuxIT: A Luxembourgish Instruction Tuning Dataset from Monolingual Seed Data
Julian Valline, Cedric Lothritz, Siwen Guo, Jordi Cabot
https://arxiv.org/abs/2510.24434 https://mastoxiv.page/@arXiv_csCL_bot/115457025096322944
- Surfacing Subtle Stereotypes: A Multilingual, Debate-Oriented Evaluation of Modern LLMs
Muhammed Saeed, Muhammad Abdul-mageed, Shady Shehata
https://arxiv.org/abs/2511.01187 https://mastoxiv.page/@arXiv_csCL_bot/115491321130591723
toXiv_bot_toot