Another of my forays into AI ethics is just out! This time the focus is on the ethics (or lack thereof) of Reinforcement Learning Feedback (RLF) techniques aimed at increasing the 'alignment' of LLMs.
The paper is fruit of the joint work of a great team of collaborators, among whom @… and @…
Towards Language-Augmented Multi-Agent Deep Reinforcement Learning
Maxime Toquebiau, Jae-Yun Jun, Fa\"iz Benamar, Nicolas Bredeche
https://arxiv.org/abs/2506.05236
This https://arxiv.org/abs/2506.01016 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csLG_…
This https://arxiv.org/abs/2501.09622 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_qu…
Reason-to-Recommend: Using Interaction-of-Thought Reasoning to Enhance LLM Recommendation
Keyu Zhao, Fengli Xu, Yong Li
https://arxiv.org/abs/2506.05069 ht…
Autonomous Vehicle Lateral Control Using Deep Reinforcement Learning with MPC-PID Demonstration
Chengdong Wu, Sven Kirchner, Nils Purschke, Alois C. Knoll
https://arxiv.org/abs/2506.04040
This https://arxiv.org/abs/2505.23703 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csAI_…
This https://arxiv.org/abs/2505.23585 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csLG_…
CORA: Coalitional Rational Advantage Decomposition for Multi-Agent Policy Gradients
Mengda Ji, Genjiu Xu, Liying Wang
https://arxiv.org/abs/2506.04265 http…