
2025-07-16 10:08:21
ARMOR: Aligning Secure and Safe Large Language Models via Meticulous Reasoning
Zhengyue Zhao, Yingzi Ma, Somesh Jha, Marco Pavone, Chaowei Xiao
https://arxiv.org/abs/2507.11500
ARMOR: Aligning Secure and Safe Large Language Models via Meticulous Reasoning
Zhengyue Zhao, Yingzi Ma, Somesh Jha, Marco Pavone, Chaowei Xiao
https://arxiv.org/abs/2507.11500
Flocking as a second-order phase transition in self-aligning active crystals
Marco Musacchio, Alexander P. Antonov, Hartmut L\"owen, Lorenzo Caprini
https://arxiv.org/abs/2506.12967
Internal Value Alignment in Large Language Models through Controlled Value Vector Activation
Haoran Jin, Meng Li, Xiting Wang, Zhihao Xu, Minlie Huang, Yantao Jia, Defu Lian
https://arxiv.org/abs/2507.11316
The Space Between Us: A Methodological Framework for Researching Bonding and Proxemics in Situated Group-Agent Interactions
Ana M\"uller, Anja Richert
https://arxiv.org/abs/2506.11829
LLM-Driven Dual-Level Multi-Interest Modeling for Recommendation
Ziyan Wang, Yingpeng Du, Zhu Sun, Jieyi Bi, Haoyan Chua, Tianjun Wei, Jie Zhang
https://arxiv.org/abs/2507.10917
""[…] Red Hat Enterprise Linux for Business Developers […] provides self-serve, no-cost access to Red Hat Enterprise Linux [#RHEL] for enterprise development use.""
From Outcomes to Processes: Guiding PRM Learning from ORM for Inference-Time Alignment
Bin Xie, Bingbing Xu, Yige Yuan, Shengmao Zhu, Huawei Shen
https://arxiv.org/abs/2506.12446 …
FMC: Formalization of Natural Language Mathematical Competition Problems
Jiaxuan Xie, Chengwu Liu, Ye Yuan, Siqi Li, Zhiping Xiao, Ming Zhang
https://arxiv.org/abs/2507.11275
Aligning Proteins and Language: A Foundation Model for Protein Retrieval
Qifeng Wu, Zhengzhe Liu, Han Zhu, Yizhou Zhao, Daisuke Kihara, Min Xu
https://arxiv.org/abs/2506.08023
Mallows Model with Learned Distance Metrics: Sampling and Maximum Likelihood Estimation
Yeganeh Alimohammadi, Kiana Asgari
https://arxiv.org/abs/2507.08108
Multi-Task Reward Learning from Human Ratings
Mingkang Wu, Devin White, Evelyn Rose, Vernon Lawhern, Nicholas R Waytowich, Yongcan Cao
https://arxiv.org/abs/2506.09183
Visual Semantic Description Generation with MLLMs for Image-Text Matching
Junyu Chen, Yihua Gao, Mingyong Li
https://arxiv.org/abs/2507.08590 https://
Stable Preference Optimization for LLMs: A Bilevel Approach Beyond Direct Preference Optimization
Chengtao Jian, Kai Yang, Ye Ouyang, Xiaozhou Ye
https://arxiv.org/abs/2507.07723 …
Trump threatens 10% tariff on countries backing BRICS 'anti-American policy': https://benborges.xyz/2025/07/07/trump-threatens-tariff-on-countries.html
Approaching Dialogue State Tracking via Aligning Speech Encoders and LLMs
\v{S}imon Sedl\'a\v{c}ek, Bolaji Yusuf, J\'an \v{S}vec, Pradyoth Hegde, Santosh Kesiraju, Old\v{r}ich Plchot, Jan \v{C}ernock\'y
https://arxiv.org/abs/2506.08633
Joint Optimization-based Targetless Extrinsic Calibration for Multiple LiDARs and GNSS-Aided INS of Ground Vehicles
Junhui Wang, Yan Qiao, Chao Gao, Naiqi Wu
https://arxiv.org/abs/2507.08349
AR2: Attention-Guided Repair for the Robustness of CNNs Against Common Corruptions
Fuyuan Zhang, Qichen Wang, Jianjun Zhao
https://arxiv.org/abs/2507.06332
This https://arxiv.org/abs/2505.07270 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csSE_…
Aligning visual imagery to the operator improves geospatial situation awareness in a single-display 360-degree periscope concept https://cognitiveresearchjournal.springeropen.com/articles/10.1186/s41235-025-00646-1
From Threat to Tool: Leveraging Refusal-Aware Injection Attacks for Safety Alignment
Kyubyung Chae, Hyunbin Jin, Taesup Kim
https://arxiv.org/abs/2506.10020
AMRScan: A hybrid R and Nextflow toolkit for rapid antimicrobial resistance gene detection from sequencing data
Kaitao Lai
https://arxiv.org/abs/2507.08062
SpecCLIP: Aligning and Translating Spectroscopic Measurements for Stars
Xiaosheng Zhao, Yang Huang, Guirong Xue, Xiao Kong, Jifeng Liu, Xiaoyu Tang, Timothy C. Beers, Yuan-Sen Ting, A-Li Luo
https://arxiv.org/abs/2507.01939
Quantile Reward Policy Optimization: Alignment with Pointwise Regression and Exact Partition Functions
Simon Matrenok, Skander Moalla, Caglar Gulcehre
https://arxiv.org/abs/2507.08068 https://arxiv.org/pdf/2507.08068 https://arxiv.org/html/2507.08068
arXiv:2507.08068v1 Announce Type: new
Abstract: Aligning large language models with pointwise absolute rewards has so far required online, on-policy algorithms such as PPO and GRPO. In contrast, simpler methods that can leverage offline or off-policy data, such as DPO and REBEL, are limited to learning from preference pairs or relative signals. To bridge this gap, we introduce \emph{Quantile Reward Policy Optimization} (QRPO), which learns from pointwise absolute rewards while preserving the simplicity and offline applicability of DPO-like methods. QRPO uses quantile rewards to enable regression to the closed-form solution of the KL-regularized RL objective. This reward yields an analytically tractable partition function, removing the need for relative signals to cancel this term. Moreover, QRPO scales with increased compute to estimate quantile rewards, opening a new dimension for pre-computation scaling. Empirically, QRPO consistently achieves top performance on chat and coding evaluations -- reward model scores, AlpacaEval 2, and LeetCode -- compared to DPO, REBEL, and SimPO across diverse datasets and 8B-scale models. Finally, we find that training with robust rewards instead of converting them to preferences induces less length bias.
toXiv_bot_toot
This is the next stages of Columbus’s long-running effort to update its 1950s-era zoning code. This phase will incorporate a housing component to address affordability issues by better aligning work and residential areas.
We also need to build more houses.
https://
ChainEdit: Propagating Ripple Effects in LLM Knowledge Editing through Logical Rule-Guided Chains
Zilu Dong, Xiangqing Shen, Zinong Yang, Rui Xia
https://arxiv.org/abs/2507.08427 …
Q&A with Hugging Face Chief Ethics Scientist Margaret Mitchell on aligning AI development with human needs, the "illusion of consensus" around AGI, and more (Melissa Heikkilä/Financial Times)
https://www.ft.com/content/7089bff2-25fc-4a25-98bf-8828ab24…
CoMemo: LVLMs Need Image Context with Image Memory
Shi Liu, Weijie Su, Xizhou Zhu, Wenhai Wang, Jifeng Dai
https://arxiv.org/abs/2506.06279 https://…
# Philosophical test fails ChatGPT: AI coherence isn’t enough to prove human mind
The research reveals that #ChatGPT does exhibit proficiency in basic coherence building. It maintains consistent dictional and intentional lines by reusing phrases and aligning responses with contextual topics. It also demonstrates some ability to construct rational coherence by offering logically consistent replies…
Leveraging machine learning features for linear optical interferometer control
Sergei S. Kuzmin, Ivan V. Dyakonov, Stanislav S. Straupe
https://arxiv.org/abs/2505.24032
Stochastically Dominant Peer Prediction
Yichi Zhang, Shengwei Xu, David Pennock, Grant Schoenebeck
https://arxiv.org/abs/2506.02259 https://
This https://arxiv.org/abs/2503.07217 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csSD_…
ROS-related Robotic Systems Development with V-model-based Application of MeROS Metamodel
Tomasz Winiarski, Jan Kaniuka, Daniel Gie{\l}dowski, Jakub Ostrysz, Krystian Radlak, Dmytro Kushnir
https://arxiv.org/abs/2506.08706
KERAG_R: Knowledge-Enhanced Retrieval-Augmented Generation for Recommendation
Zeyuan Meng, Zixuan Yi, Iadh Ounis
https://arxiv.org/abs/2507.05863 https://
Aligning Protein Conformation Ensemble Generation with Physical Feedback
Jiarui Lu, Xiaoyin Chen, Stephen Zhewen Lu, Aur\'elie Lozano, Vijil Chenthamarakshan, Payel Das, Jian Tang
https://arxiv.org/abs/2505.24203
This https://arxiv.org/abs/2402.17732 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_mat…
This https://arxiv.org/abs/2501.07071 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csAI_…
HYFuse: Aligning Heterogeneous Speech Pre-Trained Representations in Hyperbolic Space for Speech Emotion Recognition
Orchid Chetia Phukan, Girish, Mohd Mujtaba Akhtar, Swarup Ranjan Behera, Pailla Balakrishna Reddy, Arun Balaji Buduru, Rajesh Sharma
https://arxiv.org/abs/2506.03403
This https://arxiv.org/abs/2410.05605 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csSE_…
Offline Map Matching Based on Localization Error Distribution Modeling
Ruilin Xu, Yuchen Song, Kaijie Li, Xitong Gao, Kejiang Ye, Fan Zhang, Juanjuan Zhao
https://arxiv.org/abs/2505.23123
I got the tow hitch installed on the Crosstrek today. Holding a tow hitch bar up, while aligning a bolt, and trying to get the nut on said bolt before I lose the bolt inside the frame OR drop the hitch on my face can be quite exciting. Then torquing those to 110 ft/lbs while lying under all of that can be hard on your shoulder. I'm exhausted now.
MPF: Aligning and Debiasing Language Models post Deployment via Multi Perspective Fusion
Xin Guan, PeiHsin Lin, Zekun Wu, Ze Wang, Ruibo Zhang, Emre Kazim, Adriano Koshiyama
https://arxiv.org/abs/2507.02595
Insights from Educators on Building a More Cohesive Quantum Information Science and Engineering Education Ecosystem
Shams El-Adawy, A. R. Pi\~na, Benjamin M. Zwickl, H. J. Lewandowski
https://arxiv.org/abs/2507.01578
Deformable Medical Image Registration with Effective Anatomical Structure Representation and Divide-and-Conquer Network
Xinke Ma, Yongsheng Pan, Qingjie Zeng, Mengkang Lu, Bolysbek Murat Yerzhanuly, Bazargul Matkerim, Yong Xia
https://arxiv.org/abs/2506.19222
Fast entropy-regularized SDP relaxations for permutation synchronization
Michael Lindsey, Yunpeng Shi
https://arxiv.org/abs/2506.20191 https://
CTR-Guided Generative Query Suggestion in Conversational Search
Erxue Min, Hsiu-Yuan Huang, Xihong Yang, Min Yang, Xin Jia, Yunfang Wu, Hengyi Cai, Junfeng Wang, Shuaiqiang Wang, Dawei Yin
https://arxiv.org/abs/2507.04072
Case Study for Developing a UXR Point of View for FinOps Product Innovation
Jason Dong, Anna Wu
https://arxiv.org/abs/2506.15314 https://
Gradient-Adaptive Policy Optimization: Towards Multi-Objective Alignment of Large Language Models
Chengao Li, Hanyu Zhang, Yunkun Xu, Hongyan Xue, Xiang Ao, Qing He
https://arxiv.org/abs/2507.01915
Optimal alignment of Lorentz orientation and generalization to matrix Lie groups
Congzhou M Sha
https://arxiv.org/abs/2506.14994 https://
BulletGen: Improving 4D Reconstruction with Bullet-Time Generation
Denys Rozumnyi, Jonathon Luiten, Numair Khan, Johannes Sch\"onberger, Peter Kontschieder
https://arxiv.org/abs/2506.18601
This https://arxiv.org/abs/2505.10640 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_csSE_…
SafePTR: Token-Level Jailbreak Defense in Multimodal LLMs via Prune-then-Restore Mechanism
Beitao Chen, Xinyu Lyu, Lianli Gao, Jingkuan Song, Heng Tao Shen
https://arxiv.org/abs/2507.01513
Hybrid-Sep: Language-queried audio source separation via pre-trained Model Fusion and Adversarial Diffusion Training
Jianyuan Feng, Guangzheng Li, Yangfei Xu
https://arxiv.org/abs/2506.16833
Robust Alignment via Partial Gromov-Wasserstein Distances
Xiaoyun Gong, Sloan Nietert, Ziv Goldfeld
https://arxiv.org/abs/2506.21507 https://
AgentDistill: Training-Free Agent Distillation with Generalizable MCP Boxes
Jiahao Qiu, Xinzhe Juan, Yimin Wang, Ling Yang, Xuan Qi, Tongcheng Zhang, Jiacheng Guo, Yifu Lu, Zixin Yao, Hongru Wang, Shilong Liu, Xun Jiang, Liu Leqi, Mengdi Wang
https://arxiv.org/abs/2506.14728
Replaced article(s) found for cs.CL. https://arxiv.org/list/cs.CL/new
[3/3]:
- Aligning Frozen LLMs by Reinforcement Learning: An Iterative Reweight-then-Optimize Approach
Zhang, Li, Zeng, Li, Wang, Lin, Lu, Garcia, Hong
SAFER: Probing Safety in Reward Models with Sparse Autoencoder
Sihang Li, Wei Shi, Ziyuan Xie, Tao Liang, Guojun Ma, Xiang Wang
https://arxiv.org/abs/2507.00665
CoVE: Compressed Vocabulary Expansion Makes Better LLM-based Recommender Systems
Haochen Zhang, Tianyi Zhang, Junze Yin, Oren Gal, Anshumali Shrivastava, Vladimir Braverman
https://arxiv.org/abs/2506.19993
Not All Jokes Land: Evaluating Large Language Models Understanding of Workplace Humor
Moahmmadamin Shafiei, Hamidreza Saffari
https://arxiv.org/abs/2506.01819
Defining the Game Producer: A Mapping of Key Characteristics and Differentiators of the Professional Behind Digital Game Production
Rafael C. Lopes, Danilo M. Ribeiro
https://arxiv.org/abs/2506.14409
Massive Supervised Fine-tuning Experiments Reveal How Data, Layer, and Training Factors Shape LLM Alignment Quality
Yuto Harada, Yusuke Yamauchi, Yusuke Oda, Yohei Oseki, Yusuke Miyao, Yu Takagi
https://arxiv.org/abs/2506.14681