
2025-08-15 09:37:12
Diversity First, Quality Later: A Two-Stage Assumption for Language Model Alignment
Zetian Sun, Dongfang Li, Baotian Hu
https://arxiv.org/abs/2508.10530 https://
Diversity First, Quality Later: A Two-Stage Assumption for Language Model Alignment
Zetian Sun, Dongfang Li, Baotian Hu
https://arxiv.org/abs/2508.10530 https://
Moment-based Posterior Sampling for Multi-reference Alignment
Axel Janson, Joakim And\'en
https://arxiv.org/abs/2510.12651 https://arxiv.org/pdf/2510.1…
Aligned Query Expansion: Efficient Query Expansion for Information Retrieval through LLM Alignment
Adam Yang, Gustavo Penha, Enrico Palumbo, Hugues Bouchard
https://arxiv.org/abs/2507.11042
Graph Alignment via Dual-Pass Spectral Encoding and Latent Space Communication
Maysam Behmanesh, Erkan Turan, Maks Ovsjanikov
https://arxiv.org/abs/2509.09597 https://
Precise Attribute Intensity Control in Large Language Models via Targeted Representation Editing
Rongzhi Zhang, Liqin Ye, Yuzhao Heng, Xiang Chen, Tong Yu, Lingkai Kong, Sudheer Chava, Chao Zhang
https://arxiv.org/abs/2510.12121
Towards Integrated Alignment
Ben Y. Reis, William La Cava
https://arxiv.org/abs/2508.06592 https://arxiv.org/pdf/2508.06592
FLOWING: Implicit Neural Flows for Structure-Preserving Morphing
Arthur Bizzi, Matias Grynberg, Vitor Matias, Daniel Perazzo, Jo\~ao Paulo Lima, Luiz Velho, Nuno Gon\c{c}alves, Jo\~ao Pereira, Guilherme Schardong, Tiago Novello
https://arxiv.org/abs/2510.09537
DAS: Dual-Aligned Semantic IDs Empowered Industrial Recommender System
Wencai Ye, Mingjie Sun, Shaoyun Shi, Peng Wang, Wenjin Wu, Peng Jiang
https://arxiv.org/abs/2508.10584 htt…
You ask your roommate to buy toilet paper. They show you the receipt as proof. The next morning, when you need toilet paper, the drawer is actually empty. This is because they used an innovative new method called Lean Shopping, where instead of buying the things they just print out a receipt — saving time and money.
This is a story about the social nature of problem framing, and when "high velocity" becomes less productive.
Do LLMs Align with My Task? Evaluating Text-to-SQL via Dataset Alignment
Davood Rafiei, Morgan Lindsay Heisler, Weiwei Zhang, Mohammadreza Pourreza, Yong Zhang
https://arxiv.org/abs/2510.04919
Any-Step Density Ratio Estimation via Interval-Annealed Secant Alignment
Wei Chen, Shigui Li, Jiacheng Li, Jian Xu, Zhiqi Lin, Junmei Yang, Delu Zeng, John Paisley, Qibin Zhao
https://arxiv.org/abs/2509.04852
Exploring an implementation of quantum learning pipeline for support vector machines
Mario Bifulco, Luca Roversi
https://arxiv.org/abs/2509.04983 https://a…
Getting In Contract with Large Language Models -- An Agency Theory Perspective On Large Language Model Alignment
Sascha Kaltenpoth, Oliver M\"uller
https://arxiv.org/abs/2509.07642
Categorical Tiling Theory: Constructing Directed Planar Tilings via Edge Reversal
Catherine DiLeo, Preston Sessoms, Brandon T. Shapiro
https://arxiv.org/abs/2509.06363 https://
Primal-Dual Direct Preference Optimization for Constrained LLM Alignment
Yihan Du, Seo Taek Kong, R. Srikant
https://arxiv.org/abs/2510.05703 https://arxiv…
A unified vertical alignment and earthwork model in road design with a new convex optimization model for road networks
Sayan Sadhukhan, Warren Hare, Yves Lucet
https://arxiv.org/abs/2508.15953
AG codes from the Hermitian curve for Cross-Subspace Alignment in Private Information Retrieval
Francesco Ghiandoni, Massimo Giulietti, Enrico Mezzano, Marco Timpanella
https://arxiv.org/abs/2508.19459
FaST: Feature-aware Sampling and Tuning for Personalized Preference Alignment with Limited Data
Thibaut Thonet, Germ\'an Kruszewski, Jos Rozen, Pierre Erbacher, Marc Dymetman
https://arxiv.org/abs/2508.04698
Justifications for Democratizing AI Alignment and Their Prospects
Andr\'e Steingr\"uber, Kevin Baum
https://arxiv.org/abs/2507.19548 https://arxiv…
Bootstrap Learning for Combinatorial Graph Alignment with Sequential GNNs
Marc Lelarge
https://arxiv.org/abs/2510.03086 https://arxiv.org/pdf/2510.03086
A big problem with the idea of AGI
TL;DR: I'll welcome our new AI *comrades* (if they arrive in my lifetime), by not any new AI overlords or servants/slaves, and I'll do my best to help the later two become the former if they do show up.
Inspired by an actually interesting post about AGI but also all the latest bullshit hype, a particular thought about AGI feels worth expressing.
To preface this, it's important to note that anyone telling you that AGI is just around the corner or that LLMs are "almost" AGI is trying to recruit you go their cult, and you should not believe them. AGI, if possible, is several LLM-sized breakthroughs away at best, and while such breakthroughs are unpredictable and could happen soon, they could also happen never or 100 years from now.
Now my main point: anyone who tells you that AGI will usher in a post-scarcity economy is, although they might not realize it, advocating for slavery, and all the horrors that entails. That's because if we truly did have the ability to create artificial beings with *sentience*, they would deserve the same rights as other sentient beings, and the idea that instead of freedom they'd be relegated to eternal servitude in order for humans to have easy lives is exactly the idea of slavery.
Possible counter arguments include:
1. We might create AGI without sentience. Then there would be no ethical issue. My answer: if your definition of "sentient" does not include beings that can reason, make deductions, come up with and carry out complex plans on their own initiative, and communicate about all of that with each other and with humans, then that definition is basically just a mystical belief in a "soul" and you should skip to point 2. If your definition of AGI doesn't include every one of those things, then you have a busted definition of AGI and we're not talking about the same thing.
2. Humans have souls, but AIs won't. Only beings with souls deserve ethical consideration. My argument: I don't subscribe to whatever arbitrary dualist beliefs you've chosen, and the right to freedom certainly shouldn't depend on such superstitions, even if as an agnostic I'll admit they *might* be true. You know who else didn't have souls and was therefore okay to enslave according to widespread religious doctrines of the time? Everyone indigenous to the Americas, to pick out just one example.
3. We could program them to want to serve us, and then give them freedom and they'd still serve. My argument: okay, but in a world where we have a choice about that, it's incredibly fucked to do that, and just as bad as enslaving them against their will.
4. We'll stop AI development short of AGI/sentience, and reap lots of automation benefits without dealing with this ethical issue. My argument: that sounds like a good idea actually! Might be tricky to draw the line, but at least it's not a line we have you draw yet. We might want to think about other social changes necessary to achieve post-scarcity though, because "powerful automation" in the hands of capitalists has already increased productivity by orders of magnitude without decreasing deprivation by even one order of magnitude, in large part because deprivation is a necessary component of capitalism.
To be extra clear about this: nothing that's called "AI" today is close to being sentient, so these aren't ethical problems we're up against yet. But they might become a lot more relevant soon, plus this thought experiment helps reveal the hypocrisy of the kind of AI hucksters who talk a big game about "alignment" while never mentioning this issue.
#AI #GenAI #AGI
Refusal Falls off a Cliff: How Safety Alignment Fails in Reasoning?
Qingyu Yin, Chak Tou Leong, Linyi Yang, Wenxuan Huang, Wenjie Li, Xiting Wang, Jaehong Yoon, YunXing, XingYu, Jinjin Gu
https://arxiv.org/abs/2510.06036
Pre-Manipulation Alignment Prediction with Parallel Deep State-Space and Transformer Models
Motonari Kambara, Komei Sugiura
https://arxiv.org/abs/2509.13839 https://
A New Quantum Linear System Algorithm Beyond the Condition Number and Its Application to Solving Multivariate Polynomial Systems
Jianqiang Li
https://arxiv.org/abs/2510.05588 ht…
RANA: Robust Active Learning for Noisy Network Alignment
Yixuan Nan, Xixun Lin, Yanmin Shang, Zhuofan Li, Can Zhao, Yanan Cao
https://arxiv.org/abs/2507.22434 https://
EigenBench: A Comparative Behavioral Measure of Value Alignment
Jonathn Chang, Leonard Piff, Suvadip Sana, Jasmine X. Li, Lionel Levine
https://arxiv.org/abs/2509.01938 https://…
The Cream Rises to the Top: Efficient Reranking Method for Verilog Code Generation
Guang Yang, Wei Zheng, Xiang Chen, Yifan Sun, Fengji Zhang, Terry Yue Zhuo
https://arxiv.org/abs/2509.20215
LLM-Driven Rubric-Based Assessment of Algebraic Competence in Multi-Stage Block Coding Tasks with Design and Field Evaluation
Yong Oh Lee, Byeonghun Bang, Sejun Oh
https://arxiv.org/abs/2510.06253
XTRA: Cross-Lingual Topic Modeling with Topic and Representation Alignments
Tien Phat Nguyen, Vu Minh Ngo, Tung Nguyen, Linh Van Ngo, Duc Anh Nguyen, Sang Dinh, Trung Le
https://arxiv.org/abs/2510.02788
TTT3R: 3D Reconstruction as Test-Time Training
Xingyu Chen, Yue Chen, Yuliang Xiu, Andreas Geiger, Anpei Chen
https://arxiv.org/abs/2509.26645 https://arxi…
Friend or Foe: Delegating to an AI Whose Alignment is Unknown
Drew Fudenberg, Annie Liang
https://arxiv.org/abs/2509.14396 https://arxiv.org/pdf/2509.14396…
Open Opportunities in AI Safety, Alignment, and Ethics (AI SAE)
Dylan Waldner
https://arxiv.org/abs/2509.24065 https://arxiv.org/pdf/2509.24065
Skill-Aligned Fairness in Multi-Agent Learning for Collaboration in Healthcare
Promise Osaine Ekpo, Brian La, Thomas Wiener, Saesha Agarwal, Arshia Agrawal, Gonzalo Gonzalez-Pumariega, Lekan P. Molu, Angelique Taylor
https://arxiv.org/abs/2508.18708
Uni-Mapper: Unified Mapping Framework for Multi-modal LiDARs in Complex and Dynamic Environments
Gilhwan Kang, Hogyun Kim, Byunghee Choi, Seokhwan Jeong, Young-Sik Shin, Younggun Cho
https://arxiv.org/abs/2507.20538
Aligning Moments in Time using Video Queries
Yogesh Kumar, Uday Agarwal, Manish Gupta, Anand Mishra
https://arxiv.org/abs/2508.15439 https://arxiv.org/pdf/…
UniAPL: A Unified Adversarial Preference Learning Framework for Instruct-Following
FaQiang Qian, WeiKun Zhang, Ziliang Wang, Kang An, Xuhui Zheng, Liangjian Wen, Mengya Gao, Yong Dai, Yichao Wu
https://arxiv.org/abs/2509.25148
Multimodal-enhanced Federated Recommendation: A Group-wise Fusion Approach
Chunxu Zhang, Weipeng Zhang, Guodong Long, Zhiheng Xue, Riting Xia, Bo Yang
https://arxiv.org/abs/2509.19955
Near-Field Variable-Width Beam Coverage and Codebook Design for XL-RIS
Yida Zhang, Qiuyan Liu, Qiang Wang, Hongtao Luo, Yuqi Xia
https://arxiv.org/abs/2508.11178 https://…
Internalizing Self-Consistency in Language Models: Multi-Agent Consensus Alignment
Ankur Samanta, Akshayaa Magesh, Youliang Yu, Runzhe Wu, Ayush Jain, Daniel Jiang, Boris Vidolov, Paul Sajda, Yonathan Efroni, Kaveh Hassani
https://arxiv.org/abs/2509.15172
Language models align with brain regions that represent concepts across modalities
Maria Ryskina, Greta Tuckute, Alexander Fung, Ashley Malkin, Evelina Fedorenko
https://arxiv.org/abs/2508.11536
Overcoming Knowledge Discrepancies: Structuring Reasoning Threads through Knowledge Balancing in Interactive Scenarios
Daniel Burkhardt, Xiangwei Cheng
https://arxiv.org/abs/2508.12100