Are LLMs Reliable Rankers? Rank Manipulation via Two-Stage Token OptimizationTiancheng Xing, Jerry Li, Yixuan Du, Xiyang Huhttps://arxiv.org/abs/2510.06732 https://
Are LLMs Reliable Rankers? Rank Manipulation via Two-Stage Token OptimizationLarge language models (LLMs) are increasingly used as rerankers in information retrieval, yet their ranking behavior can be steered by small, natural-sounding prompts. To expose this vulnerability, we present Rank Anything First (RAF), a two-stage token optimization method that crafts concise textual perturbations to consistently promote a target item in LLM-generated rankings while remaining hard to detect. Stage 1 uses Greedy Coordinate Gradient to shortlist candidate tokens at the current po…