2025-09-18 06:41:06
OpenAI and Apollo Research trained o3 and o4-mini versions to not engage in "scheming", or secretly pursuing undesirable goals, reducing "covert actions" ~30X (Radhika Rajkumar/ZDNET)
https://www.zdnet.com/article/ai-models-kn
OpenAI and Apollo Research trained o3 and o4-mini versions to not engage in "scheming", or secretly pursuing undesirable goals, reducing "covert actions" ~30X (Radhika Rajkumar/ZDNET)
https://www.zdnet.com/article/ai-models-kn
https://www.axios.com/2025/12/16/ai-models-hacking-stanford-openai-warnings
Leaders from Amazon and Anthropic are testifying before two house committees today.
A Study on Thinking Patterns of Large Reasoning Models in Code Generation
Kevin Halim, Sin G. Teo, Ruitao Feng, Zhenpeng Chen, Yang Gu, Chong Wang, Yang Liu
https://arxiv.org/abs/2509.13758
Can Current AI Models Count What We Mean, Not What They See? A Benchmark and Systematic Evaluation
Gia Khanh Nguyen, Yifeng Huang, Minh Hoai
https://arxiv.org/abs/2509.13939 htt…
💔 Are Foundation Models Ready for Industrial Defect Recognition? A Reality Check on Real-World Data
#computing
Defending Diffusion Models Against Membership Inference Attacks via Higher-Order Langevin Dynamics
Benjamin Sterling, Yousef El-Laham, M\'onica F. Bugallo
https://arxiv.org/abs/2509.14225
Evaluation Awareness Scales Predictably in Open-Weights Large Language Models
Maheep Chaudhary, Ian Su, Nikhil Hooda, Nishith Shankar, Julia Tan, Kevin Zhu, Ashwinee Panda, Ryan Lagasse, Vasu Sharma
https://arxiv.org/abs/2509.13333
Adobe updates Firefly's app to add a prompt-based video editor using Runway's Aleph model, and adds image and video models from Black Forest Labs and others (Ivan Mehta/TechCrunch)
https://techcrunch.com/2025/12/16/adob
Can Large Language Models Robustly Perform Natural Language Inference for Japanese Comparatives?
Yosuke Mikami, Daiki Matsuoka, Hitomi Yanaka
https://arxiv.org/abs/2509.13695 ht…
All Models Are Wrong, But Can They Be Useful? Lessons from COVID-19 Agent-Based Models: A Systematic Review
Emma Von Hoene, Sara Von Hoene, Szandra Peter, Ethan Hopson, Emily Csizmadia, Faith Fenyk, Kai Barner, Timothy Leslie, Hamdi Kavak, Andreas Zufle, Amira Roess, Taylor Anderson
https://arxiv.org/abs/2509.13346
Dual-Actor Fine-Tuning of VLA Models: A Talk-and-Tweak Human-in-the-Loop Approach
Piaopiao Jin, Qi Wang, Guokang Sun, Ziwen Cai, Pinjia He, Yangwei You
https://arxiv.org/abs/2509.13774
Comprehensive Evaluation of CNN-Based Audio Tagging Models on Resource-Constrained Devices
Jordi Grau-Haro, Ruben Ribes-Serrano, Javier Naranjo-Alcazar, Marta Garcia-Ballesteros, Pedro Zuccarello
https://arxiv.org/abs/2509.14049
Accuracy Paradox in Large Language Models: Regulating Hallucination Risks in Generative AI
Zihao Li, Weiwei Yi, Jiahong Chen
https://arxiv.org/abs/2509.13345 https://
Ensemble of Pre-Trained Models for Long-Tailed Trajectory Prediction
Divya Thuremella, Yi Yang, Simon Wanna, Lars Kunze, Daniele De Martini
https://arxiv.org/abs/2509.13914 http…
PREDICT-GBM: Platform for Robust Evaluation and Development of Individualized Computational Tumor Models in Glioblastoma
L. Zimmer, J. Weidner, M. Balcerak, F. Kofler, I. Ezhov, B. Menze, B. Wiestler
https://arxiv.org/abs/2509.13360
Sample Size Calculations for the Development of Risk Prediction Models that Account for Performance Variability
Menelaos Pavlou, Rumana Z. Omar, Gareth Ambler
https://arxiv.org/abs/2509.14028
HYCO: Hybrid-Cooperative Learning for Data-Driven PDE Modeling
Lorenzo Liverani, Matthys Steynberg, Enrique Zuazua
https://arxiv.org/abs/2509.14123 https://
Parallelizable Feynman-Kac Models for Universal Probabilistic Programming
Michele Boreale (University of Florence), Luisa Collodi (University of Florence)
https://arxiv.org/abs/2509.14092
Thermal states emerging from low-entanglement background in disordered spin models
Yule Ma, Qianqian Chen, Mingyang Li, Zlatko Papi\'c, Zheng Zhu
https://arxiv.org/abs/2509.13420
The impact of modeling approaches on controlling safety-critical, highly perturbed systems: the case for data-driven models
Piotr {\L}aszkiewicz, Maria Carvalho, Cl\'audia Soares, Pedro Louren\c{c}o
https://arxiv.org/abs/2509.13531
Evaluating the Limits of QAOA Parameter Transfer at High-Rounds on Sparse Ising Models With Geometrically Local Cubic Terms
Elijah Pelofske, Marek Rams, Andreas B\"artschi, Piotr Czarnik, Paolo Braccia, Lukasz Cincio, Stephan Eidenbenz
https://arxiv.org/abs/2509.13528
@… we need better licensing models across the board
Quelle surprise: seriöse Webseiten blocken den Zugang für KI-Trainingszugriff eher als Seiten, deren Zweck der Desinformation dient.
https://arxiv.org/abs/2510.10315
Cross-diffusion limits in multispecies kinetic models
Ansgar J\"ungel, Annamaria Pollino, Satoshi Taguchi
https://arxiv.org/abs/2509.14046 https://arx…
Scrub It Out! Erasing Sensitive Memorization in Code Language Models via Machine Unlearning
Zhaoyang Chu, Yao Wan, Zhikun Zhang, Di Wang, Zhou Yang, Hongyu Zhang, Pan Zhou, Xuanhua Shi, Hai Jin, David Lo
https://arxiv.org/abs/2509.13755
A small number of samples can poison LLMs of any size:
https://www.anthropic.com/research/small-samples-poison
"In a joint study with the UK AI Security Institute and the Alan Turing Institute, we found that as few as 250 malicious documents can produce a "
RepCaM : Exploring Transparent Visual Prompt With Inference-Time Re-Parameterization for Neural Video Delivery
Rongyu Zhang, Xize Duan, Jiaming Liu, Li Du, Yuan Du, Dan Wang, Shanghang Zhang, Fangxin Wang
https://arxiv.org/abs/2509.14002
Differential Privacy in Federated Learning: Mitigating Inference Attacks with Randomized Response
Ozer Ozturk, Busra Buyuktanir, Gozde Karatas Baydogmus, Kazim Yildiz
https://arxiv.org/abs/2509.13987
"People should not "blindly trust" everything AI tools tell them, the boss of Google's parent company Alphabet told the BBC.
In an exclusive interview, chief executive Sundar Pichai said that AI models are "prone to errors" and urged people to use them alongside other tools.
Mr Pichai said it highlighted the importance of having a rich information ecosystem, rather than solely relying on AI technology."
Always fun/challenging to read new AI (pre)papers like this. "Base models know how to reason, thinking models learn when".
#AI #Google #reasoning
From Next Token Prediction to (STRIPS) World Models -- Preliminary Results
Carlos N\'u\~nez-Molina, Vicen\c{c} G\'omez, Hector Geffner
https://arxiv.org/abs/2509.13389 h…
Designing Psychometric Bias Measures for ChatBots: An Application to Racial Bias Measurement
Mouhacine Benosman
https://arxiv.org/abs/2509.13324 https://ar…
"By examining several widely used AI models designed for cancer detection, the researchers found that performance varied depending on patients’ self-reported gender, race, and age. They also identified multiple reasons why these disparities occur."
What AI Learned From Cancer Slides Shocked Researchers
https://scitechdaily.com/what-ai-learned-from-cancer-slides-shocked-researchers/
Diving into Mitigating Hallucinations from a Vision Perspective for Large Vision-Language Models
Weihang Wang, Xinhao Li, Ziyue Wang, Yan Pang, Jielei Zhang, Peiyi Li, Qiang Zhang, Longwen Gao
https://arxiv.org/abs/2509.13836
Do Large Language Models Understand Word Senses?
Domenico Meconi, Simone Stirpe, Federico Martelli, Leonardo Lavalle, Roberto Navigli
https://arxiv.org/abs/2509.13905 https://…
TICL: Text-Embedding KNN For Speech In-Context Learning Unlocks Speech Recognition Abilities of Large Multimodal Models
Haolong Zheng, Yekaterina Yegorova, Mark Hasegawa-Johnson
https://arxiv.org/abs/2509.13395
New OpenAI models likely to pose "high" cybersecurity risk
OpenAI says the cyber capabilities of its frontier AI models are accelerating and warns Wednesday that upcoming models are likely to pose a "high" risk […]
😶🌫️ https://www.axios.com/2025/12/10/openai-ne
The threat of analytic flexibility in using large language models to simulate human data: A call to attention
Jamie Cummins
https://arxiv.org/abs/2509.13397 https://
A look at Apple's iPhone roadmap; sources say the iPhone 18 Pro lineup will get under-screen Face ID, and the front camera will move to the top-left corner (The Information)
https://www.theinformation.com/articles/inside-apples-ip…
Are Prompts All You Need? Evaluating Prompt-Based Large Language Models (LLM)s for Software Requirements Classification
Manal Binkhonain, Reem Alfayaz
https://arxiv.org/abs/2509.13868
Language models' activations linearly encode training-order recency
Dmitrii Krasheninnikov, Richard E. Turner, David Krueger
https://arxiv.org/abs/2509.14223 https://…
Early Stopping Chain-of-thoughts in Large Language Models
Minjia Mao, Bowen Yin, Yu Zhu, Xiao Fang
https://arxiv.org/abs/2509.14004 https://arxiv.org/pdf/2…
OpenHA: A Series of Open-Source Hierarchical Agentic Models in Minecraft
Zihao Wang, Muyao Li, Kaichen He, Xiangyu Wang, Zhancun Mu, Anji Liu, Yitao Liang
https://arxiv.org/abs/2509.13347
A tree-based Polynomial Chaos expansion for surrogate modeling and sensitivity analysis of complex numerical models
Faten Ben Said (CERMICS, EDF R\&D LNHE), Aur\'elien Alfonsi (CERMICS, MATHRISK), Anne Dutfoy (EDF R\&D PERICLES), C\'edric Goeury (EDF R\&D LNHE, LHSV), Magali Jodeau (EDF R\&D LNHE, LHSV), Julien Reygner (CERMICS, RT-UQ), Fabrice Zaoui (EDF R\&D LNHE)
PhysicalAgent: Towards General Cognitive Robotics with Foundation World Models
Artem Lykov, Jeffrin Sam, Hung Khang Nguyen, Vladislav Kozlovskiy, Yara Mahmoud, Valerii Serpiva, Miguel Altamirano Cabrera, Mikhail Konenkov, Dzmitry Tsetserukou
https://arxiv.org/abs/2509.13903
End-to-End Complexity Analysis for Quantum Simulation of the Extended Jaynes-Cummings Models
Nam Nguyen, Michael Yu, Alan Robertson, Hiromichi Nishimura, Samuel J. Elman, Benjamin Koltenbah
https://arxiv.org/abs/2509.13546
Masked Diffusion Models as Energy Minimization
Sitong Chen, Shen Nie, Jiacheng Sun, Zijin Feng, Zhenguo Li, Ji-Rong Wen, Chongxuan Li
https://arxiv.org/abs/2509.13866 https://…
Integrating Text and Time-Series into (Large) Language Models to Predict Medical Outcomes
Iyadh Ben Cheikh Larbi, Ajay Madhavan Ravichandran, Aljoscha Burchardt, Roland Roller
https://arxiv.org/abs/2509.13696
Consistent View Alignment Improves Foundation Models for 3D Medical Image Segmentation
Puru Vaish, Felix Meister, Tobias Heimann, Christoph Brune, Jelmer M. Wolterink
https://arxiv.org/abs/2509.13846
Crash Report Enhancement with Large Language Models: An Empirical Study
S M Farah Al Fahim (Peter), Md Nakhla Rafi (Peter), Zeyang Ma (Peter), Dong Jae Kim (Peter), Tse-Hsun (Peter), Chen
https://arxiv.org/abs/2509.13535
Graph-Regularized Learning of Gaussian Mixture Models
Shamsiiat Abdurakhmanova, Alex Jung
https://arxiv.org/abs/2509.13855 https://arxiv.org/pdf/2509.13855…
FlightDiffusion: Revolutionising Autonomous Drone Training with Diffusion Models Generating FPV Video
Valerii Serpiva, Artem Lykov, Faryal Batool, Vladislav Kozlovskiy, Miguel Altamirano Cabrera, Dzmitry Tsetserukou
https://arxiv.org/abs/2509.14082
Large Language Models Discriminate Against Speakers of German Dialects
Minh Duc Bui, Carolin Holtermann, Valentin Hofmann, Anne Lauscher, Katharina von der Wense
https://arxiv.org/abs/2509.13835
Interleaving Natural Language Prompting with Code Editing for Solving Programming Tasks with Generative AI Models
Victor-Alexandru P\u{a}durean, Paul Denny, Andrew Luxton-Reilly, Alkis Gotovos, Adish Singla
https://arxiv.org/abs/2509.14088
NIRVANA: Structured pruning reimagined for large language models compression
Mengting Ai, Tianxin Wei, Sirui Chen, Jingrui He
https://arxiv.org/abs/2509.14230 https://
Inspections by Taiwan's National Security Bureau (NSB) of five Chinese generative AI apps -- Deepseek, Doubao (豆包), Yiyan (文心一言), Tongyi (通義千問), and Yuanbao (騰訊元寶) -- found violations of users' communication security across several indicators.
https://focustaiwan.tw/cross-strait/2025…
Evolution of meta's llama models and parameter-efficient fine-tuning of large language models: a survey
Abdulhady Abas Abdullah, Arkaitz Zubiaga, Seyedali Mirjalili, Amir H. Gandomi, Fatemeh Daneshfar, Mohammadsadra Amini, Alan Salam Mohammed, Hadi Veisi
https://arxiv.org/abs/2510.12178
Hala Technical Report: Building Arabic-Centric Instruction & Translation Models at Scale
Hasan Abed Al Kader Hammoud, Mohammad Zbeeb, Bernard Ghanem
https://arxiv.org/abs/2509.14008
OpenAI launches FrontierScience, a benchmark to measure models' expert-level scientific reasoning with 700 questions, finding GPT-5.2 is its strongest model (OpenAI)
https://openai.com/index/frontierscience/
ST-LINK: Spatially-Aware Large Language Models for Spatio-Temporal Forecasting
Hyotaek Jeon, Hyunwook Lee, Juwon Kim, Sungahn Ko
https://arxiv.org/abs/2509.13753 https://…
Pre-Manipulation Alignment Prediction with Parallel Deep State-Space and Transformer Models
Motonari Kambara, Komei Sugiura
https://arxiv.org/abs/2509.13839 https://
Benchmarking foundation models for hyperspectral image classification: Application to cereal crop type mapping
Walid Elbarz, Mohamed Bourriz, Hicham Hajji, Hamd Ait Abdelali, Fran\c{c}ois Bourzeix
https://arxiv.org/abs/2510.11576
Automated Triaging and Transfer Learning of Incident Learning Safety Reports Using Large Language Representational Models
Peter Beidler, Mark Nguyen, Kevin Lybarger, Ola Holmberg, Eric Ford, John Kang
https://arxiv.org/abs/2509.13706
TFMAdapter: Lightweight Instance-Level Adaptation of Foundation Models for Forecasting with Covariates
Afrin Dange, Sunita Sarawagi
https://arxiv.org/abs/2509.13906 https://
Source: OpenAI rolled back ChatGPT's model router, which sent some queries to reasoning models, for Free and $5/month Go tiers, as it was costly and hurt DAUs (Maxwell Zeff/Wired)
https://www.wired.com/story/openai-router-relaunch-gpt-5-sam-altman/
Explicit Reasoning Makes Better Judges: A Systematic Study on Accuracy, Efficiency, and Robustness
Pratik Jayarao, Himanshu Gupta, Neeraj Varshney, Chaitanya Dwivedi
https://arxiv.org/abs/2509.13332
Evaluating Classical Software Process Models as Coordination Mechanisms for LLM-Based Software Generation
Duc Minh Ha, Phu Trac Kien, Tho Quan, Anh Nguyen-Duc
https://arxiv.org/abs/2509.13942
Crosslisted article(s) found for cs.CY. https://arxiv.org/list/cs.CY/new
[1/1]:
- All Models Are Wrong, But Can They Be Useful? Lessons from COVID-19 Agent-Based Models: A Systema...
Von Hoene, Von Hoene, Peter, Hopson, Csizmadia, Fenyk, Barner, Leslie, Kavak, Zufle, Roess, Anderson
…
Multi-robot Multi-source Localization in Complex Flows with Physics-Preserving Environment Models
Benjamin Shaffer, Victoria Edwards, Brooks Kinch, Nathaniel Trask, M. Ani Hsieh
https://arxiv.org/abs/2509.14228
State Space Models over Directed Graphs
Junzhi She, Xunkai Li, Rong-Hua Li, Guoren Wang
https://arxiv.org/abs/2509.13735 https://arxiv.org/pdf/2509.13735…
Apertus: Democratizing Open and Compliant LLMs for Global Language Environments
Alejandro Hern\'andez-Cano, Alexander H\"agele, Allen Hao Huang, Angelika Romanou, Antoni-Joan Solergibert, Barna Pasztor, Bettina Messmer, Dhia Garbaya, Eduard Frank \v{D}urech, Ido Hakimi, Juan Garc\'ia Giraldo, Mete Ismayilzada, Negar Foroutan, Skander Moalla, Tiancheng Chen, Vinko Sabol\v{c}ec, Yixuan Xu, Michael Aerni, Badr AlKhamissi, Ines Altemir Marinas, Mohammad Hossein Amani, Matin An…
CrowdAgent: Multi-Agent Managed Multi-Source Annotation System
Maosheng Qin, Renyu Zhu, Mingxuan Xia, Chenkai Chen, Zhen Zhu, Minmin Lin, Junbo Zhao, Lu Xu, Changjie Fan, Runze Wu, Haobo Wang
https://arxiv.org/abs/2509.14030
The Allen Institute of AI launches Bolmo 7B and Bolmo 1B, claiming they are "the first fully open byte-level language models", built on its Olmo 3 models (Emilia David/VentureBeat)
https://venturebeat.com/ai/bolmos-architecture-unlock…
From Distributional to Quantile Neural Basis Models: the case of Electricity Price Forecasting
Alessandro Brusaferri, Danial Ramin, Andrea Ballarino
https://arxiv.org/abs/2509.14113
DriveVLA-W0: World Models Amplify Data Scaling Law in Autonomous Driving
Yingyan Li, Shuyao Shang, Weisong Liu, Bing Zhan, Haochen Wang, Yuqi Wang, Yuntao Chen, Xiaoman Wang, Yasong An, Chufeng Tang, Lu Hou, Lue Fan, Zhaoxiang Zhang
https://arxiv.org/abs/2510.12796
TGPO: Tree-Guided Preference Optimization for Robust Web Agent Reinforcement Learning
Ziyuan Chen, Zhenghui Zhao, Zhangye Han, Miancan Liu, Xianhang Ye, Yiqing Li, Hongbo Min, Jinkui Ren, Xiantao Zhang, Guitao Cao
https://arxiv.org/abs/2509.14172
Teaching LLMs to Plan: Logical Chain-of-Thought Instruction Tuning for Symbolic Planning
Pulkit Verma, Ngoc La, Anthony Favier, Swaroop Mishra, Julie A. Shah
https://arxiv.org/abs/2509.13351
Mind the Ethics! The Overlooked Ethical Dimensions of GenAI in Software Modeling Education
Shalini Chakraborty, Lola Burgue\~no, Nathalie Moreno, Javier Troya, Paula Mu\~noz
https://arxiv.org/abs/2509.13896
Sources: Amazon is in talks to invest $10B in OpenAI at a $500B valuation, with OpenAI using AWS Trainium chips; Microsoft keeps rights to sell OpenAI models (The Information)
https://www.theinformation.com/articles/openai-talks-raise-least-1…
Imagined Autocurricula
Ahmet H. G\"uzel, Matthew Thomas Jackson, Jarek Luca Liesen, Tim Rockt\"aschel, Jakob Nicolaus Foerster, Ilija Bogunovic, Jack Parker-Holder
https://arxiv.org/abs/2509.13341
Personalized Federated Fine-Tuning of Vision Foundation Models for Healthcare
Adam Tupper, Christian Gagn\'e
https://arxiv.org/abs/2510.12741 https://a…
Google releases Cell2Sentence-Scale 27B (C2S-Scale), a 27B-parameter foundation model for single-cell analysis built on its Gemma family of open models (The Keyword)
https://blog.google/technology/ai/google-gemma-ai-cancer-therapy-discovery/
Canary-1B-v2 & Parakeet-TDT-0.6B-v3: Efficient and High-Performance Models for Multilingual ASR and AST
Monica Sekoyan, Nithin Rao Koluguri, Nune Tadevosyan, Piotr Zelasko, Travis Bartley, Nick Karpov, Jagadeesh Balam, Boris Ginsburg
https://arxiv.org/abs/2509.14128
Evaluating End-User Device Energy Models in Sustainability Reporting of Browser-Based Web Services
Maja H. Kirkeby, Timmie Lagermann
https://arxiv.org/abs/2510.12566 https://
Anthropic's Skills for Claude, which are conceptually very simple, may become a bigger deal than MCP, whose high token usage is its most significant limitation (Simon Willison/Simon Willison's Weblog)
https://simonwillison.net/2025/Oct/16/claude-skills/
DSCC-HS: A Dynamic Self-Reinforcing Framework for Hallucination Suppression in Large Language Models
Xiao Zheng
https://arxiv.org/abs/2509.13702 https://ar…
On the Use of Hierarchical Vision Foundation Models for Low-Cost Human Mesh Recovery and Pose Estimation
Shuhei Tarashima, Yushan Wang, Norio Tagawa
https://arxiv.org/abs/2510.12660
You Are What You Train: Effects of Data Composition on Training Context-aware Machine Translation Models
Pawe{\l} M\k{a}ka, Yusuf Can Semerci, Jan Scholtes, Gerasimos Spanakis
https://arxiv.org/abs/2509.14031
EvoCAD: Evolutionary CAD Code Generation with Vision Language Models
Tobias Preintner, Weixuan Yuan, Adrian K\"onig, Thomas B\"ack, Elena Raponi, Niki van Stein
https://arxiv.org/abs/2510.11631
OpenAI says the cyber capabilities of its frontier AI models are accelerating and warns that upcoming models are likely to pose a "high" risk (Ina Fried/Axios)
https://www.axios.com/2025/12/10/openai-new-models-cybersecurity-risks
An End-to-End Differentiable, Graph Neural Network-Embedded Pore Network Model for Permeability Prediction
Qingqi Zhao, Heng Xiao
https://arxiv.org/abs/2509.13841 https://
Improving Context Fidelity via Native Retrieval-Augmented Reasoning
Suyuchen Wang, Jinlin Wang, Xinyu Wang, Shiqi Li, Xiangru Tang, Sirui Hong, Xiao-Wen Chang, Chenglin Wu, Bang Liu
https://arxiv.org/abs/2509.13683
ThinkPilot: Steering Reasoning Models via Automated Think-prefixes Optimization
Sunzhu Li, Zhiyu Lin, Shuling Yang, Jiale Zhao, Wei Chen
https://arxiv.org/abs/2510.12063 https:/…
Hybrid Quantum-Classical Neural Networks for Few-Shot Credit Risk Assessment
Zheng-an Wang, Yanbo J. Wang, Jiachi Zhang, Qi Xu, Yilun Zhao, Jintao Li, Yipeng Zhang, Bo Yang, Xinkai Gao, Xiaofeng Cao, Kai Xu, Pengpeng Hao, Xuan Yang, Heng Fan
https://arxiv.org/abs/2509.13818
Teaching According to Talents! Instruction Tuning LLMs with Competence-Aware Curriculum Learning
Yangning Li, Tingwei Lu, Yinghui Li, Yankai Chen, Wei-Chieh Huang, Wenhao Jiang, Hui Wang, Hai-Tao Zheng, Philip S. Yu
https://arxiv.org/abs/2509.13790
Multitask finetuning and acceleration of chemical pretrained models for small molecule drug property prediction
Matthew Adrian, Yunsie Chung, Kevin Boyd, Saee Paliwal, Srimukh Prasad Veccham, Alan C. Cheng
https://arxiv.org/abs/2510.12719
DiffEM: Learning from Corrupted Data with Diffusion Models via Expectation Maximization
Danial Hosseintabar, Fan Chen, Giannis Daras, Antonio Torralba, Constantinos Daskalakis
https://arxiv.org/abs/2510.12691
Mitigating the Noise Shift for Denoising Generative Models via Noise Awareness Guidance
Jincheng Zhong, Boyuan Jiang, Xin Tao, Pengfei Wan, Kun Gai, Mingsheng Long
https://arxiv.org/abs/2510.12497
Representation-Based Exploration for Language Models: From Test-Time to Post-Training
Jens Tuyls, Dylan J. Foster, Akshay Krishnamurthy, Jordan T. Ash
https://arxiv.org/abs/2510.11686