Tootfinder

Opt-in global Mastodon full text search. Join the index!

@heiseonline@social.heise.de
2025-09-18 16:36:00

Deepseek-R1: KI-Training hat sogar weniger als 300.000 US-Dollar gekostet
Die Konkurrenzfähigkeit der KI-Modelle von Deepseek hat Anfang des Jahres die KI-Branche schockiert. Jetzt gibt es erstmals konkrete Informationen zum Training.

@Techmeme@techhub.social
2025-09-20 13:41:48

Huawei says DeepSeek-R1-Safe, which was trained on 1,000 of its Ascend AI chips, is "nearly 100% successful" in preventing politically sensitive topics (Eduardo Baptista/Reuters)
reuters.com/business/media-tel

@Techmeme@techhub.social
2025-08-21 09:35:58

DeepSeek details V3.1 and says it surpasses R1 on key benchmarks and is customized to work with next-gen Chinese-made AI chips, after unveiling it on August 19 (Bloomberg)
bloomberg.com/news/articles/20

@arXiv_csCY_bot@mastoxiv.page
2025-07-21 08:11:50

The Emperor's New Chain-of-Thought: Probing Reasoning Theater Bias in Large Reasoning Models
Qian Wang, Yubo Fan, Zhenheng Tang, Nuo Chen, Wenxuan Wang, Bingsheng He
arxiv.org/abs/2507.13758

@Techmeme@techhub.social
2025-09-18 13:10:41

In a peer-reviewed Nature article, DeepSeek says it has spent $294,000 on training its R1 model and used 512 Nvidia H800 chips (Eduardo Baptista/Reuters)
reuters.com/world/china/chinas

@thomasrenkert@hcommons.social
2025-08-14 14:23:51

The geopolitical #aiarmsrace seems largely unimpressed by people proclaiming #LLMs have plateaued and #AGI is never coming.
Such assessments are only relevant for the market, but not so much for count…

Chinese artificial intelligence company DeepSeek delayed the release of its new model after failing to train it using Huawei’s chips, highlighting the limits of Beijing’s push to replace US technology.

DeepSeek was encouraged by authorities to adopt Huawei’s Ascend processor rather than use Nvidia’s systems after releasing its R1 model in January, according to three people familiar with the matter.
@arXiv_csSE_bot@mastoxiv.page
2025-09-18 09:38:01

A Study on Thinking Patterns of Large Reasoning Models in Code Generation
Kevin Halim, Sin G. Teo, Ruitao Feng, Zhenpeng Chen, Yang Gu, Chong Wang, Yang Liu
arxiv.org/abs/2509.13758

@arXiv_csAI_bot@mastoxiv.page
2025-08-14 08:59:12

Mathematical Computation and Reasoning Errors by Large Language Models
Liang Zhang, Edith Aurora Graf
arxiv.org/abs/2508.09932 arxiv.org/pd…

@arXiv_csLG_bot@mastoxiv.page
2025-10-14 13:40:48

MATH-Beyond: A Benchmark for RL to Expand Beyond the Base Model
Prasanna Mayilvahanan, Ricardo Dominguez-Olmedo, Thadd\"aus Wiedemer, Wieland Brendel
arxiv.org/abs/2510.11653

@arXiv_csCL_bot@mastoxiv.page
2025-09-30 14:06:31

MobileLLM-R1: Exploring the Limits of Sub-Billion Language Model Reasoners with Open Training Recipes
Changsheng Zhao, Ernie Chang, Zechun Liu, Chia-Jung Chang, Wei Wen, Chen Lai, Rick Cao, Yuandong Tian, Raghuraman Krishnamoorthi, Yangyang Shi, Vikas Chandra
arxiv.org/abs/2509.24945

@timfoster@mastodon.social
2025-08-30 17:44:18

Compare and buy tulips! One monthly subscription to get the very best tulip colours for all your tulip needs!
store.boingboing.net/sales/cha

@arXiv_csLG_bot@mastoxiv.page
2025-09-11 10:14:13

Sharing is Caring: Efficient LM Post-Training with Collective RL Experience Sharing
Jeffrey Amico, Gabriel Passamani Andrade, John Donaghy, Ben Fielding, Tristin Forbus, Harry Grieve, Semih Kara, Jari Kolehmainen, Yihua Lou, Christopher Nies, Edward Phillip Flores Nu\~no, Diogo Ortega, Shikhar Rastogi, Austin Virts, Matthew J. Wright

@arXiv_csAI_bot@mastoxiv.page
2025-10-10 07:33:08

Base Models Know How to Reason, Thinking Models Learn When
Constantin Venhoff, Iv\'an Arcuschin, Philip Torr, Arthur Conmy, Neel Nanda
arxiv.org/abs/2510.07364

@Techmeme@techhub.social
2025-10-10 20:26:02

SemiAnalysis launches InferenceMAX, an open-source benchmark that automatically tracks LLM inference performance across AI models and frameworks every night (Kimbo Chen/SemiAnalysis)
newsletter.semianalysis.com/p/

@arXiv_csCL_bot@mastoxiv.page
2025-10-09 10:26:21

Overview of the Plagiarism Detection Task at PAN 2025
Andr\'e Greiner-Petter, Maik Fr\"obe, Jan Philip Wahle, Terry Ruas, Bela Gipp, Akiko Aizawa, Martin Potthast
arxiv.org/abs/2510.06805

@arXiv_csHC_bot@mastoxiv.page
2025-09-30 11:31:31

Diamonds in the rough: Transforming SPARCs of imagination into a game concept by leveraging medium sized LLMs
Julian Geheeb, Farhan Abid Ivan, Daniel Dyrda, Miriam Ansch\"utz, Georg Groh
arxiv.org/abs/2509.24730

@arXiv_csAI_bot@mastoxiv.page
2025-10-10 10:31:19

R-Horizon: How Far Can Your Large Reasoning Model Really Go in Breadth and Depth?
Yi Lu, Jianing Wang, Linsen Guo, Wei He, Hongyin Tang, Tao Gui, Xuanjing Huang, Xuezhi Cao, Wei Wang, Xunliang Cai
arxiv.org/abs/2510.08189

@arXiv_physicsmedph_bot@mastoxiv.page
2025-09-11 08:33:13

An Iterative LLM Framework for SIBT utilizing RAG-based Adaptive Weight Optimization
Zhuo Xiao (Image Processing Center, Beihang University, Beijing, China), Qinglong Yao (Image Processing Center, Beihang University, Beijing, China), Jingjing Wang (Image Processing Center, Beihang University, Beijing, China), Fugen Zhou (Image Processing Center, Beihang University, Beijing, China), Bo Liu (Image Processing Center, Beihang University, Beijing, China), Haitao Sun (Department of Radiation…

@arXiv_csNI_bot@mastoxiv.page
2025-09-29 08:36:17

Evaluating Open-Source Large Language Models for Technical Telecom Question Answering
Arina Caraus, Alessio Buscemi, Sumit Kumar, Ion Turcanu
arxiv.org/abs/2509.21949

@arXiv_csCL_bot@mastoxiv.page
2025-08-26 11:57:26

DRQA: Dynamic Reasoning Quota Allocation for Controlling Overthinking in Reasoning Large Language Models
Kaiwen Yan, Xuanqing Shi, Hongcheng Guo, Wenxuan Wang, Zhuosheng Zhang, Chengwei Qin
arxiv.org/abs/2508.17803

@arXiv_csAI_bot@mastoxiv.page
2025-08-28 07:36:40

SLIM: Subtrajectory-Level Elimination for More Effective Reasoning
Xifeng Yao, Chengyuan Ma, Dongyu Lang, Yinhao Ni, Zhiwei Xu, Huarui Xie, Zihao Chen, Guang Shen, Dandan Tu, Yi Bai, Changzheng Zhang
arxiv.org/abs/2508.19502

@arXiv_csCL_bot@mastoxiv.page
2025-07-22 12:24:50

The Impact of Language Mixing on Bilingual LLM Reasoning
Yihao Li, Jiayi Xin, Miranda Muqing Miao, Qi Long, Lyle Ungar
arxiv.org/abs/2507.15849