Nature: Cortical circuits for cross-modal generalization (in mice) https://www.nature.com/articles/s41467-025-59342-9 "Optogenetic sensory substitution and systematic silencing of these associative areas revealed that a single area in the dorsal stream is necessary an…
Schön war die heutige äh gestrige #CriticalMass in #Zürich. Und mit 25km auch etwas länger als sonst. Mit An- und Abreise waren's für mich 33.5km, die wegen Home-Office heute gut aufs #BikeToWork…
Judge rules DOGE can access personal data as unions claim invasion of privacy | Courthouse News Service
https://www.courthousenews.com/judge-rules-doge-can-access-personal-data-as-unions-claim-invasion-of-privacy/
Double-Checker: Enhancing Reasoning of Slow-Thinking LLMs via Self-Critical Fine-Tuning
Xin Xu, Tianhao Chen, Fan Zhang, Wanlong Liu, Pengxiang Li, Ajay Kumar Jaiswal, Yuchen Yan, Jishan Hu, Yang Wang, Hao Chen, Shiwei Liu, Shizhe Diao, Can Yang, Lu Yin
https://arxiv.org/abs/2506.21285 https://arxiv.org/pdf/2506.21285 https://arxiv.org/html/2506.21285
arXiv:2506.21285v1 Announce Type: new
Abstract: While slow-thinking large language models (LLMs) exhibit reflection-like reasoning, commonly referred to as the "aha moment:, their ability to generate informative critiques and refine prior solutions remains limited. In this paper, we introduce Double-Checker, a principled framework designed to enhance the reasoning capabilities of slow-thinking LLMs by fostering explicit self-critique and iterative refinement of their previous solutions. By fine-tuning on our curated 1,730 self-critical instances, Double-Checker empowers long-CoT LLMs to iteratively critique and refine their outputs during inference until they evaluate their solutions as correct under self-generated critiques. We validate the efficacy of Double-Checker across a comprehensive suite of reasoning benchmarks, demonstrating that iterative self-critique significantly enhances the reasoning capabilities of long-CoT LLMs. Notably, our Double-Checker increases the pass@1 performance on challenging AIME benchmarks from 4.4% to 18.2% compared to the original long-CoT LLMs. These results highlight a promising direction for developing more trustworthy and effective LLMs capable of structured self-critique.
toXiv_bot_toot
Workers at Google, TikTok, Adobe, Dropbox, CrowdStrike, and other tech firms recount how managers used AI to justify firing them, speed up their work, and more (Brian Merchant/Blood in the Machine)
https://www.bloodinthemachine.com/p/how-ai-is-killing-jo…
Global and Local Contrastive Learning for Joint Representations from Cardiac MRI and ECG
Alexander Selivanov, Philip M\"uller, \"Ozg\"un Turgut, Nil Stolt-Ans\'o, Daniel R\"uckert
https://arxiv.org/abs/2506.20683
It's Friday and don't stop work for the weekend before you check out today's Metacurity for the most critical infosec developments you should know, including
--Cyber incident disrupts Hawaiian Airlines, but flights are unaffected
--Food distributor UNFI restores operations,
--Danish gov't wants people to own copyrights to their bodies, faces and voices,
--N. Korea is automating crypto theft with AI tools,
--Pro-Iranian hacktivists leaked Saudi Ga…
Object knowledge representation in the human visual cortex requires a connection with the language system https://journals.plos.org/plosbiology/article?id=10.1371/journal.pbio.3003161 "Our experiments reveal the contribution of the vision-la…
All that killing for nothing:
Court rejects Netanyahu’s call to postpone graft trial hearings | Courthouse News Service
https://www.courthousenews.com/court-rejects-netanyahus-call-to-postpone-graft-trial-hearings/