[OT] The force is with us, always? Tuning quantum vacuum forces from attractive to repulsive https://news.asu.edu/20190304-force-us-always-tuning-quantum-vacuum-forces-attractive-repulsive Now can we tune The vOICe soundscapes f…
Fine-Tuning LLMs For ‘Good’ Behavior Makes Them More Likely To Say No https://www.404media.co/fine-tuning-llms-cognitive-bias-decision-making-study/
Correlating instruction-tuning (in multimodal models) with vision-language processing (in the brain)
Subba Reddy Oota, Akshett Jindal, Ishani Mondal, Khushbu Pahwa, Satya Sai Srinath Namburi, Manish Shrivastava, Maneesh Singh, Bapi S. Raju, Manish Gupta
https://arxiv.org/abs/2505.20029
Exoplanet Ephemerides Change Observations (ExoEcho). II. Transit timing variation analysis of Brown Dwarfs around Solar-type Stars
Wenqin Wang, Xinyue Ma, Zhangliang Chen, Cong Yu, Shangfei Liu, Bo Ma
https://arxiv.org/abs/2505.21270
Amiguitos, hoy tenemos disco nuevo de la joven y talentosa neozelandesa Ella Marija Lani Yelich-O'Connor, también conocida como Lorde.
«Virgin»
https://open.spotify.com/album/28bHj2enHkHVFLwuWmkwlQ
Aus #Löwenzahnblüten kann man übrigens Sirup kochen, auch als Löwenzahnblütenhonig bezeichnet.
Mein Rezept siehe:https://www.oe…
Unlocking my own understanding of and ability to build #Swift macros feels like a superpower.
…something something great responsibility, though.
Synthesizing boilerplate and statically-verifiable elements like custom function calls based on macro input… is magic—the good kind.
`@GET("/logs/{userId}/{timing}")`
↘️
Exploring Adapter Design Tradeoffs for Low Resource Music Generation
Atharva Mehta, Shivam Chauhan, Monojit Choudhury
https://arxiv.org/abs/2506.21298 http…
Double-Checker: Enhancing Reasoning of Slow-Thinking LLMs via Self-Critical Fine-Tuning
Xin Xu, Tianhao Chen, Fan Zhang, Wanlong Liu, Pengxiang Li, Ajay Kumar Jaiswal, Yuchen Yan, Jishan Hu, Yang Wang, Hao Chen, Shiwei Liu, Shizhe Diao, Can Yang, Lu Yin
https://arxiv.org/abs/2506.21285 https://arxiv.org/pdf/2506.21285 https://arxiv.org/html/2506.21285
arXiv:2506.21285v1 Announce Type: new
Abstract: While slow-thinking large language models (LLMs) exhibit reflection-like reasoning, commonly referred to as the "aha moment:, their ability to generate informative critiques and refine prior solutions remains limited. In this paper, we introduce Double-Checker, a principled framework designed to enhance the reasoning capabilities of slow-thinking LLMs by fostering explicit self-critique and iterative refinement of their previous solutions. By fine-tuning on our curated 1,730 self-critical instances, Double-Checker empowers long-CoT LLMs to iteratively critique and refine their outputs during inference until they evaluate their solutions as correct under self-generated critiques. We validate the efficacy of Double-Checker across a comprehensive suite of reasoning benchmarks, demonstrating that iterative self-critique significantly enhances the reasoning capabilities of long-CoT LLMs. Notably, our Double-Checker increases the pass@1 performance on challenging AIME benchmarks from 4.4% to 18.2% compared to the original long-CoT LLMs. These results highlight a promising direction for developing more trustworthy and effective LLMs capable of structured self-critique.
toXiv_bot_toot
Learnable Adaptive Time-Frequency Representation via Differentiable Short-Time Fourier Transform
Maxime Leiber, Yosra Marnissi, Axel Barrau, Sylvain Meignen, Laurent Massouli\'e
https://arxiv.org/abs/2506.21440