Tootfinder

Opt-in global Mastodon full text search. Join the index!

No exact results. Similar results found.
@arXiv_csCV_bot@mastoxiv.page
2025-06-30 10:16:50

MiCo: Multi-image Contrast for Reinforcement Visual Reasoning
Xi Chen, Mingkang Zhu, Shaoteng Liu, Xiaoyang Wu, Xiaogang Xu, Yu Liu, Xiang Bai, Hengshuang Zhao
arxiv.org/abs/2506.22434

@wfryer@mastodon.cloud
2025-05-27 18:20:20

“America’s Pivot to the Pacific” by #AngryPlanet [PODCAST]
pca.st/episode/535d72f1-3347-4

The image features a colorful illustration with the title "Angry Planet: Pivot to the Pacific." It includes cartoon-style representations of various elements: a person wearing headphones, military soldiers engaged in exercises, boats on water, and a depiction of the "Nine
@arXiv_csIR_bot@mastoxiv.page
2025-06-30 09:55:10

HLTCOE at LiveRAG: GPT-Researcher using ColBERT retrieval
Kevin Duh, Eugene Yang, Orion Weller, Andrew Yates, Dawn Lawrie
arxiv.org/abs/2506.22356

@arXiv_mathGT_bot@mastoxiv.page
2025-06-30 08:07:50

Toroidal graph manifolds with small homology are not SU(2)-abelian
Giacomo Bascape
arxiv.org/abs/2506.21729 arxiv.org…

@rberger@hachyderm.io
2025-04-27 19:48:27

"I think it is a huge mistake for people to assume that they can trust AI when they do not trust each other. The safest way to develop superintelligence is to first strengthen trust between humans, and then cooperate with each other to develop superintelligence in a safe manner. But what we are doing now is exactly the opposite. Instead, all efforts are being directed toward developing a superintelligence."
#AGI #AI
wired.com/story/questions-answ

@arXiv_csAI_bot@mastoxiv.page
2025-06-24 10:50:20

Action Language BC
Joseph Babb, Joohyung Lee
arxiv.org/abs/2506.18044 arxiv.org/pdf/2506.18044

@arXiv_astrophEP_bot@mastoxiv.page
2025-05-30 07:29:07

Detecting Atmospheric CO2 Trends as Population-Level Signatures for Long-Term Stable Water Oceans and Biotic Activity on Temperate Terrestrial Exoplanets
Janina Hansen, Daniel Angerhausen, Sascha P. Quanz, Derek Vance, Bj\"orn S. Konrad, Emily O. Garvin, Eleonora Alei, Jens Kammerer, Felix A. Dannert
arxiv.org/abs/2…

@arXiv_csCL_bot@mastoxiv.page
2025-06-27 09:59:49

Potemkin Understanding in Large Language Models
Marina Mancoridis, Bec Weeks, Keyon Vafa, Sendhil Mullainathan
arxiv.org/abs/2506.21521 arxiv.org/pdf/2506.21521 arxiv.org/html/2506.21521
arXiv:2506.21521v1 Announce Type: new
Abstract: Large language models (LLMs) are regularly evaluated using benchmark datasets. But what justifies making inferences about an LLM's capabilities based on its answers to a curated set of questions? This paper first introduces a formal framework to address this question. The key is to note that the benchmarks used to test LLMs -- such as AP exams -- are also those used to test people. However, this raises an implication: these benchmarks are only valid tests if LLMs misunderstand concepts in ways that mirror human misunderstandings. Otherwise, success on benchmarks only demonstrates potemkin understanding: the illusion of understanding driven by answers irreconcilable with how any human would interpret a concept. We present two procedures for quantifying the existence of potemkins: one using a specially designed benchmark in three domains, the other using a general procedure that provides a lower-bound on their prevalence. We find that potemkins are ubiquitous across models, tasks, and domains. We also find that these failures reflect not just incorrect understanding, but deeper internal incoherence in concept representations.
toXiv_bot_toot

@arXiv_mathAG_bot@mastoxiv.page
2025-06-27 09:20:29

The Pythagoras number of fields of transcendence degree $1$ over $\mathbb{Q}$
Olivier Benoist
arxiv.org/abs/2506.21380

@arXiv_mathGR_bot@mastoxiv.page
2025-06-27 07:59:29

Connections between hyperlinearity, stability and character rigidity for higher rank lattices
Alon Dogon, Itamar Vigdorovich
arxiv.org/abs/2506.20843