Tootfinder

Opt-in global Mastodon full text search. Join the index!

No exact results. Similar results found.
@theodric@social.linux.pizza
2025-06-09 10:13:53

What's the angle here? California belongs to Mexico? In Communism, we will drive our own taxis? Baiting the people who love Evola with the defeat of martial order by the unwashed and unchecked proletariat? Kali Yuga fashion show? Just there for the violence? Strange times we live in.

A shirtless dude in skull pants and a hat with a gater around his face standing on top of a smashed-up Waymo car (with "evola" graffitied on the back) and waving a Mexican flag while surrounded by smoke and fire (just normal California things, yawn)
@arXiv_eessSP_bot@mastoxiv.page
2025-07-01 11:46:43

Integrated Polarimetric Sensing and Communication with Polarization-Reconfigurable Arrays
Byunghyun Lee, Rang Liu, David J. Love, James V. Krogmeier, A. Lee Swindlehurst
arxiv.org/abs/2506.23410

@arXiv_csLG_bot@mastoxiv.page
2025-07-11 10:23:51

Skip a Layer or Loop it? Test-Time Depth Adaptation of Pretrained LLMs
Ziyue Li, Yang Li, Tianyi Zhou
arxiv.org/abs/2507.07996 arxiv.org/pdf/2507.07996 arxiv.org/html/2507.07996
arXiv:2507.07996v1 Announce Type: new
Abstract: Can a pretrained neural network adapt its architecture to different inputs without any finetuning? Do we need all layers for simple tasks, and are they adequate for challenging tasks? We found that the layers of a pretrained large language model (LLM) can be manipulated as separate modules to build a better and even shallower model customized for each test sample. In particular, each layer from the pretrained model can be skipped/pruned or repeated multiple times as recurrent neural networks (RNN), and stacked with others in arbitrary orders, yielding a chain-of-layers (CoLa) per sample. This compositional space greatly expands the scope of existing works on looped/recurrent pretrained modules, layer pruning, or early-exit networks. We develop a Monte Carlo Tree Search (MCTS) protocol to explore and identify the optimal CoLa for each sample from math and commonsense reasoning benchmarks. Compared to a static model of a fixed depth, CoLa allows shortcut paths (fast thinking), recurrence of the same layer(s) (slow thinking), and combining both, offering more flexible, dynamic architectures for different inputs. We conduct an extensive analysis of the MCTS-optimized CoLa, which leads to two key findings: (1) For >75% of samples with correct predictions by the original LLM, we can find shorter CoLa, suggesting a large space for improving inference efficiency; (2) For >60% of samples with originally incorrect predictions, we can identify CoLa achieving correct predictions, suggesting a large space of performance enhancement. Our results highlight the shortcomings of using a fixed architecture of pre-trained LLMs for inference on different samples and pave the way to unlock the generalization power of test-time depth adaptation.
toXiv_bot_toot

@arXiv_grqc_bot@mastoxiv.page
2025-07-04 09:51:11

Primordial regular black holes as all the dark matter. III. Covariant canonical quantum gravity models
Marco Calz\`a, Davide Pedrotti, Guan-Wen Yuan, Sunny Vagnozzi
arxiv.org/abs/2507.02396

@arXiv_csRO_bot@mastoxiv.page
2025-08-08 09:56:22

Mixed-Initiative Dialog for Human-Robot Collaborative Manipulation
Albert Yu, Chengshu Li, Luca Macesanu, Arnav Balaji, Ruchira Ray, Raymond Mooney, Roberto Mart\'in-Mart\'in
arxiv.org/abs/2508.05535

@arXiv_condmatmtrlsci_bot@mastoxiv.page
2025-06-03 08:05:10

Adsorbate phase transitions on nanoclusters from nested sampling
Thanawitch Chatbipho, Ray Yang, Robert B. Wexler, Livia B. P\'artay
arxiv.org/abs/2506.01295

@arXiv_eessSP_bot@mastoxiv.page
2025-06-03 07:35:10

Physics-based Generative Models for Geometrically Consistent and Interpretable Wireless Channel Synthesis
Satyavrat Wagle, Akshay Malhotra, Shahab Hamidi-Rad, Aditya Sant, David J. Love, Christopher G. Brinton
arxiv.org/abs/2506.00374