Tootfinder

Opt-in global Mastodon full text search. Join the index!

No exact results. Similar results found.
@arXiv_csDC_bot@mastoxiv.page
2025-06-13 07:38:30

Graph-based Gossiping for Communication Efficiency in Decentralized Federated Learning
Huong Nguyen, Hong-Tri Nguyen, Praveen Kumar Donta, Susanna Pirttikangas, Lauri Lov\'en
arxiv.org/abs/2506.10607

@arXiv_csLG_bot@mastoxiv.page
2025-06-12 08:42:01

FLoRIST: Singular Value Thresholding for Efficient and Accurate Federated Fine-Tuning of Large Language Models
Hariharan Ramesh, Jyotikrishna Dass
arxiv.org/abs/2506.09199

@arXiv_csCR_bot@mastoxiv.page
2025-06-12 07:30:51

Physical Layer-Based Device Fingerprinting for Wireless Security: From Theory to Practice
Junqing Zhang, Francesco Ardizzon, Mattia Piana, Guanxiong Shen, Stefano Tomasin
arxiv.org/abs/2506.09807

@arXiv_quantph_bot@mastoxiv.page
2025-06-10 18:53:50

This arxiv.org/abs/2505.16457 has been replaced.
initial toot: mastoxiv.page/@arXiv_qu…

@arXiv_mathOC_bot@mastoxiv.page
2025-06-11 09:59:15

Gradient flow in the kernel learning problem
Yang Li, Feng Ruan
arxiv.org/abs/2506.08550 arxiv.org/pdf/2506.08550

@arXiv_csIT_bot@mastoxiv.page
2025-07-09 09:51:22

An Effective Equivalence Model of Analyzing PLS of Multiple Eavesdroppers Facing Low-altitude Communication Systems
Yujia Zhao, Zhiyong Feng, Kan Yu, Qixun Zhang, Dong Li
arxiv.org/abs/2507.05878

@arXiv_csLG_bot@mastoxiv.page
2025-07-11 10:23:51

Skip a Layer or Loop it? Test-Time Depth Adaptation of Pretrained LLMs
Ziyue Li, Yang Li, Tianyi Zhou
arxiv.org/abs/2507.07996 arxiv.org/pdf/2507.07996 arxiv.org/html/2507.07996
arXiv:2507.07996v1 Announce Type: new
Abstract: Can a pretrained neural network adapt its architecture to different inputs without any finetuning? Do we need all layers for simple tasks, and are they adequate for challenging tasks? We found that the layers of a pretrained large language model (LLM) can be manipulated as separate modules to build a better and even shallower model customized for each test sample. In particular, each layer from the pretrained model can be skipped/pruned or repeated multiple times as recurrent neural networks (RNN), and stacked with others in arbitrary orders, yielding a chain-of-layers (CoLa) per sample. This compositional space greatly expands the scope of existing works on looped/recurrent pretrained modules, layer pruning, or early-exit networks. We develop a Monte Carlo Tree Search (MCTS) protocol to explore and identify the optimal CoLa for each sample from math and commonsense reasoning benchmarks. Compared to a static model of a fixed depth, CoLa allows shortcut paths (fast thinking), recurrence of the same layer(s) (slow thinking), and combining both, offering more flexible, dynamic architectures for different inputs. We conduct an extensive analysis of the MCTS-optimized CoLa, which leads to two key findings: (1) For >75% of samples with correct predictions by the original LLM, we can find shorter CoLa, suggesting a large space for improving inference efficiency; (2) For >60% of samples with originally incorrect predictions, we can identify CoLa achieving correct predictions, suggesting a large space of performance enhancement. Our results highlight the shortcomings of using a fixed architecture of pre-trained LLMs for inference on different samples and pave the way to unlock the generalization power of test-time depth adaptation.
toXiv_bot_toot

@arXiv_csCR_bot@mastoxiv.page
2025-07-10 09:35:11

A Survey on Artificial Noise for Physical Layer Security: Opportunities, Technologies, Guidelines, Advances, and Trends
Hong Niu, Yue Xiao, Xia Lei, Jiangong Chen, Zhihan Xiao, Mao Li, Chau Yuen
arxiv.org/abs/2507.06500

@arXiv_csIT_bot@mastoxiv.page
2025-07-09 09:45:22

Does Movable Antenna Present A Dual-edged Nature? From the Perspective of Physical Layer Security: A Joint Design of Fixed-position Antenna and Movable Antenna
Kan Yu, Wenxu Wang, Xiaowu Liu, Yujia Zhao, Qixun Zhang, Zhiyong Feng, Dong Li
arxiv.org/abs/2507.05784

@arXiv_csDC_bot@mastoxiv.page
2025-07-08 10:15:40

Demystifying NCCL: An In-depth Analysis of GPU Communication Protocols and Algorithms
Zhiyi Hu, Siyuan Shen, Tommaso Bonato, Sylvain Jeaugey, Cedell Alexander, Eric Spada, Jeff Hammond, Torsten Hoefler
arxiv.org/abs/2507.04786