Tootfinder

Opt-in global Mastodon full text search. Join the index!

No exact results. Similar results found.
@jae@mastodon.me.uk
2025-05-28 10:01:36

This makes me a bit nervous as I really love this game. However some speculation I’ve seen is the Sasquatch could become a much-needed mascot for Apple Arcade.
macobserver.com/news/apple-acq

@lysander07@sigmoid.social
2025-05-28 05:10:40

Last week, we continued our #ISE2025 lecture on distributional semantics with the introduction of neural language models (NLMs) and compared them to traditional statistical n-gram models.
Benefits of NLMs:
- Capturing Long-Range Dependencies
- Computational and Statistical Tractability
- Improved Generalisation
- Higher Accuracy
@…

The image illustrates the architecture of a Neural Language Model, specifically focusing on Word Vectors II - Neural Language Models. It is part of a presentation on Natural Language Processing, created by the Karlsruhe Institute of Technology (KIT) and FIZ Karlsruhe, as indicated by their logos in the top right corner.

The diagram shows a neural network processing an input word embedding, represented by the phrase "to be or not to." The input is transformed into a d-sized vector representatio…
@arXiv_csDL_bot@mastoxiv.page
2025-05-27 13:28:35

This arxiv.org/abs/2502.19679 has been replaced.
initial toot: mastoxiv.page/@arXiv_csDL_…

@arXiv_qbioNC_bot@mastoxiv.page
2025-05-28 07:36:56

Resonance Complexity Theory and the Architecture of Consciousness: A Field-Theoretic Model of Resonant Interference and Emergent Awareness
Michael Arnold Bruna
arxiv.org/abs/2505.20580

@arXiv_csCL_bot@mastoxiv.page
2025-06-27 09:58:09

TopK Language Models
Ryosuke Takahashi, Tatsuro Inaba, Kentaro Inui, Benjamin Heinzerling
arxiv.org/abs/2506.21468 arxiv.org/pdf/2506.21468 arxiv.org/html/2506.21468
arXiv:2506.21468v1 Announce Type: new
Abstract: Sparse autoencoders (SAEs) have become an important tool for analyzing and interpreting the activation space of transformer-based language models (LMs). However, SAEs suffer several shortcomings that diminish their utility and internal validity. Since SAEs are trained post-hoc, it is unclear if the failure to discover a particular concept is a failure on the SAE's side or due to the underlying LM not representing this concept. This problem is exacerbated by training conditions and architecture choices affecting which features an SAE learns. When tracing how LMs learn concepts during training, the lack of feature stability also makes it difficult to compare SAEs features across different checkpoints. To address these limitations, we introduce a modification to the transformer architecture that incorporates a TopK activation function at chosen layers, making the model's hidden states equivalent to the latent features of a TopK SAE. This approach eliminates the need for post-hoc training while providing interpretability comparable to SAEs. The resulting TopK LMs offer a favorable trade-off between model size, computational efficiency, and interpretability. Despite this simple architectural change, TopK LMs maintain their original capabilities while providing robust interpretability benefits. Our experiments demonstrate that the sparse representations learned by TopK LMs enable successful steering through targeted neuron interventions and facilitate detailed analysis of neuron formation processes across checkpoints and layers. These features make TopK LMs stable and reliable tools for understanding how language models learn and represent concepts, which we believe will significantly advance future research on model interpretability and controllability.
toXiv_bot_toot

@arXiv_csAR_bot@mastoxiv.page
2025-05-29 07:15:28

Refining Datapath for Microscaling ViTs
Can Xiao, Jianyi Cheng, Aaron Zhao
arxiv.org/abs/2505.22194 arxiv.org/pdf/250…

Some wear balaclavas. Some wearneck gators, sunglasses and hats. Some wear masks and casual clothes.
Across the country, armed federal immigration officers have increasingly hidden their identities while carrying out immigration raids, arresting protesters and roughing upprominent Democratic critics.
It’s a trend that has sparked alarm among civil rights and law enforcement experts alike.
Mike German, a former FBIagent, said officers’ widespread use of masks was unprecedent…

@arXiv_csDS_bot@mastoxiv.page
2025-06-27 07:32:48

Adaptive Hybrid Sort: Dynamic Strategy Selection for Optimal Sorting Across Diverse Data Distributions
Shrinivass Arunachalam Balasubramanian
arxiv.org/abs/2506.20677

@arXiv_csCV_bot@mastoxiv.page
2025-06-27 10:20:39

WAFT: Warping-Alone Field Transforms for Optical Flow
Yihan Wang, Jia Deng
arxiv.org/abs/2506.21526 arxiv.org/pdf/250…

@arXiv_csAR_bot@mastoxiv.page
2025-05-27 13:26:06

This arxiv.org/abs/2505.16968 has been replaced.
initial toot: mastoxiv.page/@arXiv_csAR_…