Tootfinder

Opt-in global Mastodon full text search. Join the index!

No exact results. Similar results found.
@datascience@genomic.social
2025-05-28 10:00:01

Use multi level models with {parsnip}: multilevelmod.tidymodels.org/ #rstats #ML

@awinkler@openbiblio.social
2025-05-27 20:44:46

via theguardian.com/technology/202:
Matteo Valleriani: "It is time to build public, open-access LLMs for the humanities…

@Techmeme@techhub.social
2025-06-26 16:45:51

Source: Meta has hired highly influential OpenAI researcher Trapit Bansal to work on its AI reasoning models under the company's new AI superintelligence unit (Maxwell Zeff/TechCrunch)
techcrunch.com/2025/06/26/meta

@arXiv_qfinTR_bot@mastoxiv.page
2025-05-27 07:51:55

Hybrid Models for Financial Forecasting: Combining Econometric, Machine Learning, and Deep Learning Models
Dominik Stempie\'n, Robert \'Slepaczuk
arxiv.org/abs/2505.19617

@lysander07@sigmoid.social
2025-05-28 05:10:40

Last week, we continued our #ISE2025 lecture on distributional semantics with the introduction of neural language models (NLMs) and compared them to traditional statistical n-gram models.
Benefits of NLMs:
- Capturing Long-Range Dependencies
- Computational and Statistical Tractability
- Improved Generalisation
- Higher Accuracy
@…

The image illustrates the architecture of a Neural Language Model, specifically focusing on Word Vectors II - Neural Language Models. It is part of a presentation on Natural Language Processing, created by the Karlsruhe Institute of Technology (KIT) and FIZ Karlsruhe, as indicated by their logos in the top right corner.

The diagram shows a neural network processing an input word embedding, represented by the phrase "to be or not to." The input is transformed into a d-sized vector representatio…
@arXiv_csCR_bot@mastoxiv.page
2025-06-27 08:35:29

CodeGuard: A Generalized and Stealthy Backdoor Watermarking for Generative Code Models
Haoxuan Li, Jiale Zhang, Xiaobing Sun, Xiapu Luo
arxiv.org/abs/2506.20926

@arXiv_csSE_bot@mastoxiv.page
2025-06-27 08:47:59

$T^3$: Multi-level Tree-based Automatic Program Repair with Large Language Models
Quanming Liu, Xupeng Bu, Zhichao Yan, Ru Li
arxiv.org/abs/2506.21211

@arXiv_csRO_bot@mastoxiv.page
2025-06-27 08:58:29

Parallels Between VLA Model Post-Training and Human Motor Learning: Progress, Challenges, and Trends
Tian-Yu Xiang, Ao-Qun Jin, Xiao-Hu Zhou, Mei-Jiang Gui, Xiao-Liang Xie, Shi-Qi Liu, Shuang-Yi Wang, Sheng-Bin Duan, Fu-Chao Xie, Wen-Kai Wang, Si-Cheng Wang, Ling-Yun Li, Tian Tu, Zeng-Guang Hou
arxiv.org/abs/2506.20966…

@pbloem@sigmoid.social
2025-06-26 10:56:22

After training, we finetune on real-world data. We observe that the models that have been pre-trained with noise converge very quickly compared to a baseline which is trained from scratch.
Moreover, on the other datasets, the UP models retain their zero-shot performance during finetuning. This suggests that there may be a generalization benefit to using a UP model.
All this is at the expense of much longer training, but that cost can be amortized over many tasks.

The results for the finetuning experiment. Six datasets (linux, code, dyck, wp, german and ndfa) and the performance of four models: the baseline and UP trained models and two finetuning datasets. 

The results show that the UP models converge quicker, and that they retain most of their zero-shot performance on the other datasets.
@arXiv_condmatmtrlsci_bot@mastoxiv.page
2025-06-27 09:46:59

Crystallization of metallic glass as a grain-boundary nucleated process: experimental and theoretical evidence for the grain structure of metallic glasses
Nikolay V. Alekseechkin
arxiv.org/abs/2506.21261