Tootfinder

Opt-in global Mastodon full text search. Join the index!

@arXiv_csDC_bot@mastoxiv.page
2025-05-30 07:17:07

MemAscend: System Memory Optimization for SSD-Offloaded LLM Fine-Tuning
Yong-Cheng Liaw, Shuo-Han Chen
arxiv.org/abs/2505.23254

@arXiv_csDL_bot@mastoxiv.page
2025-05-26 07:17:10

Towards Industrial Convergence : Understanding the evolution of scientific norms and practices in the field of AI
Antoine Houssard
arxiv.org/abs/2505.17945

@berlinbuzzwords@floss.social
2025-05-14 14:00:33

LLMs are now part of our daily work, making coding easier. Join Ivan Dolgov at this year's Berlin Buzzwords to learn how they built an in-house LLM for AI code completion in JetBrains products, covering design choices, data preparation, training and model evaluation.
Learn more:

Session title: How to train a fast LLM for coding tasks
Ivan Dolgov
Join us from June 15-17 in Berlin or participate online / berlinbuzzwords.de
@jeang3nie@social.linux.pizza
2025-05-19 20:37:00

This morning I null routed another dozen IP addresses for scraping my personal git server using repeated http requests. As per usual, a quick inspection reveals that at least some of them are scraping for LLM data. As always, I have not consented to this use of my non-maintained code, experiments, college coursework, and miscellaneous crap that I for whatever reason decided to self host rather than pushing it to Codeberg.
I mean, if you really want to feed your LLM on a diet that inclu…

@arXiv_qbioGN_bot@mastoxiv.page
2025-05-21 07:36:42

OmniGenBench: A Modular Platform for Reproducible Genomic Foundation Models Benchmarking
Heng Yang, Jack Cole, Yuan Li, Renzhi Chen, Geyong Min, Ke Li
arxiv.org/abs/2505.14402