Yes please! Supporting nonlinear careers to diversify science https://journals.plos.org/plosbiology/article?id=10.1371/journal.pbio.3002291
Series D, Episode 12 - Warlord
ORAC: Why not ask them?
TARRANT: [Into comm] This is Xenon base, identify yourselves. [static] I repeat, identify yourselves. [more static] You have ten seconds or I'll open fire.
https://blake.torpidity.net/m/412/23 B7B3
Excited about the new xLSTM model release. There are many well-though designs compared to transformers: recurrence (which should allows composability), gating (like Mamba & LSTM which is based on, which allows time complexity independent of the input size), state tracking (unlike Mamba & transformers). For now, these advantage aren’t apparent on benchmarks, but most training techniques are secrets, and the recent advances of LLMs evidenced that they matter a lot.