
2025-06-19 08:07:08
Private Continual Counting of Unbounded Streams
Ben Jacobsen, Kassem Fawaz
https://arxiv.org/abs/2506.15018 https://arxiv.org/pdf/250…
Private Continual Counting of Unbounded Streams
Ben Jacobsen, Kassem Fawaz
https://arxiv.org/abs/2506.15018 https://arxiv.org/pdf/250…
Heisenberg limited multiple eigenvalue estimation via off-the-grid compressed sensing
Davide Castaldo, Stefano Corni
https://arxiv.org/abs/2507.12438 https…
Eigenstate Thermalization Hypothesis (ETH) for off-diagonal matrix elements in integrable spin chains
Federico Rottoli, Vincenzo Alba
https://arxiv.org/abs/2505.23602
Low-rank Momentum Factorization for Memory Efficient Training
Pouria Mahdavinia, Mehrdad Mahdavi
https://arxiv.org/abs/2507.08091 https://arxiv.org/pdf/2507.08091 https://arxiv.org/html/2507.08091
arXiv:2507.08091v1 Announce Type: new
Abstract: Fine-tuning large foundation models presents significant memory challenges due to stateful optimizers like AdamW, often requiring several times more GPU memory than inference. While memory-efficient methods like parameter-efficient fine-tuning (e.g., LoRA) and optimizer state compression exist, recent approaches like GaLore bridge these by using low-rank gradient projections and subspace moment accumulation. However, such methods may struggle with fixed subspaces or computationally costly offline resampling (e.g., requiring full-matrix SVDs). We propose Momentum Factorized SGD (MoFaSGD), which maintains a dynamically updated low-rank SVD representation of the first-order momentum, closely approximating its full-rank counterpart throughout training. This factorization enables a memory-efficient fine-tuning method that adaptively updates the optimization subspace at each iteration. Crucially, MoFaSGD leverages the computed low-rank momentum factors to perform efficient spectrally normalized updates, offering an alternative to subspace moment accumulation. We establish theoretical convergence guarantees for MoFaSGD, proving it achieves an optimal rate for non-convex stochastic optimization under standard assumptions. Empirically, we demonstrate MoFaSGD's effectiveness on large language model alignment benchmarks, achieving a competitive trade-off between memory reduction (comparable to LoRA) and performance compared to state-of-the-art low-rank optimization methods. Our implementation is available at https://github.com/pmahdavi/MoFaSGD.
toXiv_bot_toot
This https://arxiv.org/abs/2506.03074 has been replaced.
initial toot: https://mastoxiv.page/@arXiv_sta…
This https://arxiv.org/abs/2202.02835 has been replaced.
link: https://scholar.google.com/scholar?q=a
Efficient Training for Optical Computing
Manon P. Bart, Nick Sparks, Ryan T. Glasser
https://arxiv.org/abs/2506.20833 https://arxiv.o…