
2025-06-13 08:31:00
Data-driven balanced truncation for second-order systems with generalized proportional damping
Sean Reiter, Steffen W. R. Werner
https://arxiv.org/abs/2506.10118
Data-driven balanced truncation for second-order systems with generalized proportional damping
Sean Reiter, Steffen W. R. Werner
https://arxiv.org/abs/2506.10118
Terabyte-Scale Analytics in the Blink of an Eye
Bowen Wu, Wei Cui, Carlo Curino, Matteo Interlandi, Rathijit Sen
https://arxiv.org/abs/2506.09226 https://
“Move fast and break things.”
The things:
- Human rights
- Our habitat
- Democracy
#BigTech #SiliconValley #ventureCapital
Replaced article(s) found for cs.CL. https://arxiv.org/list/cs.CL/new
[1/5]:
- Cheap Learning: Maximising Performance of Language Models for Social Data Science Using Minimal Data
Castro-Gonzalez, Chung, Kirk, Francis, Williams, Johansson, Bright
Wow "individualised pricing", such fun. I guess this boils down to: what's the most we can charge you, personally, based on what dirt we have on you:
https://
#Microsoft outsourced administration of classified #DoD data to cheap workers in #China. 🇨🇳 🕵️
My latest update on
Dehazing Light Microscopy Images with Guided Conditional Flow Matching: finding a sweet spot between fidelity and realism
Anirban Ray, Ashesh, Florian Jug
https://arxiv.org/abs/2506.22397
To add a single example here (feel free to chime in with your own):
Problem: editing code is sometimes tedious because external APIs require boilerplate.
Solutions:
- Use LLM-generated code. Downsides: energy use, code theft, potential for legal liability, makes mistakes, etc. Upsides: popular among some peers, seems easy to use.
- Pick a better library (not always possible).
- Build internal functions to centralize boilerplate code, then use those (benefits: you get a better understanding of the external API, and a more-unit-testable internal code surface; probably less amortized effort).
- Develop a non-LLM system that actually reasons about code at something like the formal semantics level and suggests boilerplate fill-ins based on rules, while foregrounding which rules it's applying so you can see the logic behind the suggestions (needs research).
Obviously LLM use in coding goes beyond this single issue, but there are similar analyses for each potential use of LLMs in coding. I'm all cases there are:
1. Existing practical solutions that require more effort (or in many cases just seem to but are less-effort when amortized).
2. Near-term researchable solutions that directly address the problem and which would be much more desirable in the long term.
Thus in addition to disastrous LLM effects on the climate, on data laborers, and on the digital commons, they tend to suck us into cheap-seeming but ultimately costly design practices while also crowding out better long-term solutions. Next time someone suggests how useful LLMs are for some task, try asking yourself (or them) what an ideal solution for that task would look like, and whether LLM use moves us closer to or father from a world in which that solution exists.