A systematic comparison of Large Language Models for automated assignment assessment in programming education: Exploring the importance of architecture and vendor
Marcin Jukiewicz
https://arxiv.org/abs/2509.26483
Visual Analytics for Causal Reasoning from Real-World Health Data
Arran Zeyu Wang, David Borland, David Gotz
https://arxiv.org/abs/2508.17474 https://arxiv…
An Anthropic report details how Claude usage varies by country and US states, finding 36% use it for coding, 77% of enterprise use is for automation, and more (Anthropic)
https://www.anthropic.com/research/anthropic-economic-index-september-2025-report
Learning Short-Term and Long-Term Patterns of High-Order Dynamics in Real-World Networks
Yunyong Ko, Da Eun Lee, Song Kyung Yu, Sang-Wook Kim
https://arxiv.org/abs/2508.17236 ht…
ExtGraph: A Fast Extraction Method of User-intended Graphs from a Relational Database
Jeongho Park, Geonho Lee, Min-Soo Kim
https://arxiv.org/abs/2509.18534 https://
Patterns in the Transition From Founder-Leadership to Community Governance of Open Source
Mobina Noori, Mahasweta Chakraborti, Amy X Zhang, Seth Frey
https://arxiv.org/abs/2509.16295
Finally, what Xia & Lindell call a "separation problem" is, in our view, a feature of our approach and not a bug.
If, e.g., all languages in a family are polysynthetic (or none are), that’s not a statistical artefact – it’s the signal. The outcome is well associated with genealogy, showing that family membership captures someth genuinely informative about the process. When the model finds that family explains a large share of the variance, that's not a failure–it's evidence that phylogenetic structure dominates the pattern.
So while Xia & Lindell insist that "autocorrelation due to relationships and distance cannot be captured in family or regional-level analyses", we see that as an empirical question – and we treated it as one.
The real test is whether a mixed model that explicitly represents phylogeny and geography performs worse than their alternative, where the entire shared history of languages and environments is effectively collapsed into a single dimension (an eigenvector).
In other words: we model relationships – Xia & Lindell summarise them into one number per language.
Mutual Information Tracks Policy Coherence in Reinforcement Learning
Cameron Reid, Wael Hafez, Amirhossein Nazeri
https://arxiv.org/abs/2509.10423 https://…
Planet Earth in reflected and polarized light -- III. Modeling and analysis of a decade-long catalog of Earthshine observations
Giulia Roccetti, Michael F. Sterzik, Claudia Emde, Mihail Manev, Stefano Bagnulo, Julia V. Seidel
https://arxiv.org/abs/2509.13415
"My Boyfriend is AI": A Computational Analysis of Human-AI Companionship in Reddit's AI Community
Pat Pataranutaporn, Sheer Karny, Chayapatr Archiwaranguprok, Constanze Albrecht, Auren R. Liu, Pattie Maes
https://arxiv.org/abs/2509.11391