Lifted TASEP: A Solvable Paradigm for Speeding up Many-Particle Markov Chains

Physical Review X American Physical Society (APS) 14:4 (2024) 041035

Authors:

Fabian HL Essler, Werner Krauth

Thermodynamic inference of correlations in nonequilibrium collective dynamics

Physical Review Research American Physical Society (APS) 6:4 (2024) l042012

Authors:

Michalis Chatzittofi, Ramin Golestanian, Jaime Agudo-Canalejo

Inertial Focusing Dynamics of Spherical Particles in Curved Microfluidic Ducts with a Trapezoidal Cross Section

SIAM Journal on Applied Dynamical Systems Society for Industrial & Applied Mathematics (SIAM) 23:3 (2024) 1805-1835

Authors:

Brendan Harding, Yvonne M Stokes, Rahil N Valani

Isovolumetric dividing active matter

(2024)

Authors:

Samantha R Lish, Lukas Hupe, Ramin Golestanian, Philip Bittihn

An exactly solvable model for emergence and scaling laws in the multitask sparse parity problem

Advances in Neural Information Processing Systems 37 (NeurIPS 2024) Curran Associates 37 (2024)

Authors:

Yoonsoo Nam, Nayara Fonseca, Sh Lee, Christopher Mingard, Ard A Louis

Abstract:

Deep learning models can exhibit what appears to be a sudden ability to solve a new problem as training time, training data, or model size increases, a phenomenon known as emergence. In this paper, we present a framework where each new ability (a skill) is represented as a basis function. We solve a simple multi-linear model in this skill-basis, finding analytic expressions for the emergence of new skills, as well as for scaling laws of the loss with training time, data size, model size, and optimal compute. We compare our detailed calculations to direct simulations of a two-layer neural network trained on multitask sparse parity, where the tasks in the dataset are distributed according to a power-law. Our simple model captures, using a single fit parameter, the sigmoidal emergence of multiple new skills as training time, data size or model size increases in the neural network.