The scaling of machine learning models, driven by massive parallel hardware and extensive datasets, has historically been hampered by sequential bottlenecks in core algorithms. Dynamical systems, crucial for models like RNNs and MCMC, were a prime example of this limitation. This PhD dissertation, by Xavier Gonzalez, fundamentally challenges this paradigm.
Beyond Sequential Bottlenecks: Parallelizing Dynamics
This work demonstrates that dynamical systems can indeed be parallelized across their sequence length. The core innovation reframes sequential computations as nonlinear equations solvable via Newton's method, leveraging parallel associative scans. This approach bypasses the inherent sequential dependency, a significant theoretical and practical leap. The research presented in Gonzalez's dissertation addresses critical limitations of prior parallel Newton methods, namely inefficiency and instability.