Scaling Python for Machine Learning: Beyond Data Parallelism
This video is also available in the GOTO Play video app! Download it to enjoy offline access to our conference videos while on the move.
Data Parallelism can be amazing and it frees us from so many fiddly complicated tasks (like dealing with locks). On the other hand, as training large machine learning models becomes increasingly popular, we're seeing the need to move beyond purely data-parallel techniques. Depending on recompute exclusively for failure is no longer sufficient as our operations are not idempotent.
In this talk we will look at Spark, Dask, and Ray in the context of scaling machine learning models and how you can take advantage of other types of distributed parallelism (including the actor model for managing model weights during training).
-
It's a Noisy World Out ThereLinda RisingMonday May 22 @ 5:10 PM
-
One Rule to Rule Them AllDave ThomasTuesday May 23 @ 9:30 AM
-
The Psychology of UXFabio Nudge PereiraTuesday May 23 @ 1:50 PM
-
The Universe, Unfolded: NASA Webb Space TelescopeKenneth Harris IIMonday May 22 @ 1:50 PM
-
Practical Magic: The Resilience Potion and Security Chaos EngineeringKelly ShortridgeWednesday May 24 @ 9:30 AM
-
What We Talk About When We Talk About ResilienceCourtney NashWednesday May 24 @ 1:50 PM
-
Large Language Models: Friend, Foe, or OtherwiseAlex CastrounisMonday May 22 @ 9:30 AM
-
Sailing Solo: One Man's Journey Through the World's Loneliest RaceIan Herbert-JonesTuesday May 23 @ 5:10 PM