Feb 23, 2022–Feb 23, 2022 from 9:00am–10:00am
Exploiting Parallelism in Large Scale Deep Learning Model Training: From Chips to Systems to Algorithms
We live in a world where hyperscale systems for machine intelligence are increasingly being used to solve complex problems ranging from natural language processing to computer vision to molecular modeling, drug discovery and recommendation systems. A convergence of breakthrough research in machine learning models and algorithms, increased accessibility to hardware systems at cloud scale for research and thriving software ecosystems are paving the way for an exponential increase in model sizes. Effective parallel processing and model decomposition techniques and large clusters of accelerators will be required to train these models of the future economically.
Attend this session to learn about how Graphcore aims to address scale challenges associated with training large models. Get to know our Intelligent Processing Unit (IPU) – a purpose-built hardware accelerator with a unique MIMD architecture – designed to address the most demanding compute and memory bandwidth needs of modern ML models. Our network disaggregated architecture uniquely positions us to build highly scalable systems (IPU-PODs) with thousands of accelerators aimed at exploiting various dimensions of parallelism.
Feb 23, 2022–Feb 23, 2022
from 9:00am–10:00am
Virtual via Zoom
Registration for this event is required
by .
Visit the registration page for details.
Free
Susan Rathbun • events@sdsc.edu
Faculty, Staff, Students, The General Public
San Diego Supercomputer Center (SDSC)