<?xml version="1.0" encoding="utf-8" ?>
<!DOCTYPE FL_Course SYSTEM "https://www.flane.de/dtd/fl_course095.dtd"><?xml-stylesheet type="text/xsl" href="https://portal.flane.ch/css/xml-course.xsl"?><course productid="34496" language="fr" source="https://portal.flane.ch/swisscom/fr/xml-course/nvidia-dphtdlm" lastchanged="2025-07-29T12:18:27+02:00" parent="https://portal.flane.ch/swisscom/fr/xml-courses"><title>Data Parallelism: How to Train Deep Learning Models on Multiple GPUs</title><productcode>DPHTDLM</productcode><vendorcode>NV</vendorcode><vendorname>Nvidia</vendorname><fullproductcode>NV-DPHTDLM</fullproductcode><version>1.0</version><objective>&lt;p&gt;By participating in this workshop, you&amp;rsquo;ll:
&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Understand how data parallel deep learning training is performed using multiple GPUs&lt;/li&gt;&lt;li&gt;Achieve maximum throughput when training, for the best use of multiple GPUs&lt;/li&gt;&lt;li&gt;Distribute training to multiple GPUs using Pytorch Distributed Data Parallel&lt;/li&gt;&lt;li&gt;Understand and utilize algorithmic considerations specific to multi-GPU training performance and accuracy&lt;/li&gt;&lt;/ul&gt;</objective><essentials>&lt;p&gt;Experience with deep learning training using Python&lt;/p&gt;</essentials><outline>&lt;p&gt;&lt;strong&gt;Introduction&lt;/strong&gt;	
&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Meet the instructor.&lt;/li&gt;&lt;li&gt;Create an account at courses.nvidia.com/join&lt;/li&gt;&lt;/ul&gt;&lt;p&gt;&lt;strong&gt;Stochastic Gradient Descent and the Effects of Batch Size&lt;/strong&gt;	
&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Learn the significance of stochastic gradient descent when training on multiple GPUs&lt;/li&gt;&lt;li&gt;Understand the issues with sequential single-thread data processing and the theory behind speeding up applications with parallel processing.&lt;/li&gt;&lt;li&gt;Understand loss function, gradient descent, and stochastic gradient descent (SGD).&lt;/li&gt;&lt;li&gt;Understand the effect of batch size on accuracy and training time with an eye towards its use on multi-GPU systems.&lt;/li&gt;&lt;/ul&gt;&lt;p&gt;&lt;strong&gt;Training on Multiple GPUs with PyTorch Distributed Data Parallel (DDP)&lt;/strong&gt;	
&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Learn to convert single GPU training to multiple GPUs using PyTorch Distributed Data Parallel&lt;/li&gt;&lt;li&gt;Understand how DDP coordinates training among multiple GPUs.&lt;/li&gt;&lt;li&gt;Refactor single-GPU training programs to run on multiple GPUs with DDP.&lt;/li&gt;&lt;/ul&gt;&lt;p&gt;&lt;strong&gt;Maintaining Model Accuracy when Scaling to Multiple GPUs&lt;/strong&gt;	
&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Understand and apply key algorithmic considerations to retain accuracy when training on multiple GPUs&lt;/li&gt;&lt;li&gt;Understand what might cause accuracy to decrease when parallelizing training on multiple GPUs.&lt;/li&gt;&lt;li&gt;Learn and understand techniques for maintaining accuracy when scaling training to multiple GPUs.&lt;/li&gt;&lt;/ul&gt;&lt;p&gt;&lt;strong&gt;Workshop Assessment&lt;/strong&gt;	
&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Use what you have learned during the workshop: complete the workshop assessment to earn a certificate of competency&lt;/li&gt;&lt;/ul&gt;&lt;p&gt;&lt;strong&gt;Final Review&lt;/strong&gt;	
&lt;/p&gt;
&lt;ul&gt;
&lt;li&gt;Review key learnings and wrap up questions.&lt;/li&gt;&lt;li&gt;Take the workshop survey.&lt;/li&gt;&lt;/ul&gt;</outline><objective_plain>By participating in this workshop, you’ll:



- Understand how data parallel deep learning training is performed using multiple GPUs
- Achieve maximum throughput when training, for the best use of multiple GPUs
- Distribute training to multiple GPUs using Pytorch Distributed Data Parallel
- Understand and utilize algorithmic considerations specific to multi-GPU training performance and accuracy</objective_plain><essentials_plain>Experience with deep learning training using Python</essentials_plain><outline_plain>Introduction	



- Meet the instructor.
- Create an account at courses.nvidia.com/join
Stochastic Gradient Descent and the Effects of Batch Size	



- Learn the significance of stochastic gradient descent when training on multiple GPUs
- Understand the issues with sequential single-thread data processing and the theory behind speeding up applications with parallel processing.
- Understand loss function, gradient descent, and stochastic gradient descent (SGD).
- Understand the effect of batch size on accuracy and training time with an eye towards its use on multi-GPU systems.
Training on Multiple GPUs with PyTorch Distributed Data Parallel (DDP)	



- Learn to convert single GPU training to multiple GPUs using PyTorch Distributed Data Parallel
- Understand how DDP coordinates training among multiple GPUs.
- Refactor single-GPU training programs to run on multiple GPUs with DDP.
Maintaining Model Accuracy when Scaling to Multiple GPUs	



- Understand and apply key algorithmic considerations to retain accuracy when training on multiple GPUs
- Understand what might cause accuracy to decrease when parallelizing training on multiple GPUs.
- Learn and understand techniques for maintaining accuracy when scaling training to multiple GPUs.
Workshop Assessment	



- Use what you have learned during the workshop: complete the workshop assessment to earn a certificate of competency
Final Review	



- Review key learnings and wrap up questions.
- Take the workshop survey.</outline_plain><duration unit="d" days="1">1 jour</duration><pricelist><price country="US" currency="USD">500.00</price><price country="DE" currency="EUR">500.00</price><price country="AT" currency="EUR">500.00</price><price country="SE" currency="EUR">500.00</price><price country="SI" currency="EUR">500.00</price><price country="GB" currency="GBP">420.00</price><price country="IT" currency="EUR">500.00</price><price country="CA" currency="CAD">690.00</price></pricelist><miles/></course>