Skip to content

yetanothernaveed/ParallelNN

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Evaluating Distributed Model Training using PyTorch APIs

This repo contains the necessary code and bash scripts used for evaluating PyTorch's built-in distributed training module.

Modules covered:

  • Fully Sharded Data Parallel (FSDP)
  • Pipeline Parallel (PP)

Setup

Four nodes, each consisting of a single GPU, Nvidia 1650, with 4GB VRAM.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors