Skip to content

aashrith-madasu/Train-Large-Models-Distributed

Repository files navigation

Training large models on distributed setupts

About

Training large models (>=7B) on multi-gpu distributed setups using technologies like FSDP, DeepSpeed, HF Accelerate

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors