Repository navigation

#

model-parallelism

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Python
38001
2 天前

飞桨大模型开发套件,提供大语言模型、跨模态大模型、生物计算大模型等领域的全流程开发工具链。

Python
465
1 年前

Slicing a PyTorch Tensor Into Parallel Shards

Python
298
4 年前

Easy Parallel Library (EPL) is a general and efficient deep learning framework for distributed model training.

Python
267
2 年前

A curated list of awesome projects and papers for distributed training or inference

231
6 个月前

Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*

Python
82
1 年前
Python
64
4 个月前

SC23 Deep Learning at Scale Tutorial Material

Python
43
7 个月前

Distributed training of DNNs • C++/MPI Proxies (GPT-2, GPT-3, CosmoFlow, DLRM)

C++
42
1 年前
Python
18
2 个月前

WIP. Veloce is a low-code Ray-based parallelization library that makes machine learning computation novel, efficient, and heterogeneous.

Python
18
3 年前

PyTorch implementation of 3D U-Net with model parallel in 2GPU for large model

Python
9
5 年前

Adaptive Tensor Parallelism for Foundation Models

Python
9
2 年前

Performance Estimates for Transformer AI Models in Science

Jupyter Notebook
7
7 个月前

Official implementation of DynPartition: Automatic Optimal Pipeline Parallelism of Dynamic Neural Networks over Heterogeneous GPU Systems for Inference Tasks

Python
6
2 年前