Repository navigation

#

pipeline-parallelism

DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.

Python
38001
2 天前

飞桨大模型开发套件,提供大语言模型、跨模态大模型、生物计算大模型等领域的全流程开发工具链。

Python
465
1 年前

Personal Project: MPP-Qwen14B & MPP-Qwen-Next(Multimodal Pipeline Parallel based on Qwen-LM). Support [video/image/multi-image] {sft/conversations}. Don't let the poverty limit your imagination! Train your own 8B/14B LLaVA-training-like MLLM on RTX3090/4090 24GB.

Jupyter Notebook
437
1 个月前

InternEvo is an open-sourced lightweight training framework aims to support model pre-training without the need for extensive dependencies.

Python
380
2 天前

Easy Parallel Library (EPL) is a general and efficient deep learning framework for distributed model training.

Python
267
2 年前

A curated list of awesome projects and papers for distributed training or inference

231
6 个月前

Decentralized LLMs fine-tuning and inference with offloading

Python
88
1 个月前

Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*

Python
82
1 年前

An Efficient Pipelined Data Parallel Approach for Training Large Model

Python
75
4 年前

Chimera: bidirectional pipeline parallelism for efficiently training large-scale models.

Python
62
1 个月前
Python
41
2 年前

Implementation of autoregressive language model using improved Transformer and DeepSpeed pipeline parallelism.

Python
32
3 年前

Official implementation of DynPartition: Automatic Optimal Pipeline Parallelism of Dynamic Neural Networks over Heterogeneous GPU Systems for Inference Tasks

Python
6
2 年前

Model parallelism for NN architectures with skip connections (eg. ResNets, UNets)

Python
5
3 年前

Docs for torchpipe: https://github.com/torchpipe/torchpipe

MDX
4
8 个月前