Repository navigation
#
multi-gpu-inference
- Website
- Wikipedia
Inferflow is an efficient and highly configurable inference engine for large language models (LLMs).
C++
242
1 年前
A script for PyTorch multi-GPU multi-process testing
Python
21
1 年前
Distributed Reinforcement Learning for LLM Fine-Tuning with multi-GPU utilization
Python
11
1 个月前