Repository navigation

#

model-serving

Python
45285
3 小时前
bentoml/BentoML

The easiest way to serve AI apps and models - Build Model Inference APIs, Job queues, LLM apps, Multi-model pipelines, and more!

Python
7634
2 天前

In this repository, I will share some useful notes and references about deploying deep learning-based models in production.

4332
5 个月前

FEDML - The unified and scalable ML library for large-scale distributed training, model serving, and federated learning. FEDML Launch, a cross-cloud scheduler, further enables running any AI jobs on any GPU cloud or on-premise cluster. Built on this library, TensorOpera AI (https://TensorOpera.ai) is your generative AI platform at scale.

Python
3840
1 个月前

LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.

Python
3139
1 天前

Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs

Python
2954
18 小时前

🚀 Awesome System for Machine Learning ⚡️ AI System Papers and Industry Practice. ⚡️ System for Machine Learning, LLM (Large Language Model), GenAI (Generative AI). 🍻 OSDI, NSDI, SIGCOMM, SoCC, MLSys, etc. 🗃️ Llama3, Mistral, etc. 🧑‍💻 Video Tutorials.

2871
8 个月前
beclab/Olares
Shell
1991
20 小时前

MLRun is an open source MLOps platform for quickly building and managing continuous ML applications across their lifecycle. MLRun integrates into your development and CI/CD environment and automates the delivery of production data, ML pipelines, and online applications.

Python
1516
3 天前

High-performance inference framework for large language models, focusing on efficiency, flexibility, and availability.

Python
1091
7 小时前

A highly optimized LLM inference acceleration engine for Llama and its variants.

C++
885
5 天前

A high-performance ML model serving framework, offers dynamic batching and CPU/GPU pipelines to fully exploit your compute machine

Python
838
5 天前

Model Deployment at Scale on Kubernetes 🦄️

TypeScript
807
1 年前

A throughput-oriented high-performance serving framework for LLMs

Cuda
796
7 个月前