Repository navigation

#

inference-server

The goal of RamaLama is to make working with AI boring.

Python
1538
3 小时前

A REST API for Caffe using Docker and Go

C++
419
7 年前

This is a repository for an nocode object detection inference API using the Yolov3 and Yolov4 Darknet framework.

Python
280
3 年前

Work with LLMs on a local environment using containers

TypeScript
215
2 天前

Serving AI/ML models in the open standard formats PMML and ONNX with both HTTP (REST API) and gRPC endpoints

Scala
157
6 个月前

ONNX Runtime Server: The ONNX Runtime Server is a server that provides TCP and HTTP/HTTPS REST APIs for ONNX inference.

C++
155
1 个月前

K3ai is a lightweight, fully automated, AI infrastructure-in-a-box solution that allows anyone to experiment quickly with Kubeflow pipelines. K3ai is perfect for anything from Edge to laptops.

PowerShell
101
3 年前

Advanced inference pipeline using NVIDIA Triton Inference Server for CRAFT Text detection (Pytorch), included converter from Pytorch -> ONNX -> TensorRT, Inference pipelines (TensorRT, Triton server - multi-format). Supported model format for Triton inference: TensorRT engine, Torchscript, ONNX

Python
33
4 年前