Repository navigation

#

cpu-inference

Runs LLaMA with Extremely HIGH speed

C++
89
1 年前

Portable LLM - A rust library for LLM inference

Rust
9
1 年前

Wrapper for simplified use of Llama2 GGUF quantized models.

Python
6
1 年前
Visual Basic .NET
4
5 个月前

Simple bot that transcribes Telegram voice messages. Powered by go-telegram-bot-api & whisper.cpp Go bindings.

Go
2
2 年前

🧠 A comprehensive toolkit for benchmarking, optimizing, and deploying local Large Language Models. Includes performance testing tools, optimized configurations for CPU/GPU/hybrid setups, and detailed guides to maximize LLM performance on your hardware.

Shell
1
24 天前