Repository navigation
llama-cpp
- Website
- Wikipedia
A C#/.NET library to run LLM (🦙LLaMA/LLaVA) on your local device efficiently.
Maid is a cross-platform Flutter app for interfacing with GGUF / llama.cpp models locally, and with Ollama and OpenAI models remotely.
prima.cpp: Speeding up 70B-scale LLM inference on low-resource everyday home clusters
This repo is to showcase how you can run a model locally and offline, free of OpenAI dependencies.
Local ML voice chat using high-end models.
Making offline AI models accessible to all types of edge devices.
workbench for learing&practising AI tech in real scenario on Android device, powered by GGML(Georgi Gerganov Machine Learning) and FFmpeg
LLaMA Server combines the power of LLaMA C++ with the beauty of Chatbot UI.
Your customized AI assistant - Personal assistants on any hardware! With llama.cpp, whisper.cpp, ggml, LLaMA-v2.