Repository navigation
localllm
- Website
- Wikipedia
Use your locally running AI models to assist you in your web browsing
A generalized information-seeking agent system with Large Language Models (LLMs).
Run local LLMs like llama, deepseek-distill, kokoro and more inside your browser
[ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization
Model swapping for llama.cpp (or any local OpenAPI compatible server)
[NeurIPS 2024] KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
The .NET library to consume 100+ APIs: OpenAI, Anthropic, Google, DeepSeek, Cohere, Mistral, Azure, xAI, Perplexity, Groq, Ollama, LocalAi, and many more!
The PyVisionAI Official Repo
Run Open Source/Open Weight LLMs locally with OpenAI compatible APIs
MVP of an idea using multiple local LLM models to simulate and play D&D
Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on different ports and loading/unloading them on demand
MVP of an idea using multiple local LLM models to simulate and play D&D
Chat with your pdf using your local LLM, OLLAMA client.(incomplete)
The client for the Symmetry peer-to-peer inference network. Enabling users to connect with each other, share computational resources, and collect valuable machine learning data.
A local chatbot for managing docs