Run Open Source/Open Weight LLMs locally with OpenAI compatible APIs
-
Updated
Apr 9, 2026 - Rust
Run Open Source/Open Weight LLMs locally with OpenAI compatible APIs
Modern desktop application (Rust + Tauri v2 + Svelte 5 + Candle (HF)) for communicating with AI models that runs completely locally on your computer. No subscriptions, no data sent to the internet — just you and your personal AI assistant
Your Very Own Agent: The Ultimate, Complete Edition
caro: fast Rust CLI that turns natural‑language tasks into a safe POSIX command. Built for macOS (MLX/Metal) with a built‑in model; supports vLLM/Ollama/LM Studio. JSON‑only output, safety checks, confirmation, multi‑step goals, devcontainer included.
Rust AI agent where every context token earns its place. Self-learning skills, temporal graph memory, cascade quality routing, OWASP AI security. Hybrid inference: Ollama · Claude · Gemini · OpenAI · GGUF. MCP + ACP. One binary.
Analyze how "surprised" LLMs are when reading a piece of text
Nucleus is a modular Rust library for building advanced local AI tools and applications, built-in RAG for private documents, and an extensible plugin system for custom workflows.
Attempt to summarize text from `stdin`, using a large language model (locally and offline), to `stdout`
LLM-based code analysis tool for detecting suspicious or inconsistent code regions written in Rust 🦀
A terminal shell built with AI at the core
Smart HTTP router for local LLMs (Ollama, LM Studio, llama.cpp). Rule-based + LLM-powered routing, health checks, load balancing, Prometheus metrics. Rust-native, zero-overhead.
Open-source AI wearable companion. Local-first multimodal perception (VLM & Audio) on Edge NPU. Built with Rust & Flutter.
A fast, cross-platform TavernAI-like client for advanced roleplay. Native Android & Windows support, offline storage, and modern AI backends.
ChatLogger is a Python package written in RUST for managing chat conversations with support for branching dialogue, user management, and integration with language models.
ImpForge — AI Workstation Builder. Your complete AI stack. One desktop app. Offline-first with Ollama + OpenRouter cloud.
A high-level Rust interface for language models powered by the Candle ML framework. It provides ergonomic and efficient APIs for intuitive language model interactions.
Add a description, image, and links to the local-llm topic page so that developers can more easily learn about it.
To associate your repository with the local-llm topic, visit your repo's landing page and select "manage topics."