⚡ Python-free Rust inference server — OpenAI-API compatible. GGUF + SafeTensors, hot model swap, auto-discovery, single binary. FREE now, FREE forever.
-
Updated
Oct 23, 2025 - Rust
8000
⚡ Python-free Rust inference server — OpenAI-API compatible. GGUF + SafeTensors, hot model swap, auto-discovery, single binary. FREE now, FREE forever.
Olares: An Open-Source Personal Cloud to Reclaim Your Data
Maid is a cross-platform Flutter app for interfacing with GGUF / llama.cpp models locally, and with Ollama and OpenAI models remotely.
MLX-VLM is a package for inference and fine-tuning of Vision Language Models (VLMs) on your Mac using MLX.
A privacy-preserving home security camera that uses end-to-end encryption. (Secluso was previously named Privastead.)
NativeMind: Your fully private, open-source, on-device AI assistant
MemoryCache is an experimental development project to turn a local desktop environment into an on-device AI agent
Open‑WebUI Tools is a modular toolkit designed to extend and enrich your Open WebUI instance, turning it into a powerful AI workstation. With a suite of over 15 specialized tools, function pipelines, and filters, this project supports academic research, agentic autonomy, multimodal creativity, workflows, and more
🦙 Ollama Telegram bot, with advanced configuration
Shinkai is a two click install App that allows you to create Local AI agents in 5 minutes or less using a simple UI. Supports: MCPs, Remote and Local AI, Crypto and Payments.
A curated list of awesome platforms, tools, practices and resources that helps run LLMs locally
Like ChatGPT's voice conversations with an AI, but entirely offline/private/trade-secret-friendly, using local AI models such as LLama 2 and Whisper
High-performance lightweight proxy and load balancer for LLM infrastructure. Intelligent routing, automatic failover and unified model discovery across local and remote inference backends.
Blueprint by Mozilla.ai for generating podcasts from documents using local AI
AgC is the open-core platform that powers Open Agentic Compute — a new compute substrate purpose-built for deploying, running, and orchestrating AI agents at scale.
Privacy-first AI ecosystem for Android. Run GGUF models offline or access 100+ cloud models via OpenRouter. Features 11 premium offline voices, extensible plugins, and dynamic DataHub for context injection. No subscriptions, no data harvesting—just AI on your terms.
A flexible free and unlimited PDF Translator for Human with Local-LLM or ChatGPT
Add a description, image, and links to the local-ai topic page so that developers can more easily learn about it.
To associate your repository with the local-ai topic, visit your repo's landing page and select "manage topics."