Ecosystem
Software
Everything you need to run agents, serve models, and build on open-source AI.
AI Agent Platforms
Personal software for running agentic AI services locally.
Run AI agents locally with tool use, memory, and multi-channel messaging. CLI-first, self-hosted.
Multi-agent orchestration framework. Role-based agents with task delegation.
Microsoft's multi-agent conversation framework. Agent-to-agent workflows.
Stateful multi-agent workflows from LangChain. Graph-based agent orchestration.
End-to-end NLP/AI framework by deepset. Pipelines, RAG, agents.
Distributed & Cluster Inference
Pool multiple devices to run models too large for one machine.
Inference Engines
Run models locally on your hardware.
One-command model runner. Pull and run any GGUF model.
C/C++ inference engine. The foundation most local tools build on.
High-throughput serving with PagedAttention. Production-grade.
Apple Silicon native. Best performance on Mac hardware.
Optimized GPTQ/EXL2 inference for NVIDIA GPUs.
Fast serving framework with RadixAttention. Structured generation.
Interfaces & Frontends
UIs for interacting with local models.
ChatGPT-style interface for Ollama and OpenAI-compatible APIs.
Desktop app for discovering and running local models. GUI-first.
Offline-first desktop client. Clean UI, local-only by default.
All-in-one RAG application. Documents → local LLM → answers.
Cross-platform desktop client from Nomic. Simple and reliable.
Native Mac/Windows app for running local models. Clean, fast.
Fine-Tuning
Train and adapt models on your own data.
Quantization
Compress models to fit smaller hardware.