
Open-source, self-hosted AI coding assistant for private code completion and chat
Tabby is an open-source, self-hosted AI coding assistant that lets development teams run their own LLM-powered code completion and chat server entirely on-premises. Built in Rust by former Google engineers, it supports major coding LLMs like CodeLlama, StarCoder, and DeepSeek Coder, and integrates with VS Code, JetBrains, Vim, and other editors via lightweight extensions. Tabby features an Answer Engine for codebase Q&A, inline chat for contextual help, and Data Connectors that ingest project documentation, Git repositories, and web pages for retrieval-augmented generation. With adaptive caching for sub-second completions, enterprise SSO/LDAP, audit logging, and the ability to run on consumer-grade GPUs, Tabby gives teams full control over their AI coding workflow without sending code to external services.
Run your own LLM-powered code completion server on-premises with sub-second response times via adaptive caching and streaming.
Ask questions about your codebase and get contextual answers powered by retrieval-augmented generation over your repositories and docs.
Get AI assistance directly in your editor with inline chat that understands your current code context for refactoring, explaining, and generating code.
Connect Git repositories, project documentation, web pages, and other sources so the AI has full context about your codebase and standards.
Choose from major coding LLMs including CodeLlama, StarCoder, DeepSeek Coder, and others via a curated models registry.
Lightweight extensions for VS Code, JetBrains IDEs, Vim/Neovim, and other editors with seamless code completion and chat.
Organizations in regulated industries (finance, healthcare, defense) that need AI coding assistance without sending proprietary code to external cloud services.
Teams working in secure or offline environments who need AI code completion running entirely on local infrastructure with no internet dependency.
Developers and teams seeking a free, customizable AI coding assistant without vendor lock-in or recurring per-seat cloud subscription costs.
Small to mid-size engineering teams wanting shared AI code completion with team management, context from their repositories, and control over model selection.
LDAP and SSO authentication support with user/group management and audit logging for enterprise compliance requirements.
Deploy via Docker, Homebrew, Hugging Face Spaces, or cloud infrastructure. Runs efficiently on consumer-grade GPUs without external database dependencies.
ML-savvy teams who want to fine-tune or swap coding models (CodeLlama, StarCoder, DeepSeek) to optimize completions for their specific codebase and languages.

Private email from the makers of Startpage