Core Features
Everything you need to build AI-powered applications
User Experience
Intuitive chat interface with full markdown and settings.
Run everything locally on your machine with complete data control.
One-click downloads from HuggingFace with real-time progress tracking.
Use local GGUF models alongside API providers (OpenAI, Anthropic, Groq) in one unified interface.
Server-Sent Events provide instant response feedback with live token streaming.
12+ parameters for fine-tuning: temperature, top-p, frequency penalty, and more.
Technical Capabilities
Drop-in replacement for OpenAI APIs. Use your existing code and tools.
Run models on your hardware for enhanced privacy and control.
Optimized inference with llama.cpp. 8-12x speedup with GPU acceleration (CUDA, ROCm).
Save and switch between inference configurations instantly without restarts.
Real-time statistics showing tokens per second and processing speed.
Download models asynchronously with progress tracking and auto-resumption.
Enterprise & Team Ready
Built for secure collaboration with enterprise-grade authentication and comprehensive user management
User Management Dashboard
Comprehensive admin interface for managing users, roles, and access requests.
Role-Based Access Control
4 role levels (User, PowerUser, Manager, Admin) with granular permission management.
Self-service access requests with admin approval gates and audit trail.
Enterprise-grade authentication with PKCE, session management, and token lifecycle control.
Secure team collaboration with session invalidation and role change enforcement.
Developer Tools & SDKs
Everything developers need to integrate AI into applications with production-ready tools
Production-ready npm package @bodhiapp/ts-client for seamless integration.
Scope-based permissions with SHA-256 hashing and database-backed security.
Interactive API documentation with auto-generated specs and live testing.
Drop-in replacement for OpenAI APIs - use existing libraries and tools seamlessly.
Additional API format support for Ollama chat and models endpoints.
Flexible Deployment Options
Deploy anywhere - from desktop to cloud, with hardware-optimized variants for maximum performance
Native desktop apps for Windows, macOS (Intel/ARM), and Linux with Tauri.
CPU (AMD64/ARM64), CUDA, ROCm, and Vulkan optimized images for every hardware.
RunPod auto-configuration and support for any Docker-compatible cloud platform.
8-12x speedup with CUDA/ROCm GPU support for NVIDIA and AMD graphics cards.
Persistent storage with backup/restore strategies and migration support.
Health checks, monitoring, log management, and automatic database migrations.
Download for your platform
Choose your operating system to download BodhiApp. All platforms support running LLMs locally with full privacy.
Package Managers
brew install BodhiSearch/apps/bodhiLoading Docker releases...
