WaddleAI Documentation
Enterprise-grade AI proxy and management system with OpenAI-compatible APIs, advanced routing, security, and token management.
Quick Start
WaddleAI provides a comprehensive AI proxy solution that acts as a drop-in replacement for OpenAI's API while adding advanced features like multi-LLM support, security scanning, and enterprise-grade management.
Key Features
✨ OpenAI-Compatible API - Drop-in replacement for OpenAI API
🔀 Multi-LLM Support - Route to OpenAI, Anthropic, Ollama, and more
👥 Role-Based Access Control - Admin, Resource Manager, Reporter, User roles
🪙 Dual Token System - WaddleAI tokens for billing, LLM tokens for analytics
🛡️ Security Scanning - Prompt injection and jailbreak detection
📊 Token Management - Quota enforcement and usage tracking
🏢 Multi-Tenant - Organization-based isolation
📈 Monitoring - Prometheus metrics and health checks
🧠 Memory Integration - Conversation memory with mem0/ChromaDB
Getting Started
Docker Compose (Recommended)
# Clone repository
git clone https://github.com/penguintechinc/waddleai.git
cd waddleai
# Create environment file
echo "JWT_SECRET=$(openssl rand -hex 32)" > .env
echo "POSTGRES_PASSWORD=$(openssl rand -hex 16)" >> .env
# Start all services
docker-compose up -d
# Check status
docker-compose ps
Services
- Proxy Server: http://localhost:8000 (OpenAI-compatible API)
- Management Portal: http://localhost:8001 (Admin interface)
- Documentation: http://localhost:8080 (MkDocs site)
First Steps
-
Get Admin API Key: Check logs for the admin API key created on first startup
-
Test the API:
-
Access Management Portal: Visit http://localhost:8001 and login with
admin/admin123
Architecture Overview
WaddleAI consists of two main components:
Proxy Server (Stateless)
- OpenAI-compatible API endpoints
- Request routing and load balancing
- Security scanning and prompt injection detection
- Token counting and quota enforcement
- Prometheus metrics and health checks
Management Server (Stateful)
- Web-based administration portal
- User and organization management
- API key management with RBAC
- Usage analytics and reporting
- LLM provider configuration
Integration Guide
For detailed integration instructions with various tools and platforms, see the Claude Integration guide, which provides comprehensive examples for:
- Python applications with OpenAI SDK
- Node.js applications
- cURL/HTTP requests
- VS Code extension integration
- Management API usage
- Role-based access control
Navigation
- Getting Started - Installation and setup
- API Reference - Complete API documentation
- Deployment - Cloudflare deployment
- Integrations - VS Code extension integration
Ready to get started? Follow the installation guide or try the Docker Compose setup above!