Beyond Retrieval v2¶
Enterprise-grade Retrieval-Augmented Generation platform for document intelligence.
Upload documents. Ask questions. Get cited answers grounded in your actual content.
What is Beyond Retrieval?¶
Beyond Retrieval v2 is a NotebookLM-inspired knowledge management system that transforms raw documents into searchable, AI-powered knowledge bases. Organize your knowledge into notebooks, ingest files in many formats, and have intelligent conversations grounded in your actual documents — with precise citations back to the source text.
Key Features¶
Document Ingestion¶
Upload PDF, DOCX, TXT, XLSX, CSV. Parse via Docling or Mistral OCR. Chunk with Recursive, Hybrid, or Agentic strategies. Embed with OpenAI, OpenRouter, or Ollama.
RAG Chat¶
Ask natural-language questions and get answers with numbered citations. 6 personas, 10 languages, smart language mode, cache-first pipeline with semantic similarity.
AI Enhancement¶
Kanban board for chunk-level context augmentation. Parallel processing with per-notebook concurrency control. Publish enhanced chunks to the vector store.
Search Playground¶
7 retrieval strategies: Fusion, Semantic, Full-Text, Cache, Contextual, Agentic, Smart Router. Side-by-side A/B comparison mode and search history.
LLM Judge¶
Background quality evaluation after every RAG response. Cache eligibility scoring. Response quality logging for continuous improvement.
Multi-Provider LLM¶
OpenRouter as gateway (25+ models), OpenAI Direct, Ollama (local/air-gapped). Per-notebook provider selection with dynamic API keys from the UI.
Flexible Deployment¶
Cloud Supabase or fully self-hosted local Supabase (Docker). 4 storage backends: Supabase Storage, S3, Local filesystem, None. 4 deployment modes.
Auth & Sharing¶
Clerk JWT authentication with bypass mode for development. Invite links for notebook sharing with role-based access control (admin / chat_only).
At a Glance¶
| Metric | Value |
|---|---|
| API Endpoints | 83+ across 15 route modules |
| Unit Tests | 1213+ with end-to-end coverage |
| LLM Providers | 3 (OpenRouter, OpenAI, Ollama) |
| Retrieval Strategies | 7 |
| Supported Languages | 10 |
| Chat Personas | 6 |
| Deployment Modes | 4 |
| Docker Services | 16 (across base + profiles) |
Tech Stack¶
| Layer | Technology |
|---|---|
| Backend | Python 3.12, FastAPI 0.115, Pydantic v2, Pydantic AI |
| Frontend | React 19, Vite 7, Tailwind CSS 4, Lucide Icons |
| Database | PostgreSQL 15 + pgvector (HNSW indexes), Supabase |
| Auth | Clerk (JWT, JWKS verification) with bypass mode |
| LLM Providers | OpenRouter, OpenAI, Ollama |
| Document Parsing | Docling, Mistral OCR |
| Embeddings | OpenAI text-embedding-3-small (1536d), 21+ OpenRouter models, Ollama nomic-embed-text |
| Storage | Supabase Storage, S3, Local filesystem |
| Reverse Proxy | Caddy 2 (auto-HTTPS via Let's Encrypt) |
| Containerization | Docker Compose |
Architecture¶
graph TB
User[User / Browser]
subgraph Caddy ["Caddy (Reverse Proxy, Auto-HTTPS)"]
CaddyProxy[Caddy 2]
end
subgraph App ["Application"]
Frontend["Frontend<br/>React 19 + Vite 7"]
Backend["Backend<br/>FastAPI + Pydantic AI"]
end
subgraph Database ["Database Layer"]
Supabase["Supabase<br/>PostgreSQL 15 + pgvector"]
Storage["Supabase Storage<br/>/ S3 / Local FS"]
end
subgraph LLM ["LLM Providers"]
OpenRouter["OpenRouter<br/>25+ models"]
OpenAI["OpenAI Direct"]
Ollama["Ollama<br/>Local / Air-Gapped"]
end
subgraph Parsing ["Document Parsing"]
Docling["Docling Serve"]
MistralOCR["Mistral OCR"]
end
User --> CaddyProxy
CaddyProxy --> Frontend
CaddyProxy --> Backend
Backend --> Supabase
Backend --> Storage
Backend --> OpenRouter
Backend --> OpenAI
Backend --> Ollama
Backend --> Docling
Backend --> MistralOCR
Frontend --> Backend Next Steps¶
- Installation — Get up and running in minutes
- Quick Start — Create your first notebook and ask a question
- API Reference — Connect to every endpoint
- Architecture — Understand the system design