Deploy CortexDB entirely on your own infrastructure with no external dependencies.
On-Premises Deployment
CortexDB can run entirely on your infrastructure with no external dependencies. This guide covers fully air-gapped deployments.
Architecture
An on-prem CortexDB deployment consists of:
- CortexDB cluster -- 3+ nodes running the CortexDB binary
- Local LLM (optional) -- Ollama or vLLM for entity extraction
- Local embeddings (optional) -- Local embedding model for vector search
No data leaves your network.
Air-Gapped Deployment
1. Load Docker Images
Transfer the CortexDB and Ollama images to your environment:
# On a machine with internet access
docker pull cortexdb/cortexdb:latest
docker pull ollama/ollama:latest
docker save cortexdb/cortexdb:latest | gzip > cortexdb.tar.gz
docker save ollama/ollama:latest | gzip > ollama.tar.gz
# Transfer files to air-gapped environment
# On the air-gapped machine
docker load < cortexdb.tar.gz
docker load < ollama.tar.gz
2. Load Models
# Pull models on an internet-connected machine
ollama pull llama3.2
ollama pull nomic-embed-text
# Export models
# Copy ~/.ollama/models to the air-gapped environment
3. Deploy
version: "3.8"
services:
cortexdb:
image: cortexdb/cortexdb:latest
environment:
CORTEX_LLM_PROVIDER: ollama
CORTEX_LLM_BASE_URL: http://ollama:11434
CORTEX_LLM_MODEL: llama3.2
CORTEX_EMBEDDING_PROVIDER: ollama
CORTEX_EMBEDDING_BASE_URL: http://ollama:11434
CORTEX_EMBEDDING_MODEL: nomic-embed-text
CORTEX_ENCRYPTION_AT_REST: "true"
CORTEX_ENCRYPTION_KEY: "${CORTEX_ENCRYPTION_KEY}"
CORTEX_API_KEY: "${CORTEX_API_KEY}"
volumes:
- cortexdb-data:/data
ports:
- "8080:8080"
ollama:
image: ollama/ollama:latest
volumes:
- ollama-models:/root/.ollama
volumes:
cortexdb-data:
ollama-models:
No-LLM Mode
CortexDB can operate without any LLM. Entity extraction and graph features are disabled, but remember/recall with keyword and vector search still works:
CORTEX_EXTRACTION_ENABLED=false
CORTEX_EMBEDDING_PROVIDER=local # Uses built-in embedding model
Hardware Requirements
| Component | Minimum | Recommended | |---|---|---| | CortexDB node | 4 CPU, 4 GB RAM, 50 GB SSD | 8 CPU, 16 GB RAM, 200 GB NVMe | | Ollama (LLM) | 8 CPU, 16 GB RAM | GPU with 24 GB VRAM | | Ollama (embed only) | 2 CPU, 4 GB RAM | 4 CPU, 8 GB RAM |
Support
Enterprise on-prem deployments include:
- Dedicated support channel
- Deployment assistance
- Custom integration development
- SLA guarantees
Contact [email protected] for details.