Skip to main content

Deployment Options

Airweave can be deployed in several ways depending on your infrastructure and scalability requirements:

Docker Compose

Single-server deployment for development and small production workloads

Kubernetes

Scalable production deployment with orchestration and high availability

Airweave Cloud

Fully managed hosting - no infrastructure required

Configuration

Environment variables and advanced configuration options

Architecture Components

Airweave consists of several interconnected services:

Core Services

Technology: FastAPI (Python 3.11)The main application server that handles:
  • REST API endpoints
  • Authentication and authorization
  • Business logic and orchestration
  • Database migrations (Alembic)
Ports:
  • 8001 - API server
  • 9090 - Metrics/monitoring
Health check: http://localhost:8001/health/ready
Technology: React/TypeScript with ViteWeb-based user interface for:
  • Managing collections and source connections
  • Configuring integrations
  • Monitoring sync jobs
  • Testing search queries
Port: 8080 (default)Build: Multi-stage Docker build with production optimization
Version: 16Primary data store for:
  • User accounts and permissions
  • Collection and source metadata
  • Sync job history and state
  • Webhook subscriptions
Port: 5432Optimized settings:
  • max_connections=200
  • shared_buffers=256MB
  • effective_cache_size=1GB
Version: 7-alpineUsed for:
  • Pub/sub messaging
  • Session storage
  • Webhook queue management (via Svix)
Port: 6379

Search & Indexing

Version: 8Vector search engine for:
  • Dense vector embeddings (configurable dimensions)
  • Sparse BM25 embeddings
  • Hybrid search ranking
  • Real-time document indexing
Ports:
  • 8081 - Query/Document API
  • 19071 - Config server
Features:
  • Dynamic schema templating based on embedding dimensions
  • Automatic deployment via init container
  • Persistent storage with Docker volumes
Image: semitechnologies/transformers-inferenceLocal embedding generation using:
  • Model: sentence-transformers-all-MiniLM-L6-v2
  • Dimensions: 384
Port: 9878Note: Automatically skipped if OPENAI_API_KEY is provided. Uses ~2GB of memory.

Workflow Orchestration

Version: 1.24.2Durable workflow engine for:
  • Data sync orchestration
  • Retry logic and error handling
  • Long-running background tasks
  • Scheduled sync jobs
Ports:
  • 7233 - gRPC API
  • 8233 - Internal metrics
Version: 2.26.2Web interface for:
  • Workflow monitoring
  • Task queue inspection
  • Debugging failed workflows
Port: 8088
Executes workflow tasks including:
  • Connector sync activities
  • Document processing
  • Embedding generation
  • Cleanup tasks
Uses the same backend image as the API server.

Additional Services

Webhook delivery systemManages:
  • Webhook subscriptions
  • Event delivery with retries
  • Delivery logs and debugging
Port: 8071Features:
  • Redis-based queue
  • Subnet whitelisting for testing
  • Automatic database initialization

Resource Requirements

Minimum (Development)

  • CPU: 2 cores
  • Memory: 4GB RAM
  • Storage: 10GB
  • Docker: 20.10+ with Docker Compose
  • CPU: 4-8 cores
  • Memory: 8-16GB RAM
  • Storage: 50GB+ (depends on data volume)
  • Network: Stable internet for connector syncs
The local embeddings service requires ~2GB of memory. If you have limited resources, use OpenAI or Mistral embeddings instead.

Service Dependencies

The startup sequence is managed automatically via health checks:

Storage Options

Airweave supports multiple storage backends for file attachments:
BackendUse CaseConfiguration
FilesystemLocal development, K8s PVCSTORAGE_BACKEND=filesystem
Azure BlobAzure-based deploymentsSTORAGE_BACKEND=azure
AWS S3AWS deployments or S3-compatibleSTORAGE_BACKEND=aws
GCP StorageGoogle Cloud deploymentsSTORAGE_BACKEND=gcp
By default, local deployments use filesystem storage at ./local_storage. See Configuration for cloud storage setup.

Next Steps

1

Choose your deployment method

2

Configure environment variables

Review Configuration for all available options
3

Set up integrations

Configure connector credentials and embedding providers
4

Monitor your deployment

Use Temporal UI and backend health endpoints to verify services