RecordEngine requires an NVIDIA GPU with at least 16 GB VRAM. The GPU runs all AI inference locally — document extraction, summarisation, confidence scoring, and chat all happen on the GPU.
GPU
VRAM
Status
NVIDIA RTX 5090
32 GB
✅ Recommended
NVIDIA RTX 4090
24 GB
✅ Supported — standard production config
NVIDIA RTX 3090 / 3090 Ti
24 GB
✅ Supported
NVIDIA A10G (AWS g5.xlarge)
24 GB
✅ Supported — cloud deployment
NVIDIA RTX 4080
16 GB
⚠️ Minimum — limited headroom
AMD GPUs
—
❌ Not supported
Apple Silicon
—
❌ Not supported
RecordEngine will not run without a supported NVIDIA GPU. CPU-only operation is not supported — inference speed on CPU is 20–50x slower than GPU and not viable for production use.
Storage grows with your document archive. Plan for approximately 1–5 MB per document including the original file and extracted data. At 500 documents per month, 200 GB lasts several years — but NVMe SSD is strongly recommended over SATA SSD for faster model loading on startup.
All other dependencies (Python, FastAPI, Streamlit, Ollama, etc.) are included inside the Docker containers — nothing needs to be installed separately on the host OS.