Install RecordEngine on a Linux server in under an hour — step by step.
This guide walks through a complete RecordEngine installation on a fresh Ubuntu server. By the end you’ll have a running instance accessible from your browser over HTTPS.
Before starting, confirm your server meets the System Requirements — particularly the NVIDIA GPU requirement. RecordEngine will not run without a supported GPU.
docker run --rm --gpus all nvidia/cuda:11.8.0-base-ubuntu22.04 nvidia-smi
You should see the same GPU information as from the host. If this fails, the NVIDIA Container Toolkit is not configured correctly — do not proceed until it works.
Set the SERVER_URL environment variable in both the ui and api service sections. This must be your server’s public URL — the one users will access in their browser:
YAML indentation is significant. Every environment variable must be indented exactly 2 spaces under the environment: key. A single extra space will silently prevent the variable from loading.
The AI model variables are pre-configured correctly — leave them as-is unless you have a specific reason to change them.
The AI model needs to be downloaded once into the Ollama container. This is a ~6.6 GB download:
docker exec ollama ollama pull qwen3.5:9b
This takes 5–20 minutes depending on your internet connection. Once the download completes, warm up the model to load it into GPU memory:
docker exec ollama ollama run qwen3.5:9b "Ready" --keepalive -1
The first run takes 1–3 minutes as the model loads. You’ll see a response when it’s ready. Press Ctrl+D to exit the interactive session.
This warmup step is important. The first real document a user processes will otherwise trigger the model load, causing an apparent delay of several minutes. Running the warmup now means the first user experience is fast.