Ollama is the inference engine that powers the AI layer of NextAudit AI. It serves large language models through a local HTTP API, meaning all inference requests from Flowise stay within the stack’s network boundary. No prompts, audit findings, or host data are transmitted to external AI providers.Documentation Index
Fetch the complete documentation index at: https://mintlify.com/Kevin2523/nextAuditAi/llms.txt
Use this file to discover all available pages before exploring further.
Image and build
The Ollama service uses different sources depending on the environment:- Development
- Production / Test
In development, Ollama is built from a local
./ollama context. This lets you customize the image — for example, to pre-bake specific models or apply configuration changes — and rebuild quickly without pulling from a registry.Service configuration
The full service definition (shown here for prod/test) exposes the Ollama API onOLLAMA_PORT on the host, mapping to the fixed internal port 11434:
Environment variables
OLLAMA_MODELS
OLLAMA_MODELS specifies which models Ollama pre-loads when the container starts. Set this to a comma-separated list of model names (using Ollama’s model tag syntax, e.g. llama3.2,nomic-embed-text). Pre-loading avoids the latency of on-demand pulls when Flowise first requests inference.
VERSION
TheVERSION variable is used in the prod and test compose files to pin the jjsotom2k4/ollama-ai image tag:
VERSION in your environment file (prod.env, test.env) to control which image revision is deployed. This ensures that infrastructure updates are explicit and auditable.
Volume
ollama_data volume stores downloaded model weights. Models can be several gigabytes each; persisting this volume prevents re-downloading on container restart.
Internal API for Flowise
Ollama listens on port11434 inside the container. Flowise connects to it using the Docker service hostname:
http://ollama:11434.
The host-facing
OLLAMA_PORT is available for direct API calls during development (e.g., testing a model with curl), but production traffic between Flowise and Ollama uses the internal Docker network exclusively.