Deploy RAGapp with Ollama and Qdrant easily in your own infrastructure.
Using the MODEL
environment variable, you can specify which model to use, e.g. llama3
:
MODEL=llama3 docker-compose up
If you don't specify the MODEL
variable, the default model used is phi3
, which is less capable than llama3
but faster to download.
Note: The
setup
container in thedocker-compose.yml
file will download the selected model into theollama
folder - this will take a few minutes.
Using the OLLAMA_BASE_URL
environment variables, you can specify which Ollama host to use.
If you don't specify the OLLAMA_BASE_URL
variable, the default points to the Ollama instance started by Docker Compose (http://ollama:11434
).
If you're running a local Ollama instance, you can choose to connect it to RAGapp by setting the OLLAMA_BASE_URL
variable to http://host.docker.internal:11434
:
MODEL=llama3 OLLAMA_BASE_URL=http://host.docker.internal:11434 docker-compose up
Note:
host.docker.internal
is not available on Linux machines, you'll have to use172.17.0.1
instead. For details see Issue #78.
Using a local Ollama instance is necessary if you're running RAGapp on macOS, as Docker for Mac does not support GPU acceleration.
To enable Docker access to NVIDIA GPUs on Linux, install the NVIDIA Container Toolkit.
- Chat UI: http://localhost:8000
- Admin UI: http://localhost:8000/admin