@@ -5,7 +5,7 @@ This document provides step-by-step instructions for setting up the RAG-enhanced
55## Prerequisites
66
77- Docker and Docker Compose
8- - Docker's Model Runner with llama3.2 model
8+ - Docker model pull permissions for llama3.2 model
99
1010## Setup Steps
1111
@@ -17,36 +17,49 @@ This document provides step-by-step instructions for setting up the RAG-enhanced
1717 docker model pull ai/llama3.2:1B-Q8_0
1818 ```
1919
20- 2 . ** Start the Model Runner **
20+ 2 . ** Start all services including model-runner **
2121
22- Run the model using Docker's Model Runner functionality :
22+ Start the application with one command :
2323
2424 ``` bash
25- docker model run ai/llama3.2:1B-Q8_0 --port 12434 --server
25+ docker compose -f docker-compose.rag.yml up -d --build
2626 ```
2727
28- This exposes the model on port 12434.
29-
30- 3 . ** Start the RAG application**
28+ 3 . ** Run the model separately**
3129
32- In a separate terminal, start the RAG application using :
30+ After the containers are up, you need to start the model separately because the model-runner container doesn't execute the model by default :
3331
3432 ``` bash
35- docker compose -f docker-compose.rag.yml up -d --build
33+ # First, find the container ID
34+ docker ps | grep model-runner
35+
36+ # Then exec into the container and run the model
37+ docker exec -it CONTAINER_ID /bin/bash
38+
39+ # Once inside the container, run:
40+ model run --server --port 12434
3641 ```
3742
43+ Leave this terminal open to keep the model running.
44+
38454 . ** Access the application**
3946
40- Once all containers are running, access the application at:
47+ Once all containers are running and the model is active , access the application at:
4148
4249 - Frontend: [ http://localhost:3000 ] ( http://localhost:3000 )
4350 - Grafana: [ http://localhost:3001 ] ( http://localhost:3001 ) (admin/admin)
4451 - Jaeger: [ http://localhost:16686 ] ( http://localhost:16686 )
4552
4653## Troubleshooting
4754
48- - ** Connection issues to Model Runner** : Make sure the model is running with ` docker model ls ` and that it's accessible on port 12434
55+ - ** Model runner errors** : If you have issues starting the model inside the container, you can try running it on your host system:
56+ ``` bash
57+ docker model run ai/llama3.2:1B-Q8_0 --port 12434 --server
58+ ```
59+ Then update ` BASE_URL ` in docker-compose.rag.yml to: ` http://host.docker.internal:12434/engines/llama.cpp/v1/ `
60+
4961- ** Backend connectivity issues** : Check logs with ` docker compose -f docker-compose.rag.yml logs backend `
62+
5063- ** Vector database errors** : Verify ChromaDB is running with ` docker compose -f docker-compose.rag.yml logs vectordb `
5164
5265## Architecture
0 commit comments