Skip to content

Commit f0a4df1

Browse files
committed
Update setup instructions to match the revised docker-compose structure
1 parent a7b1016 commit f0a4df1

File tree

1 file changed

+24
-11
lines changed

1 file changed

+24
-11
lines changed

SETUP.md

Lines changed: 24 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -5,7 +5,7 @@ This document provides step-by-step instructions for setting up the RAG-enhanced
55
## Prerequisites
66

77
- Docker and Docker Compose
8-
- Docker's Model Runner with llama3.2 model
8+
- Docker model pull permissions for llama3.2 model
99

1010
## Setup Steps
1111

@@ -17,36 +17,49 @@ This document provides step-by-step instructions for setting up the RAG-enhanced
1717
docker model pull ai/llama3.2:1B-Q8_0
1818
```
1919

20-
2. **Start the Model Runner**
20+
2. **Start all services including model-runner**
2121

22-
Run the model using Docker's Model Runner functionality:
22+
Start the application with one command:
2323

2424
```bash
25-
docker model run ai/llama3.2:1B-Q8_0 --port 12434 --server
25+
docker compose -f docker-compose.rag.yml up -d --build
2626
```
2727

28-
This exposes the model on port 12434.
29-
30-
3. **Start the RAG application**
28+
3. **Run the model separately**
3129

32-
In a separate terminal, start the RAG application using:
30+
After the containers are up, you need to start the model separately because the model-runner container doesn't execute the model by default:
3331

3432
```bash
35-
docker compose -f docker-compose.rag.yml up -d --build
33+
# First, find the container ID
34+
docker ps | grep model-runner
35+
36+
# Then exec into the container and run the model
37+
docker exec -it CONTAINER_ID /bin/bash
38+
39+
# Once inside the container, run:
40+
model run --server --port 12434
3641
```
3742

43+
Leave this terminal open to keep the model running.
44+
3845
4. **Access the application**
3946

40-
Once all containers are running, access the application at:
47+
Once all containers are running and the model is active, access the application at:
4148

4249
- Frontend: [http://localhost:3000](http://localhost:3000)
4350
- Grafana: [http://localhost:3001](http://localhost:3001) (admin/admin)
4451
- Jaeger: [http://localhost:16686](http://localhost:16686)
4552

4653
## Troubleshooting
4754

48-
- **Connection issues to Model Runner**: Make sure the model is running with `docker model ls` and that it's accessible on port 12434
55+
- **Model runner errors**: If you have issues starting the model inside the container, you can try running it on your host system:
56+
```bash
57+
docker model run ai/llama3.2:1B-Q8_0 --port 12434 --server
58+
```
59+
Then update `BASE_URL` in docker-compose.rag.yml to: `http://host.docker.internal:12434/engines/llama.cpp/v1/`
60+
4961
- **Backend connectivity issues**: Check logs with `docker compose -f docker-compose.rag.yml logs backend`
62+
5063
- **Vector database errors**: Verify ChromaDB is running with `docker compose -f docker-compose.rag.yml logs vectordb`
5164

5265
## Architecture

0 commit comments

Comments
 (0)