This repository was archived by the owner on Jul 4, 2025. It is now read-only.
This repository was archived by the owner on Jul 4, 2025. It is now read-only.
bug: some models failed to load if many GPU are selected #1458
Open
Description
Will be fixed by
- Add Multi-GPU Support for LlamaCpp Engine #1391
- planning: prioritize GPUs with CUDA_VISIBLE_DEVICES #1679
Original Bug report:
Jan version
0.5.3
Describe the Bug
I imported many models and for some of them, they are failing to load if I selected my both graphic cards (RTX 3060 12Go).
If I unselect one of them, the model is loaded.
It will be great if the models list could indicate if the models are supporting multi-GPU
Steps to Reproduce
- Go to Settings -> Advanced Settings
- In Choose device(s), select 2 GPUs
- Go in "My Models"
- Select "Meta-Llama-3.1-8B-Instruct-128k-Q4_0" and start it -> NOT loaded !!!
- Go in Advanced Settings
- Unselect one GPU from "Choose device(s)"
- Go in "My Models"
- Select "Meta-Llama-3.1-8B-Instruct-128k-Q4_0" and start it -> loaded !!!
Screenshots / Logs
No response
What is your OS?
- MacOS
- Windows
- Linux
Metadata
Metadata
Assignees
Labels
Type
Projects
Status
Completed