You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/content/getting_started/_index.en.md
+41-20Lines changed: 41 additions & 20 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -143,39 +143,60 @@ Note: this feature currently is available only on master builds.
143
143
You can run `local-ai` directly with a model name, and it will download the model and start the API with the model loaded.
144
144
145
145
> Don't need GPU acceleration? use the CPU images which are lighter and do not have Nvidia dependencies
146
+
> To know which version of CUDA do you have available, you can check with `nvidia-smi` or `nvcc --version`
147
+
146
148
147
149
{{< tabs >}}
148
150
{{% tab name="CPU-only" %}}
149
151
150
-
| Model | Docker command |
151
-
| --- | --- |
152
-
| phi2 |```docker run -ti -p 8080:8080 localai/localai:{{< version >}}-ffmpeg-core phi-2```|
153
-
| llava |```docker run -ti -p 8080:8080 localai/localai:{{< version >}}-ffmpeg-core llava```|
154
-
| mistral-openorca |```docker run -ti -p 8080:8080 localai/localai:{{< version >}}-ffmpeg-core mistral-openorca```|
155
-
152
+
| Model | Category | Docker command |
153
+
| --- | --- | --- |
154
+
|[phi-2](https://huggingface.co/microsoft/phi-2)| LLM |```docker run -ti -p 8080:8080 localai/localai:{{< version >}}-ffmpeg-core phi-2```|
155
+
|[llava](https://github.com/SkunkworksAI/BakLLaVA)| Multimodal LLM |```docker run -ti -p 8080:8080 localai/localai:{{< version >}}-ffmpeg-core llava```|
156
+
|[mistral-openorca](https://huggingface.co/Open-Orca/Mistral-7B-OpenOrca)| LLM |```docker run -ti -p 8080:8080 localai/localai:{{< version >}}-ffmpeg-core mistral-openorca```|
157
+
|[bert-cpp](https://github.com/skeskinen/bert.cpp)| Embeddings |```docker run -ti -p 8080:8080 localai/localai:{{< version >}}-ffmpeg-core bert-cpp```|
158
+
| all-minilm-l6-v2 | Embeddings |```docker run -ti -p 8080:8080 localai/localai:{{< version >}}-ffmpeg all-minilm-l6-v2```|
159
+
| whisper-base | Audio to Text |```docker run -ti -p 8080:8080 localai/localai:{{< version >}}-ffmpeg-core whisper-base```|
160
+
| rhasspy-voice-en-us-amy | Text to Audio |```docker run -ti -p 8080:8080 localai/localai:{{< version >}}-ffmpeg-core rhasspy-voice-en-us-amy```|
161
+
| coqui | Text to Audio |```docker run -ti -p 8080:8080 localai/localai:{{< version >}}-ffmpeg coqui```|
162
+
| bark | Text to Audio |```docker run -ti -p 8080:8080 localai/localai:{{< version >}}-ffmpeg bark```|
163
+
| vall-e-x | Text to Audio |```docker run -ti -p 8080:8080 localai/localai:{{< version >}}-ffmpeg vall-e-x```|
156
164
157
165
{{% /tab %}}
158
166
{{% tab name="GPU (CUDA 11)" %}}
159
167
160
-
> To know which version of CUDA do you have available, you can check with `nvidia-smi` or `nvcc --version`
161
168
162
-
| Model | Docker command |
163
-
| --- | --- |
164
-
| phi-2 |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda11-core phi-2```|
165
-
| llava |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda11-core llava```|
166
-
| mistral-openorca |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda11-core mistral-openorca```|
169
+
170
+
| Model | Category | Docker command |
171
+
| --- | --- | --- |
172
+
|[phi-2](https://huggingface.co/microsoft/phi-2)| LLM |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda11-core phi-2```|
173
+
|[llava](https://github.com/SkunkworksAI/BakLLaVA)| Multimodal LLM |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda11-core llava```|
174
+
|[mistral-openorca](https://huggingface.co/Open-Orca/Mistral-7B-OpenOrca)| LLM |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda11-core mistral-openorca```|
175
+
|[bert-cpp](https://github.com/skeskinen/bert.cpp)| Embeddings |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda11-core bert-cpp```|
176
+
|[all-minilm-l6-v2](https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2)| Embeddings |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda11 all-minilm-l6-v2```|
177
+
| whisper-base | Audio to Text |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda11-core whisper-base```|
178
+
| rhasspy-voice-en-us-amy | Text to Audio |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda11-core rhasspy-voice-en-us-amy```|
179
+
| coqui | Text to Audio |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda11 coqui```|
180
+
| bark | Text to Audio |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda11 bark```|
181
+
| vall-e-x | Text to Audio |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda11 vall-e-x```|
167
182
168
183
{{% /tab %}}
169
184
170
-
{{% tab name="GPU (CUDA 12)" %}}
171
185
172
-
> To know which version of CUDA do you have available, you can check with `nvidia-smi` or `nvcc --version`
186
+
{{% tab name="GPU (CUDA 12)" %}}
173
187
174
-
| Model | Docker command |
175
-
| --- | --- |
176
-
| phi-2 |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda12-core phi-2```|
177
-
| llava |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda12-core llava```|
178
-
| mistral-openorca |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda12-core mistral-openorca```|
188
+
| Model | Category | Docker command |
189
+
| --- | --- | --- |
190
+
|[phi-2](https://huggingface.co/microsoft/phi-2)| LLM |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda12-core phi-2```|
191
+
|[llava](https://github.com/SkunkworksAI/BakLLaVA)| Multimodal LLM |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda12-core llava```|
192
+
|[mistral-openorca](https://huggingface.co/Open-Orca/Mistral-7B-OpenOrca)| LLM |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda12-core mistral-openorca```|
193
+
| bert-cpp | Embeddings |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda12-core bert-cpp```|
194
+
| all-minilm-l6-v2 | Embeddings |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda12 all-minilm-l6-v2```|
195
+
| whisper-base | Audio to Text |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda12-core whisper-base```|
196
+
| rhasspy-voice-en-us-amy | Text to Audio |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda12-core rhasspy-voice-en-us-amy```|
197
+
| coqui | Text to Audio |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda12 coqui```|
198
+
| bark | Text to Audio |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda12 bark```|
199
+
| vall-e-x | Text to Audio |```docker run -ti -p 8080:8080 --gpus all localai/localai:{{< version >}}-cublas-cuda12 vall-e-x```|
179
200
180
201
{{% /tab %}}
181
202
@@ -201,7 +222,7 @@ For example, to start localai with phi-2, it's possible for instance to also use
201
222
docker run -p 8080:8080 localai/localai:{{< version >}}-ffmpeg-core https://gist.githubusercontent.com/mudler/ad601a0488b497b69ec549150d9edd18/raw/a8a8869ef1bb7e3830bf5c0bae29a0cce991ff8d/phi-2.yaml
202
223
```
203
224
204
-
The file should be a valid YAML configuration file, for the full syntax see [advanced]({{%relref "advanced" %}}).
225
+
The file should be a valid LocalAI YAML configuration file, for the full syntax see [advanced]({{%relref "advanced" %}}).
Copy file name to clipboardExpand all lines: docs/content/model-compatibility/_index.en.md
+5-2Lines changed: 5 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -43,15 +43,18 @@ Besides llama based models, LocalAI is compatible also with other architectures.
43
43
|[langchain-huggingface](https://github.com/tmc/langchaingo)| Any text generators available on HuggingFace through API | yes | GPT | no | no | N/A |
44
44
|[piper](https://github.com/rhasspy/piper) ([binding](https://github.com/mudler/go-piper)) | Any piper onnx model | no | Text to voice | no | no | N/A |
45
45
|[falcon](https://github.com/cmp-nct/ggllm.cpp/tree/c12b2d65f732a0d8846db2244e070f0f3e73505c) ([binding](https://github.com/mudler/go-ggllm.cpp)) | Falcon ***| yes | GPT | no | yes | CUDA |
46
-
|`huggingface-embeddings`[sentence-transformers](https://github.com/UKPLab/sentence-transformers)| BERT | no | Embeddings only | yes | no | N/A |
46
+
|[sentencetransformers](https://github.com/UKPLab/sentence-transformers)| BERT | no | Embeddings only | yes | no | N/A |
47
47
|`bark`| bark | no | Audio generation | no | no | yes |
48
-
|`AutoGPTQ`| GPTQ | yes | GPT | yes | no | N/A |
48
+
|`autogptq`| GPTQ | yes | GPT | yes | no | N/A |
49
49
|`exllama`| GPTQ | yes | GPT only | no | no | N/A |
50
50
|`diffusers`| SD,... | no | Image generation | no | no | N/A |
51
51
|`vall-e-x`| Vall-E | no | Audio generation and Voice cloning | no | no | CPU/CUDA |
52
52
|`vllm`| Various GPTs and quantization formats | yes | GPT | no | no | CPU/CUDA |
53
53
|`exllama2`| GPTQ | yes | GPT only | no | no | N/A |
54
54
|`transformers-musicgen`|| no | Audio generation | no | no | N/A |
55
+
|[tinydream](https://github.com/symisc/tiny-dream#tiny-dreaman-embedded-header-only-stable-diffusion-inference-c-librarypixlabiotiny-dream)| stablediffusion | no | Image | no | no | N/A |
56
+
|`coqui`| Coqui | no | Audio generation and Voice cloning | no | no | CPU/CUDA |
57
+
|`petals`| Various GPTs and quantization formats | yes | GPT | no | no | CPU/CUDA |
55
58
56
59
Note: any backend name listed above can be used in the `backend` field of the model configuration file (See [the advanced section]({{%relref "advanced" %}})).
0 commit comments