Ready to run LLM stack configuration for Kuberenetes using Open Web UI / Ollama / HuggingFace TGI
kubectl create ns k8s-llm-stack
kubectl apply -f https://raw.githubusercontent.com/kubernetes/ingress-nginx/controller-v1.8.2/deploy/static/provider/cloud/deploy.yaml
kubectl apply -f nvidia-runtime.yaml
kubectl apply -f open-webui.yaml
kubectl apply -f ollama.yaml
kubectl apply -f tgi.yaml
kubectl apply -f vllm.yaml