Replies: 2 comments 1 reply
-
Hi, LMDeploy does support Llama 2 both in ckpt and huggingface transformers model format. You can try the following command to convert it to LMDeploy required format python lmdeploy/serve/turbomind/deploy.py llama2 /the/path/of/original/llama2/model --model-format llama --tokenizer-path /the/path/of/tokenizer/model/of/original/llama2 |
Beta Was this translation helpful? Give feedback.
1 reply
-
Hi, @realhaik |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I am reading this article and unfortunately I see that the article uses hugging face llama2 model. Does this mean that the original meta llama2 is not supported?
https://openmmlab.medium.com/deploy-llama-2-models-easily-with-lmdeploy-1cb001d70290
This is a real deal breaker, because the hf model is defective. The results that hf model produces are a joke, completely broken. The original meta model works quite nice.
I feel really sorry for anyone who is wasting his time with the hf model.
Beta Was this translation helpful? Give feedback.
All reactions