File tree Expand file tree Collapse file tree 2 files changed +9
-3
lines changed
Expand file tree Collapse file tree 2 files changed +9
-3
lines changed Original file line number Diff line number Diff line change @@ -11,9 +11,8 @@ checkpointer:
1111 pytorch_model-00002-of-00003.bin,
1212 pytorch_model-00003-of-00003.bin
1313 ]
14- recipe_checkpoint : null
1514 output_dir : /tmp/Llama-2-13b-hf/
16- model_type : MISTRAL
15+ model_type : LLAMA2
1716
1817device : cuda
1918dtype : bf16
@@ -22,7 +21,7 @@ seed: 1234
2221
2322# Tokenizer arguments
2423tokenizer :
25- _component_ : torchtune.models.mistral.mistral_tokenizer
24+ _component_ : torchtune.models.llama2.llama2_tokenizer
2625 path : /tmp/Llama-2-13b-hf/tokenizer.model
2726
2827# Generation arguments; defaults taken from gpt-fast
Original file line number Diff line number Diff line change 1818
1919
2020class InferenceRecipe :
21+ """
22+ Recipe for generating tokens from a dense Transformer-based LLM.
23+
24+ Currently this recipe support single-GPU generation only. Speculative
25+ decoding is not supported.
26+ """
27+
2128 def __init__ (self , cfg : DictConfig ) -> None :
2229 self ._device = utils .get_device (device = cfg .device )
2330 self ._dtype = utils .get_dtype (dtype = cfg .dtype )
You can’t perform that action at this time.
0 commit comments