- "- This code uses the RoPE **split-halves** style, which matches the Hugging Face Transformers implementation ([modeling\\_llama.py](https://github.com/huggingface/transformers/blob/e42587f596181396e1c4b63660abf0c736b10dae/src/transformers/models/llama/modeling_llama.py#L173-L188)).<br> The original RoPE paper and Meta’s official LLaMA-2 repo, however, use the **interleaved (even/odd)** version ([llama/model.py](https://github.com/meta-llama/llama/blob/6c7fe276574e78057f917549435a2554000a876d/llama/model.py#L64-L74)).<br> Both variants work fine — just be aware of this difference to avoid confusion."
0 commit comments