@@ -109,23 +109,22 @@ Served models can then be benchmarked using [GuideLLM](https://github.com/vllm-p
109
109
110
110
The following models are currently supported or are planned to be supported in the short term.
111
111
112
- | Model (Verifier) | Training via speculators | Deployment in vLLM | Pretrained Checkpoints |
113
- | ---------------------------------- | ---------------------------| -------------------| -------------------------|
114
- | Llama-3.1-8B-Instruct | EAGLE-3 ✅ \| HASS ✅ | ✅ | ✅ [ EAGLE-3] ( https://huggingface.co/yuhuili/EAGLE3-LLaMA3.1-Instruct-8B )
115
- | Llama-3.1-70B-Instruct | EAGLE-3 ⏳ | ✅ | ✅ [ EAGLE-3] ( https://huggingface.co/yuhuili/EAGLE3-LLaMA3.1-Instruct-70B ) |
116
- | DeepSeek-R1-Distill-LLama-8B | EAGLE-3 ❌ | ✅ | ✅ [ EAGLE-3] ( https://huggingface.co/yuhuili/EAGLE3-DeepSeek-R1-Distill-LLaMA-8B )
117
- | Qwen3-8B | EAGLE-3 ✅ | ✅ | ❌
118
- | Qwen3-14B | EAGLE-3 ❌ | ✅ | ❌
119
- | Qwen3-32B | EAGLE-3 ❌ | ✅ | ❌
120
- | Qwen3-30B-A3B | EAGLE-3 ❌ | ❌ | ❌
121
- | Qwen3-235B-A22B | EAGLE-3 ❌ | ❌ | ✅ [ EAGLE-3] ( https://huggingface.co/nvidia/Qwen3-235B-A22B-Eagle3 )
122
- | Llama-4-Scout-17B-16E-Instruct | EAGLE-3 ❌ | ❌ | ❌
123
- | Llama-4-Maverick-17B-128E-Eagle3 | EAGLE-3 ❌ | ❌ | ✅ [ EAGLE-3] ( https://huggingface.co/nvidia/Llama-4-Maverick-17B-128E-Eagle3 )
124
- | DeepSeek-R1 | EAGLE-3 ❌ | ❌ | ✅ [ HASS] ( https://huggingface.co/HArmonizedSS/HASS-DeepSeek-R1 )
112
+ | Model (Verifier) | Training via speculators | Deployment in vLLM | Pretrained Checkpoints |
113
+ | -------------------------------- | -- ---------------------- | ------------------ | -------------------------------------------------------------------------------- |
114
+ | Llama-3.1-8B-Instruct | EAGLE-3 ✅ \| HASS ✅ | ✅ | ✅ [ EAGLE-3] ( https://huggingface.co/yuhuili/EAGLE3-LLaMA3.1-Instruct-8B ) |
115
+ | Llama-3.1-70B-Instruct | EAGLE-3 ⏳ | ✅ | ✅ [ EAGLE-3] ( https://huggingface.co/yuhuili/EAGLE3-LLaMA3.1-Instruct-70B ) |
116
+ | DeepSeek-R1-Distill-LLama-8B | EAGLE-3 ❌ | ✅ | ✅ [ EAGLE-3] ( https://huggingface.co/yuhuili/EAGLE3-DeepSeek-R1-Distill-LLaMA-8B ) |
117
+ | Qwen3-8B | EAGLE-3 ✅ | ✅ | ❌ |
118
+ | Qwen3-14B | EAGLE-3 ❌ | ✅ | ❌ |
119
+ | Qwen3-32B | EAGLE-3 ❌ | ✅ | ❌ |
120
+ | Qwen3-30B-A3B | EAGLE-3 ❌ | ❌ | ❌ |
121
+ | Qwen3-235B-A22B | EAGLE-3 ❌ | ❌ | ✅ [ EAGLE-3] ( https://huggingface.co/nvidia/Qwen3-235B-A22B-Eagle3 ) |
122
+ | Llama-4-Scout-17B-16E-Instruct | EAGLE-3 ❌ | ❌ | ❌ |
123
+ | Llama-4-Maverick-17B-128E-Eagle3 | EAGLE-3 ❌ | ❌ | ✅ [ EAGLE-3] ( https://huggingface.co/nvidia/Llama-4-Maverick-17B-128E-Eagle3 ) |
124
+ | DeepSeek-R1 | EAGLE-3 ❌ | ❌ | ✅ [ HASS] ( https://huggingface.co/HArmonizedSS/HASS-DeepSeek-R1 ) |
125
125
126
126
✅ = Supported, ⏳ = In Progress, ❌ = Not Yet Supported
127
127
128
-
129
128
## License
130
129
131
130
Speculators is licensed under the [ Apache License 2.0] ( https://github.com/neuralmagic/speculators/blob/main/LICENSE ) .
0 commit comments