Skip to content

Conversation

anmarques
Copy link
Collaborator

This PR lists the model architectures currently supported and planned to be supported with the speculators format in vLLM. It specifies model architectures that are currently supported and others for which there are trained checkpoints, from Red Hat AI, EAGLE or HASS.

@anmarques anmarques requested a review from markurtz August 19, 2025 19:41
Copy link

github-actions bot commented Aug 19, 2025

📦 Build Artifacts Available
The build artifacts (`.whl` and `.tar.gz`) have been successfully generated and are available for download: https://github.com/vllm-project/speculators/actions/runs/17306160423/artifacts/3877147489.
They will be retained for up to 30 days.
Commit: 8d8729f

…vLLM Inference session

Signed-off-by: Alexandre Marques <[email protected]>
This update introduces a new section detailing the currently supported model architectures, including Llama-3 and Qwen3, along with their trained checkpoints. This addition enhances the documentation by providing users with clear information on available models.

Signed-off-by: Alexandre Marques <[email protected]>
Signed-off-by: Alexandre Marques <[email protected]>
@anmarques anmarques force-pushed the model_architectures branch from 75f13f5 to 2a13cc4 Compare August 19, 2025 22:20
Copy link
Collaborator

@markurtz markurtz left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@anmarques I'd like to adapt this further to ideally balance discoverability, clarity, and depth. Can we change this into a table format and either add it as a new sub section in either the Overview (after key features) or in Resources (### Models and then ### Research Implementations).

I'm thinking something like the following for the table format so it is compact and informative as a really quick example:

Model / Verifier Training & Creation Deployment (vLLM) Pretrained Checkpoints
Qwen3 (8B) Eagle 3 ✔️, HASS ⏳ Eagle 3 ✔️ Eagle 3 ✔️
Llama-3.1 (70B) Eagle 3 ⏳ Eagle 3 ⏳ ✖️

✔️ = Supported,⏳ = In Progress,✖️ = Not Available

@rahul-tuli
Copy link
Collaborator

I like having this central place for everything supported, LGTM once the existing comments are resolved!

@dsikka
Copy link
Collaborator

dsikka commented Aug 28, 2025

I agree with the suggested changes. Rahul and I are putting together a similar table for in-depth vLLM overview that we'll add in the next week or so.

@anmarques anmarques requested review from markurtz and dsikka August 28, 2025 19:40
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants