Skip to content

Conversation

aarnphm
Copy link
Collaborator

@aarnphm aarnphm commented Aug 13, 2025

Continuation of #17420

This PR introduces the args --structured-output-config as a way to unify all related structured outputs config in one CLI field.
This would help simplify general UX for specifying custom options with backends.

I also remove all previous guided_decoding options

This would also be considered breaking. There will be no --guided-decoding-* option anymore. Instead, you should use --structured-outputs-config '{...}' or --structured-outputs-config.backend outlines

Signed-off-by: Aaron Pham [email protected]
Signed-off-by: Harry Mellor [email protected]
Co-authored-by: Nick Hill [email protected]
Co-authored-by: Harry Mellor [email protected]

Copy link

👋 Hi! Thank you for contributing to the vLLM project.

💬 Join our developer Slack at https://slack.vllm.ai to discuss your PR in #pr-reviews, coordinate on features in #feat- channels, or join special interest groups in #sig- channels.

Just a reminder: PRs would not trigger full CI run by default. Instead, it would only run fastcheck CI which starts running only a small and essential subset of CI tests to quickly catch errors. You can run other CI tests on top of those by going to your fastcheck build on Buildkite UI (linked in the PR checks section) and unblock them. If you do not have permission to unblock, ping simon-mo or khluu to add you in our Buildkite org.

Once the PR is approved and ready to go, your PR reviewer(s) can run CI to test the changes comprehensively before merging.

To run CI, PR reviewers can either: Add ready label to the PR or enable auto-merge.

🚀

@mergify mergify bot added documentation Improvements or additions to documentation frontend performance Performance-related issues structured-output v1 labels Aug 13, 2025
Copy link

mergify bot commented Aug 13, 2025

This pull request has merge conflicts that must be resolved before it can be
merged. Please rebase the PR, @aarnphm.

https://docs.github.com/en/pull-requests/collaborating-with-pull-requests/working-with-forks/syncing-a-fork

@aarnphm aarnphm force-pushed the feat/decoding-args-rename-all branch from d3ac885 to 69068cd Compare August 13, 2025 00:47
@hmellor
Copy link
Member

hmellor commented Sep 17, 2025

Kernel and model executor failures appear to have come from main.


The entrypoints failure is legitimate. For some reason, tool calling is no longer using structured outputs. We can see this is the case by running pytest -vsx tests/entrypoints/openai/test_chat.py::test_named_tool_use[True].

Before this PR, the schema was included in the user message which would allow the model to cheat in the event that structured outputs stopped working.

The modification I've made to test_named_tool_use() (removing the schema from the user message) passes on main, meaning that structured outputs were being used for this test before this PR.

Signed-off-by: Harry Mellor <[email protected]>
@DarkLight1337 DarkLight1337 merged commit 29283e8 into vllm-project:main Sep 18, 2025
79 checks passed
xuechendi pushed a commit to vllm-project/vllm-gaudi that referenced this pull request Sep 18, 2025
Culprit commit: vllm-project/vllm#22772

---------

Signed-off-by: Agata Dobrzyniewicz <[email protected]>
debroy-rh pushed a commit to debroy-rh/vllm that referenced this pull request Sep 19, 2025
pytorchmergebot pushed a commit to pytorch/pytorch that referenced this pull request Sep 20, 2025
Yikun pushed a commit to vllm-project/vllm-ascend that referenced this pull request Sep 20, 2025
…2907)

### What this PR does / why we need it?
1. This pr bump vllm commit to
vllm-project/vllm@6d8246a
2. fix upstream changes vllm-project/vllm#24548
abort multi-modal kwargs, make vllm main and `v0.10.2` both adaptable
3. fix metadata_builder changes introduced by
vllm-project/vllm#23693
4. fix `structured_outputs_config` changes introduced by
vllm-project/vllm#22772
5. fix `moe_config` changes introduced by
vllm-project/vllm#22537

Co-authored-by:  MengqingCao <[email protected]>
Co-authored-by:  Yikun Jiang <[email protected]>


- vLLM version: v0.10.2
- vLLM main:
vllm-project/vllm@c60e613

---------

Signed-off-by: wangli <[email protected]>
Signed-off-by: MengqingCao <[email protected]>
Co-authored-by: MengqingCao <[email protected]>
weijinqian0 pushed a commit to weijinqian0/vllm-ascend that referenced this pull request Sep 22, 2025
…llm-project#2907)

### What this PR does / why we need it?
1. This pr bump vllm commit to
vllm-project/vllm@6d8246a
2. fix upstream changes vllm-project/vllm#24548
abort multi-modal kwargs, make vllm main and `v0.10.2` both adaptable
3. fix metadata_builder changes introduced by
vllm-project/vllm#23693
4. fix `structured_outputs_config` changes introduced by
vllm-project/vllm#22772
5. fix `moe_config` changes introduced by
vllm-project/vllm#22537

Co-authored-by:  MengqingCao <[email protected]>
Co-authored-by:  Yikun Jiang <[email protected]>

- vLLM version: v0.10.2
- vLLM main:
vllm-project/vllm@c60e613

---------

Signed-off-by: wangli <[email protected]>
Signed-off-by: MengqingCao <[email protected]>
Co-authored-by: MengqingCao <[email protected]>
mansiag05 pushed a commit to mansiag05/pytorch that referenced this pull request Sep 22, 2025
Mercykid-bash pushed a commit to Mercykid-bash/vllm-ascend that referenced this pull request Sep 22, 2025
…llm-project#2907)

### What this PR does / why we need it?
1. This pr bump vllm commit to
vllm-project/vllm@6d8246a
2. fix upstream changes vllm-project/vllm#24548
abort multi-modal kwargs, make vllm main and `v0.10.2` both adaptable
3. fix metadata_builder changes introduced by
vllm-project/vllm#23693
4. fix `structured_outputs_config` changes introduced by
vllm-project/vllm#22772
5. fix `moe_config` changes introduced by
vllm-project/vllm#22537

Co-authored-by:  MengqingCao <[email protected]>
Co-authored-by:  Yikun Jiang <[email protected]>

- vLLM version: v0.10.2
- vLLM main:
vllm-project/vllm@c60e613

---------

Signed-off-by: wangli <[email protected]>
Signed-off-by: MengqingCao <[email protected]>
Co-authored-by: MengqingCao <[email protected]>
Signed-off-by: Che Ruan <[email protected]>
Mercykid-bash pushed a commit to Mercykid-bash/vllm-ascend that referenced this pull request Sep 22, 2025
…llm-project#2907)

### What this PR does / why we need it?
1. This pr bump vllm commit to
vllm-project/vllm@6d8246a
2. fix upstream changes vllm-project/vllm#24548
abort multi-modal kwargs, make vllm main and `v0.10.2` both adaptable
3. fix metadata_builder changes introduced by
vllm-project/vllm#23693
4. fix `structured_outputs_config` changes introduced by
vllm-project/vllm#22772
5. fix `moe_config` changes introduced by
vllm-project/vllm#22537

Co-authored-by:  MengqingCao <[email protected]>
Co-authored-by:  Yikun Jiang <[email protected]>

- vLLM version: v0.10.2
- vLLM main:
vllm-project/vllm@c60e613

---------

Signed-off-by: wangli <[email protected]>
Signed-off-by: MengqingCao <[email protected]>
Co-authored-by: MengqingCao <[email protected]>
Signed-off-by: Che Ruan <[email protected]>
cleonard530 pushed a commit to cleonard530/pytorch that referenced this pull request Sep 22, 2025
@simon-mo
Copy link
Collaborator

@aarnphm can we add backward compatibility for one version so people know how to migrate?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ci/build documentation Improvements or additions to documentation frontend performance Performance-related issues ready ONLY add when PR is ready to merge/full CI is needed structured-output tool-calling v1
Projects
Status: Done
Status: Done
Development

Successfully merging this pull request may close these issues.

5 participants