issue-2394: use deployment variable instead of model for embeddings API call#2395
Merged
danielaskdd merged 1 commit intoHKUDS:mainfrom Nov 20, 2025
Merged
issue-2394: use deployment variable instead of model for embeddings API call#2395danielaskdd merged 1 commit intoHKUDS:mainfrom
danielaskdd merged 1 commit intoHKUDS:mainfrom
Conversation
Collaborator
|
@codex review |
|
Codex Review: Didn't find any major issues. Can't wait for the next one! ℹ️ About Codex in GitHubYour team has set up Codex to review pull requests in this repo. Reviews are triggered when you
If Codex has suggestions, it will comment; otherwise it will react with 👍. Codex can also answer questions or update the PR. Try commenting "@codex address that feedback". |
Collaborator
|
Using deployment directly as the model parameter leads to:
We should prioritize the response = await openai_async_client.embeddings.create(
model=model or deployment, input=texts, encoding_format="float"
) |
2 tasks
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Description
This pull request fixes a critical bug in azure_openai_embed() where the Azure embeddings API is called with the wrong model value.
The function correctly resolves the Azure deployment name into the deployment variable, but it mistakenly passes the unused model parameter (typically None) to embeddings.create(). This causes Azure to look for a deployment literally named "None" and results in a 404 DeploymentNotFound error.
This PR updates the call to:
model=deployment
which is the correct behavior and aligns with how Azure OpenAI routing works.
Related Issues
Fixes: #
(If you haven’t created the issue yet, you can add it afterward.)
Changes Made
Updated lightrag/llm/azure_openai.py in azure_openai_embed():
Replaced model=model with model=deployment
Added in-line comment explaining why Azure requires model=deployment
Verified the change through local testing with Azure embeddings
Checklist
Changes tested locally
Code reviewed
Documentation updated (not required for this change)
Unit tests added (optional — no tests currently exist for this module)
Additional Notes
This fix resolves a blocking issue for all Azure OpenAI embedding users.
Without this change, embedding requests always fail with 404, even with correct Azure configuration.
Happy to make any adjustments requested by maintainers.