-
Notifications
You must be signed in to change notification settings - Fork 218
Update onnxruntime-gpu #697
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
base: main
Are you sure you want to change the base?
Conversation
…updated onnxruntime-gpu in whisper example Signed-off-by: Hrishith Thadicherla <[email protected]>
Signed-off-by: Hrishith Thadicherla <[email protected]>
|
ONNX unit tests failing: https://github.com/NVIDIA/Model-Optimizer/actions/runs/20257328002/job/58162067752?pr=697 Is this because of bumping ort from 1.22 to 1.23? |
Yes, the torch tests are failing. I'm looking into what exactly the issue is |
i figured out what the issue is, essentially in tests/unit/torch/quantization/test_onnx_export_cpu.py. we are setting a seed here If we look at onnx_export_tester it fails at this line Changing atol and rtol to 1e-3 made the tests pass. Sounds like a floating point error |
|
Setting seed to set_seed(90) also makes the tests pass. This does look like a floating point error. Will change the seed for now i guess. Should we raise a bug for this though? |
@ajrasane thoughts? |
Signed-off-by: Hrishith Thadicherla <[email protected]>
Signed-off-by: Hrishith Thadicherla <[email protected]>
Codecov Report✅ All modified and coverable lines are covered by tests. Additional details and impacted files@@ Coverage Diff @@
## main #697 +/- ##
==========================================
- Coverage 74.72% 74.70% -0.03%
==========================================
Files 192 192
Lines 18828 18828
==========================================
- Hits 14070 14066 -4
- Misses 4758 4762 +4 ☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
|
All checks are passing now. |
|
@kevalmorabia97 @ajrasane the github CI/CD tests are passing but i have personally only tested this in windows side. So is this okay or should there be some more validation from linux side with the latest version of onnxruntime-gpu 1.23.2 ? |
What does this PR do?
Type of change: Bug fix
Overview: Updated setup.py to use only onnxruntime-gpu and removed onnxruntime-directml as dependency.
Also changed onnxruntime-gpu version in examples.
Testing
Tested int4 quantization and MMLU benchmark with updated onnxruntime-gpu , working as expected