-
Notifications
You must be signed in to change notification settings - Fork 3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[DO NOT UNPIN] ORT 1.18.0 Release Candidates available for testing #20558
Comments
Is there GPU python package for CUDA 12? |
OrtApi last entry from 1.18 is not protected by the usual https://github.com/microsoft/onnxruntime/blob/rel-1.18.0/onnxruntime/core/session/onnxruntime_c_api.cc#L2729 |
Thanks for catching that! @yihonglyu would you mind looking into this? |
@Hopobcn @sophies927 A fix is here #20671 |
Hi, I just found out that ScatterND nodes are now placed on CPU. They were placed on GPU before. See below:
With ORT 1.17.3:
Is it because of the upgrade to TensorRT 10? |
|
|
@JulienTheron can you provide us a repro test case or point us to a public model that can be used to investigate further? Are you also updating TensorRT version from 8.6 to 10.0 along with the ORT version update? +@chilo-ms |
Yes I've used the default TensorRT version for each ORT release, so 8.6 for 1.17.3 and 10 for 1.18.0. |
Here is a version of our model that we can share: To repro, just create a TensorRT session like this:
Thanks. |
There will be for the release. We had to make some updates to make the CUDA 12 package publication a bit easier (see this PR), which is why the packages weren't included in the first round release candidates. |
Here's a follow-up to the TensorRT issue I mentioned. Line 41 in 4573740
onnx_tensorrt;https://github.com/onnx/onnx-tensorrt/archive/bacfaaa951653cd4e72efe727a543567cb38f7de.zip;26434329612e804164ab7baa6ae629ada56c1b26 , then built with TensorRT 8.6.
The issue is now gone. I haven't quite figured why ScatterND is not placed on the TensorRT EP with TRT 10, but I guess the issue has nothing to do with the ONNX Runtime? |
Thanks for calling this out! Appears to be an issue with the model input caused by an ONNX update, rather than an issue with ORT 1.18. |
@sophies927, Regarding to #20737, shall we achieve older pypi package before release (also publish new packages to aiinfra.pkgs.visualstudio.com as a backup)? It seems that every new release will cause an older version be deleted. |
Yes, I think that should definitely be the plan moving forward |
The ORT 1.18.0 release is now available here: https://github.com/microsoft/onnxruntime/releases |
@sophies927 Thank you for the release! Unfortunately, Python's |
ORT 1.18 will be released soon, and release candidate builds are available now for testing. If you encounter issues, please report them by responding to this issue.
Release branch: rel-1.18.0
Release manager: @yihonglyu
GPU: 1.18.0.dev20240430005
GPU (CUDA/TRT): 1.18.0-dev-20240430-2214-204f1f59b9
DirectML: 1.18.0-dev-20240501-0503-204f1f59b9
WindowsAI: 1.18.0-dev-20240430-1128-204f1f5
onnxruntime-react-native: 1.18.0-dev.20240430-204f1f59b9
onnxruntime-web: 1.18.0-dev.20240430-204f1f59b9
GPU: 1.18.0-rc1
The text was updated successfully, but these errors were encountered: