-
Notifications
You must be signed in to change notification settings - Fork 3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature Request] Support CUDA 12 ? #18850
Comments
I believe it is not supporting CUDA 12, I have tried 12.2 and getting below error. '[ErrorCode:RuntimeException] D:\a_work\1\s\onnxruntime\core\session\provider_bridge_ort.cc:1193 onnxruntime::ProviderLibrary::Get [ONNXRuntimeError] : 1 : FAIL : LoadLibrary failed with error 126 "" when trying to load |
Using the nightly builds via
|
Does anyone know how many days need to support CUDA12? |
no idea |
current version 1.16.x support cuda 12, but must build it from source |
In my case it didn't help to use the latest nightly (...002). It's still looking for cuda11/libcublasLt.so.11 while cuda12 has libcublasLt.so.12.
|
here my instructions to build onnxruntime with cuda 12 on colab
take ~2h, no need gpu session (remove tensorrt if not needed) wheel file created: easier solution but only temporary: googlecolab/colabtools#4214 (comment) |
For people who will come after me, you can download whl here |
Thank you! Alternatively, you can install the package via pip: Previously, I used a slightly different package, maybe someone will find it useful too: |
Nice to see you Oleksandr. What is the actual difference between both packages? I assume the one on azure is from the onnxruntime team and based on the latest build. The other by one of the onnxruntime devs and locked to a version? Using a locked version feels more suitable. |
Hi Henry, good to see you too. In my case, at the moment both packages work in Colab with Cuda 12, so it seems to me that there is no big difference which one to use. At least for now there are no critical bugs in the package based on the latest build. |
Thank you so much for addressing and fixing the existing defects in onnxruntime-gpu. Your efforts are truly appreciated! |
So the nightly onnxruntime-gpu (v 1.17) should work? |
@AndreyOrb, I tested it on Ubuntu 22.04.3 LTS and CUDA 12.2 (Google Colab). @henryruhs Henry, after small tests of both packages, it turned out that Codeformer converted to .onnx does not work with
|
@netrunner-exe I can report have issues with |
@netrunner-exe Thanks for sharing the details. |
The CUDA 12 support will be official in upcoming Onnxrtime 1.17 realase, in a week. Official instruction on how to download it will be provided then. |
Any updates? (Looking for support 12.3 on local machine(ubuntu)) |
solution from https://onnxruntime.ai/docs/install/
|
Describe the feature request
Is ONNXRUNTIME support CUDA 12 ?
Describe scenario use case
i got error on google colab using onnxruntime but it doesnt want to use T4 GPU
The text was updated successfully, but these errors were encountered: