-
Notifications
You must be signed in to change notification settings - Fork 177
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: Running Olive with ROCMExecutionProvider. #667
Comments
Hi, Thanks for bringing this up! "ROCmExecutionProvider" is a typo for "ROCMExecutionProvider". With regard to the GroupNorm error, this is because the options for the unet example were set for the DML EP which supports Can you try the example again after setting We haven't tested the example with Rocm ep so there might be other incompatibilities with the rocm ep. |
## Describe your changes As described in #667, the name for rocm ep has a typo. onnxruntime uses `ROCMExecutionProvider` as the name. ## Checklist before requesting a review - [ ] Add unit tests for this change. - [ ] Make sure all tests can pass. - [ ] Update documents if necessary. - [ ] Format your code by running `pre-commit run --all-files` - [ ] Is this a user-facing change? If yes, give a description of this change to be included in the release notes. ## (Optional) Issue link
Thank you for your kind reply. Its official name is ROCm so I think onnxruntime's is typo but I understand for now. I now get the following error.
And I'm getting the error below since I first tried optimization with ROCMExecutionProvider. This message occurs not only in UNet but also in other models, but does not stop optimization.
|
I found it is because of float16. I changed float16 to false and I got this error on loading ort model after optimization.
Where the error occurred: submodels = ("text_encoder", "unet", "vae_encoder", "vae_decoder",)
for submodel in submodels:
kwargs[submodel] = diffusers.OnnxRuntimeModel.from_pretrained(
os.path.dirname(optimized_model_paths[submodel]),
) |
This looks like some other transformer optimization options in the example that are not compatible with ROCm EP. Because the example was only tested with DML EP, I am not aware of which. |
I did. It took longer time than before, and I got the same error.
|
I found microsoft/onnxruntime#17837 and added
The optimized model was larger in size than unoptimized one, the generation speed was slower than usual, and the results were corrupted. |
did you solve the issue? I meet a simliar issue. even thtough provider=DMLProvider in my enviroment. |
I'm getting this error nowadays.
This error occurs when I'm trying to optimize unet.
I get another error.
EnvironmentWindows 11 23H2 torch==2.5.0.dev20240706+rocm6.1 |
What happened?
I was able to get
onnxruntime-training 1.16.1+rocm56
from onnxruntime.ai and it includesROCMExecutionProvider
. But I found out that Olive needs aROCmExecutionProvider
. I addedROCMExecutionProvider
toAcceleratorLookup.EXECUTION_PROVIDERS
, but I got the error below when optimizing unet. What is the difference betweenROCmExecutionProvider
andROCMExecutionProvider
? IsROCMExectionProvider
not supported?Version?
torch==2.2.0.dev20231024+rocm5.6
torchvision==0.17.0.dev20231024+rocm5.6
olive-ai==0.3.3
onnxruntime==1.16.1
onnxruntime-training==1.16.1+rocm56
The text was updated successfully, but these errors were encountered: