Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

localcudacluster with onnxruntime inference #1399

Open
kanglcn opened this issue Oct 22, 2024 · 0 comments
Open

localcudacluster with onnxruntime inference #1399

kanglcn opened this issue Oct 22, 2024 · 0 comments

Comments

@kanglcn
Copy link

kanglcn commented Oct 22, 2024

Hi,

I am working on using a trained deep-learning model for image denoising.
The model is saved in onnx format and I successfully deployed this model with onnxruntime.
The workflow is:

  1. convert numpy array to cupy array
  2. do some preprocessing on cupy array
  3. create the onnxruntime session with gpu support
  4. run the model inference with input and out binding to cupy array
  5. do some afterprocessing on cupy array
  6. convert cupy array back to numpy array

Since I have many images to denoise and a signal-node-multi-gpu machine,
I wrap the above workflow to one function and I want to use dask-cuda to automatically
distribute these tasks.
However, the worker always died unreasonably.

I did one test on other cupy-only processing workflow and it works.
But with onnxruntime, it never works.
I would appreciate it if anybody can help!

Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant