[Documentation] how to modularize ONNXRT on CPU first , then on CPU with OpenVino EP then on Nvidia GPU with TRT EP simply by adding new provider libraries and all their dependencies #23104
Labels
documentation
improvements or additions to documentation; typically submitted using template
ep:OpenVINO
issues related to OpenVINO execution provider
ep:TensorRT
issues related to TensorRT execution provider
Describe the documentation issue
The goal is to create modular deployment profiles (such as docker image layers) that you pile up when you add the provider capability . See the diagram below:
This is supposed to work already like said in Build with different EPs - onnxruntime
Quoting :
"Execution Provider Shared Libraries
Loading the shared providers
However , it does not seems to work because when building ONNXRT with TRT EP , I am getting both CUDA EP and TRT EP, but , if I remove the libonnxruntime_providers_cuda.so, then my Client code does got runtime link error looking for this library despite my client code does not add (depend) on OrtSessionOptionsAppendExecutionProvider_CUDA. See [Build] Cuda Execution Provider library is needed despite we only use TensoRT Execution provider · Issue #22960 · microsoft/onnxruntime
Here is how I built ONNRT :
CC=gcc-11 CXX=g++-11 ./build.sh
--skip_submodule_sync --nvcc_threads 2
--config $ORT_BUILD_MODE --use_cuda
--cudnn_home /usr/local/cuda/lib64
--cuda_home /usr/local/cuda/
--use_tensorrt --use_tensorrt_oss_parser --tensorrt_home /usr/local/TensorRT
--build_shared_lib --parallel --skip_tests
--allow_running_as_root --cmake_extra_defines "CMAKE_CUDA_ARCHITECTURES=75"
--cmake_extra_defines "CMAKE_CUDA_HOST_COMPILER=/usr/bin/gcc-11"
So can you tell how to build all provider libs and deploy them separately based on my CPU / GPU host configuration ?
Onnxruntime shall run even if a provider library is not there as the client code does not use it.
Shall I use all provider in the build such as ?
CC=gcc-11 CXX=g++-11 ./build.sh
--skip_submodule_sync --nvcc_threads 2
--config $ORT_BUILD_MODE --use_cuda
--cudnn_home /usr/local/cuda/lib64
--cuda_home /usr/local/cuda/
--use_tensorrt --use_tensorrt_oss_parser --tensorrt_home /usr/local/TensorRT
--use_openvino --openvino_home /usr/local/OpenVino
--build_shared_lib --parallel --skip_tests
--allow_running_as_root --cmake_extra_defines "CMAKE_CUDA_ARCHITECTURES=75"
--cmake_extra_defines "CMAKE_CUDA_HOST_COMPILER=/usr/bin/gcc-11"
Please advise
Page / URL
No response
The text was updated successfully, but these errors were encountered: