-
Notifications
You must be signed in to change notification settings - Fork 0
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
INVALID_ARGUMENT: getPluginCreator could not find plugin CustomEmbLayerNormPluginDynamic version 1 #2
Comments
Hi Vilmara, how you pass LD_PRELOAD to tritonserver inside the container? |
Hi @TrojanXu, I have tried several methods without success: Method 1: -e LD_PRELOAD=/opt/tritonserver/libbert_plugins.so:/opt/tritontserver/libcommon.so Error:
Method 2: once inside the docker export LD_PRELOAD=/opt/tritonserver/libbert_plugins.so:/opt/tritontserver/libcommon.so method 2 does nothing just starts the docker image and exits immediately without errors:
Further questions: |
Yes, my question is how you put file libbert_plugins.so to /opt/tritonserver/ inside the docker container? Have you modified nvcr.io/nvidia/tensorrtserver image? From the command you provided, I only see you were using unmodified image nvcr.io/nvidia/tritonserver:20.03.1-py3 from ngc. But have you put libbert_plugins.so itself into the docker image? |
hi @TrojanXu, what is the best way to put the files libbert_plugins.so and libcommon.so to /opt/tritonserver/ inside the docker container?. I modified the nvcr.io/nvidia/tensorrtserver image editing the Dockerfile as shown below but I got errors rebuilding the image: Dockerfile addition:
Error when building the image:
|
One best way is to launch this docker container interactively, and pass lib directory into the container via '-v' option. Actually this the way described in the method 2 you mentioned. But you need to add some additional flags, the full cmd should be, |
Hi @TrojanXu, am trying to deploy the BERT TensorRT model with Triton following your steps but getting the below error:
Previously I built and copied the plugins to /opt/tritonserver
export LD_PRELOAD=/opt/tritonserver/libbert_plugins.so:/opt/tritontserver/libcommon.so
then I ran the triton server
sudo docker run --gpus all --rm --shm-size=1g --ulimit memlock=-1 --ulimit stack=67108864 -p443:443 -p8080:8080 -v /home/triton_server/docs/examples/model_repository/:/models nvcr.io/nvidia/tritonserver:20.03.1-py3 tritonserver --model-repository=/models
Error:
E1009 01:05:16.971516 1 logging.cc:43] INVALID_ARGUMENT: getPluginCreator could not find plugin CustomEmbLayerNormPluginDynamic version 1 E1009 01:05:16.971577 1 logging.cc:43] safeDeserializationUtils.cpp (293) - Serialization Error in load: 0 (Cannot deserialize plugin since corresponding IPluginCreator not found in Plugin Registry) E1009 01:05:16.971903 1 logging.cc:43] INVALID_STATE: std::exception E1009 01:05:16.971987 1 logging.cc:43] INVALID_CONFIG: Deserialize the cuda engine failed. I1009 01:05:16.972723 1 onnx_backend.cc:203] Creating instance densenet_onnx_0_gpu1 on GPU 1 (7.5) using model.onnx E1009 01:05:17.004133 1 model_repository_manager.cc:891] failed to load 'bert_trt' version 1: Internal: unable to create TensorRT engine
Some recommendations on how to fix this issue or if there is a recent version of the SW that simplified the deployment?
Versions used:
TensorRT: release 6.0
Triton docker image: tritonserver:20.03.1-py3
The text was updated successfully, but these errors were encountered: