Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Error when running triton server with whisper model #522

Open
jackNhat opened this issue Dec 19, 2023 · 5 comments
Open

Error when running triton server with whisper model #522

jackNhat opened this issue Dec 19, 2023 · 5 comments

Comments

@jackNhat
Copy link

When i ran client.py, i got errror message:
tritonclient.utils.InferenceServerException: [StatusCode.INTERNAL] in ensemble 'whisper', Failed to process the request(s) for model instance 'scorer_0', message: AssertionError: <EMPTY MESSAGE>
How to fix?
I ran triton server with whisper model verson large-v2

@jwkyeongzz
Copy link

jwkyeongzz commented Jan 12, 2024

I got same issue. but it work properly.

  • Error env:
    windows (ubuntu 20.04)
    worksation ( intel xeon gold 6246 / rtx 3090 )

  • success pc ::
    centox 7.9
    server ( intel xeon gold 5218 / v100 )

  • Up to 7 channels can be operated simultaneously. ( v100 32G)

@csukuangfj
Copy link
Collaborator

@yuekaizhang

Could you have a look at this issue?

@yuekaizhang
Copy link
Collaborator

yuekaizhang commented Jan 12, 2024

I got same issue. but it work properly.

  • Error env:
    windows (ubuntu 20.04)
    worksation ( intel xeon gold 6246 / rtx 3090 )

  • success pc ::
    centox 7.9
    server ( intel xeon gold 5218 / v100 )

  • Up to 7 channels can be operated simultaneously. ( v100 32G)

@jwkyeongzz You mean using V100 is good. The issue only happened with RTX3090 GPU ?

@yuekaizhang
Copy link
Collaborator

yuekaizhang commented Jan 12, 2024

When i ran client.py, i got errror message: tritonclient.utils.InferenceServerException: [StatusCode.INTERNAL] in ensemble 'whisper', Failed to process the request(s) for model instance 'scorer_0', message: AssertionError: <EMPTY MESSAGE> How to fix? I ran triton server with whisper model verson large-v2

@jackNhat May I ask what's your GPU's name? Also, would you mind attaching more details? e.g. how to reproduce the error.

@jwkyeongzz
Copy link

I got same issue. but it work properly.

  • Error env:
    windows (ubuntu 20.04)
    worksation ( intel xeon gold 6246 / rtx 3090 )
  • success pc ::
    centox 7.9
    server ( intel xeon gold 5218 / v100 )
  • Up to 7 channels can be operated simultaneously. ( v100 32G)

@jwkyeongzz You mean using V100 is good. The issue only happened with RTX3090 GPU ?

I thought the test environment might be the problem. At first, since the error environment was in Windows' virtual Ubuntu 20.04, it was assumed that there was a problem with cuda memory allocation. In addition, it seems that it may have occurred due to insufficient memory of the RTX 3090. Therefore, it seems that the rtx3090 is not necessarily the problem.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants