-
Notifications
You must be signed in to change notification settings - Fork 3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
T5 BeamSearch model loading failing if initializers are moved on the outer graph #23043
Labels
model:transformer
issues related to a transformer model: BERT, GPT2, Hugging Face, Longformer, T5, etc.
Comments
github-actions
bot
added
the
model:transformer
issues related to a transformer model: BERT, GPT2, Hugging Face, Longformer, T5, etc.
label
Dec 6, 2024
tianleiwu
pushed a commit
that referenced
this issue
Dec 9, 2024
### Description This PR adds the logic needed to consider only the needed implicit inputs on BeamSearch op in case of T5 model (encoder/decoder, 2 graphs). The logic added is similar to what happens in the _If_ kernel setup. ### Motivation and Context Fixes #23043
guschmue
pushed a commit
that referenced
this issue
Dec 9, 2024
### Description This PR adds the logic needed to consider only the needed implicit inputs on BeamSearch op in case of T5 model (encoder/decoder, 2 graphs). The logic added is similar to what happens in the _If_ kernel setup. ### Motivation and Context Fixes #23043
ankitm3k
pushed a commit
to intel/onnxruntime
that referenced
this issue
Dec 11, 2024
### Description This PR adds the logic needed to consider only the needed implicit inputs on BeamSearch op in case of T5 model (encoder/decoder, 2 graphs). The logic added is similar to what happens in the _If_ kernel setup. ### Motivation and Context Fixes microsoft#23043
ankitm3k
pushed a commit
to intel/onnxruntime
that referenced
this issue
Dec 11, 2024
### Description This PR adds the logic needed to consider only the needed implicit inputs on BeamSearch op in case of T5 model (encoder/decoder, 2 graphs). The logic added is similar to what happens in the _If_ kernel setup. ### Motivation and Context Fixes microsoft#23043
ankitm3k
pushed a commit
to intel/onnxruntime
that referenced
this issue
Dec 11, 2024
### Description This PR adds the logic needed to consider only the needed implicit inputs on BeamSearch op in case of T5 model (encoder/decoder, 2 graphs). The logic added is similar to what happens in the _If_ kernel setup. ### Motivation and Context Fixes microsoft#23043
tarekziade
pushed a commit
to tarekziade/onnxruntime
that referenced
this issue
Jan 10, 2025
### Description This PR adds the logic needed to consider only the needed implicit inputs on BeamSearch op in case of T5 model (encoder/decoder, 2 graphs). The logic added is similar to what happens in the _If_ kernel setup. ### Motivation and Context Fixes microsoft#23043
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Labels
model:transformer
issues related to a transformer model: BERT, GPT2, Hugging Face, Longformer, T5, etc.
Describe the issue
Seems there's a bug in subgraph setup for BeamSearch op in case of T5 models in which the initializers are moved into outer graph. Essentially, the setup of the encoder/decoder subgraphs is not taking into consideration which outer scope initializer is used inside the graph or not (similarly to what happens in the If kernel)
To reproduce
Following python script creates a dummy T5 like model, runs it, performs another pass to move the initializers in the outer graphs, then runs the new model.
Urgency
No response
Platform
Linux
OS Version
Ubuntu 20.04
ONNX Runtime Installation
Built from Source
ONNX Runtime Version or Commit ID
09c9843
ONNX Runtime API
Python
Architecture
X64
Execution Provider
Other / Unknown
Execution Provider Library Version
No response
The text was updated successfully, but these errors were encountered: