From 216dff7549386f3f9ce0b79fd058343b6883ede9 Mon Sep 17 00:00:00 2001 From: fleance Date: Thu, 28 Sep 2023 11:27:43 +0200 Subject: [PATCH] Do not warn about unexpected decoder weights when loading T5EncoderModel and LongT5EncoderModel (#26211) Ignore decoder weights when using T5EncoderModel and LongT5EncoderModel Both T5EncoderModel and LongT5EncoderModel do not have any decoder layers, so loading a pretrained model checkpoint such as t5-small will give warnings about keys found in the model checkpoint that are not in the model itself. To prevent this log warning, r"decoder" has been added to _keys_to_ignore_on_load_unexpected for both T5EncoderModel and LongT5EncoderModel --- src/transformers/models/longt5/modeling_longt5.py | 1 + src/transformers/models/t5/modeling_t5.py | 1 + 2 files changed, 2 insertions(+) diff --git a/src/transformers/models/longt5/modeling_longt5.py b/src/transformers/models/longt5/modeling_longt5.py index 303755ae4338dc..d08ed83af07ea1 100644 --- a/src/transformers/models/longt5/modeling_longt5.py +++ b/src/transformers/models/longt5/modeling_longt5.py @@ -2152,6 +2152,7 @@ def _reorder_cache(self, past_key_values, beam_idx): ) class LongT5EncoderModel(LongT5PreTrainedModel): _tied_weights_keys = ["encoder.embed_tokens.weight"] + _keys_to_ignore_on_load_unexpected = [r"decoder"] def __init__(self, config: LongT5Config): super().__init__(config) diff --git a/src/transformers/models/t5/modeling_t5.py b/src/transformers/models/t5/modeling_t5.py index 33f29298dcca5c..e6d9deefa14639 100644 --- a/src/transformers/models/t5/modeling_t5.py +++ b/src/transformers/models/t5/modeling_t5.py @@ -1866,6 +1866,7 @@ def _reorder_cache(self, past_key_values, beam_idx): ) class T5EncoderModel(T5PreTrainedModel): _tied_weights_keys = ["encoder.embed_tokens.weight"] + _keys_to_ignore_on_load_unexpected = [r"decoder"] def __init__(self, config: T5Config): super().__init__(config)