-
Notifications
You must be signed in to change notification settings - Fork 515
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Missing Keys in state_dict #172
Comments
Actually Lines 60 to 81 in c9f683e
But I think you can replace |
Thank you for this comment.
Get Outlook for Android<https://aka.ms/AAb9ysg>
…________________________________
From: Yanbo Liang ***@***.***>
Sent: Sunday, September 15, 2024 11:31:41 PM
To: pytorch-labs/gpt-fast ***@***.***>
Cc: John B Olan ***@***.***>; Author ***@***.***>
Subject: Re: [pytorch-labs/gpt-fast] Missing Keys in state_dict (Issue #172)
Actually Llama3-ChatQA-1.5-8B is not supported, please check all supported models at:
https://github.com/pytorch-labs/gpt-fast/blob/c9f683edd4f89d3e81ed8f52387e866a245e3226/model.py#L60-L81
But I think you can replace llama-3-8b in the list with Llama3-ChatQA-1.5-8B and play around it. They should have the same architecture.
—
Reply to this email directly, view it on GitHub<#172 (comment)>, or unsubscribe<https://github.com/notifications/unsubscribe-auth/APJRHRK4Q2GP3ON2G2WQZTLZWZNK3AVCNFSM6AAAAABHIFYWXCVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDGNJSGAYDCOJVGY>.
You are receiving this because you authored the thread.Message ID: ***@***.***>
|
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I downloaded
nvidia/Llama3-ChatQA-1.5-8B
manually from HF into local. I ranscripts/convert_hf_checkpoint.py
Then I wanted to run generate.py using the local checkpoint dir:raise RuntimeError('Error(s) in loading state_dict for {}:\n\t{}'.format( RuntimeError: Error(s) in loading state_dict for Transformer: Missing key(s) in state_dict: "tok_embeddings.weight", "layers.0.attention.wqkv.weight", "layers.0.attention.wo.weight", "layers.0.feed_forward.w1.weight", "layers.0.feed_forward.w3.weight", "layers.0.feed_forward.w2.weight", "layers.0.ffn_norm.weight", "layers.0.attention_norm.weight",
Here is my weight directory:
The text was updated successfully, but these errors were encountered: