Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

gte-Qwen2-7b-instruct and related Qwen2 models have incorrect max_token size #6

Open
tosaddler opened this issue Jul 10, 2024 · 4 comments · Fixed by #7 or #13
Open

gte-Qwen2-7b-instruct and related Qwen2 models have incorrect max_token size #6

tosaddler opened this issue Jul 10, 2024 · 4 comments · Fixed by #7 or #13

Comments

@tosaddler
Copy link

Not sure where the info is pulled from, but it is showing 131072 rather than 32768.

image image
@orionw orionw mentioned this issue Jul 11, 2024
@orionw
Copy link
Collaborator

orionw commented Jul 11, 2024

Thanks @tosaddler for noticing, will fix!

@orionw orionw closed this as completed in #7 Jul 11, 2024
@orionw
Copy link
Collaborator

orionw commented Jul 11, 2024

Seems like it may be re-appearing. Will take a deeper look later.

@orionw orionw reopened this Jul 11, 2024
@orionw
Copy link
Collaborator

orionw commented Jul 11, 2024

I can make a manual fix for this but would have to write it in.

FWIW, this is caused by extracting it out their their config.json file. I think it's a bug on their part. We can either provide some code to override the extraction or we can leave it.

Thoughts @KennethEnevoldsen?

@orionw orionw reopened this Jul 11, 2024
@KennethEnevoldsen
Copy link
Contributor

Ahh that is frustrating - I would probably create an overwrite on our end (create some sort of file which takes priority).

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
3 participants