Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bug fix #1249

Merged
merged 175 commits into from
Nov 6, 2024
Merged

Bug fix #1249

merged 175 commits into from
Nov 6, 2024

Conversation

danielhanchen
Copy link
Contributor

No description provided.

danielhanchen and others added 30 commits October 21, 2024 01:02
* Fix DPO, ORPO (#1177)

* Fix TRL

* Update mistral.py

* Patch processing_class

* Update tokenizer_utils.py

* Update tokenizer_utils.py

* Update tokenizer_utils.py

* Update tokenizer_utils.py

* Update tokenizer_utils.py

* Update tokenizer_utils.py

* Installation guide (#1165)

* chore: update chat_templates.py (#1166)

orginal -> original

* Disable Flex Attention

* Update tokenizer_utils.py

* Update _utils.py

* n_items

* Update cross_entropy_loss.py

* Fix DPO, ORPO

* Update _utils.py

---------

Co-authored-by: timothelaborie <[email protected]>
Co-authored-by: Ikko Eltociear Ashimine <[email protected]>

* Add warning for missing Unpack and KwargsForCausalLM in older Transformers versions

---------

Co-authored-by: Daniel Han <[email protected]>
Co-authored-by: timothelaborie <[email protected]>
Co-authored-by: Ikko Eltociear Ashimine <[email protected]>
* Enhance rotary embedding handling in LlamaAttention and LongRopeRotaryEmbedding

* Typo

* Improve rotary embedding handling in LlamaAttention to prevent errors with short KV cache

* Update llama.py

* Update llama.py

---------

Co-authored-by: Daniel Han <[email protected]>
@danielhanchen
Copy link
Contributor Author

danielhanchen commented Nov 6, 2024

Fixes #1248

@danielhanchen danielhanchen merged commit 3ea7044 into main Nov 6, 2024
1 check passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

5 participants