-
Notifications
You must be signed in to change notification settings - Fork 60
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
CI/CD - Upgrade dependency versions in pipeline #671
Conversation
Upgrade dependency versions in Azure pipeline: * Remove Python 3.6 and add Python 3.10 for cpu-unit-test * Upgrade CUDA from 11.1 to 12.4 for cuda-unit-test * Update labels accordingly
Codecov ReportAll modified and coverable lines are covered by tests ✅
Additional details and impacted files@@ Coverage Diff @@
## main #671 +/- ##
==========================================
- Coverage 85.75% 85.58% -0.18%
==========================================
Files 98 98
Lines 7043 7046 +3
==========================================
- Hits 6040 6030 -10
- Misses 1003 1016 +13
Flags with carried forward coverage won't be shown. Click here to find out more. ☔ View full report in Codecov by Sentry. |
Fix sudo issue inside container.
7689d22
to
1d2c652
Compare
442a8a6
to
2a240b6
Compare
Fix unit test.
455f99b
to
957983b
Compare
Update. Co-authored-by: Dilip Patlolla <[email protected]>
Fix unit test.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
Fix cache tests issues with cuda 12.4 and pytorch 2 https://pytorch.org/docs/stable/notes/cuda.html#cuda-memory-management: ``` PyTorch uses a caching memory allocator to speed up memory allocations. This allows fast memory deallocation without device synchronizations. However, the unused memory managed by the allocator will still show as if used in nvidia-smi. You can use memory_allocated() and max_memory_allocated() to monitor memory occupied by tensors, and use memory_reserved() and max_memory_reserved() to monitor the total amount of memory managed by the caching allocator. Calling empty_cache() releases all unused cached memory from PyTorch so that those can be used by other GPU applications. However, the occupied GPU memory by tensors will not be freed so it can not increase the amount of GPU memory available for PyTorch. ```
Revert changes for transformer.
Upgrade dependency versions in Azure pipeline: