Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Memory tool chain improvement #18890

Closed
wants to merge 9 commits into from
Closed

Memory tool chain improvement #18890

wants to merge 9 commits into from

Conversation

pengwa
Copy link
Contributor

@pengwa pengwa commented Dec 20, 2023

Display recomputable graph reusing buffers - For some subgraph whose ending node is reusing other buffers, currently it is not displayed in the recomputation plans, this PR fix this bug.

Extract memory logging into torch_profile_utils.py, which can be used in any other places we want to inspect memory during training.

Add MemoryStatisticsSubscriber to print memory stats, used to compare PyTorch and ORT memory peak.

@pengwa pengwa added the training issues related to ONNX Runtime training; typically submitted using template label Dec 20, 2023
@pengwa pengwa changed the title An improvement on recompute graph detection Memory tool chain improvement Dec 21, 2023
@pengwa pengwa marked this pull request as draft February 23, 2024 01:57
@pengwa
Copy link
Contributor Author

pengwa commented Feb 23, 2024

Covered by other PR and #19620

@pengwa pengwa closed this Feb 23, 2024
@pengwa pengwa deleted the pengwa/update_recompute branch April 22, 2024 09:57
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
training issues related to ONNX Runtime training; typically submitted using template
Projects
None yet
Development

Successfully merging this pull request may close these issues.

1 participant