-
Notifications
You must be signed in to change notification settings - Fork 680
Move memory from utils to training #1456
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
🔗 Helpful Links🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/torchtune/1456
Note: Links to docs will display an error until the docs builds have been completed. ✅ No FailuresAs of commit 3cdfe03 with merge base e959321 ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
torchtune/utils/__init__.py
Outdated
"DEFAULT_TRACE_OPTS", | ||
"DummyProfiler", | ||
"PROFILER_KEY", | ||
"setup_torch_profiler", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
do you mind double checking those? i dont understand what is being added here or why
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
ah good catch, these should be removed
there are conflicts. I can approve after that. |
docs/source/api_ref_training.rst
Outdated
|
||
.. _ac_label: | ||
|
||
Memory Management |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
This is a duplicate of the section immediately following
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
One comment, other than that looks good
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
comment below
log_memory_stats, | ||
OptimizerInBackwardWrapper, | ||
register_optim_in_bwd_hooks, | ||
set_activation_checkpointing, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
some of the deleted imports are still in __ all __
Context
What is the purpose of this PR? Is it to
Addresses #1437.
Changelog
What are the changes made in this PR?
Test plan
Please make sure to do each of the following if applicable to your PR. (If you're not sure about any one of these just ask and we will happily help. We also have a contributing page for some guidance on contributing.)
pre-commit install
)pytest tests
pytest tests -m integration_test
UX
If your function changed a public API, please add a dummy example of what the user experience will look like when calling it.
Example of docstring: https://github.com/pytorch/torchtune/blob/6a7951f1cdd0b56a9746ef5935106989415f50e3/torchtune/modules/vision_transformer.py#L285
Example in our docs: https://pytorch.org/torchtune/main/tutorials/qat_finetune.html#applying-qat-to-llama3-models