CARVIEW |
Navigation Menu
-
Notifications
You must be signed in to change notification settings - Fork 24.7k
[pt2e][quant] Make move_exported_model_to_train/eval idempotent #142239
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. Weβll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Summary: Before we would recompile the model unnecessarily even if the model is already in the desired mode. For training frameworks that assume `model.train()` is idempotent and calls this before every single training step, this led to a bunch of tiny graphs and poor performance. This commit makes these calls no-ops if we're already in the target train/eval mode. Test Plan: python test/test_quantization -k TestQuantizePT2E.test_allow_exported_model_train_eval_idempotent [ghstack-poisoned]
π Helpful Linksπ§ͺ See artifacts and rendered test results at hud.pytorch.org/pr/142239
Note: Links to docs will display an error until the docs builds have been completed. β No FailuresAs of commit b0d946e with merge base 34033cc ( This comment was automatically generated by Dr. CI and updates every 15 minutes. |
β¦otent" Summary: Before we would recompile the model unnecessarily even if the model is already in the desired mode. For training frameworks that assume `model.train()` is idempotent and calls this before every single training step, this led to a bunch of tiny graphs and poor performance. This commit makes these calls no-ops if we're already in the target train/eval mode. Test Plan: python test/test_quantization -k TestQuantizePT2E.test_allow_exported_model_train_eval_idempotent [ghstack-poisoned]
β¦otent" Summary: Before we would recompile the model unnecessarily even if the model is already in the desired mode. For training frameworks that assume `model.train()` is idempotent and calls this before every single training step, this led to a bunch of tiny graphs and poor performance. This commit makes these calls no-ops if we're already in the target train/eval mode. Test Plan: python test/test_quantization -k TestQuantizePT2E.test_allow_exported_model_train_eval_idempotent [ghstack-poisoned]
Summary: Before we would recompile the model unnecessarily even if the model is already in the desired mode. For training frameworks that assume `model.train()` is idempotent and calls this before every single training step, this led to a bunch of tiny graphs and poor performance. This commit makes these calls no-ops if we're already in the target train/eval mode. Test Plan: python test/test_quantization -k TestQuantizePT2E.test_allow_exported_model_train_eval_idempotent ghstack-source-id: f0e6c78 Pull Request resolved: #142239
β¦otent" Summary: Before we would recompile the model unnecessarily even if the model is already in the desired mode. For training frameworks that assume `model.train()` is idempotent and calls this before every single training step, this led to a bunch of tiny graphs and poor performance. This commit makes these calls no-ops if we're already in the target train/eval mode. Test Plan: python test/test_quantization -k TestQuantizePT2E.test_allow_exported_model_train_eval_idempotent [ghstack-poisoned]
Summary: Before we would recompile the model unnecessarily even if the model is already in the desired mode. For training frameworks that assume `model.train()` is idempotent and calls this before every single training step, this led to a bunch of tiny graphs and poor performance. This commit makes these calls no-ops if we're already in the target train/eval mode. Test Plan: python test/test_quantization -k TestQuantizePT2E.test_allow_exported_model_train_eval_idempotent ghstack-source-id: 6848f32 Pull Request resolved: #142239
@pytorchbot merge |
Merge startedYour change will be merged once all checks pass (ETA 0-4 Hours). Learn more about merging in the wiki. Questions? Feedback? Please reach out to the PyTorch DevX Team |
Stack from ghstack (oldest at bottom):
Summary: Before we would recompile the model unnecessarily even
if the model is already in the desired mode. For training
frameworks that assume
model.train()
is idempotent and callsthis before every single training step, this led to a bunch of
tiny graphs and poor performance. This commit makes these calls
no-ops if we're already in the target train/eval mode.
Test Plan:
python test/test_quantization -k TestQuantizePT2E.test_allow_exported_model_train_eval_idempotent