-
Notifications
You must be signed in to change notification settings - Fork 218
Added support for TElinear ops #632
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
Codecov Report❌ Patch coverage is
Additional details and impacted files@@ Coverage Diff @@
## main #632 +/- ##
=======================================
Coverage 74.50% 74.50%
=======================================
Files 183 183
Lines 18400 18444 +44
=======================================
+ Hits 13709 13742 +33
- Misses 4691 4702 +11 ☔ View full report in Codecov by Sentry. 🚀 New features to boost your workflow:
|
mxinO
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks great to me! Do we still need a special modelopt layer spec after this MR?
iCo-authored-by: Asma Kuriparambil Thekkumpate <[email protected]> Co-authored-by: Kinjal Patel <[email protected]> Signed-off-by: Kinjal Patel <[email protected]>
Signed-off-by: Kinjal Patel <[email protected]>
Signed-off-by: Kinjal Patel <[email protected]>
Signed-off-by: Kinjal Patel <[email protected]>
Signed-off-by: Kinjal Patel <[email protected]>
Signed-off-by: Kinjal Patel <[email protected]>
d7c4802 to
021718d
Compare
meenchen
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do you know how much im provement we can get by enabling TE?
Signed-off-by: Kinjal Patel <[email protected]>
Signed-off-by: Kinjal Patel <[email protected]>
Signed-off-by: Kinjal Patel <[email protected]>
meenchen
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM. I am wondering if you have some data points of enabling TE in terms of perf and accuracy. Also, do we need to update the modelopt example in megatron to enable this by default?
|
Thanks @meenchen
@realAsma, do you have data points in terms of perf and accuracy for TE?
We will enable TE by default in megatron and remove local spec to keep things consistent between megatron-core and megatron + modelopt example |
Signed-off-by: Kinjal Patel <[email protected]>
|
Signed-off-by: Kinjal Patel <[email protected]>
Signed-off-by: Kinjal Patel <[email protected]>
iCo-authored-by: Asma Kuriparambil Thekkumpate <[email protected]> ## What does this PR do? **Type of change:** New Feature **Overview:** This MR adds support for quantizing TE Ops in megatron, specifically TERowParallelLinear, TEColumnParallelLinear and TELayerNormColumnParallelLinear. ## Usage It can be used by enabling TE spec in megatron ## Testing Added unit tests for testing functionality `test_homogeneous_sharded_state_dict_te_spec` `test_convert_mcore_te_gpt_model` `test_quantize_forward_backward` ## Before your PR is "*Ready for review*" <!-- If you haven't finished some of the above items you can still open `Draft` PR. --> - **Make sure you read and follow [Contributor guidelines](https://github.com/NVIDIA/TensorRT-Model-Optimizer/blob/main/CONTRIBUTING.md)** and your commits are signed. - **Is this change backward compatible?**: Yes - **Did you write any new necessary tests?**: Yes - **Did you add or update any necessary documentation?**: Yes - **Did you update [Changelog](https://github.com/NVIDIA/TensorRT-Model-Optimizer/blob/main/CHANGELOG.rst)?**: Yes ## Additional Information <!-- E.g. related issue. --> --------- Signed-off-by: Kinjal Patel <[email protected]>
iCo-authored-by: Asma Kuriparambil Thekkumpate [email protected]
What does this PR do?
Type of change: New Feature
Overview:
This MR adds support for quantizing TE Ops in megatron, specifically TERowParallelLinear, TEColumnParallelLinear and TELayerNormColumnParallelLinear.
Usage
It can be used by enabling TE spec in megatron
Testing
Added unit tests for testing functionality
test_homogeneous_sharded_state_dict_te_spectest_convert_mcore_te_gpt_modeltest_quantize_forward_backwardBefore your PR is "Ready for review"
Additional Information