-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Deepseek v2 16b enablement #34
base: rocm_dev
Are you sure you want to change the base?
Conversation
ramp_func = torch.clamp(linear_func, 0, 1) | ||
return ramp_func | ||
|
||
class DeepseekV2YarnRotaryEmbedding(DeepseekV2RotaryEmbedding): |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can we use YarnRotaryEmbedding
from megatron.core.models.common.embeddings.yarn_rotary_pos_embedding
instead?
@hakankiymaz-amd Could you also make sure the copyright statements are all applied properly? If files are copied over from PAI-megatron-patch without any change, you should keep the original copyright statement there. If it is incorporated with your change, You need to add additional copyright statement for AMD and specify what are the changes. |
Copyright statements are added proper to changes from PAI and changes from AMD. |
Deepseek v2 16b enablement
test_report.csv
MI300X-DeepSeek-V2-Lite-bf16-seq2048-tp1pp1ep8-mbsgbs-ac_sel-do_true-fa_true-sp_true-20241210_173933.log
throughput per GPU: 611.822
All tests passed