Skip to content

Fix softmax dim of Residual MoE implementation in moe/layer.py#2110

Merged
yaozhewei merged 2 commits intodeepspeedai:masterfrom
hero007feng:master
Jul 20, 2022
Merged

Fix softmax dim of Residual MoE implementation in moe/layer.py#2110
yaozhewei merged 2 commits intodeepspeedai:masterfrom
hero007feng:master

Conversation

@hero007feng
Copy link

We found that the shape of hidden_states and coef is [batch_size, seq_len, 2], and the softmax function should be applied to the last dimensions, which can be used to normalize the weights of expert and mlp.

@yaozhewei yaozhewei enabled auto-merge (squash) July 20, 2022 01:12
@yaozhewei yaozhewei merged commit b4513f6 into deepspeedai:master Jul 20, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants