Skip to content

Conversation

@limin2021
Copy link
Contributor

@limin2021 limin2021 commented Sep 28, 2021

PR types

Function optimization

PR changes

OPs

Describe

In fused_attention op and fused_ffn op, the fused bias_add+dropout+residual+layernorm kernel or bias_add+dropout+residual kernel is used. To ease the use of this kernel, we provide a wrapper in this PR.
1.To reuse the increment computing code, we exact the corresponding code to "GetSeedDataAndIncrement" routine in dropout_impl_util.h.
2.The fused_dropout_helper.h provides the fused dropout kernel wrapper.

Note: the test of this warper will be provided in the following fused_attention_op and fused_ffn PRs.

@paddle-bot-old
Copy link

Thanks for your contribution!
Please wait for the result of CI firstly. See Paddle CI Manual for details.

@CLAassistant
Copy link

CLAassistant commented Sep 28, 2021

CLA assistant check
All committers have signed the CLA.

@limin2021 limin2021 force-pushed the add-fused-dropout-helper branch from d8f0ac7 to fca8ae7 Compare September 28, 2021 12:31
@xingfeng01
Copy link
Contributor

LGTM

@AnnaTrainingG AnnaTrainingG merged commit 092d45c into PaddlePaddle:develop Sep 29, 2021
AnnaTrainingG pushed a commit to AnnaTrainingG/Paddle that referenced this pull request Sep 29, 2021
limin2021 added a commit to limin2021/Paddle that referenced this pull request Oct 22, 2021
lanxianghit pushed a commit that referenced this pull request Oct 25, 2021
In fused_attention op and fused_ffn op, the fused bias_add+dropout+residual+layernorm kernel or bias_add+dropout+residual kernel is used. To ease the use of this kernel, we provide a wrapper in this PR. 1.To reuse the increment computing code, we exact the corresponding code to "GetSeedDataAndIncrement" routine in dropout_impl_util.h. 2.The fused_dropout_helper.h provides the fused dropout kernel wrapper. Note: the test of this warper will be provided in the following fused_attention_op and fused_ffn PRs.
wangxicoding added a commit to wangxicoding/Paddle that referenced this pull request Oct 25, 2021
wangxicoding pushed a commit to wangxicoding/Paddle that referenced this pull request Oct 25, 2021
fuyinno4 pushed a commit that referenced this pull request Oct 25, 2021
… RandomSeedGenerator (#36682) * Revert "Add fused_dropout wrapper to ease use. (#36185) (#36640)" This reverts commit 05d7e2f. * [hybrid] seed and dropout op support force-cpu (#35820) * [HIP] fix op not support AMD GPU bug, the flag PADDLE_WITH_ROCM is invalid * [HIP] fix op not support AMD GPU bug, the flag PADDLE_WITH_ROCM is invalid * [HIP] fix op not support AMD GPU bug * [hybrid] seed and dropout op support force-cpu * [hybrid] seed and dropout op support force-cpu * [hybrid] seed and dropout op support force-cpu * [hybrid] seed and dropout op support force-cpu * [hybrid] seed and dropout op support force-cpu * [hybrid] fix seed ci failed issue * add AsExtra for force_cpu of seed op * Add fused_dropout wrapper to ease use. (#36185) * [hybrid] static model parallel dropout support deterministic RandomSeedGenerator (#36228) Co-authored-by: xiayanming <41795079@qq.com> Co-authored-by: Li Min <11663212+limin2021@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

6 participants