|
|
|
@ -15,126 +15,80 @@ limitations under the License. */
|
|
|
|
|
#include <vector>
|
|
|
|
|
#include "paddle/fluid/framework/op_registry.h"
|
|
|
|
|
#include "paddle/fluid/operators/detail/safe_ref.h"
|
|
|
|
|
#include "paddle/fluid/platform/errors.h"
|
|
|
|
|
|
|
|
|
|
namespace paddle {
|
|
|
|
|
namespace operators {
|
|
|
|
|
|
|
|
|
|
class MultiHeadMatMulOp : public framework::OperatorWithKernel {
|
|
|
|
|
class MultiHeadMatMulV2Op : public framework::OperatorWithKernel {
|
|
|
|
|
public:
|
|
|
|
|
using framework::OperatorWithKernel::OperatorWithKernel;
|
|
|
|
|
|
|
|
|
|
protected:
|
|
|
|
|
void InferShape(framework::InferShapeContext *context) const override {
|
|
|
|
|
PADDLE_ENFORCE_EQ(context->HasInput("Q"), true,
|
|
|
|
|
"Input(Q) of MultiheadOp should not be null.");
|
|
|
|
|
PADDLE_ENFORCE_EQ(context->HasInput("K"), true,
|
|
|
|
|
"Input(K) of MultiheadOp should not be null.");
|
|
|
|
|
PADDLE_ENFORCE_EQ(context->HasInput("V"), true,
|
|
|
|
|
"Input(V) of MultiheadOp should not be null.");
|
|
|
|
|
PADDLE_ENFORCE_EQ(context->HasInput("BiasQ"), true,
|
|
|
|
|
"Input(BiasQ) of MultiheadOp should not be null.");
|
|
|
|
|
PADDLE_ENFORCE_EQ(context->HasInput("BiasK"), true,
|
|
|
|
|
"Input(BiasQ) of MultiheadOp should not be null.");
|
|
|
|
|
PADDLE_ENFORCE_EQ(context->HasInput("BiasV"), true,
|
|
|
|
|
"Input(BiasQ) of MultiheadOp should not be null.");
|
|
|
|
|
PADDLE_ENFORCE_EQ(context->HasInput("BiasQK"), true,
|
|
|
|
|
"Input(BiasQK) of MultiheadOp should not be null.");
|
|
|
|
|
PADDLE_ENFORCE_EQ(context->HasOutput("Out"), true,
|
|
|
|
|
"Output(Out) of MatMulOp should not be null.");
|
|
|
|
|
|
|
|
|
|
auto dim_q = context->GetInputDim("Q");
|
|
|
|
|
PADDLE_ENFORCE_GT(dim_q.size(), 2,
|
|
|
|
|
"Multihead input should be at least 3-D tensor.");
|
|
|
|
|
|
|
|
|
|
auto dim_k = context->GetInputDim("K");
|
|
|
|
|
PADDLE_ENFORCE_GT(dim_q.size(), 2,
|
|
|
|
|
"Multihead input should be at least 3-D tensor.");
|
|
|
|
|
|
|
|
|
|
auto dim_v = context->GetInputDim("V");
|
|
|
|
|
PADDLE_ENFORCE_GT(dim_q.size(), 2,
|
|
|
|
|
"Multihead input should be at least 3-D tensor.");
|
|
|
|
|
|
|
|
|
|
PADDLE_ENFORCE_EQ(dim_q[0], dim_k[0],
|
|
|
|
|
"Multihead input should have same batch size");
|
|
|
|
|
PADDLE_ENFORCE_EQ(dim_q[0], dim_v[0],
|
|
|
|
|
"Multihead input should have same batch size");
|
|
|
|
|
|
|
|
|
|
PADDLE_ENFORCE_EQ(dim_q[1], dim_k[1],
|
|
|
|
|
"Multihead input should have same size");
|
|
|
|
|
PADDLE_ENFORCE_EQ(dim_q[1], dim_v[1],
|
|
|
|
|
"Multihead input should have same size");
|
|
|
|
|
|
|
|
|
|
PADDLE_ENFORCE_EQ(dim_q[2], dim_k[2],
|
|
|
|
|
"Multihead input should have same size");
|
|
|
|
|
PADDLE_ENFORCE_EQ(dim_q[2], dim_v[2],
|
|
|
|
|
"Multihead input should have same size");
|
|
|
|
|
|
|
|
|
|
auto dim_bias_q = context->GetInputDim("BiasQ");
|
|
|
|
|
PADDLE_ENFORCE_GT(dim_bias_q.size(), 0,
|
|
|
|
|
"Multihead input should be at least 1-D tensor.");
|
|
|
|
|
auto dim_bias_k = context->GetInputDim("BiasK");
|
|
|
|
|
PADDLE_ENFORCE_GT(dim_bias_k.size(), 0,
|
|
|
|
|
"Multihead input should be at least 1-D tensor.");
|
|
|
|
|
auto dim_bias_v = context->GetInputDim("BiasV");
|
|
|
|
|
PADDLE_ENFORCE_GT(dim_bias_v.size(), 0,
|
|
|
|
|
"Multihead input should be at least 1-D tensor.");
|
|
|
|
|
|
|
|
|
|
PADDLE_ENFORCE_EQ(dim_bias_q[0], dim_bias_k[0],
|
|
|
|
|
"Multihead input bias should have same batch size");
|
|
|
|
|
PADDLE_ENFORCE_EQ(dim_bias_q[0], dim_bias_v[0],
|
|
|
|
|
"Multihead input bias should have same batch size");
|
|
|
|
|
|
|
|
|
|
auto dim_bias_qk = context->GetInputDim("BiasQK");
|
|
|
|
|
PADDLE_ENFORCE_GT(dim_bias_qk.size(), 3,
|
|
|
|
|
"Multihead input bias qk should be at least 4-D tensor.");
|
|
|
|
|
|
|
|
|
|
int b_indx = dim_bias_q.size() - 1;
|
|
|
|
|
int indx = dim_q.size() - 1;
|
|
|
|
|
|
|
|
|
|
PADDLE_ENFORCE_EQ(
|
|
|
|
|
dim_bias_q[b_indx], dim_q[indx],
|
|
|
|
|
context->HasInput("Input"), true,
|
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
|
"bias_q's last dim size should equal to"
|
|
|
|
|
" q last dim size, but received bias_q's size is:%d q is:%d",
|
|
|
|
|
dim_bias_q[b_indx], dim_q[indx]));
|
|
|
|
|
"Input(Input) of MultiHeadMatMul should not be null."));
|
|
|
|
|
PADDLE_ENFORCE_EQ(context->HasInput("W"), true,
|
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
|
"Input(W) of MultiHeadMatMul should not be null."));
|
|
|
|
|
PADDLE_ENFORCE_EQ(
|
|
|
|
|
dim_bias_k[b_indx], dim_k[indx],
|
|
|
|
|
context->HasInput("Bias"), true,
|
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
|
"bias_k's last dim size should equal to"
|
|
|
|
|
" k last dim size, but received bias_k's size is:%d k is:%d",
|
|
|
|
|
dim_bias_k[b_indx], dim_k[indx]));
|
|
|
|
|
"Input(Bias) of MultiHeadMatMul should not be null."));
|
|
|
|
|
PADDLE_ENFORCE_EQ(
|
|
|
|
|
dim_bias_v[b_indx], dim_v[indx],
|
|
|
|
|
context->HasInput("BiasQK"), true,
|
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
|
"bias_v's last dim size should equal to"
|
|
|
|
|
" v last dim size, but received bias_v's size is:%d v is:%d",
|
|
|
|
|
dim_bias_v[b_indx], dim_v[indx]));
|
|
|
|
|
"Input(BiasQK) of MultiHeadMatMul should not be null."));
|
|
|
|
|
PADDLE_ENFORCE_EQ(
|
|
|
|
|
context->HasOutput("Out"), true,
|
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
|
"Output(Out) of MultiHeadMatMul should not be null."));
|
|
|
|
|
|
|
|
|
|
PADDLE_ENFORCE_EQ(dim_q[0], dim_bias_qk[0],
|
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
|
"q should have same batch size"
|
|
|
|
|
"with bias_qk, but received q's batch size is:%d "
|
|
|
|
|
"bias_qk's batch size is:%d",
|
|
|
|
|
dim_q[0], dim_bias_qk[0]));
|
|
|
|
|
auto dim_w = context->GetInputDim("W");
|
|
|
|
|
PADDLE_ENFORCE_GT(
|
|
|
|
|
dim_w.size(), 2,
|
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
|
"Multihead input is expected at least a 3-D tensor, but "
|
|
|
|
|
"it's %d-D tensor now.",
|
|
|
|
|
dim_w.size()));
|
|
|
|
|
|
|
|
|
|
int head_number = context->Attrs().Get<int>("head_number");
|
|
|
|
|
PADDLE_ENFORCE_GT(head_number, 1,
|
|
|
|
|
"Multihead input head number should be at least 1.");
|
|
|
|
|
auto dim_bias_q = context->GetInputDim("Bias");
|
|
|
|
|
PADDLE_ENFORCE_GT(
|
|
|
|
|
dim_bias_q.size(), 1,
|
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
|
"Multihead input should be at least 2-D tensor, but it's "
|
|
|
|
|
"%d-D tensor now.",
|
|
|
|
|
dim_bias_q.size()));
|
|
|
|
|
|
|
|
|
|
auto dim_bias_qk = context->GetInputDim("BiasQK");
|
|
|
|
|
PADDLE_ENFORCE_GT(
|
|
|
|
|
dim_bias_qk.size(), 3,
|
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
|
"Multihead input bias qk should be at least 4-D tensor, "
|
|
|
|
|
"but it's %d-D tensor now.",
|
|
|
|
|
dim_bias_qk.size()));
|
|
|
|
|
|
|
|
|
|
context->SetOutputDim("Out", dim_q);
|
|
|
|
|
context->ShareLoD("Q", /*->*/ "Out");
|
|
|
|
|
int head_number = context->Attrs().Get<int>("head_number");
|
|
|
|
|
PADDLE_ENFORCE_GT(
|
|
|
|
|
head_number, 1,
|
|
|
|
|
platform::errors::InvalidArgument(
|
|
|
|
|
"Multihead input head number should be at least 1, but it %d now.",
|
|
|
|
|
head_number));
|
|
|
|
|
// modify this
|
|
|
|
|
auto dim_input = context->GetInputDim("Input");
|
|
|
|
|
context->SetOutputDim("Out", dim_input);
|
|
|
|
|
context->ShareLoD("Input", /*->*/ "Out");
|
|
|
|
|
}
|
|
|
|
|
};
|
|
|
|
|
|
|
|
|
|
class MultiHeadMatMulOpMaker : public framework::OpProtoAndCheckerMaker {
|
|
|
|
|
class MultiHeadMatMulV2OpMaker : public framework::OpProtoAndCheckerMaker {
|
|
|
|
|
public:
|
|
|
|
|
void Make() override {
|
|
|
|
|
AddInput("Q", "The first input of MultiHeadMatMul op");
|
|
|
|
|
AddInput("K", "The second input of MMultiHeadMatMul op");
|
|
|
|
|
AddInput("V", "The third input of MultiHeadMatMul op");
|
|
|
|
|
AddInput("BiasQ", "The first bias input of MultiHeadMatMul op");
|
|
|
|
|
AddInput("BiasK", "The second bias input of MultiHeadMatMul op");
|
|
|
|
|
AddInput("BiasV", "The third bias input of MultiHeadMatMul op");
|
|
|
|
|
AddInput("Input", "The input of MultiHeadMatMul op");
|
|
|
|
|
AddInput("W", "The weight input of MultiHeadMatMul op");
|
|
|
|
|
AddInput("Bias", "The bias input of MultiHeadMatMul op");
|
|
|
|
|
AddInput("BiasQK", "The QK bias input of MultiHeadMatMul op");
|
|
|
|
|
AddOutput("Out", "The output of MultiHeadMatMul op");
|
|
|
|
|
AddAttr<bool>("transpose_Q",
|
|
|
|
@ -161,10 +115,6 @@ Not suggest to use in other case except has same structure as ernie.
|
|
|
|
|
Example of matrix multiplication with head_number of B
|
|
|
|
|
- X: [B, M, K], Y: [B, K, N] => Out: [B, M, N]
|
|
|
|
|
|
|
|
|
|
Both the input `Q` and `K` can carry the LoD (Level of Details) information,
|
|
|
|
|
or not. But the output only shares the LoD information with input `Q`, because
|
|
|
|
|
they are the same.
|
|
|
|
|
|
|
|
|
|
)DOC");
|
|
|
|
|
}
|
|
|
|
|
};
|
|
|
|
@ -173,5 +123,5 @@ they are the same.
|
|
|
|
|
} // namespace paddle
|
|
|
|
|
|
|
|
|
|
namespace ops = paddle::operators;
|
|
|
|
|
REGISTER_OP_WITHOUT_GRADIENT(multihead_matmul, ops::MultiHeadMatMulOp,
|
|
|
|
|
ops::MultiHeadMatMulOpMaker);
|
|
|
|
|
REGISTER_OP_WITHOUT_GRADIENT(multihead_matmul, ops::MultiHeadMatMulV2Op,
|
|
|
|
|
ops::MultiHeadMatMulV2OpMaker);
|
|
|
|
|