You can not select more than 25 topics
Topics must start with a letter or number, can include dashes ('-') and can be up to 35 characters long.
211 lines
6.7 KiB
211 lines
6.7 KiB
/**
|
|
* Copyright 2020 Huawei Technologies Co., Ltd
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
#ifndef MINDSPORE_CCSRC_OPTIMIZER_IRPASS_SPECIAL_OP_ELIMINATE_H_
|
|
#define MINDSPORE_CCSRC_OPTIMIZER_IRPASS_SPECIAL_OP_ELIMINATE_H_
|
|
|
|
#include <securec.h>
|
|
#include <vector>
|
|
#include <memory>
|
|
#include <algorithm>
|
|
|
|
#include "optimizer/optimizer.h"
|
|
#include "optimizer/irpass.h"
|
|
#include "ir/optimizer_caller.h"
|
|
#include "optimizer/irpass/prim_eliminate.h"
|
|
#include "ir/visitor.h"
|
|
#include "operator/ops.h"
|
|
#include "ir/pattern_matcher.h"
|
|
|
|
namespace mindspore {
|
|
namespace opt {
|
|
namespace irpass {
|
|
class SpecialOpEliminater {
|
|
public:
|
|
SpecialOpEliminater()
|
|
: insert_gradient_of_(prim::kPrimInsertGradientOf),
|
|
stop_gradient_(prim::kPrimStopGradient),
|
|
hook_backward_(prim::kPrimHookBackward),
|
|
print_shape_type_(prim::kPrimPrintShapeType),
|
|
get_ref_value_(prim::kPrimGetRefValue),
|
|
mirror_(prim::kPrimMirror),
|
|
virtual_div_(prim::kPrimVirtualDiv) {
|
|
eliminaters_.emplace_back(insert_gradient_of_);
|
|
eliminaters_.emplace_back(stop_gradient_);
|
|
eliminaters_.emplace_back(hook_backward_);
|
|
eliminaters_.emplace_back(print_shape_type_);
|
|
eliminaters_.emplace_back(get_ref_value_);
|
|
eliminaters_.emplace_back(mirror_);
|
|
eliminaters_.emplace_back(virtual_div_);
|
|
}
|
|
~SpecialOpEliminater() = default;
|
|
|
|
AnfNodePtr operator()(const OptimizerPtr &optimizer, const AnfNodePtr &node) {
|
|
AnfNodePtr new_node;
|
|
for (auto &eliminater : eliminaters_) {
|
|
new_node = eliminater(optimizer, node);
|
|
if (new_node != nullptr) {
|
|
return new_node;
|
|
}
|
|
}
|
|
return nullptr;
|
|
}
|
|
|
|
private:
|
|
PrimEliminater insert_gradient_of_, stop_gradient_, hook_backward_, print_shape_type_, get_ref_value_, mirror_,
|
|
virtual_div_;
|
|
std::vector<TransformFuncType> eliminaters_{};
|
|
};
|
|
|
|
// {PrimVirtualDataset, X} -> X
|
|
// {PrimVirtualDataset, Xs} -> {prim::kPrimMakeTuple, Xs}
|
|
class VirtualDatasetEliminater : public AnfVisitor {
|
|
public:
|
|
AnfNodePtr operator()(const OptimizerPtr &, const AnfNodePtr &node) override {
|
|
if (!IsPrimitiveCNode(node, prim::kPrimVirtualDataset) || node->func_graph() == nullptr) {
|
|
return nullptr;
|
|
}
|
|
|
|
auto &inputs = node->cast<CNodePtr>()->inputs();
|
|
if (inputs.size() < 1) {
|
|
return nullptr;
|
|
}
|
|
|
|
std::vector<AnfNodePtr> args;
|
|
(void)std::copy(inputs.begin() + 1, inputs.end(), std::back_inserter(args));
|
|
if (args.size() == 1) {
|
|
return args.front();
|
|
}
|
|
|
|
(void)args.insert(args.begin(), NewValueNode(prim::kPrimMakeTuple));
|
|
|
|
return node->func_graph()->NewCNode(args);
|
|
}
|
|
|
|
void Visit(const AnfNodePtr &) override {}
|
|
};
|
|
|
|
// {prim::kPrimSameTypeShape, X, Y} -> X
|
|
class SameEliminater : public AnfVisitor {
|
|
public:
|
|
AnfNodePtr operator()(const OptimizerPtr &, const AnfNodePtr &node) override {
|
|
x_ = nullptr;
|
|
AnfVisitor::Match(prim::kPrimSameTypeShape, {IsNode, IsNode})(node);
|
|
return x_;
|
|
}
|
|
|
|
void Visit(const AnfNodePtr &node) override {
|
|
if (x_ == nullptr) {
|
|
x_ = node;
|
|
}
|
|
}
|
|
|
|
private:
|
|
AnfNodePtr x_{nullptr};
|
|
};
|
|
|
|
// {prim::kPrimCheckBprop, X, Y} -> X
|
|
class CheckBpropEliminater : public AnfVisitor {
|
|
public:
|
|
AnfNodePtr operator()(const OptimizerPtr &, const AnfNodePtr &node) override {
|
|
x_ = nullptr;
|
|
AnfVisitor::Match(prim::kPrimCheckBprop, {IsNode, IsNode})(node);
|
|
return x_;
|
|
}
|
|
|
|
void Visit(const AnfNodePtr &node) override {
|
|
if (x_ == nullptr) {
|
|
x_ = node;
|
|
}
|
|
}
|
|
|
|
private:
|
|
AnfNodePtr x_{nullptr};
|
|
};
|
|
|
|
// Reset defer_inline flag
|
|
class ResetDeferInline : public AnfVisitor {
|
|
public:
|
|
AnfNodePtr operator()(const OptimizerPtr &, const AnfNodePtr &node) override {
|
|
if (IsValueNode<FuncGraph>(node)) {
|
|
auto fg = GetValueNode<FuncGraphPtr>(node);
|
|
fg->set_flag(FUNC_GRAPH_FLAG_DEFER_INLINE, false);
|
|
}
|
|
return nullptr;
|
|
}
|
|
};
|
|
|
|
// {PrimZerosLike, Y} ->
|
|
// {PrimFill, {PrimDType, Y}, {PrimShape, Y}, 0}
|
|
class ZeroLikeFillZero : public AnfVisitor {
|
|
public:
|
|
ZeroLikeFillZero()
|
|
: PrimFill_(prim::GetPythonOps("fill", "mindspore.ops.functional")->cast<PrimitivePtr>()),
|
|
PrimShape_(prim::GetPythonOps("shape", "mindspore.ops.functional")->cast<PrimitivePtr>()),
|
|
PrimDType_(prim::GetPythonOps("dtype", "mindspore.ops.functional")->cast<PrimitivePtr>()) {}
|
|
~ZeroLikeFillZero() override = default;
|
|
|
|
AnfNodePtr operator()(const OptimizerPtr &, const AnfNodePtr &node) override {
|
|
y_ = nullptr;
|
|
AnfVisitor::Match(prim::kPrimZerosLike, {IsNode})(node);
|
|
if (y_ == nullptr || node->func_graph() == nullptr) {
|
|
return nullptr;
|
|
}
|
|
if ((y_->abstract() == nullptr) || !y_->abstract()->isa<abstract::AbstractTensor>()) {
|
|
auto fg = node->func_graph();
|
|
auto dtype = fg->NewCNode({NewValueNode(PrimDType_), y_});
|
|
auto shape = fg->NewCNode({NewValueNode(PrimShape_), y_});
|
|
return fg->NewCNode({NewValueNode(PrimFill_), dtype, shape, NewValueNode(MakeValue(0))});
|
|
}
|
|
|
|
abstract::AbstractTensorPtr tensor_abstract = y_->abstract()->cast<abstract::AbstractTensorPtr>();
|
|
|
|
TypePtr tensor_type_ptr = tensor_abstract->element()->BuildType();
|
|
std::vector<int> tensor_shape = tensor_abstract->shape()->shape();
|
|
|
|
tensor::TensorPtr new_tensor_ptr = std::make_shared<tensor::Tensor>(tensor_type_ptr->type_id(), tensor_shape);
|
|
size_t mem_size = GetTypeByte(tensor_type_ptr) * IntToSize(new_tensor_ptr->ElementsNum());
|
|
char *data = reinterpret_cast<char *>(new_tensor_ptr->data_c());
|
|
(void)memset_s(data, mem_size, 0, mem_size);
|
|
|
|
auto new_cnode = NewValueNode(new_tensor_ptr);
|
|
new_cnode->set_abstract(new_tensor_ptr->ToAbstract());
|
|
|
|
return new_cnode;
|
|
}
|
|
|
|
void Visit(const AnfNodePtr &node) override { y_ = node; }
|
|
|
|
private:
|
|
AnfNodePtr y_{nullptr};
|
|
PrimitivePtr PrimFill_, PrimShape_, PrimDType_;
|
|
};
|
|
|
|
// {prim::kPrimDepend, X, ValueCond}->X
|
|
class DependValueElim : public OptimizerCaller {
|
|
public:
|
|
AnfNodePtr operator()(const OptimizerPtr &, const AnfNodePtr &node) override {
|
|
PatternNode<AnfNodePtr> x, cond;
|
|
MATCH_REPLACE_IF(node, PPrimitive(prim::kPrimDepend, x, cond), x, IsVNode(cond.GetNode(node)));
|
|
return nullptr;
|
|
}
|
|
};
|
|
|
|
} // namespace irpass
|
|
} // namespace opt
|
|
} // namespace mindspore
|
|
#endif // MINDSPORE_CCSRC_OPTIMIZER_IRPASS_SPECIAL_OP_ELIMINATE_H_
|