diff --git a/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/derelu_fusion.cc b/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/derelu_fusion.cc index b8b475e981d..3130de9f2ea 100644 --- a/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/derelu_fusion.cc +++ b/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/derelu_fusion.cc @@ -107,10 +107,6 @@ const AnfNodePtr DereluFusion::Process(const FuncGraphPtr &graph, const AnfNodeP MS_EXCEPTION_IF_NULL(graph); MS_EXCEPTION_IF_NULL(node); - if (!LicManager::GetInstance().GetPassSwitch(OptPassEnum::DereluFusion)) { - return nullptr; - } - auto relu_grad = node->cast(); MS_EXCEPTION_IF_NULL(relu_grad); auto relu = GetRelu(relu_grad); diff --git a/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/fused_batch_norm_fusion.cc b/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/fused_batch_norm_fusion.cc index bed8fcaa806..9adaa2a1fc3 100644 --- a/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/fused_batch_norm_fusion.cc +++ b/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/fused_batch_norm_fusion.cc @@ -195,10 +195,6 @@ const AnfNodePtr FusedBatchNormFusion::Process(const FuncGraphPtr &func_graph, c MS_EXCEPTION_IF_NULL(equiv); MS_EXCEPTION_IF_NULL(node); - if (!LicManager::GetInstance().GetPassSwitch(OptPassEnum::FusedBatchNormFusion)) { - return nullptr; - } - AnfNodePtr bn_training_reduce = CreateBNTrainingReduce(func_graph, node, equiv); std::vector bn_training_reduce_outputs; CreateMultipleOutputsOfAnfNode(func_graph, bn_training_reduce, kBNTrainingReduceOutputNum, diff --git a/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/momentum_lossscale_fusion.cc b/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/momentum_lossscale_fusion.cc index d80922f497a..c7c8d4dc21d 100644 --- a/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/momentum_lossscale_fusion.cc +++ b/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/momentum_lossscale_fusion.cc @@ -54,10 +54,6 @@ const AnfNodePtr MomentumLossscaleFusion::Process(const FuncGraphPtr &func_graph MS_EXCEPTION_IF_NULL(func_graph); MS_EXCEPTION_IF_NULL(node); - if (!LicManager::GetInstance().GetPassSwitch(OptPassEnum::MomentumLossscaleFusion)) { - return nullptr; - } - auto cnode = node->cast(); MS_EXCEPTION_IF_NULL(cnode); CheckCNodeInputSize(cnode, kApplyMomentumInputTensorNum); diff --git a/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/softmax_grad_ext_fusion.cc b/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/softmax_grad_ext_fusion.cc index fe8534bb393..b59687d2036 100644 --- a/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/softmax_grad_ext_fusion.cc +++ b/mindspore/ccsrc/backend/optimizer/ascend/ir_fusion/softmax_grad_ext_fusion.cc @@ -56,10 +56,6 @@ const AnfNodePtr SoftmaxGradExtFusion::Process(const FuncGraphPtr &graph, const MS_EXCEPTION_IF_NULL(equiv); MS_EXCEPTION_IF_NULL(node); - if (!LicManager::GetInstance().GetPassSwitch(OptPassEnum::SoftmaxGradExtFusion)) { - return nullptr; - } - auto input0 = GetAnfNodeByVar(equiv, input0_); auto input1 = GetAnfNodeByVar(equiv, input1_); auto input2 = GetAnfNodeByVar(equiv, input2_); diff --git a/mindspore/ccsrc/runtime/device/ascend/lic_manager.cc b/mindspore/ccsrc/runtime/device/ascend/lic_manager.cc index 823f922c332..8a7c76da184 100644 --- a/mindspore/ccsrc/runtime/device/ascend/lic_manager.cc +++ b/mindspore/ccsrc/runtime/device/ascend/lic_manager.cc @@ -29,10 +29,7 @@ constexpr auto kAllOpen = "ALL"; static const std::map kPassCodeMap = { {std::to_string(3), OptPassEnum::MatmulBiasaddFusion}, - {std::to_string(8), OptPassEnum::DereluFusion}, {std::to_string(9), OptPassEnum::TransposeReshapeFusion}, - {std::to_string(10), OptPassEnum::MomentumLossscaleFusion}, - {std::to_string(12), OptPassEnum::FusedBatchNormFusion}, {std::to_string(15), OptPassEnum::BnupdateEltwiseEltwiseFusionPass}, {std::to_string(16), OptPassEnum::BnupdateEltwiseFusionPass}, {std::to_string(17), OptPassEnum::Conv2DBackpropEltwiseFusionPass}, @@ -44,7 +41,6 @@ static const std::map kPassCodeMap = { {std::to_string(34), OptPassEnum::EltwiseFusionPass}, {std::to_string(36), OptPassEnum::MultiOutputFusionPass}, {std::to_string(37), OptPassEnum::MulAddFusion}, - {std::to_string(38), OptPassEnum::SoftmaxGradExtFusion}, {std::to_string(39), OptPassEnum::ClipByNormNoDivSquareSumFusion}, {std::to_string(42), OptPassEnum::MulAddNPass}, {std::to_string(43), OptPassEnum::Resnet50DbnDwFusionPass},