diff --git a/mindspore/ccsrc/backend/graph_compiler/backend.cc b/mindspore/ccsrc/backend/graph_compiler/backend.cc index 268cde713d6..404409c5ad5 100644 --- a/mindspore/ccsrc/backend/graph_compiler/backend.cc +++ b/mindspore/ccsrc/backend/graph_compiler/backend.cc @@ -784,8 +784,8 @@ void GetControlOpInput(const std::shared_ptr &graph_compiler, con ++back_index; } } - if (args_tuple_num) { - args_tuple.emplace_back(value); + if (args_tuple_num != 0) { + (void)args_tuple.emplace_back(value); if (args_tuple.size() == args_tuple_num) { value = std::make_shared(args_tuple); args_tuple_num = 0; diff --git a/mindspore/ccsrc/frontend/optimizer/ad/dfunctor.cc b/mindspore/ccsrc/frontend/optimizer/ad/dfunctor.cc index 369b60dde1b..815c3b82097 100644 --- a/mindspore/ccsrc/frontend/optimizer/ad/dfunctor.cc +++ b/mindspore/ccsrc/frontend/optimizer/ad/dfunctor.cc @@ -198,12 +198,12 @@ static AnfNodePtr SkipHookNodeInBackProp(const AnfNodePtr &node) { // Replace hook node with make tuple node. abstract::AbstractBasePtrList multi_output_abs; std::vector multi_output_nodes{NewValueNode(prim::kPrimMakeTuple)}; - std::for_each(output_cnode->inputs().cbegin() + 1, output_cnode->inputs().cend(), - [&multi_output_nodes, &multi_output_abs](const AnfNodePtr &inp) { - MS_EXCEPTION_IF_NULL(inp); - (void)multi_output_nodes.emplace_back(inp); - (void)multi_output_abs.emplace_back(inp->abstract()); - }); + (void)std::for_each(output_cnode->inputs().cbegin() + 1, output_cnode->inputs().cend(), + [&multi_output_nodes, &multi_output_abs](const AnfNodePtr &inp) { + MS_EXCEPTION_IF_NULL(inp); + (void)multi_output_nodes.emplace_back(inp); + (void)multi_output_abs.emplace_back(inp->abstract()); + }); auto primal_graph = node->func_graph(); MS_EXCEPTION_IF_NULL(primal_graph); auto make_tuple = primal_graph->NewCNode(std::move(multi_output_nodes)); diff --git a/mindspore/ccsrc/frontend/optimizer/auto_monad_eliminate.cc b/mindspore/ccsrc/frontend/optimizer/auto_monad_eliminate.cc index c57df9226e9..01c60ec5dc4 100644 --- a/mindspore/ccsrc/frontend/optimizer/auto_monad_eliminate.cc +++ b/mindspore/ccsrc/frontend/optimizer/auto_monad_eliminate.cc @@ -355,7 +355,7 @@ bool AutoMonadEliminator::ReplaceAutoMonadNode(const FuncGraphManagerPtr &manage auto &group = load_group.second; const auto ¶m_user_indexes = param_users[ref_key]; auto groups = SplitGroup(group, param_user_indexes, special_op_indexes); - need_merge_loads.insert(need_merge_loads.cend(), groups.cbegin(), groups.cend()); + (void)need_merge_loads.insert(need_merge_loads.cend(), groups.cbegin(), groups.cend()); } for (auto &group : need_merge_loads) { bool replaced = ReplaceSameGroupLoad(manager, fg, toposet, group); diff --git a/mindspore/ccsrc/frontend/optimizer/clean.cc b/mindspore/ccsrc/frontend/optimizer/clean.cc index 3ea953a02f2..24f8e33cabc 100644 --- a/mindspore/ccsrc/frontend/optimizer/clean.cc +++ b/mindspore/ccsrc/frontend/optimizer/clean.cc @@ -345,7 +345,7 @@ class SimplifyDataStructuresRewriter : public BaseRewriter { } // dict(k0:v0, k1:v1, ...) --> tuple(v0, v1, ...) - ValueTuplePtr DictToTuple(const ValueDictionaryPtr &dict) { + ValueTuplePtr DictToTuple(const ValueDictionaryPtr &dict) const { const auto &elements = dict->value(); std::vector values; values.reserve(elements.size()); diff --git a/mindspore/ccsrc/frontend/optimizer/graph_transform.cc b/mindspore/ccsrc/frontend/optimizer/graph_transform.cc index a68b404b790..288b868f826 100644 --- a/mindspore/ccsrc/frontend/optimizer/graph_transform.cc +++ b/mindspore/ccsrc/frontend/optimizer/graph_transform.cc @@ -56,7 +56,7 @@ std::vector TransformTupleArgument(const FuncGraphPtr &fg, const Anf elem_node->set_abstract(elements[i]); if (elements[i]->isa()) { auto nodes = TransformTupleArgument(fg, elem_node, elements[i]->cast()); - tuple_node_expanded.insert(tuple_node_expanded.cend(), nodes.cbegin(), nodes.cend()); + (void)tuple_node_expanded.insert(tuple_node_expanded.cend(), nodes.cbegin(), nodes.cend()); } else { tuple_node_expanded.push_back(elem_node); } diff --git a/mindspore/ccsrc/frontend/optimizer/irpass/environ_eliminate.h b/mindspore/ccsrc/frontend/optimizer/irpass/environ_eliminate.h index c41d7fe6568..e26ffc3b6df 100644 --- a/mindspore/ccsrc/frontend/optimizer/irpass/environ_eliminate.h +++ b/mindspore/ccsrc/frontend/optimizer/irpass/environ_eliminate.h @@ -340,7 +340,7 @@ class IncorporateEnvironGet : public AnfVisitor { ~IncorporateEnvironGet() override = default; AnfNodePtr operator()(const OptimizerPtr &, const AnfNodePtr &node) override { - static bool enable_closure = common::GetEnv("MS_DEV_ENABLE_CLOSURE") != "0"; + static const bool enable_closure = common::GetEnv("MS_DEV_ENABLE_CLOSURE") != "0"; if (enable_closure) { return nullptr; } diff --git a/mindspore/ccsrc/frontend/optimizer/irpass/ge/sparse_softmax_cross_entropy_with_logits_split.cc b/mindspore/ccsrc/frontend/optimizer/irpass/ge/sparse_softmax_cross_entropy_with_logits_split.cc index b53ec7aeef2..942825e182d 100644 --- a/mindspore/ccsrc/frontend/optimizer/irpass/ge/sparse_softmax_cross_entropy_with_logits_split.cc +++ b/mindspore/ccsrc/frontend/optimizer/irpass/ge/sparse_softmax_cross_entropy_with_logits_split.cc @@ -556,7 +556,7 @@ void SparseSoftmaxCrossEntropyWithLogitsSplit::Visit(const AnfNodePtr &node) { is_match_ = true; } -AnfNodePtr SparseSoftmaxCrossEntropyWithLogitsSplit::operator()(const OptimizerPtr &opt, const AnfNodePtr &node) { +AnfNodePtr SparseSoftmaxCrossEntropyWithLogitsSplit::operator()(const OptimizerPtr &, const AnfNodePtr &node) { MS_LOG(INFO) << "SparseSoftmaxCrossEntropyWithLogits split start."; auto IsSparseSoftmaxNode = [](const AnfNodePtr &node) -> bool { return IsPrimitiveCNode(node, prim::kPrimSparseSoftmaxCrossEntropyWithLogits) && diff --git a/mindspore/ccsrc/frontend/optimizer/irpass/vmap_eliminate.cc b/mindspore/ccsrc/frontend/optimizer/irpass/vmap_eliminate.cc index edba632c0c9..8a8d9c1f7c4 100644 --- a/mindspore/ccsrc/frontend/optimizer/irpass/vmap_eliminate.cc +++ b/mindspore/ccsrc/frontend/optimizer/irpass/vmap_eliminate.cc @@ -196,7 +196,7 @@ int GetAxisSizeByAbs(const AbstractBasePtr &abs, ValuePtr *const in_axes) { } axis = axis < 0 ? shape_len + axis : axis; *in_axes = std::make_shared(axis); - axis_size = orig_shape[LongToSize(axis)]; + axis_size = LongToInt(orig_shape[LongToSize(axis)]); return axis_size; } return axis_size; diff --git a/mindspore/ccsrc/frontend/optimizer/optimizer.h b/mindspore/ccsrc/frontend/optimizer/optimizer.h index 64f3221a78a..8f3614420e9 100644 --- a/mindspore/ccsrc/frontend/optimizer/optimizer.h +++ b/mindspore/ccsrc/frontend/optimizer/optimizer.h @@ -251,7 +251,7 @@ class Optimizer : public std::enable_shared_from_this { void set_is_first_order_j(bool is_first_order_j) { is_first_order_j_ = is_first_order_j; } struct { - int64_t counter; + int64_t counter = 0; std::string name; } CurPass_;