forked from mindspore-Ecosystem/mindspore
!265 [bug]fix InsertGradientOf with Parameter.
Merge pull request !265 from vlne-v1/I1EE1P-insertgradiendof-with-parameter
This commit is contained in:
commit
e231127cd8
|
@ -40,6 +40,7 @@
|
||||||
#include "optimizer/irpass/incorporate_getitem.h"
|
#include "optimizer/irpass/incorporate_getitem.h"
|
||||||
#include "optimizer/irpass/incorporate_call.h"
|
#include "optimizer/irpass/incorporate_call.h"
|
||||||
#include "optimizer/irpass/grad_var_prepare.h"
|
#include "optimizer/irpass/grad_var_prepare.h"
|
||||||
|
#include "optimizer/irpass/param_replace.h"
|
||||||
|
|
||||||
namespace mindspore {
|
namespace mindspore {
|
||||||
namespace opt {
|
namespace opt {
|
||||||
|
@ -81,6 +82,7 @@ OptimizeIRPassLib::OptimizeIRPassLib() {
|
||||||
get_make_ref_eliminate_ =
|
get_make_ref_eliminate_ =
|
||||||
MakeSubstitution(GetMakeRefEliminater(), "get_make_ref_eliminate", {prim::kPrimGetRefKey, prim::kPrimGetRefValue});
|
MakeSubstitution(GetMakeRefEliminater(), "get_make_ref_eliminate", {prim::kPrimGetRefKey, prim::kPrimGetRefValue});
|
||||||
replace_refkey_by_param_ = MakeSubstitution(ReplaceRefkeyByParam(), "replace_refkey_by_param", IsValueNode<RefKey>);
|
replace_refkey_by_param_ = MakeSubstitution(ReplaceRefkeyByParam(), "replace_refkey_by_param", IsValueNode<RefKey>);
|
||||||
|
replace_old_param_ = MakeSubstitution(ReplaceOldParam(), "replace_old_param", IsParam);
|
||||||
|
|
||||||
// Gradient transforms
|
// Gradient transforms
|
||||||
expand_jprim_ = MakeSubstitution(ExpandJPrim(), "expand_jprim", prim::kPrimJ);
|
expand_jprim_ = MakeSubstitution(ExpandJPrim(), "expand_jprim", prim::kPrimJ);
|
||||||
|
|
|
@ -58,6 +58,7 @@ class OptimizeIRPassLib {
|
||||||
SubstitutionPtr make_ref_eliminate_;
|
SubstitutionPtr make_ref_eliminate_;
|
||||||
SubstitutionPtr get_make_ref_eliminate_;
|
SubstitutionPtr get_make_ref_eliminate_;
|
||||||
SubstitutionPtr replace_refkey_by_param_;
|
SubstitutionPtr replace_refkey_by_param_;
|
||||||
|
SubstitutionPtr replace_old_param_;
|
||||||
|
|
||||||
// Branch culling
|
// Branch culling
|
||||||
SubstitutionPtr switch_simplify_;
|
SubstitutionPtr switch_simplify_;
|
||||||
|
|
|
@ -0,0 +1,60 @@
|
||||||
|
/**
|
||||||
|
* Copyright 2020 Huawei Technologies Co., Ltd
|
||||||
|
*
|
||||||
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||||
|
* you may not use this file except in compliance with the License.
|
||||||
|
* You may obtain a copy of the License at
|
||||||
|
*
|
||||||
|
* http://www.apache.org/licenses/LICENSE-2.0
|
||||||
|
*
|
||||||
|
* Unless required by applicable law or agreed to in writing, software
|
||||||
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||||
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||||
|
* See the License for the specific language governing permissions and
|
||||||
|
* limitations under the License.
|
||||||
|
*/
|
||||||
|
|
||||||
|
#ifndef MINDSPORE_CCSRC_OPTIMIZER_IRPASS_PARAM_REPLACE_H_
|
||||||
|
#define MINDSPORE_CCSRC_OPTIMIZER_IRPASS_PARAM_REPLACE_H_
|
||||||
|
|
||||||
|
#include <memory>
|
||||||
|
|
||||||
|
#include "optimizer/optimizer.h"
|
||||||
|
#include "optimizer/irpass.h"
|
||||||
|
#include "ir/visitor.h"
|
||||||
|
#include "operator/ops.h"
|
||||||
|
#include "pipeline/parse/parse.h"
|
||||||
|
|
||||||
|
namespace mindspore {
|
||||||
|
namespace opt {
|
||||||
|
namespace irpass {
|
||||||
|
class ReplaceOldParam : public AnfVisitor {
|
||||||
|
public:
|
||||||
|
AnfNodePtr operator()(const OptimizerPtr &optimizer, const AnfNodePtr &node) override {
|
||||||
|
if (!IsParam(node)) {
|
||||||
|
return nullptr;
|
||||||
|
}
|
||||||
|
auto resource = std::dynamic_pointer_cast<pipeline::Resource>(optimizer->resource());
|
||||||
|
MS_EXCEPTION_IF_NULL(resource);
|
||||||
|
|
||||||
|
auto top_graph = resource->func_graph(); // parse::Parser::GetTopFuncGraph();
|
||||||
|
MS_EXCEPTION_IF_NULL(top_graph);
|
||||||
|
|
||||||
|
auto param_node = node->cast<ParameterPtr>();
|
||||||
|
if (!param_node->has_default() || node->func_graph() == top_graph) {
|
||||||
|
return nullptr;
|
||||||
|
}
|
||||||
|
auto para_name = param_node->name();
|
||||||
|
for (const auto &tnode : top_graph->parameters()) {
|
||||||
|
auto para = tnode->cast<ParameterPtr>();
|
||||||
|
if (para != nullptr && para->name() == para_name) {
|
||||||
|
return para;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return nullptr;
|
||||||
|
}
|
||||||
|
};
|
||||||
|
} // namespace irpass
|
||||||
|
} // namespace opt
|
||||||
|
} // namespace mindspore
|
||||||
|
#endif // MINDSPORE_CCSRC_OPTIMIZER_IRPASS_PARAM_REPLACE_H_
|
|
@ -88,6 +88,7 @@ FuncGraphPtr Renormalize(const ResourcePtr& res, const FuncGraphPtr& func_graph,
|
||||||
double t2 = GetTime();
|
double t2 = GetTime();
|
||||||
#endif
|
#endif
|
||||||
auto ret = ProgramSpecialize(res, func_graph, result.context);
|
auto ret = ProgramSpecialize(res, func_graph, result.context);
|
||||||
|
res->set_func_graph(ret);
|
||||||
#ifdef ENABLE_PROFILE
|
#ifdef ENABLE_PROFILE
|
||||||
double t3 = GetTime();
|
double t3 = GetTime();
|
||||||
MsProfile::StatTime("renormalize.infer", t2 - t1);
|
MsProfile::StatTime("renormalize.infer", t2 - t1);
|
||||||
|
|
|
@ -114,11 +114,9 @@ OptPassGroupMap GetOptPassesA(const opt::irpass::OptimizeIRPassLib& irpass) {
|
||||||
opt::OptPassConfig grad = opt::OptPassConfig({irpass.expand_jprim_}, true);
|
opt::OptPassConfig grad = opt::OptPassConfig({irpass.expand_jprim_}, true);
|
||||||
opt::irpass::ResolveIRPassLib resolve_irpass;
|
opt::irpass::ResolveIRPassLib resolve_irpass;
|
||||||
|
|
||||||
opt::OptPassConfig resolve_pass = opt::OptPassConfig({
|
opt::OptPassConfig resolve_pass =
|
||||||
resolve_irpass.resolver_resolve_,
|
opt::OptPassConfig({resolve_irpass.resolver_resolve_, resolve_irpass.resolver_getattr_,
|
||||||
resolve_irpass.resolver_getattr_,
|
irpass.get_make_ref_eliminate_, irpass.replace_old_param_});
|
||||||
irpass.get_make_ref_eliminate_,
|
|
||||||
});
|
|
||||||
|
|
||||||
OptPassGroupMap map_a({{"a_1", a_1},
|
OptPassGroupMap map_a({{"a_1", a_1},
|
||||||
{"a_2", a_2},
|
{"a_2", a_2},
|
||||||
|
|
|
@ -129,7 +129,7 @@ def test_cell_assign():
|
||||||
self.matrix_g = mindspore.Parameter(Tensor(np.ones([2, 2], np.float32)), name="matrix_g")
|
self.matrix_g = mindspore.Parameter(Tensor(np.ones([2, 2], np.float32)), name="matrix_g")
|
||||||
|
|
||||||
def save_gradient(self, dout):
|
def save_gradient(self, dout):
|
||||||
self.matrix_g = dout
|
self.matrix_g = dout + self.matrix_g
|
||||||
return dout
|
return dout
|
||||||
|
|
||||||
def construct(self, x, y):
|
def construct(self, x, y):
|
||||||
|
|
Loading…
Reference in New Issue