From 68ba6532c43735376185b107f471598842094e68 Mon Sep 17 00:00:00 2001 From: yao_yf Date: Sat, 15 Aug 2020 10:41:23 +0800 Subject: [PATCH] add field in stra ckpt --- .../strategy_checkpoint/parallel_strategy_checkpoint.cc | 1 + mindspore/ccsrc/utils/node_strategy.proto | 1 + .../official/recommend/wide_and_deep/src/wide_and_deep.py | 6 +----- 3 files changed, 3 insertions(+), 5 deletions(-) diff --git a/mindspore/ccsrc/frontend/parallel/strategy_checkpoint/parallel_strategy_checkpoint.cc b/mindspore/ccsrc/frontend/parallel/strategy_checkpoint/parallel_strategy_checkpoint.cc index 9ab55e6915a..5ac82218438 100644 --- a/mindspore/ccsrc/frontend/parallel/strategy_checkpoint/parallel_strategy_checkpoint.cc +++ b/mindspore/ccsrc/frontend/parallel/strategy_checkpoint/parallel_strategy_checkpoint.cc @@ -129,6 +129,7 @@ Status StrategyCheckpoint::Save(const StrategyMap &strategy_map, const TensorInf param_split_shape->add_dim(dim_pair.first); indices_offset->add_dim(dim_pair.second); } + parallel_layouts->set_field(tensor_layout.get_field_size()); } std::fstream output(save_file_, std::ios::out | std::ios::trunc | std::ios::binary); diff --git a/mindspore/ccsrc/utils/node_strategy.proto b/mindspore/ccsrc/utils/node_strategy.proto index dc9d65407d7..daa51880d63 100644 --- a/mindspore/ccsrc/utils/node_strategy.proto +++ b/mindspore/ccsrc/utils/node_strategy.proto @@ -53,6 +53,7 @@ message ParallelLayouts { repeated TensorMap tensor_map = 2; repeated ParamSplitShape param_split_shape = 3; repeated IndicesOffset indices_offset = 4; + required int32 field = 5; } message ParallelLayoutItem { diff --git a/model_zoo/official/recommend/wide_and_deep/src/wide_and_deep.py b/model_zoo/official/recommend/wide_and_deep/src/wide_and_deep.py index 151ecbdeb64..8a956803b95 100644 --- a/model_zoo/official/recommend/wide_and_deep/src/wide_and_deep.py +++ b/model_zoo/official/recommend/wide_and_deep/src/wide_and_deep.py @@ -161,13 +161,9 @@ class WideDeepModel(nn.Cell): self.layer_dims = self.deep_layer_dims_list + [1] self.all_dim_list = [self.deep_input_dims] + self.layer_dims - init_acts = [('Wide_w', [self.vocab_size, 1], self.emb_init), - ('V_l2', [self.vocab_size, self.emb_dim], self.emb_init), - ('Wide_b', [1], self.emb_init)] + init_acts = [('Wide_b', [1], self.emb_init)] var_map = init_var_dict(self.init_args, init_acts) - self.wide_w = var_map["Wide_w"] self.wide_b = var_map["Wide_b"] - self.embedding_table = var_map["V_l2"] if parameter_server: self.wide_w.set_param_ps() self.embedding_table.set_param_ps()