From 5250b327aedc215abba851e99e655f567a7257b5 Mon Sep 17 00:00:00 2001 From: alex-yuyue Date: Tue, 8 Dec 2020 16:59:14 -0500 Subject: [PATCH] Fix some non-minddata typos Signed-off-by: alex-yuyue --- .../ccsrc/frontend/parallel/ops_info/reduce_method_info.cc | 4 ++-- model_zoo/official/cv/faster_rcnn/src/dataset.py | 2 +- model_zoo/official/cv/ssd/src/dataset.py | 2 +- model_zoo/official/cv/yolov3_resnet18/src/dataset.py | 2 +- model_zoo/official/nlp/lstm/src/dataset.py | 4 ++-- model_zoo/research/cv/ssd_ghostnet/src/dataset.py | 2 +- tests/st/model_zoo_tests/yolov3/src/dataset.py | 2 +- 7 files changed, 9 insertions(+), 9 deletions(-) diff --git a/mindspore/ccsrc/frontend/parallel/ops_info/reduce_method_info.cc b/mindspore/ccsrc/frontend/parallel/ops_info/reduce_method_info.cc index 393fe2f7517..db8b81643ed 100644 --- a/mindspore/ccsrc/frontend/parallel/ops_info/reduce_method_info.cc +++ b/mindspore/ccsrc/frontend/parallel/ops_info/reduce_method_info.cc @@ -199,7 +199,7 @@ Status ReduceMethod::InferForwardCommunication() { return SUCCESS; } -ForwardOp CreatReduceMeanForwardOp(const std::vector &forward_group, const TypePtr &dtype) { +ForwardOp CreateReduceMeanForwardOp(const std::vector &forward_group, const TypePtr &dtype) { // Creat AllReduceSum op Operator op0 = CreateAllReduceOp(REDUCE_OP_SUM, forward_group[0].name()); std::string group_name = forward_group[0].name(); @@ -275,7 +275,7 @@ Status ReduceMeanInfo::InferForwardCommunication() { } auto element_type = outputs_dtype_->cast()->element(); - forward_op_ = CreatReduceMeanForwardOp(forward_group, element_type); + forward_op_ = CreateReduceMeanForwardOp(forward_group, element_type); } return SUCCESS; diff --git a/model_zoo/official/cv/faster_rcnn/src/dataset.py b/model_zoo/official/cv/faster_rcnn/src/dataset.py index 2c4fd207144..131ac86c0a6 100644 --- a/model_zoo/official/cv/faster_rcnn/src/dataset.py +++ b/model_zoo/official/cv/faster_rcnn/src/dataset.py @@ -413,7 +413,7 @@ def data_to_mindrecord_byte_image(dataset="coco", is_training=True, prefix="fast def create_fasterrcnn_dataset(mindrecord_file, batch_size=2, device_num=1, rank_id=0, is_training=True, num_parallel_workers=8): - """Creatr FasterRcnn dataset with MindDataset.""" + """Create FasterRcnn dataset with MindDataset.""" cv2.setNumThreads(0) de.config.set_prefetch_size(8) ds = de.MindDataset(mindrecord_file, columns_list=["image", "annotation"], num_shards=device_num, shard_id=rank_id, diff --git a/model_zoo/official/cv/ssd/src/dataset.py b/model_zoo/official/cv/ssd/src/dataset.py index 74f6344a47f..0550c1be265 100644 --- a/model_zoo/official/cv/ssd/src/dataset.py +++ b/model_zoo/official/cv/ssd/src/dataset.py @@ -390,7 +390,7 @@ def data_to_mindrecord_byte_image(dataset="coco", is_training=True, prefix="ssd. def create_ssd_dataset(mindrecord_file, batch_size=32, repeat_num=10, device_num=1, rank=0, is_training=True, num_parallel_workers=4, use_multiprocessing=True): - """Creatr SSD dataset with MindDataset.""" + """Create SSD dataset with MindDataset.""" ds = de.MindDataset(mindrecord_file, columns_list=["img_id", "image", "annotation"], num_shards=device_num, shard_id=rank, num_parallel_workers=num_parallel_workers, shuffle=is_training) decode = C.Decode() diff --git a/model_zoo/official/cv/yolov3_resnet18/src/dataset.py b/model_zoo/official/cv/yolov3_resnet18/src/dataset.py index b6c37cfccf2..00652b73de5 100644 --- a/model_zoo/official/cv/yolov3_resnet18/src/dataset.py +++ b/model_zoo/official/cv/yolov3_resnet18/src/dataset.py @@ -291,7 +291,7 @@ def data_to_mindrecord_byte_image(image_dir, anno_path, mindrecord_dir, prefix, def create_yolo_dataset(mindrecord_dir, batch_size=32, repeat_num=1, device_num=1, rank=0, is_training=True, num_parallel_workers=8): - """Creatr YOLOv3 dataset with MindDataset.""" + """Create YOLOv3 dataset with MindDataset.""" ds = de.MindDataset(mindrecord_dir, columns_list=["image", "annotation"], num_shards=device_num, shard_id=rank, num_parallel_workers=num_parallel_workers, shuffle=is_training) decode = C.Decode() diff --git a/model_zoo/official/nlp/lstm/src/dataset.py b/model_zoo/official/nlp/lstm/src/dataset.py index 03d4276dfd0..92e28c29374 100644 --- a/model_zoo/official/nlp/lstm/src/dataset.py +++ b/model_zoo/official/nlp/lstm/src/dataset.py @@ -43,7 +43,7 @@ def lstm_create_dataset(data_home, batch_size, repeat_num=1, training=True): def _convert_to_mindrecord(data_home, features, labels, weight_np=None, training=True): """ - convert imdb dataset to mindrecoed dataset + convert imdb dataset to mindrecord dataset """ if weight_np is not None: np.savetxt(os.path.join(data_home, 'weight.txt'), weight_np) @@ -76,7 +76,7 @@ def _convert_to_mindrecord(data_home, features, labels, weight_np=None, training def convert_to_mindrecord(embed_size, aclimdb_path, preprocess_path, glove_path): """ - convert imdb dataset to mindrecoed dataset + convert imdb dataset to mindrecord dataset """ parser = ImdbParser(aclimdb_path, glove_path, embed_size) parser.parse() diff --git a/model_zoo/research/cv/ssd_ghostnet/src/dataset.py b/model_zoo/research/cv/ssd_ghostnet/src/dataset.py index c363ff5d18b..b56d333dcc3 100644 --- a/model_zoo/research/cv/ssd_ghostnet/src/dataset.py +++ b/model_zoo/research/cv/ssd_ghostnet/src/dataset.py @@ -394,7 +394,7 @@ def data_to_mindrecord_byte_image(dataset="coco", is_training=True, prefix="ssd. def create_ssd_dataset(mindrecord_file, batch_size=32, repeat_num=10, device_num=1, rank=0, is_training=True, num_parallel_workers=4): - """Creatr SSD dataset with MindDataset.""" + """Create SSD dataset with MindDataset.""" ds = de.MindDataset(mindrecord_file, columns_list=["img_id", "image", "annotation"], num_shards=device_num, shard_id=rank, num_parallel_workers=num_parallel_workers, shuffle=is_training) decode = C.Decode() diff --git a/tests/st/model_zoo_tests/yolov3/src/dataset.py b/tests/st/model_zoo_tests/yolov3/src/dataset.py index f002ea46d39..386551d0d46 100644 --- a/tests/st/model_zoo_tests/yolov3/src/dataset.py +++ b/tests/st/model_zoo_tests/yolov3/src/dataset.py @@ -294,7 +294,7 @@ def data_to_mindrecord_byte_image(image_dir, anno_path, mindrecord_dir, prefix=" def create_yolo_dataset(mindrecord_dir, batch_size=32, repeat_num=10, device_num=1, rank=0, is_training=True, num_parallel_workers=8): - """Creatr YOLOv3 dataset with MindDataset.""" + """Create YOLOv3 dataset with MindDataset.""" ds = de.MindDataset(mindrecord_dir, columns_list=["image", "annotation"], num_shards=device_num, shard_id=rank, num_parallel_workers=num_parallel_workers, shuffle=False) decode = C.Decode()