From 77fd2e841eab957a08b766a210e2db45942263a4 Mon Sep 17 00:00:00 2001 From: zhaojichen Date: Thu, 16 Apr 2020 03:54:32 -0400 Subject: [PATCH] Add Group Normalization --- mindspore/nn/layer/normalization.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/mindspore/nn/layer/normalization.py b/mindspore/nn/layer/normalization.py index b286eaae1b0..4aafaf031e1 100644 --- a/mindspore/nn/layer/normalization.py +++ b/mindspore/nn/layer/normalization.py @@ -20,7 +20,7 @@ from mindspore.common.initializer import initializer from mindspore.common.tensor import Tensor import mindspore.common.dtype as mstype import mindspore.context as context -from mindspore._checkparam import check_int_positive, check_bool, check_typename +from mindspore._checkparam import check_int_positive, check_bool, check_typename from mindspore._extends import cell_attr_register from ..cell import Cell @@ -293,7 +293,7 @@ class LayerNorm(Cell): class GroupNorm(Cell): r""" - Group Normalization over a mini-batch of inputs. + Group Normalization over a mini-batch of inputs. Group normalization is widely used in recurrent neural networks. It applies normalization over a mini-batch of inputs for each single training case as described