forked from mindspore-Ecosystem/mindspore
Revert "!16693 add Conv2dTranspose"
This reverts commita2f50fb7db
, reversing changes made to4735ec3296
.
This commit is contained in:
parent
63086109e0
commit
ea04c4304f
|
@ -26,7 +26,6 @@
|
|||
#include "backend/optimizer/pass/convert_attr_to_unify_mindir.h"
|
||||
#include "backend/optimizer/pass/add_training_attr.h"
|
||||
#include "backend/optimizer/pass/optimize_updatestate.h"
|
||||
#include "backend/optimizer/pass/conv_transpose_to_conv_bp.h"
|
||||
#include "utils/ms_context.h"
|
||||
#include "debug/anf_ir_dump.h"
|
||||
|
||||
|
@ -44,7 +43,6 @@ void BackendCommonOptimization(const std::shared_ptr<session::KernelGraph> &kern
|
|||
}
|
||||
auto optimizer = std::make_shared<GraphOptimizer>();
|
||||
auto common_pm = std::make_shared<PassManager>("common_pm");
|
||||
common_pm->AddPass(std::make_shared<ConvTransposeToConvBackpropInputPass>());
|
||||
common_pm->AddPass(std::make_shared<ConvertConstInputToAttr>());
|
||||
common_pm->AddPass(std::make_shared<ConvertAttrToUnifyMindIR>());
|
||||
common_pm->AddPass(std::make_shared<ConstToAttrStridedSliceGradPass>());
|
||||
|
|
|
@ -1,53 +0,0 @@
|
|||
/**
|
||||
* Copyright 2021 Huawei Technologies Co., Ltd
|
||||
*
|
||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||
* you may not use this file except in compliance with the License.
|
||||
* You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
#include "backend/optimizer/pass/conv_transpose_to_conv_bp.h"
|
||||
#include <memory>
|
||||
#include <vector>
|
||||
#include "ir/primitive.h"
|
||||
#include "utils/utils.h"
|
||||
#include "backend/optimizer/common/helper.h"
|
||||
|
||||
namespace mindspore {
|
||||
namespace opt {
|
||||
namespace {
|
||||
constexpr size_t kCNodePrimitiveIdx = 0;
|
||||
} // namespace
|
||||
|
||||
const BaseRef ConvTransposeToConvBackpropInputPass::DefinePattern() const {
|
||||
VarPtr Xs = std::make_shared<SeqVar>();
|
||||
auto conv_transpose = std::make_shared<Primitive>(kConv2DTransposeOpName);
|
||||
return VectorRef({conv_transpose, Xs});
|
||||
}
|
||||
|
||||
const AnfNodePtr ConvTransposeToConvBackpropInputPass::Process(const FuncGraphPtr &graph, const AnfNodePtr &node,
|
||||
const EquivPtr &) const {
|
||||
MS_EXCEPTION_IF_NULL(graph);
|
||||
MS_EXCEPTION_IF_NULL(node);
|
||||
auto conv_transpose = node->cast<CNodePtr>();
|
||||
MS_EXCEPTION_IF_NULL(conv_transpose);
|
||||
|
||||
if (conv_transpose->size() <= kCNodePrimitiveIdx) {
|
||||
MS_LOG(EXCEPTION) << "Invalid cnode " << node->DebugString() << " input size " << conv_transpose->size();
|
||||
}
|
||||
|
||||
auto prim = GetValueNode<PrimitivePtr>(conv_transpose->input(kCNodePrimitiveIdx));
|
||||
MS_EXCEPTION_IF_NULL(prim);
|
||||
prim->Named::operator=(Named(kConv2DBackpropInputOpName));
|
||||
|
||||
return node;
|
||||
}
|
||||
} // namespace opt
|
||||
} // namespace mindspore
|
|
@ -1,34 +0,0 @@
|
|||
/**
|
||||
* Copyright 2020 Huawei Technologies Co., Ltd
|
||||
*
|
||||
* Licensed under the Apache License, Version 2.0 (the "License");
|
||||
* you may not use this file except in compliance with the License.
|
||||
* You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*/
|
||||
#ifndef MINDSPORE_CCSRC_BACKEND_OPTIMIZER_PASS_CONV_TRANSPOSE_TO_CONV_BP_H_
|
||||
#define MINDSPORE_CCSRC_BACKEND_OPTIMIZER_PASS_CONV_TRANSPOSE_TO_CONV_BP_H_
|
||||
|
||||
#include <memory>
|
||||
#include "backend/optimizer/common/optimizer.h"
|
||||
|
||||
namespace mindspore {
|
||||
namespace opt {
|
||||
class ConvTransposeToConvBackpropInputPass : public PatternProcessPass {
|
||||
public:
|
||||
explicit ConvTransposeToConvBackpropInputPass(bool multigraph = true)
|
||||
: PatternProcessPass("conv_transpose_to_conv_backprop_input", multigraph) {}
|
||||
~ConvTransposeToConvBackpropInputPass() override = default;
|
||||
const BaseRef DefinePattern() const override;
|
||||
const AnfNodePtr Process(const FuncGraphPtr &, const AnfNodePtr &, const EquivPtr &) const override;
|
||||
};
|
||||
} // namespace opt
|
||||
} // namespace mindspore
|
||||
#endif // MINDSPORE_CCSRC_BACKEND_OPTIMIZER_PASS_CONST_TO_ATTR_STRIDED_SLICE_GRAD_H_
|
|
@ -49,7 +49,6 @@ ATTR_MAP(Conv2DBackpropInputD) = {
|
|||
};
|
||||
OUTPUT_MAP(Conv2DBackpropInputD) = {{0, OUTPUT_DESC(y)}};
|
||||
REG_ADPT_DESC(Conv2DBackpropInputD, prim::kPrimConv2DBackpropInput->name(), ADPT_DESC(Conv2DBackpropInputD))
|
||||
REG_ADPT_DESC(Conv2DTranspose, kConv2DTransposeOpName, ADPT_DESC(Conv2DBackpropInputD))
|
||||
|
||||
// Conv2DBackpropFilterD
|
||||
INPUT_MAP(Conv2DBackpropFilterD) = {{1, INPUT_DESC(out_backprop)}, {2, INPUT_DESC(x)}};
|
||||
|
|
|
@ -175,7 +175,6 @@ constexpr auto kSpaceToBatchOpName = "SpaceToBatch";
|
|||
constexpr auto kBatchToSpaceOpName = "BatchToSpace";
|
||||
constexpr auto kSpaceToDepthOpName = "SpaceToDepth";
|
||||
constexpr auto kPadOpName = "Pad";
|
||||
constexpr auto kConv2DTransposeOpName = "Conv2DTranspose";
|
||||
constexpr auto kConv2DBackpropInputOpName = "Conv2DBackpropInput";
|
||||
constexpr auto kConv2DBackpropFilterOpName = "Conv2DBackpropFilter";
|
||||
constexpr auto kDepthwiseConv2dNativeOpName = "DepthwiseConv2dNative";
|
||||
|
|
|
@ -963,15 +963,15 @@ class Conv2dTranspose(_Conv):
|
|||
if Validator.check_bool(has_bias):
|
||||
self.bias = Parameter(initializer(bias_init, [out_channels]), name='bias')
|
||||
|
||||
# cause Conv2DTranspose's out_channel refers to Conv2D's out_channel.
|
||||
self.conv2d_transpose = P.Conv2DTranspose(out_channel=in_channels,
|
||||
kernel_size=kernel_size,
|
||||
mode=1,
|
||||
pad_mode=pad_mode,
|
||||
pad=padding,
|
||||
stride=stride,
|
||||
dilation=dilation,
|
||||
group=group)
|
||||
# cause Conv2DBackpropInput's out_channel refers to Conv2D's out_channel.
|
||||
self.conv2d_transpose = P.Conv2DBackpropInput(out_channel=in_channels,
|
||||
kernel_size=kernel_size,
|
||||
mode=1,
|
||||
pad_mode=pad_mode,
|
||||
pad=padding,
|
||||
stride=stride,
|
||||
dilation=dilation,
|
||||
group=group)
|
||||
self.bias_add = P.BiasAdd()
|
||||
if isinstance(self.padding, int):
|
||||
self.padding_top, self.padding_bottom, self.padding_left, self.padding_right = (self.padding,) * 4
|
||||
|
|
|
@ -1076,7 +1076,6 @@ def get_bprop_roi_align(self):
|
|||
return bprop
|
||||
|
||||
|
||||
@bprop_getters.register(P.Conv2DTranspose)
|
||||
@bprop_getters.register(P.Conv2DBackpropInput)
|
||||
def get_bprop_conv2d_backprop_input(self):
|
||||
"""Grad definition for `Conv2DBackpropInput` operation."""
|
||||
|
|
|
@ -64,7 +64,7 @@ from .random_ops import (RandomChoiceWithMask, StandardNormal, Gamma, Poisson, U
|
|||
RandomCategorical, StandardLaplace, Multinomial, UniformCandidateSampler,
|
||||
LogUniformCandidateSampler)
|
||||
from .nn_ops import (LSTM, SGD, Adam, FusedSparseAdam, FusedSparseLazyAdam, AdamNoUpdateParam, ApplyMomentum, BatchNorm,
|
||||
BiasAdd, Conv2D, Conv3D, Conv2DTranspose, Conv3DTranspose,
|
||||
BiasAdd, Conv2D, Conv3D, Conv3DTranspose,
|
||||
DepthwiseConv2dNative,
|
||||
DropoutDoMask, Dropout, Dropout2D, Dropout3D, DropoutGenMask, Flatten,
|
||||
InstanceNorm, BNTrainingReduce, BNTrainingUpdate,
|
||||
|
@ -142,7 +142,6 @@ __all__ = [
|
|||
'Xlogy',
|
||||
'Conv2D',
|
||||
'Conv3D',
|
||||
'Conv2DTranspose',
|
||||
'Conv3DTranspose',
|
||||
'Flatten',
|
||||
'MaxPoolWithArgmax',
|
||||
|
|
|
@ -1997,66 +1997,6 @@ class Conv2DBackpropInput(Primitive):
|
|||
self.pad_list = pad_list
|
||||
|
||||
|
||||
class Conv2DTranspose(Conv2DBackpropInput):
|
||||
"""
|
||||
Compute a 2D transposed convolution, which is also known as a deconvolution
|
||||
(although it is not an actual deconvolution).
|
||||
|
||||
Args:
|
||||
out_channel (int): The dimensionality of the output space.
|
||||
kernel_size (Union[int, tuple[int]]): The size of the convolution window.
|
||||
pad_mode (str): Modes to fill padding. It could be "valid", "same", or "pad". Default: "valid".
|
||||
pad (Union[int, tuple[int]]): The pad value to be filled. Default: 0. If `pad` is an integer, the paddings of
|
||||
top, bottom, left and right are the same, equal to pad. If `pad` is a tuple of four integers, the
|
||||
padding of top, bottom, left and right equal to pad[0], pad[1], pad[2], and pad[3] correspondingly.
|
||||
mode (int): Modes for different convolutions. 0 Math convolutiuon, 1 cross-correlation convolution ,
|
||||
2 deconvolution, 3 depthwise convolution. Default: 1.
|
||||
stride (Union[int. tuple[int]]): The stride to be applied to the convolution filter. Default: 1.
|
||||
dilation (Union[int. tuple[int]]): Specifies the dilation rate to be used for the dilated convolution.
|
||||
Default: 1.
|
||||
group (int): Splits input into groups. Default: 1.
|
||||
data_format (str) - The format of input and output data. It should be 'NHWC' or 'NCHW',\
|
||||
default is 'NCHW'.
|
||||
|
||||
Inputs:
|
||||
- **dout** (Tensor) - the gradients w.r.t the output of the convolution. The shape conforms to the default
|
||||
data_format :math:`(N, C_{out}, H_{out}, W_{out})`.
|
||||
- **weight** (Tensor) - Set size of kernel is :math:`(K_1, K_2)`, then the shape is
|
||||
:math:`(C_{out}, C_{in}, K_1, K_2)`.
|
||||
- **input_size** (Tensor) - A tuple describes the shape of the input which conforms to the format
|
||||
:math:`(N, C_{in}, H_{in}, W_{in})`.
|
||||
|
||||
Outputs:
|
||||
Tensor, the gradients w.r.t the input of convolution. It has the same shape as the input.
|
||||
|
||||
Raises:
|
||||
TypeError: If `kernel_size`, `stride`, `pad` or `dilation` is neither an int nor a tuple.
|
||||
TypeError: If `out_channel` or `group` is not an int.
|
||||
ValueError: If `kernel_size`, `stride` or `dilation` is less than 1.
|
||||
ValueError: If `pad_mode` is not one of 'same', 'valid', 'pad'.
|
||||
ValueError: If `padding` is a tuple whose length is not equal to 4.
|
||||
ValueError: If `pad_mode` it not equal to 'pad' and `pad` is not equal to (0, 0, 0, 0).
|
||||
ValueError: If `data_format` is neither 'NCHW' not 'NHWC'.
|
||||
|
||||
Supported Platforms:
|
||||
``Ascend`` ``GPU`` ``CPU``
|
||||
|
||||
Examples:
|
||||
>>> dout = Tensor(np.ones([10, 32, 30, 30]), mindspore.float32)
|
||||
>>> weight = Tensor(np.ones([32, 32, 3, 3]), mindspore.float32)
|
||||
>>> x = Tensor(np.ones([10, 32, 32, 32]))
|
||||
>>> conv2d_transpose_input = ops.Conv2DTranspose(out_channel=32, kernel_size=3)
|
||||
>>> output = conv2d_transpose_input(dout, weight, F.shape(x))
|
||||
>>> print(output.shape)
|
||||
(10, 32, 32, 32)
|
||||
"""
|
||||
@prim_attr_register
|
||||
def __init__(self, out_channel, kernel_size, pad_mode="valid", pad=0,
|
||||
pad_list=None, mode=1, stride=1, dilation=1, group=1, data_format="NCHW"):
|
||||
super(Conv2DTranspose, self).__init__(out_channel, kernel_size, pad_mode, pad,
|
||||
pad_list, mode, stride, dilation, group, data_format)
|
||||
|
||||
|
||||
class BiasAdd(PrimitiveWithCheck):
|
||||
r"""
|
||||
Returns sum of input and bias tensor.
|
||||
|
|
|
@ -1,73 +0,0 @@
|
|||
# Copyright 2021 Huawei Technologies Co., Ltd
|
||||
#
|
||||
# Licensed under the Apache License, Version 2.0 (the "License");
|
||||
# you may not use this file except in compliance with the License.
|
||||
# You may obtain a copy of the License at
|
||||
#
|
||||
# http://www.apache.org/licenses/LICENSE-2.0
|
||||
#
|
||||
# Unless required by applicable law or agreed to in writing, software
|
||||
# distributed under the License is distributed on an "AS IS" BASIS,
|
||||
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
# See the License for the specific language governing permissions and
|
||||
# limitations under the License.
|
||||
# ============================================================================
|
||||
import numpy as np
|
||||
|
||||
import mindspore.context as context
|
||||
import mindspore.nn as nn
|
||||
from mindspore import Tensor
|
||||
from mindspore.common.api import ms_function
|
||||
from mindspore.common.initializer import initializer
|
||||
from mindspore.common.parameter import Parameter
|
||||
from mindspore.ops import operations as P
|
||||
|
||||
context.set_context(device_target="Ascend")
|
||||
|
||||
|
||||
class Net(nn.Cell):
|
||||
def __init__(self):
|
||||
super(Net, self).__init__()
|
||||
out_channel = 4
|
||||
kernel_size = 1
|
||||
self.conv_input = P.Conv2DTranspose(out_channel,
|
||||
kernel_size,
|
||||
pad_mode="valid",
|
||||
pad=0,
|
||||
mode=1,
|
||||
stride=1,
|
||||
dilation=1,
|
||||
group=1)
|
||||
self.w = Parameter(
|
||||
initializer(Tensor(np.array([[[[1, 0, -1], [1, 0, -1], [1, 0, -1]]]]).astype(np.float32)), [1, 1, 3, 3]),
|
||||
name='w')
|
||||
self.x = Parameter(initializer(Tensor(np.array([[[
|
||||
[3, 0, 1, 2, 7, 4],
|
||||
[1, 5, 8, 9, 3, 1],
|
||||
[2, 7, 2, 5, 1, 3],
|
||||
[0, 1, 3, 1, 7, 8],
|
||||
[4, 2, 1, 6, 2, 8],
|
||||
[2, 4, 5, 2, 3, 9]]]]).astype(np.float32)), [1, 1, 6, 6]), name='x')
|
||||
self.out = Parameter(initializer(Tensor(np.array([[[
|
||||
[-5, -4, 0, 8],
|
||||
[-10, -2, 2, 3],
|
||||
[0, -2, -4, -7],
|
||||
[-3, -2, -3, -16]]]]).astype(np.float32)), [1, 1, 4, 4]), name='y')
|
||||
self.get_shape = P.Shape()
|
||||
|
||||
@ms_function
|
||||
def construct(self):
|
||||
return self.conv_input(self.out, self.w, self.get_shape(self.x))
|
||||
|
||||
|
||||
def test_conv2d_backprop_input():
|
||||
conv2d_input = Net()
|
||||
output = conv2d_input()
|
||||
expect = np.array([[[[-5, -4, 5, 12, 0, -8],
|
||||
[-15, -6, 17, 17, -2, -11],
|
||||
[-15, -8, 13, 12, 2, -4],
|
||||
[-13, -6, 8, -14, 5, 20],
|
||||
[-3, -4, -4, -19, 7, 23],
|
||||
[-3, -2, 0, -14, 3, 16]]]]).astype(np.float32)
|
||||
print(output)
|
||||
assert (output.asnumpy() == expect).all()
|
Loading…
Reference in New Issue