clean pylint

This commit is contained in:
jinyaohui 2020-06-04 19:53:40 +08:00
parent a193d0977a
commit 5e43edc474
21 changed files with 44 additions and 50 deletions

View File

@ -16,8 +16,8 @@ Some basic function for nlp
"""
from enum import IntEnum
import mindspore._c_dataengine as cde
import numpy as np
import mindspore._c_dataengine as cde
from .validators import check_from_file, check_from_list, check_from_dict

View File

@ -13,16 +13,17 @@
# limitations under the License.
# ============================================================================
"""lstm"""
from mindspore.ops import operations as P
from mindspore.nn.cell import Cell
from mindspore.common.parameter import Parameter, ParameterTuple
from mindspore.common.initializer import initializer
from mindspore._checkparam import Validator as validator
from mindspore import context
import mindspore.nn as nn
from mindspore.common.tensor import Tensor
import numpy as np
import mindspore.nn as nn
from mindspore import context
from mindspore._checkparam import Validator as validator
from mindspore.common.initializer import initializer
from mindspore.common.parameter import Parameter, ParameterTuple
from mindspore.common.tensor import Tensor
from mindspore.nn.cell import Cell
from mindspore.ops import operations as P
__all__ = ['LSTM', 'LSTMCell']

View File

@ -18,8 +18,8 @@
from typing import Callable, List, Any
import mindspore._c_expression as _c_expression
import numpy as np
import mindspore._c_expression as _c_expression
from mindspore import ParameterTuple
from mindspore import Tensor

View File

@ -22,9 +22,9 @@ import os
import mindspore.common.dtype as mstype
import mindspore.context as context
from mindspore import Tensor
from model_zoo.bert.src import BertConfig, BertNetworkWithLoss, BertTrainOneStepCell, BertTrainOneStepWithLossScaleCell
from mindspore.nn.optim import AdamWeightDecayDynamicLR
from mindspore.train.loss_scale_manager import DynamicLossScaleManager
from model_zoo.bert.src import BertConfig, BertNetworkWithLoss, BertTrainOneStepCell, BertTrainOneStepWithLossScaleCell
from ...dataset_mock import MindData
from ...ops_common import nn, np, batch_tuple_tensor, build_construct_graph

View File

@ -91,10 +91,10 @@ class Net2(nn.Cell):
self.all_reduce3 = P.AllReduce(self.op2, group=NCCL_WORLD_COMM_GROUP)
def construct(self):
x = self.all_reduce1(self.x1)
y = self.all_reduce2(x)
x_ = self.all_reduce1(self.x1)
y = self.all_reduce2(x_)
z = self.all_reduce3(y)
return (x, y, z)
return (x_, y, z)
def test_AllReduce2():

View File

@ -17,8 +17,11 @@
import os
import time
import pytest
import numpy as np
import pytest
from src.bert_for_pre_training import BertNetworkWithLoss, BertTrainOneStepWithLossScaleCell
from src.bert_model import BertConfig
import mindspore.common.dtype as mstype
import mindspore.dataset.engine.datasets as de
@ -26,8 +29,6 @@ import mindspore.dataset.transforms.c_transforms as C
from mindspore import context
from mindspore import log as logger
from mindspore.common.tensor import Tensor
from src.bert_model import BertConfig
from src.bert_for_pre_training import BertNetworkWithLoss, BertTrainOneStepWithLossScaleCell
from mindspore.nn.optim import Lamb
from mindspore.train.callback import Callback
from mindspore.train.loss_scale_manager import DynamicLossScaleManager

View File

@ -13,8 +13,9 @@
# limitations under the License.
# ============================================================================
import numpy as np
import os
import numpy as np
import pytest
import mindspore.context as context

View File

@ -397,5 +397,5 @@ def test_trainTensor_amp(num_classes=10, epoch=18, batch_size=16):
loss = train_network(data, label)
losses.append(loss)
assert (losses[-1][0].asnumpy() < 1)
assert (losses[-1][1].asnumpy() == False)
assert not losses[-1][1].asnumpy()
assert (losses[-1][2].asnumpy() > 1)

View File

@ -19,9 +19,8 @@ Usage:
python test_network_main.py --net lenet --target Ascend
"""
import argparse
import numpy as np
import os
import time
from models.alexnet import AlexNet
from models.lenet import LeNet
from models.resnetv1_5 import resnet50

View File

@ -41,4 +41,4 @@ def test_clip_by_norm_const():
net = Network()
x = Tensor(np.array([[-2, 0, 0], [0, 3, 4]]).astype(np.float32))
output = net(x)
net(x)

View File

@ -14,6 +14,7 @@
# ============================================================================
""" test ops """
import functools
import numpy as np
import mindspore.nn as nn
@ -898,7 +899,7 @@ test_case_nn_ops = [
'skip': ['backward']}),
('BasicLSTMCell', {
'block': P.BasicLSTMCell(keep_prob=1.0, forget_bias=1.0, state_is_tuple=True, activation='tanh'),
'desc_inputs': [[128, 128], [128, 128], [128, 128], [512, 256, 1, 1],[512, 1, 1, 1]],
'desc_inputs': [[128, 128], [128, 128], [128, 128], [512, 256, 1, 1], [512, 1, 1, 1]],
'desc_bprop': [[128, 128], [128, 128], [128, 128], [128, 128], [128, 128], [128, 128], [128, 128]],
'skip': []}),
('TopK', {
@ -1324,7 +1325,7 @@ test_case_array_ops = [
'desc_inputs': [(Tensor(np.array([1], np.float32)),
Tensor(np.array([1], np.float32)),
Tensor(np.array([1], np.float32)))],
'desc_bprop': [[3, ]]}),
'desc_bprop': [[3,]]}),
('Pack_0', {
'block': NetForPackInput(P.Pack()),
'desc_inputs': [[2, 2], [2, 2], [2, 2]],
@ -1486,8 +1487,7 @@ test_case = functools.reduce(lambda x, y: x + y, test_case_lists)
test_exec_case = test_case
test_backward_exec_case = filter(lambda x: 'skip' not in x[1] or
'backward' not in x[1]['skip'], test_case)
test_backward_exec_case = filter(lambda x: 'skip' not in x[1] or 'backward' not in x[1]['skip'], test_case)
@non_graph_engine

View File

@ -144,7 +144,7 @@ def test_op_forward_infererror():
input_np = np.random.randn(2, 3, 4, 5).astype(np.float32)
input_me = Tensor(input_np)
net = Net3()
with pytest.raises(TypeError) as e:
with pytest.raises(TypeError):
net(input_me)
@ -162,7 +162,7 @@ def test_sequential_resolve_error():
input_np = np.random.randn(2, 3, 4, 5).astype(np.float32)
input_me = Tensor(input_np)
net = SequenceNet()
with pytest.raises(RuntimeError) as e:
with pytest.raises(RuntimeError):
net(input_me)

View File

@ -19,7 +19,6 @@ import mindspore.nn as nn
from mindspore import Tensor
from mindspore import context
from mindspore import dtype as mstype
from mindspore.common.api import _executor
context.set_context(device_target="Ascend")

View File

@ -44,7 +44,7 @@ class ChooseInitParameterWithInput(nn.Cell):
self.x = Parameter(Tensor(np.ones(2), dtype=mstype.int32), name='x')
@ms_function
def construct(self, input):
def construct(self, input_data):
return self.x
@ -57,7 +57,7 @@ def test_choose_init_param():
def test_choose_param_with_input():
choose = ChooseInitParameterWithInput()
input = Tensor(np.zeros(2), dtype=mstype.int32)
input_data = Tensor(np.zeros(2), dtype=mstype.int32)
expect = Tensor(np.ones(2), dtype=mstype.int32)
out = choose(input)
out = choose(input_data)
assert np.allclose(expect.asnumpy(), out.asnumpy())

View File

@ -1,10 +1,11 @@
import numpy as np
import mindspore.nn as nn
import mindspore.ops.operations as P
from mindspore.ops import composite as C
from mindspore import context, Tensor, ParameterTuple
from mindspore.common.initializer import TruncatedNormal
from mindspore.nn import WithLossCell, Momentum
from mindspore.ops import composite as C
context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU")
@ -45,7 +46,6 @@ class LeNet5(nn.Cell):
Lenet network
Args:
num_class (int): Num classes. Default: 10.
Returns:
Tensor, output tensor

View File

@ -21,7 +21,6 @@ from mindspore import Tensor
from mindspore import context
from mindspore.common.api import ms_function
from mindspore.ops import composite as C
from mindspore.ops import functional as F
from mindspore.ops import operations as P
from ....mindspore_test_framework.utils.bprop_util import bprop
from ....mindspore_test_framework.utils.debug_util import PrintShapeTypeCell, PrintGradShapeTypeCell

View File

@ -19,9 +19,10 @@
@Desc : test parse the object's method
"""
import logging
from dataclasses import dataclass
import numpy as np
import pytest
from dataclasses import dataclass
import mindspore.nn as nn
from mindspore import context

View File

@ -19,9 +19,7 @@ import mindspore.nn as nn
from mindspore import context
from mindspore.common.tensor import Tensor
from mindspore.nn import WithGradCell, WithLossCell
from mindspore.nn.optim import Momentum
from mindspore.ops import operations as P
from mindspore.train.model import Model
from ..ut_filter import non_graph_engine

View File

@ -13,14 +13,9 @@
# limitations under the License.
# ============================================================================
""" tests for quant """
import numpy as np
from mobilenetv2_combined import MobileNetV2
import mindspore.context as context
from mindspore import Tensor
from mindspore import nn
from mindspore.nn.layer import combined
from mindspore.train.quant import quant as qat
context.set_context(mode=context.GRAPH_MODE, device_target="GPU")

View File

@ -74,7 +74,7 @@ class LossNet(nn.Cell):
return out
def test_Model_Checkpoint_prefix_invalid():
def test_model_checkpoint_prefix_invalid():
"""Test ModelCheckpoint prefix invalid."""
with pytest.raises(ValueError):
ModelCheckpoint(123)
@ -338,9 +338,9 @@ def test_step_end_save_graph():
ckpoint_cb.begin(run_context)
# import pdb;pdb.set_trace()
ckpoint_cb.step_end(run_context)
assert os.path.exists('./test_files/test-graph.meta') == True
assert os.path.exists('./test_files/test-graph.meta')
if os.path.exists('./test_files/test-graph.meta'):
os.chmod('./test_files/test-graph.meta', stat.S_IWRITE)
os.remove('./test_files/test-graph.meta')
ckpoint_cb.step_end(run_context)
assert os.path.exists('./test_files/test-graph.meta') == False
assert not os.path.exists('./test_files/test-graph.meta')

View File

@ -71,16 +71,16 @@ def setup_module():
def test_save_graph():
""" test_exec_save_graph """
class Net(nn.Cell):
class Net1(nn.Cell):
def __init__(self):
super(Net, self).__init__()
super(Net1, self).__init__()
self.add = P.TensorAdd()
def construct(self, x, y):
z = self.add(x, y)
return z
net = Net()
net = Net1()
net.set_train()
out_me_list = []
x = Tensor(np.random.rand(2, 1, 2, 3).astype(np.float32))