clean pylint

pull/1678/head
jinyaohui 5 years ago
parent a193d0977a
commit 5e43edc474

@ -16,8 +16,8 @@ Some basic function for nlp
"""
from enum import IntEnum
import mindspore._c_dataengine as cde
import numpy as np
import mindspore._c_dataengine as cde
from .validators import check_from_file, check_from_list, check_from_dict

@ -13,15 +13,16 @@
# limitations under the License.
# ============================================================================
"""lstm"""
from mindspore.ops import operations as P
from mindspore.nn.cell import Cell
from mindspore.common.parameter import Parameter, ParameterTuple
from mindspore.common.initializer import initializer
from mindspore._checkparam import Validator as validator
from mindspore import context
import numpy as np
import mindspore.nn as nn
from mindspore import context
from mindspore._checkparam import Validator as validator
from mindspore.common.initializer import initializer
from mindspore.common.parameter import Parameter, ParameterTuple
from mindspore.common.tensor import Tensor
import numpy as np
from mindspore.nn.cell import Cell
from mindspore.ops import operations as P
__all__ = ['LSTM', 'LSTMCell']

@ -18,8 +18,8 @@
from typing import Callable, List, Any
import mindspore._c_expression as _c_expression
import numpy as np
import mindspore._c_expression as _c_expression
from mindspore import ParameterTuple
from mindspore import Tensor

@ -22,9 +22,9 @@ import os
import mindspore.common.dtype as mstype
import mindspore.context as context
from mindspore import Tensor
from model_zoo.bert.src import BertConfig, BertNetworkWithLoss, BertTrainOneStepCell, BertTrainOneStepWithLossScaleCell
from mindspore.nn.optim import AdamWeightDecayDynamicLR
from mindspore.train.loss_scale_manager import DynamicLossScaleManager
from model_zoo.bert.src import BertConfig, BertNetworkWithLoss, BertTrainOneStepCell, BertTrainOneStepWithLossScaleCell
from ...dataset_mock import MindData
from ...ops_common import nn, np, batch_tuple_tensor, build_construct_graph

@ -91,10 +91,10 @@ class Net2(nn.Cell):
self.all_reduce3 = P.AllReduce(self.op2, group=NCCL_WORLD_COMM_GROUP)
def construct(self):
x = self.all_reduce1(self.x1)
y = self.all_reduce2(x)
x_ = self.all_reduce1(self.x1)
y = self.all_reduce2(x_)
z = self.all_reduce3(y)
return (x, y, z)
return (x_, y, z)
def test_AllReduce2():

@ -17,8 +17,11 @@
import os
import time
import pytest
import numpy as np
import pytest
from src.bert_for_pre_training import BertNetworkWithLoss, BertTrainOneStepWithLossScaleCell
from src.bert_model import BertConfig
import mindspore.common.dtype as mstype
import mindspore.dataset.engine.datasets as de
@ -26,8 +29,6 @@ import mindspore.dataset.transforms.c_transforms as C
from mindspore import context
from mindspore import log as logger
from mindspore.common.tensor import Tensor
from src.bert_model import BertConfig
from src.bert_for_pre_training import BertNetworkWithLoss, BertTrainOneStepWithLossScaleCell
from mindspore.nn.optim import Lamb
from mindspore.train.callback import Callback
from mindspore.train.loss_scale_manager import DynamicLossScaleManager

@ -13,8 +13,9 @@
# limitations under the License.
# ============================================================================
import numpy as np
import os
import numpy as np
import pytest
import mindspore.context as context

@ -397,5 +397,5 @@ def test_trainTensor_amp(num_classes=10, epoch=18, batch_size=16):
loss = train_network(data, label)
losses.append(loss)
assert (losses[-1][0].asnumpy() < 1)
assert (losses[-1][1].asnumpy() == False)
assert not losses[-1][1].asnumpy()
assert (losses[-1][2].asnumpy() > 1)

@ -19,9 +19,8 @@ Usage:
python test_network_main.py --net lenet --target Ascend
"""
import argparse
import numpy as np
import os
import time
from models.alexnet import AlexNet
from models.lenet import LeNet
from models.resnetv1_5 import resnet50

@ -35,10 +35,10 @@ def test_clip_by_norm_const():
super(Network, self).__init__()
self.norm_value = Tensor(np.array([1]).astype(np.float32))
self.clip = nn.ClipByNorm()
def construct(self, x):
return self.clip(x, self.norm_value)
net = Network()
x = Tensor(np.array([[-2, 0, 0], [0, 3, 4]]).astype(np.float32))
output = net(x)
net(x)

@ -14,6 +14,7 @@
# ============================================================================
""" test ops """
import functools
import numpy as np
import mindspore.nn as nn
@ -898,7 +899,7 @@ test_case_nn_ops = [
'skip': ['backward']}),
('BasicLSTMCell', {
'block': P.BasicLSTMCell(keep_prob=1.0, forget_bias=1.0, state_is_tuple=True, activation='tanh'),
'desc_inputs': [[128, 128], [128, 128], [128, 128], [512, 256, 1, 1],[512, 1, 1, 1]],
'desc_inputs': [[128, 128], [128, 128], [128, 128], [512, 256, 1, 1], [512, 1, 1, 1]],
'desc_bprop': [[128, 128], [128, 128], [128, 128], [128, 128], [128, 128], [128, 128], [128, 128]],
'skip': []}),
('TopK', {
@ -1324,7 +1325,7 @@ test_case_array_ops = [
'desc_inputs': [(Tensor(np.array([1], np.float32)),
Tensor(np.array([1], np.float32)),
Tensor(np.array([1], np.float32)))],
'desc_bprop': [[3, ]]}),
'desc_bprop': [[3,]]}),
('Pack_0', {
'block': NetForPackInput(P.Pack()),
'desc_inputs': [[2, 2], [2, 2], [2, 2]],
@ -1486,8 +1487,7 @@ test_case = functools.reduce(lambda x, y: x + y, test_case_lists)
test_exec_case = test_case
test_backward_exec_case = filter(lambda x: 'skip' not in x[1] or
'backward' not in x[1]['skip'], test_case)
test_backward_exec_case = filter(lambda x: 'skip' not in x[1] or 'backward' not in x[1]['skip'], test_case)
@non_graph_engine

@ -144,7 +144,7 @@ def test_op_forward_infererror():
input_np = np.random.randn(2, 3, 4, 5).astype(np.float32)
input_me = Tensor(input_np)
net = Net3()
with pytest.raises(TypeError) as e:
with pytest.raises(TypeError):
net(input_me)
@ -162,7 +162,7 @@ def test_sequential_resolve_error():
input_np = np.random.randn(2, 3, 4, 5).astype(np.float32)
input_me = Tensor(input_np)
net = SequenceNet()
with pytest.raises(RuntimeError) as e:
with pytest.raises(RuntimeError):
net(input_me)

@ -19,7 +19,6 @@ import mindspore.nn as nn
from mindspore import Tensor
from mindspore import context
from mindspore import dtype as mstype
from mindspore.common.api import _executor
context.set_context(device_target="Ascend")

@ -44,7 +44,7 @@ class ChooseInitParameterWithInput(nn.Cell):
self.x = Parameter(Tensor(np.ones(2), dtype=mstype.int32), name='x')
@ms_function
def construct(self, input):
def construct(self, input_data):
return self.x
@ -57,7 +57,7 @@ def test_choose_init_param():
def test_choose_param_with_input():
choose = ChooseInitParameterWithInput()
input = Tensor(np.zeros(2), dtype=mstype.int32)
input_data = Tensor(np.zeros(2), dtype=mstype.int32)
expect = Tensor(np.ones(2), dtype=mstype.int32)
out = choose(input)
out = choose(input_data)
assert np.allclose(expect.asnumpy(), out.asnumpy())

@ -1,10 +1,11 @@
import numpy as np
import mindspore.nn as nn
import mindspore.ops.operations as P
from mindspore.ops import composite as C
from mindspore import context, Tensor, ParameterTuple
from mindspore.common.initializer import TruncatedNormal
from mindspore.nn import WithLossCell, Momentum
from mindspore.ops import composite as C
context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU")
@ -45,7 +46,6 @@ class LeNet5(nn.Cell):
Lenet network
Args:
num_class (int): Num classes. Default: 10.
Returns:
Tensor, output tensor

@ -21,7 +21,6 @@ from mindspore import Tensor
from mindspore import context
from mindspore.common.api import ms_function
from mindspore.ops import composite as C
from mindspore.ops import functional as F
from mindspore.ops import operations as P
from ....mindspore_test_framework.utils.bprop_util import bprop
from ....mindspore_test_framework.utils.debug_util import PrintShapeTypeCell, PrintGradShapeTypeCell

@ -19,9 +19,10 @@
@Desc : test parse the object's method
"""
import logging
from dataclasses import dataclass
import numpy as np
import pytest
from dataclasses import dataclass
import mindspore.nn as nn
from mindspore import context

@ -19,9 +19,7 @@ import mindspore.nn as nn
from mindspore import context
from mindspore.common.tensor import Tensor
from mindspore.nn import WithGradCell, WithLossCell
from mindspore.nn.optim import Momentum
from mindspore.ops import operations as P
from mindspore.train.model import Model
from ..ut_filter import non_graph_engine

@ -13,14 +13,9 @@
# limitations under the License.
# ============================================================================
""" tests for quant """
import numpy as np
from mobilenetv2_combined import MobileNetV2
import mindspore.context as context
from mindspore import Tensor
from mindspore import nn
from mindspore.nn.layer import combined
from mindspore.train.quant import quant as qat
context.set_context(mode=context.GRAPH_MODE, device_target="GPU")

@ -74,7 +74,7 @@ class LossNet(nn.Cell):
return out
def test_Model_Checkpoint_prefix_invalid():
def test_model_checkpoint_prefix_invalid():
"""Test ModelCheckpoint prefix invalid."""
with pytest.raises(ValueError):
ModelCheckpoint(123)
@ -338,9 +338,9 @@ def test_step_end_save_graph():
ckpoint_cb.begin(run_context)
# import pdb;pdb.set_trace()
ckpoint_cb.step_end(run_context)
assert os.path.exists('./test_files/test-graph.meta') == True
assert os.path.exists('./test_files/test-graph.meta')
if os.path.exists('./test_files/test-graph.meta'):
os.chmod('./test_files/test-graph.meta', stat.S_IWRITE)
os.remove('./test_files/test-graph.meta')
ckpoint_cb.step_end(run_context)
assert os.path.exists('./test_files/test-graph.meta') == False
assert not os.path.exists('./test_files/test-graph.meta')

@ -71,16 +71,16 @@ def setup_module():
def test_save_graph():
""" test_exec_save_graph """
class Net(nn.Cell):
class Net1(nn.Cell):
def __init__(self):
super(Net, self).__init__()
super(Net1, self).__init__()
self.add = P.TensorAdd()
def construct(self, x, y):
z = self.add(x, y)
return z
net = Net()
net = Net1()
net.set_train()
out_me_list = []
x = Tensor(np.random.rand(2, 1, 2, 3).astype(np.float32))

Loading…
Cancel
Save