From a0ff742442b28a39061a84736a3fa2d9e047584a Mon Sep 17 00:00:00 2001 From: yanghaoran Date: Fri, 27 Aug 2021 19:59:59 +0800 Subject: [PATCH] prune less effective Ascend testcases to level1 --- tests/st/auto_monad/test_auto_monad.py | 28 +++++++++---------- .../auto_monad/test_auto_monad_mindtester.py | 2 +- tests/st/auto_monad/test_effect_ops.py | 4 +-- tests/st/auto_monad/test_effect_optimizer.py | 20 ++++++------- tests/st/auto_monad/test_effect_random.py | 12 ++++---- tests/st/auto_monad/test_float_overflow.py | 2 +- tests/st/control/test_ascend_control_sink.py | 10 +++---- tests/st/control/test_cont_grad.py | 10 +++---- tests/st/control/test_multigraph_sink.py | 14 +++++----- tests/st/control/test_switch_layer.py | 2 +- tests/st/control/test_tensor_in_list.py | 2 +- tests/st/control/test_while_grad.py | 2 +- tests/st/dynamic_shape/test_ftrl.py | 2 +- .../_attribution/test_localization.py | 8 +++--- .../test_heterogeneous_excutor.py | 2 +- tests/st/high_grad/test_highgrad_param.py | 2 +- tests/st/mix_precision/test_mix_precision.py | 2 +- tests/st/numpy_native/test_math_ops.py | 4 +-- .../ops/ascend/test_conv2d_depthwiseconv2d.py | 4 +-- tests/st/ops/ascend/test_dynamic_gru_v2.py | 2 +- tests/st/ops/custom_ops_tbe/test_square.py | 2 +- tests/st/ops/graph_kernel/test_layernorm.py | 2 +- tests/st/ops/graph_kernel/test_reorder_ops.py | 2 +- tests/st/ops/graph_kernel/test_softmax.py | 2 +- .../test_entry_cmp_sparse_embedding.py | 2 +- .../st/pynative/loss_scale/test_loss_scale.py | 2 +- .../pynative/parser/test_parser_construct.py | 2 +- .../parser/test_parser_tensor_assign.py | 2 +- .../st/pynative/test_graph_param_transform.py | 4 +-- tests/st/pynative/test_pynative_hook.py | 8 +++--- tests/st/pynative/test_pynative_hook_grad.py | 12 ++++---- ...ive_layernorm_input_and_argmaxwithvalue.py | 2 +- tests/st/pynative/test_pynative_lenet.py | 4 +-- .../test_pynative_mixed_precision_cells.py | 4 +-- .../pynative/test_pynative_temporary_cell.py | 2 +- tests/st/pynative/test_tensor_augassign.py | 6 ++-- tests/st/pynative/test_tensor_getitem.py | 16 +++++------ tests/st/pynative/test_tensor_setitem.py | 18 ++++++------ 38 files changed, 113 insertions(+), 113 deletions(-) diff --git a/tests/st/auto_monad/test_auto_monad.py b/tests/st/auto_monad/test_auto_monad.py index 9acf136eeb0..e5294cb460c 100644 --- a/tests/st/auto_monad/test_auto_monad.py +++ b/tests/st/auto_monad/test_auto_monad.py @@ -320,7 +320,7 @@ def test_assign(): np.testing.assert_array_equal(out.asnumpy(), expect.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -341,7 +341,7 @@ def test_assign_implicit(): assert out.dtype == ms.float32 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -370,7 +370,7 @@ def test_assign_write_after_read(): np.testing.assert_array_equal(out.asnumpy(), expect.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -424,7 +424,7 @@ def test_assign_if(): np.testing.assert_array_equal(out.asnumpy(), expect.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -450,7 +450,7 @@ def test_if(): np.testing.assert_array_equal(out.asnumpy(), expect.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -471,7 +471,7 @@ def test_while(): np.testing.assert_array_equal(out.asnumpy(), expect.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -497,7 +497,7 @@ def test_assign_while(): np.testing.assert_array_equal(out.asnumpy(), expect.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -597,7 +597,7 @@ def test_print_assign_for(): check_output(cap.output, patterns) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -684,7 +684,7 @@ def test_if_lambda(): np.testing.assert_array_equal(out.asnumpy(), expect.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -712,7 +712,7 @@ def test_multi_assign(): np.testing.assert_array_equal(out.asnumpy(), expect.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -771,7 +771,7 @@ def test_multi_assign_print(): net.para2.data.asnumpy(), expect_para2.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -808,7 +808,7 @@ def test_matmul_assign_biasadd(): context.set_context(mode=context.GRAPH_MODE) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -874,7 +874,7 @@ def test_isolate_call(): np.testing.assert_array_equal(out.asnumpy(), expect.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -1370,7 +1370,7 @@ def test_auto_mixed_precision_controlflow_auto(pynative_save_graphs): # op_cast should be located in order_list after abstract_specialize. # Besides Ascend, it can work on CPU. -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/auto_monad/test_auto_monad_mindtester.py b/tests/st/auto_monad/test_auto_monad_mindtester.py index a5a4857d98c..f2c6afa134b 100644 --- a/tests/st/auto_monad/test_auto_monad_mindtester.py +++ b/tests/st/auto_monad/test_auto_monad_mindtester.py @@ -411,7 +411,7 @@ class SideEffectAssignAddnReluReturnParNet(Cell): return grad_out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/auto_monad/test_effect_ops.py b/tests/st/auto_monad/test_effect_ops.py index 6cbb13b3826..7029785d685 100644 --- a/tests/st/auto_monad/test_effect_ops.py +++ b/tests/st/auto_monad/test_effect_ops.py @@ -312,7 +312,7 @@ class ScatterNdUpdateNet(nn.Cell): return self.input_x -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -337,7 +337,7 @@ class ScatterNonAliasingAddNet(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/auto_monad/test_effect_optimizer.py b/tests/st/auto_monad/test_effect_optimizer.py index 4f763099367..7b53d753857 100644 --- a/tests/st/auto_monad/test_effect_optimizer.py +++ b/tests/st/auto_monad/test_effect_optimizer.py @@ -37,7 +37,7 @@ class AdamNet(nn.Cell): return self.var, self.m, self.v -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -264,7 +264,7 @@ class ApplyFtrlNet(nn.Cell): return self.var, self.accum, self.linear -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -321,7 +321,7 @@ class ApplyMomentumNet(nn.Cell): return self.var, self.accum -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -351,7 +351,7 @@ class ApplyPowerSignNet(nn.Cell): return self.var, self.m -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -439,7 +439,7 @@ class ApplyRMSPropNet(nn.Cell): return self.var -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -508,7 +508,7 @@ class FusedSparseFtrlNet(nn.Cell): return self.var, self.accum, self.linear -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -576,7 +576,7 @@ class FusedSparseProximalAdagradNet(nn.Cell): return self.var, self.accum -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -667,7 +667,7 @@ class SparseApplyFtrlNet(nn.Cell): return self.var, self.accum, self.linear -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -699,7 +699,7 @@ class SparseApplyFtrlV2Net(nn.Cell): return self.var, self.accum, self.linear -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -729,7 +729,7 @@ class SparseApplyProximalAdagradNet(nn.Cell): return self.var, self.accum -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/auto_monad/test_effect_random.py b/tests/st/auto_monad/test_effect_random.py index 7d594c10361..f79124affbe 100644 --- a/tests/st/auto_monad/test_effect_random.py +++ b/tests/st/auto_monad/test_effect_random.py @@ -42,7 +42,7 @@ class Sampling(nn.Cell): return s1, s2, s3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -68,7 +68,7 @@ class CompositeNormalNet(nn.Cell): return s1, s2, s3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -95,7 +95,7 @@ class CompositeLaplaceNet(nn.Cell): return s1, s2, s3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -122,7 +122,7 @@ class CompositeGammaNet(nn.Cell): return s1, s2, s3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -175,7 +175,7 @@ class CompositeUniformNet(nn.Cell): return s1, s2, s3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -231,7 +231,7 @@ class StandardLaplaceNet(nn.Cell): return s1, s2, s3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/auto_monad/test_float_overflow.py b/tests/st/auto_monad/test_float_overflow.py index 374cf3dffaf..b41e8bb0b79 100644 --- a/tests/st/auto_monad/test_float_overflow.py +++ b/tests/st/auto_monad/test_float_overflow.py @@ -59,7 +59,7 @@ class NpuFloatNet(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/control/test_ascend_control_sink.py b/tests/st/control/test_ascend_control_sink.py index 8c6bf19aaa0..ec8e98bb487 100644 --- a/tests/st/control/test_ascend_control_sink.py +++ b/tests/st/control/test_ascend_control_sink.py @@ -189,7 +189,7 @@ class NotOperation(nn.Cell): return not x_sum -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -207,7 +207,7 @@ def test_simple_if(): assert np.allclose(expect, output.asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -223,7 +223,7 @@ def test_simple_if_with_assign(): assert np.allclose(expect, output.asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -237,7 +237,7 @@ def test_if_in_if(): assert np.allclose(expect, output.asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -271,7 +271,7 @@ def test_simple_while(): assert np.allclose(expect, output.asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/control/test_cont_grad.py b/tests/st/control/test_cont_grad.py index ed7a8aba3a9..0fa3fc0bba6 100644 --- a/tests/st/control/test_cont_grad.py +++ b/tests/st/control/test_cont_grad.py @@ -148,7 +148,7 @@ def test_while_with_variable_grad(): assert np.allclose(graph_output[1].asnumpy(), expect_two, 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -217,7 +217,7 @@ def test_while_endless_case(): assert np.allclose(graph_output.asnumpy(), pynative_output.asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -298,7 +298,7 @@ def test_while_with_param_forward_with_const_branch(): assert np.allclose(graph_output.asnumpy(), pynative_output.asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -626,7 +626,7 @@ def test_while_with_param_basic_grad(): assert np.allclose(graph_output[0].asnumpy(), pynative_output[0].asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -1034,7 +1034,7 @@ def test_with_param_if_by_if_grad_param_excute_null(): assert np.allclose(graph_output[0].asnumpy(), pynative_output[0].asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard diff --git a/tests/st/control/test_multigraph_sink.py b/tests/st/control/test_multigraph_sink.py index ba0f45b06cf..8f67198ba30 100644 --- a/tests/st/control/test_multigraph_sink.py +++ b/tests/st/control/test_multigraph_sink.py @@ -132,7 +132,7 @@ def while_in_while_in_while(x, y, z): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @pytest.mark.env_onecard @@ -148,7 +148,7 @@ def test_if_by_if(): assert output == expect -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @pytest.mark.env_onecard @@ -158,7 +158,7 @@ def test_if_in_if(): assert output == expect -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @pytest.mark.env_onecard @@ -168,7 +168,7 @@ def test_simple_while(): assert output == expect -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @pytest.mark.env_onecard @@ -178,7 +178,7 @@ def test_while_by_while(): assert output == expect -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @pytest.mark.env_onecard @@ -188,7 +188,7 @@ def test_while_in_while(): assert output == expect -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @pytest.mark.env_onecard @@ -198,7 +198,7 @@ def test_while_by_while_in_while(): assert output == expect -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/control/test_switch_layer.py b/tests/st/control/test_switch_layer.py index 3f33b94b829..607421909db 100644 --- a/tests/st/control/test_switch_layer.py +++ b/tests/st/control/test_switch_layer.py @@ -37,7 +37,7 @@ class CaseNet(nn.Cell): return x -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/control/test_tensor_in_list.py b/tests/st/control/test_tensor_in_list.py index e1d423e719a..f769e73a2b5 100644 --- a/tests/st/control/test_tensor_in_list.py +++ b/tests/st/control/test_tensor_in_list.py @@ -35,7 +35,7 @@ class Net(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/control/test_while_grad.py b/tests/st/control/test_while_grad.py index 870d46a1efa..6d667cdc545 100644 --- a/tests/st/control/test_while_grad.py +++ b/tests/st/control/test_while_grad.py @@ -37,7 +37,7 @@ class GradNet(nn.Cell): return gradient_function(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/dynamic_shape/test_ftrl.py b/tests/st/dynamic_shape/test_ftrl.py index bc1ce5b5143..36cda79cf5d 100644 --- a/tests/st/dynamic_shape/test_ftrl.py +++ b/tests/st/dynamic_shape/test_ftrl.py @@ -74,7 +74,7 @@ def test_lazy_adam_net(): np.allclose(net.weight2.asnumpy(), np.array([[[0.9, 0.9]], [[0.9, 0.9]], [[0.9, 0.9]]])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/explainer/benchmark/_attribution/test_localization.py b/tests/st/explainer/benchmark/_attribution/test_localization.py index d456c1132f6..cd76530a3dd 100644 --- a/tests/st/explainer/benchmark/_attribution/test_localization.py +++ b/tests/st/explainer/benchmark/_attribution/test_localization.py @@ -64,7 +64,7 @@ class TestLocalization: self.saliency_gt = mock_gradient_call(self.explainer, self.data, self.target) self.num_class = 2 - @pytest.mark.level0 + @pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -81,7 +81,7 @@ class TestLocalization: res = pg.evaluate(self.explainer, self.data, targets=self.target, mask=self.masks) assert np.max(np.abs(np.array([res_gt]) - res)) < 1e-5 - @pytest.mark.level0 + @pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -100,7 +100,7 @@ class TestLocalization: assert np.allclose(np.array([res_gt]), res) - @pytest.mark.level0 + @pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -116,7 +116,7 @@ class TestLocalization: res = pg.evaluate(self.explainer, self.data, targets=self.target, mask=self.masks, saliency=self.saliency_gt) assert np.allclose(np.array([res_gt]), res) - @pytest.mark.level0 + @pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/heterogeneous_excutor/test_heterogeneous_excutor.py b/tests/st/heterogeneous_excutor/test_heterogeneous_excutor.py index e267525afcf..fbe11c4949c 100644 --- a/tests/st/heterogeneous_excutor/test_heterogeneous_excutor.py +++ b/tests/st/heterogeneous_excutor/test_heterogeneous_excutor.py @@ -57,7 +57,7 @@ class Net(nn.Cell): return x -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/high_grad/test_highgrad_param.py b/tests/st/high_grad/test_highgrad_param.py index a91f87c40d9..2add565f2dd 100644 --- a/tests/st/high_grad/test_highgrad_param.py +++ b/tests/st/high_grad/test_highgrad_param.py @@ -59,7 +59,7 @@ class GradSec(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/mix_precision/test_mix_precision.py b/tests/st/mix_precision/test_mix_precision.py index 302c90554f7..cb768bccf2f 100644 --- a/tests/st/mix_precision/test_mix_precision.py +++ b/tests/st/mix_precision/test_mix_precision.py @@ -105,7 +105,7 @@ def test_sit_auto_mix_precision_train_o3(): assert np.allclose(out.asnumpy(), out_pynative.asnumpy(), 0.001, 0.001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/numpy_native/test_math_ops.py b/tests/st/numpy_native/test_math_ops.py index 743e0cfda45..2c66bfed00c 100644 --- a/tests/st/numpy_native/test_math_ops.py +++ b/tests/st/numpy_native/test_math_ops.py @@ -1382,7 +1382,7 @@ def test_negative(): match_array(mnp_neg.asnumpy(), onp_neg, 1e-5) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -2132,7 +2132,7 @@ def test_digitize(): match_res(mnp.digitize, onp.digitize, x, bins, right=True) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/ascend/test_conv2d_depthwiseconv2d.py b/tests/st/ops/ascend/test_conv2d_depthwiseconv2d.py index 59bc586877b..8df149e9680 100644 --- a/tests/st/ops/ascend/test_conv2d_depthwiseconv2d.py +++ b/tests/st/ops/ascend/test_conv2d_depthwiseconv2d.py @@ -25,7 +25,7 @@ from mindspore import Tensor context.set_context(mode=context.GRAPH_MODE, device_target="Ascend") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @pytest.mark.env_onecard @@ -36,7 +36,7 @@ def test_conv2d_depthwiseconv2d_str(): assert output.shape == (3, 128, 32, 28) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/ops/ascend/test_dynamic_gru_v2.py b/tests/st/ops/ascend/test_dynamic_gru_v2.py index 67eef756f83..44eba6b0d89 100644 --- a/tests/st/ops/ascend/test_dynamic_gru_v2.py +++ b/tests/st/ops/ascend/test_dynamic_gru_v2.py @@ -32,7 +32,7 @@ class DynamicGRUV2(nn.Cell): return self.dynamic_gru(x, weight_i, weight_h, bias_i, bias_h, None, init_h) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/ops/custom_ops_tbe/test_square.py b/tests/st/ops/custom_ops_tbe/test_square.py index 5e9c4d8535b..0290a67969c 100644 --- a/tests/st/ops/custom_ops_tbe/test_square.py +++ b/tests/st/ops/custom_ops_tbe/test_square.py @@ -48,7 +48,7 @@ def test_net(): expect = np.array([1.0, 16.0, 81.0]).astype(np.float32) assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/ops/graph_kernel/test_layernorm.py b/tests/st/ops/graph_kernel/test_layernorm.py index aea6a510295..d9ac47f55d8 100644 --- a/tests/st/ops/graph_kernel/test_layernorm.py +++ b/tests/st/ops/graph_kernel/test_layernorm.py @@ -147,7 +147,7 @@ def test_layernorm_grad_gpu(): test_layernorm_grad([4, 32, 32], np.float32, -1, -1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/ops/graph_kernel/test_reorder_ops.py b/tests/st/ops/graph_kernel/test_reorder_ops.py index 6652a8ee0e1..280b3b90804 100644 --- a/tests/st/ops/graph_kernel/test_reorder_ops.py +++ b/tests/st/ops/graph_kernel/test_reorder_ops.py @@ -106,7 +106,7 @@ def test_castdown_gpu(): test_castdown() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/ops/graph_kernel/test_softmax.py b/tests/st/ops/graph_kernel/test_softmax.py index e7eea225f88..dbf6cdb5437 100644 --- a/tests/st/ops/graph_kernel/test_softmax.py +++ b/tests/st/ops/graph_kernel/test_softmax.py @@ -55,7 +55,7 @@ def test_softmax_gpu(): context.set_context(mode=context.GRAPH_MODE, device_target="GPU") test_softmax([4, 32, 48], np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/ps/cmp_sparse_embedding/test_entry_cmp_sparse_embedding.py b/tests/st/ps/cmp_sparse_embedding/test_entry_cmp_sparse_embedding.py index bc400c963c1..cfa45ebdb0c 100644 --- a/tests/st/ps/cmp_sparse_embedding/test_entry_cmp_sparse_embedding.py +++ b/tests/st/ps/cmp_sparse_embedding/test_entry_cmp_sparse_embedding.py @@ -16,7 +16,7 @@ import os import pytest -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/pynative/loss_scale/test_loss_scale.py b/tests/st/pynative/loss_scale/test_loss_scale.py index 3cbbaa819e7..e0e4fe0bc63 100644 --- a/tests/st/pynative/loss_scale/test_loss_scale.py +++ b/tests/st/pynative/loss_scale/test_loss_scale.py @@ -232,7 +232,7 @@ def test_loss_scale_fp16_opt_rmsprop_overflow(): assert output_1[0].asnumpy() == output_2[0].asnumpy() assert output_1[1].asnumpy() == output_2[1].asnumpy() == True -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/pynative/parser/test_parser_construct.py b/tests/st/pynative/parser/test_parser_construct.py index 1ecd1e86990..f09fe4323a7 100644 --- a/tests/st/pynative/parser/test_parser_construct.py +++ b/tests/st/pynative/parser/test_parser_construct.py @@ -25,7 +25,7 @@ from mindspore.common.parameter import Parameter def setup_module(): context.set_context(mode=context.PYNATIVE_MODE, device_target="Ascend") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/pynative/parser/test_parser_tensor_assign.py b/tests/st/pynative/parser/test_parser_tensor_assign.py index deba67e6fa4..1a4f74d96cf 100644 --- a/tests/st/pynative/parser/test_parser_tensor_assign.py +++ b/tests/st/pynative/parser/test_parser_tensor_assign.py @@ -25,7 +25,7 @@ from mindspore.ops import operations as P def setup_module(): context.set_context(mode=context.PYNATIVE_MODE, device_target="Ascend") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/pynative/test_graph_param_transform.py b/tests/st/pynative/test_graph_param_transform.py index a6975fafefa..94c48701fa0 100644 --- a/tests/st/pynative/test_graph_param_transform.py +++ b/tests/st/pynative/test_graph_param_transform.py @@ -72,7 +72,7 @@ class GradOfAllInputs(_Grad): network=network, real_inputs_count=real_inputs_count) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -120,7 +120,7 @@ def test_row_tensor_in_while(): assert dense_shape == out[2] -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/pynative/test_pynative_hook.py b/tests/st/pynative/test_pynative_hook.py index 11ead25f903..62804afde27 100644 --- a/tests/st/pynative/test_pynative_hook.py +++ b/tests/st/pynative/test_pynative_hook.py @@ -161,7 +161,7 @@ class Ms_Cell_Change_Shape(nn.Cell): return dout -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -185,7 +185,7 @@ def test_pynative_lenet_train_hook_function_print_and_save_grad(): assert success -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -198,7 +198,7 @@ def test_pynative_custom_bprop_and_Cell_MulAdd(): (Tensor(1.0, mstype.float32), Tensor(2.0, mstype.float32)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -211,7 +211,7 @@ def test_pynative_custom_bprop_and_Cell_Ms_Cell_Change_Shape(): assert "Shapes of input and parameter are different, input index" in str(ex.value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/pynative/test_pynative_hook_grad.py b/tests/st/pynative/test_pynative_hook_grad.py index 75a2a6fcec7..d240ccf1226 100644 --- a/tests/st/pynative/test_pynative_hook_grad.py +++ b/tests/st/pynative/test_pynative_hook_grad.py @@ -406,7 +406,7 @@ def pynative_hook_child_cell_record_grad(): if ms_net.grad_output_list or ms_net.grad_input_list: assert False -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -421,7 +421,7 @@ def test_pynative_hook_diff_hook_gpu(): context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") pynative_hook_diff_hook() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -466,7 +466,7 @@ def test_pynative_hook_mul_change_input_grad_gpu(): context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") pynative_hook_mul_change_input_grad() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -496,7 +496,7 @@ def test_pynative_hook_outermost_cell_change_grad_gpu(): context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") pynative_hook_outermost_cell_change_grad() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -511,7 +511,7 @@ def test_pynative_hook_outermost_cell_record_grad_gpu(): context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") pynative_hook_outermost_cell_record_grad() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -526,7 +526,7 @@ def test_pynative_hook_bprop_outermost_cell_record_grad_gpu(): context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") pynative_hook_bprop_outermost_cell_record_grad() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/pynative/test_pynative_layernorm_input_and_argmaxwithvalue.py b/tests/st/pynative/test_pynative_layernorm_input_and_argmaxwithvalue.py index 0c506d3b47b..c156ced4f4b 100644 --- a/tests/st/pynative/test_pynative_layernorm_input_and_argmaxwithvalue.py +++ b/tests/st/pynative/test_pynative_layernorm_input_and_argmaxwithvalue.py @@ -218,7 +218,7 @@ def argmaxwithvalue_input(): fact.forward_cmp() fact.grad_cmp() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/pynative/test_pynative_lenet.py b/tests/st/pynative/test_pynative_lenet.py index cc665d697e0..b2b40bbc788 100644 --- a/tests/st/pynative/test_pynative_lenet.py +++ b/tests/st/pynative/test_pynative_lenet.py @@ -128,7 +128,7 @@ class GradWrap(nn.Cell): return grad_by_list(self.network, weights)(x, label) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_cpu @@ -166,7 +166,7 @@ def test_ascend_pynative_lenet(): assert loss_output.asnumpy() > 0.003 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_cpu diff --git a/tests/st/pynative/test_pynative_mixed_precision_cells.py b/tests/st/pynative/test_pynative_mixed_precision_cells.py index f80312c1ecc..8c42dd24b13 100644 --- a/tests/st/pynative/test_pynative_mixed_precision_cells.py +++ b/tests/st/pynative/test_pynative_mixed_precision_cells.py @@ -161,7 +161,7 @@ def test_mixed_precision_multiples_cell_gpu_01(): context.set_context(device_target="GPU") mixed_precision_multiples_cell_01() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -176,7 +176,7 @@ def test_mixed_precision_multiples_cell_gpu_02(): context.set_context(device_target="GPU") mixed_precision_multiples_cell_02() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/pynative/test_pynative_temporary_cell.py b/tests/st/pynative/test_pynative_temporary_cell.py index 817c23daa30..9f9a39e23b9 100644 --- a/tests/st/pynative/test_pynative_temporary_cell.py +++ b/tests/st/pynative/test_pynative_temporary_cell.py @@ -33,7 +33,7 @@ class GradofParams(nn.Cell): out = self.grad(self.net, self.params)(*x) return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/pynative/test_tensor_augassign.py b/tests/st/pynative/test_tensor_augassign.py index 19887c61591..75b8837e243 100644 --- a/tests/st/pynative/test_tensor_augassign.py +++ b/tests/st/pynative/test_tensor_augassign.py @@ -25,7 +25,7 @@ def setup_module(): # GPU: does not supported op "FloorMod" -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -73,7 +73,7 @@ def test_tesnsor_augassign_by_slice(): # GPU: does not supported op "FloorMod" -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard @@ -134,7 +134,7 @@ def test_tesnsor_augassign_by_ellipsis(): # GPU: does not supported op "FloorMod" -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_onecard diff --git a/tests/st/pynative/test_tensor_getitem.py b/tests/st/pynative/test_tensor_getitem.py index 32a8bda0825..38da21ea774 100644 --- a/tests/st/pynative/test_tensor_getitem.py +++ b/tests/st/pynative/test_tensor_getitem.py @@ -79,7 +79,7 @@ class NetWorkSliceEllipsis(Cell): return ret0, ret1, ret2, ret3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -111,7 +111,7 @@ class NetWorkReduceDimension(Cell): return ret1, ret2, ret3, ret4 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -171,7 +171,7 @@ class TensorGetItemByThreeTensors(Cell): return ret0, ret1, ret2 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -359,7 +359,7 @@ class TensorSetItemByMixedTensors_0(Cell): return ret -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -399,7 +399,7 @@ class TensorSetItemByMixedTensors_1(Cell): return ret -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -803,7 +803,7 @@ def test_setitem_by_tensor_with_tuple_of_tensor_error(): net(index_0_ms, index_1_ms, index_2_ms, value_0_ms, value_1_ms) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -903,7 +903,7 @@ def test_tensor_assign_slice_value_1(): assert np.all(z == out.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -1099,7 +1099,7 @@ class TensorAssignWithBoolTensorIndex2Error(Cell): return a -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/pynative/test_tensor_setitem.py b/tests/st/pynative/test_tensor_setitem.py index 8c680b3eb3e..cc09b923a92 100644 --- a/tests/st/pynative/test_tensor_setitem.py +++ b/tests/st/pynative/test_tensor_setitem.py @@ -57,7 +57,7 @@ class NumpySetItemByList(): return x -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -73,7 +73,7 @@ def test_setitem_by_list(): setup_testcase(x, cases) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -90,7 +90,7 @@ def test_setitem_with_sequence(): setup_testcase(x, cases) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -107,7 +107,7 @@ def test_setitem_dtype(): setup_testcase(x, cases) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -124,7 +124,7 @@ def test_setitem_by_tuple_with_int(): setup_testcase(x, cases) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -142,7 +142,7 @@ def test_setitem_by_tuple_with_list(): setup_testcase(x, cases) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -177,7 +177,7 @@ def test_setitem_with_broadcast(): setup_testcase(x, cases) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -192,7 +192,7 @@ def test_setitem_mul_by_scalar(): setup_testcase(x, cases) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -212,7 +212,7 @@ def test_setitem_by_slice(): setup_testcase(x, cases) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training