From b5b4c7c9d95251a1d96590c48518ddfc13593324 Mon Sep 17 00:00:00 2001 From: liuruotao Date: Wed, 1 Jun 2022 20:37:11 +0800 Subject: [PATCH] takedown gpu testcases where pass >=200 && fail==0 --- .../test_tdt_data_transfer_enable_rdr.py | 2 +- tests/st/export_and_load/test_lite_mindir.py | 2 +- tests/st/networks/test_gpu_alexnet.py | 2 +- tests/st/networks/test_gpu_lenet.py | 2 +- .../st/networks/test_gpu_lenet_enable_rdr.py | 2 +- tests/st/networks/test_gpu_lstm.py | 2 +- tests/st/networks/test_gpu_resnet.py | 2 +- tests/st/numpy_native/test_array_ops.py | 2 +- tests/st/numpy_native/test_logic_ops.py | 2 +- tests/st/numpy_native/test_math_ops.py | 4 +- .../test_aicpu_ops/test_tensor_copy_slices.py | 6 +-- tests/st/ops/gpu/test_acos_grad_grad_op.py | 2 +- tests/st/ops/gpu/test_acos_grad_op.py | 4 +- tests/st/ops/gpu/test_acos_op.py | 4 +- tests/st/ops/gpu/test_acosh_grad_grad_op.py | 4 +- tests/st/ops/gpu/test_acosh_grad_op.py | 4 +- tests/st/ops/gpu/test_acosh_op.py | 4 +- tests/st/ops/gpu/test_adagrad_op.py | 2 +- tests/st/ops/gpu/test_adam_fusion.py | 2 +- tests/st/ops/gpu/test_adam_op.py | 2 +- .../st/ops/gpu/test_adaptive_avg_pool2d_op.py | 8 +-- tests/st/ops/gpu/test_add_op.py | 4 +- tests/st/ops/gpu/test_addn_op.py | 6 +-- .../ops/gpu/test_apply_gradient_descent_op.py | 4 +- tests/st/ops/gpu/test_argmax_op.py | 4 +- tests/st/ops/gpu/test_argmaxwithvalue_op.py | 10 ++-- tests/st/ops/gpu/test_argminwithvalue_op.py | 10 ++-- tests/st/ops/gpu/test_asin_grad_grad_op.py | 2 +- tests/st/ops/gpu/test_asin_grad_op.py | 4 +- tests/st/ops/gpu/test_asin_op.py | 4 +- tests/st/ops/gpu/test_asinh_grad_grad_op.py | 4 +- tests/st/ops/gpu/test_asinh_grad_op.py | 2 +- tests/st/ops/gpu/test_asinh_op.py | 4 +- tests/st/ops/gpu/test_assign_add_op.py | 2 +- tests/st/ops/gpu/test_assign_op.py | 4 +- tests/st/ops/gpu/test_atan_grad_grad_op.py | 4 +- tests/st/ops/gpu/test_atan_grad_op.py | 4 +- tests/st/ops/gpu/test_atan_op.py | 4 +- tests/st/ops/gpu/test_batch_matmul.py | 14 ++--- tests/st/ops/gpu/test_batchnorm_fold2_op.py | 2 +- .../st/ops/gpu/test_batchnorm_fold_grad_op.py | 6 +-- tests/st/ops/gpu/test_batchnorm_fold_op.py | 6 +-- tests/st/ops/gpu/test_batchnorm_op.py | 12 ++--- tests/st/ops/gpu/test_bce_with_logits_loss.py | 4 +- .../ops/gpu/test_binary_cross_entropy_op.py | 4 +- .../st/ops/gpu/test_boundingbox_decode_op.py | 2 +- .../st/ops/gpu/test_boundingbox_encode_op.py | 2 +- tests/st/ops/gpu/test_broadcast_op.py | 14 ++--- tests/st/ops/gpu/test_broadcast_to_ops.py | 4 +- tests/st/ops/gpu/test_cast_op.py | 10 ++-- tests/st/ops/gpu/test_check_valid_op.py | 4 +- tests/st/ops/gpu/test_concatv2_op.py | 24 ++++----- .../ops/gpu/test_conv2d_backprop_filter_op.py | 2 +- .../ops/gpu/test_conv2d_backprop_input_op.py | 2 +- tests/st/ops/gpu/test_conv2d_op.py | 8 +-- tests/st/ops/gpu/test_conv3dtranspose_op.py | 2 +- .../st/ops/gpu/test_correction_mul_grad_op.py | 2 +- tests/st/ops/gpu/test_correction_mul_op.py | 2 +- tests/st/ops/gpu/test_cos_op.py | 2 +- tests/st/ops/gpu/test_crop_and_resize.py | 6 +-- tests/st/ops/gpu/test_ctcloss_op.py | 2 +- tests/st/ops/gpu/test_cudnn_inplace_fusion.py | 2 +- tests/st/ops/gpu/test_cumprod_op.py | 6 +-- tests/st/ops/gpu/test_cumsum_op.py | 6 +-- tests/st/ops/gpu/test_dense_op.py | 16 +++--- tests/st/ops/gpu/test_depthtospace_op.py | 8 +-- tests/st/ops/gpu/test_div_op.py | 4 +- tests/st/ops/gpu/test_dropout.py | 4 +- tests/st/ops/gpu/test_dtype_op.py | 8 +-- tests/st/ops/gpu/test_dynamic_shape_op.py | 4 +- tests/st/ops/gpu/test_elu_grad_op.py | 4 +- tests/st/ops/gpu/test_elu_op.py | 4 +- tests/st/ops/gpu/test_embedding_lookup_op.py | 4 +- tests/st/ops/gpu/test_equal_op.py | 6 +-- tests/st/ops/gpu/test_equalcount_op.py | 2 +- tests/st/ops/gpu/test_erf_op.py | 4 +- tests/st/ops/gpu/test_erfc_op.py | 4 +- tests/st/ops/gpu/test_exp_op.py | 2 +- tests/st/ops/gpu/test_expand_dims.py | 4 +- .../ops/gpu/test_extract_image_patches_op.py | 4 +- .../st/ops/gpu/test_fake_quant_perchannel.py | 12 ++--- .../gpu/test_fake_quant_perchannel_grad.py | 4 +- tests/st/ops/gpu/test_fake_quant_perlayer.py | 6 +-- .../ops/gpu/test_fake_quant_perlayer_grad.py | 4 +- tests/st/ops/gpu/test_flatten_grad_op.py | 2 +- tests/st/ops/gpu/test_flatten_op.py | 6 +-- tests/st/ops/gpu/test_float_status_op.py | 8 +-- tests/st/ops/gpu/test_floordiv_op.py | 2 +- tests/st/ops/gpu/test_ftrl_op.py | 2 +- tests/st/ops/gpu/test_gatherV2_op.py | 16 +++--- tests/st/ops/gpu/test_gather_grad_op.py | 16 +++--- tests/st/ops/gpu/test_gather_op.py | 8 +-- tests/st/ops/gpu/test_gathernd_op.py | 12 ++--- tests/st/ops/gpu/test_gelu_grad_op.py | 4 +- tests/st/ops/gpu/test_gelu_op.py | 10 ++-- .../test_gpu_convert_to_dynamic_shape_op.py | 6 +-- tests/st/ops/gpu/test_hsigmoid_op.py | 6 +-- tests/st/ops/gpu/test_hswish_op.py | 2 +- tests/st/ops/gpu/test_identity_op.py | 4 +- tests/st/ops/gpu/test_in_top_k.py | 6 +-- tests/st/ops/gpu/test_index_add_op.py | 16 +++--- tests/st/ops/gpu/test_instancenorm2d.py | 2 +- tests/st/ops/gpu/test_iou_op.py | 2 +- tests/st/ops/gpu/test_kl_div_op.py | 4 +- tests/st/ops/gpu/test_l2loss_op.py | 6 +-- tests/st/ops/gpu/test_l2normalize_grad_op.py | 2 +- tests/st/ops/gpu/test_l2normalize_op.py | 2 +- .../ops/gpu/test_layer_norm_grad_grad_op.py | 8 +-- tests/st/ops/gpu/test_layer_norm_grad_op.py | 6 +-- tests/st/ops/gpu/test_layer_norm_op.py | 14 ++--- tests/st/ops/gpu/test_lessequal_op.py | 2 +- tests/st/ops/gpu/test_lin_space.py | 8 +-- .../st/ops/gpu/test_local_response_norm_op.py | 4 +- tests/st/ops/gpu/test_log1p_op.py | 4 +- tests/st/ops/gpu/test_log_op.py | 2 +- tests/st/ops/gpu/test_logical_op.py | 4 +- tests/st/ops/gpu/test_logsoftmax_op.py | 6 +-- tests/st/ops/gpu/test_loss.py | 8 +-- tests/st/ops/gpu/test_matmul_op.py | 6 +-- tests/st/ops/gpu/test_matrix_inverse_op.py | 2 +- tests/st/ops/gpu/test_maximum_op.py | 6 +-- tests/st/ops/gpu/test_maxpool_gpu_op.py | 12 ++--- tests/st/ops/gpu/test_maxpool_grad_gpu_op.py | 2 +- .../gpu/test_maxpool_with_argmax_gpu_op.py | 6 +-- .../test_maxpool_with_argmax_grad_gpu_op.py | 4 +- tests/st/ops/gpu/test_minimum_op.py | 8 +-- tests/st/ops/gpu/test_mirror_pad.py | 8 +-- tests/st/ops/gpu/test_momentum_fusion.py | 2 +- tests/st/ops/gpu/test_mul_op.py | 8 +-- tests/st/ops/gpu/test_multinomial_op.py | 2 +- tests/st/ops/gpu/test_neg_op.py | 4 +- tests/st/ops/gpu/test_nll_loss.py | 12 ++--- tests/st/ops/gpu/test_nms_with_mask_op.py | 8 +-- tests/st/ops/gpu/test_one_hot_op.py | 4 +- tests/st/ops/gpu/test_oneslike_op.py | 2 +- tests/st/ops/gpu/test_pack_op.py | 8 +-- tests/st/ops/gpu/test_pad.py | 14 ++--- tests/st/ops/gpu/test_prelu_op.py | 12 ++--- tests/st/ops/gpu/test_pyfunc_op.py | 6 +-- .../ops/gpu/test_random_choice_with_mask.py | 8 +-- tests/st/ops/gpu/test_randperm.py | 6 +-- tests/st/ops/gpu/test_range_op.py | 8 +-- tests/st/ops/gpu/test_realdiv_op.py | 2 +- tests/st/ops/gpu/test_reciprocal_grad_op.py | 4 +- tests/st/ops/gpu/test_reciprocal_op.py | 2 +- tests/st/ops/gpu/test_reduce_all_op.py | 4 +- tests/st/ops/gpu/test_reduce_any_op.py | 4 +- tests/st/ops/gpu/test_reduce_max_op.py | 6 +-- tests/st/ops/gpu/test_reduce_min_op.py | 4 +- tests/st/ops/gpu/test_reduce_prod_op.py | 2 +- tests/st/ops/gpu/test_reduce_sum_op.py | 4 +- tests/st/ops/gpu/test_relu6_grad_op.py | 2 +- tests/st/ops/gpu/test_relu6_op.py | 4 +- tests/st/ops/gpu/test_relu_grad_grad_op.py | 4 +- tests/st/ops/gpu/test_relu_op.py | 10 ++-- tests/st/ops/gpu/test_relu_v2.py | 6 +-- tests/st/ops/gpu/test_repeat_elements_op.py | 54 +++++++++---------- tests/st/ops/gpu/test_reshape_op.py | 4 +- .../ops/gpu/test_resize_bilinear_grad_op.py | 2 +- tests/st/ops/gpu/test_resize_bilinear_op.py | 14 ++--- .../test_resize_nearest_neighbor_grad_op.py | 4 +- tests/st/ops/gpu/test_reverse_sequence.py | 6 +-- tests/st/ops/gpu/test_reverse_v2_op.py | 4 +- tests/st/ops/gpu/test_rint_op.py | 4 +- tests/st/ops/gpu/test_rl_buffer_net.py | 2 +- tests/st/ops/gpu/test_rl_buffer_op.py | 2 +- tests/st/ops/gpu/test_rmsprop.py | 4 +- tests/st/ops/gpu/test_roi_align_grad_op.py | 2 +- tests/st/ops/gpu/test_round_op.py | 4 +- tests/st/ops/gpu/test_rsqrt_grad_op.py | 2 +- .../ops/gpu/test_sampled_softmax_loss_op.py | 4 +- tests/st/ops/gpu/test_scatter_func_op.py | 28 +++++----- tests/st/ops/gpu/test_select_op.py | 2 +- tests/st/ops/gpu/test_sequence_mask_op.py | 6 +-- tests/st/ops/gpu/test_sgd_op.py | 2 +- ...gmoid_cross_entropy_with_logits_grad_op.py | 4 +- ...st_sigmoid_cross_entropy_with_logits_op.py | 4 +- tests/st/ops/gpu/test_sigmoid_grad_grad_op.py | 2 +- tests/st/ops/gpu/test_sigmoid_grad_op.py | 2 +- tests/st/ops/gpu/test_sigmoid_op.py | 2 +- tests/st/ops/gpu/test_sign_op.py | 6 +-- tests/st/ops/gpu/test_sin_op.py | 4 +- tests/st/ops/gpu/test_slice.py | 6 +-- tests/st/ops/gpu/test_slice_grad.py | 4 +- tests/st/ops/gpu/test_smoothl1loss_op.py | 4 +- ...st_softmax_cross_entropy_with_logits_op.py | 2 +- tests/st/ops/gpu/test_softmax_op.py | 4 +- tests/st/ops/gpu/test_softplus_grad_op.py | 4 +- tests/st/ops/gpu/test_softplus_op.py | 10 ++-- tests/st/ops/gpu/test_sort.py | 16 +++--- tests/st/ops/gpu/test_spacetodepth_op.py | 12 ++--- tests/st/ops/gpu/test_sparse_apply_ftrl_op.py | 10 ++-- .../test_sparse_apply_proximal_adagrad_op.py | 8 +-- tests/st/ops/gpu/test_sparse_gather_v2_op.py | 10 ++-- ...se_softmax_cross_entropy_with_logits_op.py | 2 +- tests/st/ops/gpu/test_split.py | 14 ++--- tests/st/ops/gpu/test_sqrt_grad_grad_op.py | 4 +- tests/st/ops/gpu/test_sqrt_grad_op.py | 2 +- tests/st/ops/gpu/test_sqrt_op.py | 6 +-- tests/st/ops/gpu/test_square_op.py | 4 +- .../st/ops/gpu/test_squared_difference_op.py | 36 ++++++------- tests/st/ops/gpu/test_squeeze_op.py | 4 +- tests/st/ops/gpu/test_standard_normal.py | 2 +- tests/st/ops/gpu/test_stridedslice_grad_op.py | 24 ++++----- tests/st/ops/gpu/test_stridedslice_op.py | 6 +-- tests/st/ops/gpu/test_sub_op.py | 6 +-- tests/st/ops/gpu/test_tanh_grad_grad_op.py | 4 +- tests/st/ops/gpu/test_tanh_op.py | 4 +- tests/st/ops/gpu/test_tensordot_op.py | 8 +-- tests/st/ops/gpu/test_time_distributed_op.py | 22 ++++---- tests/st/ops/gpu/test_topk_op.py | 8 +-- tests/st/ops/gpu/test_transpose_op.py | 24 ++++----- .../gpu/test_uniform_candidate_sampler_op.py | 18 +++---- tests/st/ops/gpu/test_uniform_int.py | 2 +- tests/st/ops/gpu/test_uniform_real.py | 2 +- tests/st/ops/gpu/test_unique_op.py | 36 ++++++------- tests/st/ops/gpu/test_unpack_op.py | 8 +-- tests/st/ops/gpu/test_unsorted_segment_max.py | 16 +++--- tests/st/ops/gpu/test_unsorted_segment_min.py | 14 ++--- tests/st/ops/gpu/test_unsorted_segment_sum.py | 10 ++-- tests/st/ops/gpu/test_zeroslike_op.py | 8 +-- tests/st/ops/graph_kernel/test_argmax.py | 2 +- tests/st/ops/graph_kernel/test_atomic_add.py | 4 +- tests/st/ops/graph_kernel/test_batchnorm.py | 4 +- .../ops/graph_kernel/test_batchnorm_grad.py | 4 +- tests/st/ops/graph_kernel/test_bias_add.py | 2 +- tests/st/ops/graph_kernel/test_compare.py | 2 +- tests/st/ops/graph_kernel/test_cse.py | 2 +- .../test_dropoutgrad_reducesum_stitch.py | 2 +- tests/st/ops/graph_kernel/test_equal_count.py | 4 +- tests/st/ops/graph_kernel/test_expand_dims.py | 2 +- tests/st/ops/graph_kernel/test_fused_adam.py | 6 +-- tests/st/ops/graph_kernel/test_gelu.py | 4 +- tests/st/ops/graph_kernel/test_identity.py | 2 +- tests/st/ops/graph_kernel/test_layernorm.py | 4 +- .../ops/graph_kernel/test_layernorm_stitch.py | 2 +- tests/st/ops/graph_kernel/test_logsoftmax.py | 4 +- .../st/ops/graph_kernel/test_low_precision.py | 4 +- .../st/ops/graph_kernel/test_matmul_to_mul.py | 2 +- .../st/ops/graph_kernel/test_maximum_grad.py | 2 +- .../st/ops/graph_kernel/test_minimum_grad.py | 2 +- tests/st/ops/graph_kernel/test_oneslike.py | 4 +- .../ops/graph_kernel/test_optimize_assign.py | 4 +- tests/st/ops/graph_kernel/test_recompute.py | 6 +-- tests/st/ops/graph_kernel/test_reduce_max.py | 2 +- tests/st/ops/graph_kernel/test_reduce_mean.py | 2 +- tests/st/ops/graph_kernel/test_reduce_min.py | 2 +- tests/st/ops/graph_kernel/test_relu.py | 2 +- tests/st/ops/graph_kernel/test_relu_grad.py | 2 +- tests/st/ops/graph_kernel/test_reorder_ops.py | 4 +- tests/st/ops/graph_kernel/test_sigmoid.py | 4 +- .../test_sigmoid_cross_entropy_with_logits.py | 4 +- tests/st/ops/graph_kernel/test_simplify.py | 2 +- tests/st/ops/graph_kernel/test_slice.py | 2 +- tests/st/ops/graph_kernel/test_softmax.py | 2 +- .../test_softmax_cross_entropy_with_logits.py | 2 +- .../ops/graph_kernel/test_softmax_stitch.py | 2 +- .../ops/graph_kernel/test_square_sum_all.py | 4 +- .../graph_kernel/test_squared_difference.py | 2 +- tests/st/ops/graph_kernel/test_tanh_grad.py | 2 +- tests/st/pynative/test_pynative_hook_grad.py | 18 +++---- ...ive_layernorm_input_and_argmaxwithvalue.py | 2 +- .../test_pynative_mixed_precision_cells.py | 6 +-- tests/st/pynative/test_tensor_getitem.py | 2 +- 264 files changed, 762 insertions(+), 762 deletions(-) diff --git a/tests/st/data_transfer/test_tdt_data_transfer_enable_rdr.py b/tests/st/data_transfer/test_tdt_data_transfer_enable_rdr.py index e1a6ac2934d..73fa05bad46 100644 --- a/tests/st/data_transfer/test_tdt_data_transfer_enable_rdr.py +++ b/tests/st/data_transfer/test_tdt_data_transfer_enable_rdr.py @@ -44,7 +44,7 @@ def test_train_with_Ascend(): test_train("Ascend") @security_off_wrap -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_train_with_GPU(): diff --git a/tests/st/export_and_load/test_lite_mindir.py b/tests/st/export_and_load/test_lite_mindir.py index 0230c5cf553..6cdfc032eb2 100644 --- a/tests/st/export_and_load/test_lite_mindir.py +++ b/tests/st/export_and_load/test_lite_mindir.py @@ -29,7 +29,7 @@ from tests.models.official.cv.mobilenetv2.src.mobilenetV2 import MobileNetV2Back context.set_context(mode=context.GRAPH_MODE, device_target="GPU") ckpt_path = '/home/workspace/mindspore_dataset/checkpoint/mobilenetv2/mobilenetv2_gpu.ckpt' -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_export_mobilenetv2_gpu_mindir(): diff --git a/tests/st/networks/test_gpu_alexnet.py b/tests/st/networks/test_gpu_alexnet.py index f6644c31340..13561e7b7e2 100644 --- a/tests/st/networks/test_gpu_alexnet.py +++ b/tests/st/networks/test_gpu_alexnet.py @@ -68,7 +68,7 @@ class AlexNet(nn.Cell): return x -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_trainTensor(num_classes=10, epoch=15, batch_size=32): diff --git a/tests/st/networks/test_gpu_lenet.py b/tests/st/networks/test_gpu_lenet.py index c2e25ff5b21..342d07dcf6c 100644 --- a/tests/st/networks/test_gpu_lenet.py +++ b/tests/st/networks/test_gpu_lenet.py @@ -127,7 +127,7 @@ def multisteplr(total_steps, gap, base_lr=0.9, gamma=0.1, dtype=mstype.float32): return Tensor(np.array(lr), dtype) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_train_lenet(): diff --git a/tests/st/networks/test_gpu_lenet_enable_rdr.py b/tests/st/networks/test_gpu_lenet_enable_rdr.py index fbc8e561d83..436bcbd9b9c 100644 --- a/tests/st/networks/test_gpu_lenet_enable_rdr.py +++ b/tests/st/networks/test_gpu_lenet_enable_rdr.py @@ -31,7 +31,7 @@ def test_train_and_eval(): test_train_and_eval_lenet() @security_off_wrap -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_train_with_GPU(): diff --git a/tests/st/networks/test_gpu_lstm.py b/tests/st/networks/test_gpu_lstm.py index 32c96943fcc..4ec063278da 100644 --- a/tests/st/networks/test_gpu_lstm.py +++ b/tests/st/networks/test_gpu_lstm.py @@ -105,7 +105,7 @@ class SentimentNet(nn.Cell): batch_size = 64 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_LSTM(): diff --git a/tests/st/networks/test_gpu_resnet.py b/tests/st/networks/test_gpu_resnet.py index b2d5565854c..de67d16318d 100644 --- a/tests/st/networks/test_gpu_resnet.py +++ b/tests/st/networks/test_gpu_resnet.py @@ -328,7 +328,7 @@ def resnet50(num_classes): return ResNet(ResidualBlock, [3, 4, 6, 3], num_classes) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_trainTensor(num_classes=10, epoch=8, batch_size=1): diff --git a/tests/st/numpy_native/test_array_ops.py b/tests/st/numpy_native/test_array_ops.py index 65de6968aa3..47f9e77796e 100644 --- a/tests/st/numpy_native/test_array_ops.py +++ b/tests/st/numpy_native/test_array_ops.py @@ -1252,7 +1252,7 @@ def test_select(): match_res(mnp.select, onp.select, condlist, choicelist, default=10) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard diff --git a/tests/st/numpy_native/test_logic_ops.py b/tests/st/numpy_native/test_logic_ops.py index c537d24f21c..dab3c6c9534 100644 --- a/tests/st/numpy_native/test_logic_ops.py +++ b/tests/st/numpy_native/test_logic_ops.py @@ -274,7 +274,7 @@ def test_isscalar(): assert mnp.isscalar('numpy') == onp.isscalar('numpy') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard diff --git a/tests/st/numpy_native/test_math_ops.py b/tests/st/numpy_native/test_math_ops.py index 944437600fe..008d71004f5 100644 --- a/tests/st/numpy_native/test_math_ops.py +++ b/tests/st/numpy_native/test_math_ops.py @@ -1733,7 +1733,7 @@ def onp_convolve(mode): return a, b, c, d, e -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_convolve(): @@ -2659,7 +2659,7 @@ def onp_correlate(a, v): return a, b, c, d -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_correlate(): diff --git a/tests/st/ops/ascend/test_aicpu_ops/test_tensor_copy_slices.py b/tests/st/ops/ascend/test_aicpu_ops/test_tensor_copy_slices.py index f5ccba87cd4..f971dd789ae 100644 --- a/tests/st/ops/ascend/test_aicpu_ops/test_tensor_copy_slices.py +++ b/tests/st/ops/ascend/test_aicpu_ops/test_tensor_copy_slices.py @@ -75,7 +75,7 @@ def test_tensor_copy_slices(): test_tensor_copy_slices_net_many_dtype((10, 10, 10), (5, 10), (9, 5,), (10, 10,), (1, 1,), support_dtype) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_tensor_copy_slices_bprop(): @@ -105,14 +105,14 @@ def test_tensor_copy_slices_ascend_pynative(): context.set_context(mode=context.PYNATIVE_MODE, device_target="Ascend") test_tensor_copy_slices() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_tensor_copy_slices_gpu_graph(): context.set_context(mode=context.GRAPH_MODE, device_target="GPU") test_tensor_copy_slices() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_tensor_copy_slices_gpu_pynative(): diff --git a/tests/st/ops/gpu/test_acos_grad_grad_op.py b/tests/st/ops/gpu/test_acos_grad_grad_op.py index 89c288956bf..12e8b4acbaa 100644 --- a/tests/st/ops/gpu/test_acos_grad_grad_op.py +++ b/tests/st/ops/gpu/test_acos_grad_grad_op.py @@ -45,7 +45,7 @@ class Grad(nn.Cell): return self.grad(self.network)(x, grad, dout) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @pytest.mark.parametrize("fp_type, error_magnitude, mode", [ diff --git a/tests/st/ops/gpu/test_acos_grad_op.py b/tests/st/ops/gpu/test_acos_grad_op.py index c1f5695306a..5a20bb96ed4 100644 --- a/tests/st/ops/gpu/test_acos_grad_op.py +++ b/tests/st/ops/gpu/test_acos_grad_op.py @@ -21,7 +21,7 @@ from mindspore import Tensor import mindspore.ops.operations._grad_ops as P context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_acosgrad_fp32(): @@ -33,7 +33,7 @@ def test_acosgrad_fp32(): diff = output_ms.asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_acosgrad_fp16(): diff --git a/tests/st/ops/gpu/test_acos_op.py b/tests/st/ops/gpu/test_acos_op.py index 3d27de6fa03..84ac8c3b46a 100644 --- a/tests/st/ops/gpu/test_acos_op.py +++ b/tests/st/ops/gpu/test_acos_op.py @@ -21,7 +21,7 @@ from mindspore import Tensor from mindspore.ops import operations as P context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_acos_fp32(): @@ -30,7 +30,7 @@ def test_acos_fp32(): output_np = np.arccos(x_np) assert np.allclose(output_ms.asnumpy(), output_np) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_acos_fp16(): diff --git a/tests/st/ops/gpu/test_acosh_grad_grad_op.py b/tests/st/ops/gpu/test_acosh_grad_grad_op.py index 001ede6b804..e66fa29d523 100644 --- a/tests/st/ops/gpu/test_acosh_grad_grad_op.py +++ b/tests/st/ops/gpu/test_acosh_grad_grad_op.py @@ -72,14 +72,14 @@ def acosh_grad_grad_base(dtype, loss): assert np.allclose(dgrad_ms.asnumpy(), dgrad_np, loss, loss) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_acosh_grad_grad_float16(): acosh_grad_grad_base(np.float16, 2e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_acosh_grad_grad_float32(): diff --git a/tests/st/ops/gpu/test_acosh_grad_op.py b/tests/st/ops/gpu/test_acosh_grad_op.py index 2d4e015d406..f3c58df78c2 100644 --- a/tests/st/ops/gpu/test_acosh_grad_op.py +++ b/tests/st/ops/gpu/test_acosh_grad_op.py @@ -22,7 +22,7 @@ import mindspore.ops.operations._grad_ops as P context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") np.random.seed(1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_acoshgrad_fp32(): @@ -32,7 +32,7 @@ def test_acoshgrad_fp32(): output_np = dout_np / np.sinh(y_np) assert np.allclose(output_ms.asnumpy(), output_np, 1e-4, 1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_acoshgrad_fp16(): diff --git a/tests/st/ops/gpu/test_acosh_op.py b/tests/st/ops/gpu/test_acosh_op.py index 69f7b72707a..7891da6f242 100644 --- a/tests/st/ops/gpu/test_acosh_op.py +++ b/tests/st/ops/gpu/test_acosh_op.py @@ -22,7 +22,7 @@ from mindspore.ops import operations as P context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") np.random.seed(1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_acosh_fp32(): @@ -31,7 +31,7 @@ def test_acosh_fp32(): output_np = np.arccosh(x_np) assert np.allclose(output_ms.asnumpy(), output_np, 1e-4, 1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_acosh_fp16(): diff --git a/tests/st/ops/gpu/test_adagrad_op.py b/tests/st/ops/gpu/test_adagrad_op.py index ffe34aa96fd..adc5d25e183 100644 --- a/tests/st/ops/gpu/test_adagrad_op.py +++ b/tests/st/ops/gpu/test_adagrad_op.py @@ -40,7 +40,7 @@ class Net(nn.Cell): return z -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_apply_adagrad(): diff --git a/tests/st/ops/gpu/test_adam_fusion.py b/tests/st/ops/gpu/test_adam_fusion.py index f98465b8b00..2f055dc96a7 100644 --- a/tests/st/ops/gpu/test_adam_fusion.py +++ b/tests/st/ops/gpu/test_adam_fusion.py @@ -65,7 +65,7 @@ class Net(nn.Cell): return next_v -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_adam_fusion(): diff --git a/tests/st/ops/gpu/test_adam_op.py b/tests/st/ops/gpu/test_adam_op.py index 8c2e16e6386..15d5ab1303a 100644 --- a/tests/st/ops/gpu/test_adam_op.py +++ b/tests/st/ops/gpu/test_adam_op.py @@ -41,7 +41,7 @@ class NetAdam(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_adam(): diff --git a/tests/st/ops/gpu/test_adaptive_avg_pool2d_op.py b/tests/st/ops/gpu/test_adaptive_avg_pool2d_op.py index c3953df936f..e3f518001d2 100644 --- a/tests/st/ops/gpu/test_adaptive_avg_pool2d_op.py +++ b/tests/st/ops/gpu/test_adaptive_avg_pool2d_op.py @@ -47,7 +47,7 @@ class GradNet(nn.Cell): return self.adaptive_avg_pool2d_grad(x, dy) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net_normal(): @@ -58,7 +58,7 @@ def test_net_normal(): assert output.asnumpy().shape == expect_shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net_single(): @@ -69,7 +69,7 @@ def test_net_single(): assert output.asnumpy().shape == expect_shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net_none(): @@ -80,7 +80,7 @@ def test_net_none(): assert output.asnumpy().shape == expect_shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net_value(): diff --git a/tests/st/ops/gpu/test_add_op.py b/tests/st/ops/gpu/test_add_op.py index 7cf8522cd99..dd3b0396e97 100644 --- a/tests/st/ops/gpu/test_add_op.py +++ b/tests/st/ops/gpu/test_add_op.py @@ -129,7 +129,7 @@ def add(nptype): @pytest.mark.skip(reason='0 in shape is not support') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_add_float64(): @@ -137,7 +137,7 @@ def test_add_float64(): @pytest.mark.skip(reason='0 in shape is not support') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_add_float32(): diff --git a/tests/st/ops/gpu/test_addn_op.py b/tests/st/ops/gpu/test_addn_op.py index 1e70cdb7bf0..5045b239ab9 100644 --- a/tests/st/ops/gpu/test_addn_op.py +++ b/tests/st/ops/gpu/test_addn_op.py @@ -33,7 +33,7 @@ class Net(nn.Cell): return self.add((x, y, z)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net(): @@ -56,7 +56,7 @@ def test_net(): assert (output.asnumpy() == expect_result).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net_float64(): @@ -95,7 +95,7 @@ def test_net_float64(): assert (output.asnumpy() == expect_result).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net_int64(): diff --git a/tests/st/ops/gpu/test_apply_gradient_descent_op.py b/tests/st/ops/gpu/test_apply_gradient_descent_op.py index 6475ee7b59b..8896618f5e4 100644 --- a/tests/st/ops/gpu/test_apply_gradient_descent_op.py +++ b/tests/st/ops/gpu/test_apply_gradient_descent_op.py @@ -32,7 +32,7 @@ class Net(nn.Cell): return self.apply_gradient_descent(self.var, alpha, delta) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_apply_gradient_descent_float32(): @@ -59,7 +59,7 @@ def test_apply_gradient_descent_float32(): np.testing.assert_almost_equal(net.var.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_apply_gradient_descent_float16(): diff --git a/tests/st/ops/gpu/test_argmax_op.py b/tests/st/ops/gpu/test_argmax_op.py index b251cbe7d5f..fe9417f5097 100644 --- a/tests/st/ops/gpu/test_argmax_op.py +++ b/tests/st/ops/gpu/test_argmax_op.py @@ -34,7 +34,7 @@ class NetArgmax(nn.Cell): return self.argmax(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_argmax_1d(): @@ -48,7 +48,7 @@ def test_argmax_1d(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_argmax_2d(): diff --git a/tests/st/ops/gpu/test_argmaxwithvalue_op.py b/tests/st/ops/gpu/test_argmaxwithvalue_op.py index a814f3ef8ff..7e3ae2be15c 100644 --- a/tests/st/ops/gpu/test_argmaxwithvalue_op.py +++ b/tests/st/ops/gpu/test_argmaxwithvalue_op.py @@ -101,21 +101,21 @@ def argmaxwithvalue_3d(data_type, shape_x): assert (output[1].asnumpy() == expect2).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_argmaxwithvalue_base_float32(): argmaxwithvalue_base(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_argmaxwithvalue_base_float16(): argmaxwithvalue_base(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_argmaxwithvalue_3d_float32(): @@ -126,7 +126,7 @@ def test_argmaxwithvalue_3d_float32(): argmaxwithvalue_3d(np.float32, shape_x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_argmaxwithvalue_3d_float16(): @@ -135,7 +135,7 @@ def test_argmaxwithvalue_3d_float16(): argmaxwithvalue_3d(np.float16, shape_x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_argmaxwithvalue_3d_big_float32(): diff --git a/tests/st/ops/gpu/test_argminwithvalue_op.py b/tests/st/ops/gpu/test_argminwithvalue_op.py index 3a4fcd5ba92..ab5e5266fdb 100644 --- a/tests/st/ops/gpu/test_argminwithvalue_op.py +++ b/tests/st/ops/gpu/test_argminwithvalue_op.py @@ -101,21 +101,21 @@ def argminwithvalue_3d(data_type, shape_x): assert (output[1].asnumpy() == expect2).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_argminwithvalue_base_float32(): argminwithvalue_base(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_argminwithvalue_base_float16(): argminwithvalue_base(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_argminwithvalue_3d_float32(): @@ -126,7 +126,7 @@ def test_argminwithvalue_3d_float32(): argminwithvalue_3d(np.float32, shape_x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_argminwithvalue_3d_float16(): @@ -135,7 +135,7 @@ def test_argminwithvalue_3d_float16(): argminwithvalue_3d(np.float16, shape_x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_argminwithvalue_3d_big_float32(): diff --git a/tests/st/ops/gpu/test_asin_grad_grad_op.py b/tests/st/ops/gpu/test_asin_grad_grad_op.py index 5fb74e4db12..3e35f4a25fc 100644 --- a/tests/st/ops/gpu/test_asin_grad_grad_op.py +++ b/tests/st/ops/gpu/test_asin_grad_grad_op.py @@ -45,7 +45,7 @@ class Grad(nn.Cell): return self.grad(self.network)(x, grad, dout) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @pytest.mark.parametrize("fp_type, error_magnitude, mode", [ diff --git a/tests/st/ops/gpu/test_asin_grad_op.py b/tests/st/ops/gpu/test_asin_grad_op.py index 5d3606fb664..6ca2ec77969 100644 --- a/tests/st/ops/gpu/test_asin_grad_op.py +++ b/tests/st/ops/gpu/test_asin_grad_op.py @@ -21,7 +21,7 @@ from mindspore import Tensor import mindspore.ops.operations._grad_ops as P context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_asingrad_fp32(): @@ -33,7 +33,7 @@ def test_asingrad_fp32(): diff = output_ms.asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_asingrad_fp16(): diff --git a/tests/st/ops/gpu/test_asin_op.py b/tests/st/ops/gpu/test_asin_op.py index b034a225b5d..a824e141359 100644 --- a/tests/st/ops/gpu/test_asin_op.py +++ b/tests/st/ops/gpu/test_asin_op.py @@ -21,7 +21,7 @@ from mindspore import Tensor from mindspore.ops import operations as P context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_asin_fp32(): @@ -30,7 +30,7 @@ def test_asin_fp32(): output_np = np.arcsin(x_np) assert np.allclose(output_ms.asnumpy(), output_np) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_asin_fp16(): diff --git a/tests/st/ops/gpu/test_asinh_grad_grad_op.py b/tests/st/ops/gpu/test_asinh_grad_grad_op.py index 3dae2c8306c..280081689f6 100644 --- a/tests/st/ops/gpu/test_asinh_grad_grad_op.py +++ b/tests/st/ops/gpu/test_asinh_grad_grad_op.py @@ -72,14 +72,14 @@ def asinh_grad_grad_base(dtype, loss): assert np.allclose(dgrad_ms.asnumpy(), dgrad_np, loss, loss) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_asinh_grad_grad_float16(): asinh_grad_grad_base(np.float16, 1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_asinh_grad_grad_float32(): diff --git a/tests/st/ops/gpu/test_asinh_grad_op.py b/tests/st/ops/gpu/test_asinh_grad_op.py index a42d820386b..3a0ae1ff2f4 100644 --- a/tests/st/ops/gpu/test_asinh_grad_op.py +++ b/tests/st/ops/gpu/test_asinh_grad_op.py @@ -22,7 +22,7 @@ import mindspore.ops.operations._grad_ops as P context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") np.random.seed(1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_asinhgrad_fp32(): diff --git a/tests/st/ops/gpu/test_asinh_op.py b/tests/st/ops/gpu/test_asinh_op.py index 21b1fceece6..8cf2c1419a7 100644 --- a/tests/st/ops/gpu/test_asinh_op.py +++ b/tests/st/ops/gpu/test_asinh_op.py @@ -22,7 +22,7 @@ from mindspore.ops import operations as P context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") np.random.seed(1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_asinh_fp32(): @@ -31,7 +31,7 @@ def test_asinh_fp32(): output_np = np.arcsinh(x_np) assert np.allclose(output_ms.asnumpy(), output_np, 1e-4, 1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_asinh_fp16(): diff --git a/tests/st/ops/gpu/test_assign_add_op.py b/tests/st/ops/gpu/test_assign_add_op.py index 8fd7f6b1340..c500a385698 100644 --- a/tests/st/ops/gpu/test_assign_add_op.py +++ b/tests/st/ops/gpu/test_assign_add_op.py @@ -33,7 +33,7 @@ class AssignAdd(nn.Cell): return res -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_assign_add(): diff --git a/tests/st/ops/gpu/test_assign_op.py b/tests/st/ops/gpu/test_assign_op.py index c08337a176b..8a2a6c85fdf 100644 --- a/tests/st/ops/gpu/test_assign_op.py +++ b/tests/st/ops/gpu/test_assign_op.py @@ -36,7 +36,7 @@ x = np.array([[1.2, 1], [1, 0]]).astype(np.float32) value = np.array([[1, 2], [3, 4.0]]).astype(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_assign(): @@ -54,7 +54,7 @@ def test_assign(): assert np.all(-diff2 < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_assign_float64(): diff --git a/tests/st/ops/gpu/test_atan_grad_grad_op.py b/tests/st/ops/gpu/test_atan_grad_grad_op.py index d0373e52075..b7829e70207 100644 --- a/tests/st/ops/gpu/test_atan_grad_grad_op.py +++ b/tests/st/ops/gpu/test_atan_grad_grad_op.py @@ -73,14 +73,14 @@ def atan_grad_grad_base(dtype, loss): assert np.allclose(dgrad_ms.asnumpy(), dgrad_np, loss, loss) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_atan_grad_grad_float16(): atan_grad_grad_base(np.float16, 1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_atan_grad_grad_float32(): diff --git a/tests/st/ops/gpu/test_atan_grad_op.py b/tests/st/ops/gpu/test_atan_grad_op.py index 986709734bc..bac60679388 100644 --- a/tests/st/ops/gpu/test_atan_grad_op.py +++ b/tests/st/ops/gpu/test_atan_grad_op.py @@ -22,7 +22,7 @@ import mindspore.ops.operations._grad_ops as P context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") np.random.seed(1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_atangrad_fp32(): @@ -32,7 +32,7 @@ def test_atangrad_fp32(): output_np = dout_np / (1 + np.square(x_np)) assert np.allclose(output_ms.asnumpy(), output_np, 1e-4, 1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_atangrad_fp16(): diff --git a/tests/st/ops/gpu/test_atan_op.py b/tests/st/ops/gpu/test_atan_op.py index ea1ca25a893..5beb2d97491 100644 --- a/tests/st/ops/gpu/test_atan_op.py +++ b/tests/st/ops/gpu/test_atan_op.py @@ -22,7 +22,7 @@ from mindspore.ops import operations as P context.set_context(mode=context.PYNATIVE_MODE, device_target="GPU") np.random.seed(1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_atan_fp32(): @@ -31,7 +31,7 @@ def test_atan_fp32(): output_np = np.arctan(x_np) assert np.allclose(output_ms.asnumpy(), output_np, 1e-4, 1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_atan_fp16(): diff --git a/tests/st/ops/gpu/test_batch_matmul.py b/tests/st/ops/gpu/test_batch_matmul.py index a9340ac1573..52f4fcbae72 100644 --- a/tests/st/ops/gpu/test_batch_matmul.py +++ b/tests/st/ops/gpu/test_batch_matmul.py @@ -32,7 +32,7 @@ class BatchMatMulNet(nn.Cell): def construct(self, x, y): return self.batch_matmul(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_4d(): @@ -54,7 +54,7 @@ def test_4d(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_4d_float64(): @@ -76,7 +76,7 @@ def test_4d_float64(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_4d_transpose_a(): @@ -98,7 +98,7 @@ def test_4d_transpose_a(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_4d_transpose_b(): @@ -120,7 +120,7 @@ def test_4d_transpose_b(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_4d_transpose_ab(): @@ -142,7 +142,7 @@ def test_4d_transpose_ab(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_4D_fp16(): @@ -176,7 +176,7 @@ class BatchMatMul_d(nn.Cell): return self.batch_matmul(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_batchmatmul_dynamic(): diff --git a/tests/st/ops/gpu/test_batchnorm_fold2_op.py b/tests/st/ops/gpu/test_batchnorm_fold2_op.py index cd27f07dce8..4aa3f7c29b4 100644 --- a/tests/st/ops/gpu/test_batchnorm_fold2_op.py +++ b/tests/st/ops/gpu/test_batchnorm_fold2_op.py @@ -52,7 +52,7 @@ class Net_gnd(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_batchnrom_fold2(): diff --git a/tests/st/ops/gpu/test_batchnorm_fold_grad_op.py b/tests/st/ops/gpu/test_batchnorm_fold_grad_op.py index 8ec35d7a6ce..9f1ec759bbd 100644 --- a/tests/st/ops/gpu/test_batchnorm_fold_grad_op.py +++ b/tests/st/ops/gpu/test_batchnorm_fold_grad_op.py @@ -43,7 +43,7 @@ def np_result(d_batch_mean, d_batch_std, x, batch_mean, batch_std): return dx -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_batchnorm_fold_grad1(): @@ -61,7 +61,7 @@ def test_batchnorm_fold_grad1(): assert np.allclose(dx.asnumpy(), expect, rtol=1.e-7, atol=1.e-7) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_batchnorm_fold_grad2(): @@ -79,7 +79,7 @@ def test_batchnorm_fold_grad2(): assert np.allclose(dx.asnumpy(), expect, rtol=1.e-7, atol=1.e-7) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_batchnorm_fold_grad_freeze(): diff --git a/tests/st/ops/gpu/test_batchnorm_fold_op.py b/tests/st/ops/gpu/test_batchnorm_fold_op.py index 7225c29c465..d0cdfac7be2 100644 --- a/tests/st/ops/gpu/test_batchnorm_fold_op.py +++ b/tests/st/ops/gpu/test_batchnorm_fold_op.py @@ -52,7 +52,7 @@ def np_result(x, mean, var, momentum, epsilon): return np_mean, np_var, mean_update, var_update, delay_mean, delay_std -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_batchnorm_fold(): @@ -77,7 +77,7 @@ def test_batchnorm_fold(): assert np.allclose(delay_std.asnumpy(), expect6, rtol=1.e-7, atol=1.e-5) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_batchnorm_fold2(): @@ -100,7 +100,7 @@ def test_batchnorm_fold2(): assert np.allclose(delay_std.asnumpy(), expect6, rtol=1.e-7, atol=1.e-5) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_batchnorm_fold_freeze(): diff --git a/tests/st/ops/gpu/test_batchnorm_op.py b/tests/st/ops/gpu/test_batchnorm_op.py index 036927be5c3..e88119d4070 100644 --- a/tests/st/ops/gpu/test_batchnorm_op.py +++ b/tests/st/ops/gpu/test_batchnorm_op.py @@ -47,7 +47,7 @@ class Grad(Cell): return gout -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_train_forward(): @@ -101,7 +101,7 @@ def test_train_forward(): output = bn_net(Tensor(x)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_train_backward(): @@ -137,7 +137,7 @@ def test_train_backward(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_train_stats_false_forward(): @@ -181,7 +181,7 @@ def test_train_stats_false_forward(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_infer_backward(): @@ -229,7 +229,7 @@ class GradByListNet(Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_1d_train(): @@ -298,7 +298,7 @@ def test_1d_train(): assert np.allclose(bn_net.bn1.moving_variance.asnumpy(), data[6], atol=1.0e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_1d_eval(): diff --git a/tests/st/ops/gpu/test_bce_with_logits_loss.py b/tests/st/ops/gpu/test_bce_with_logits_loss.py index a66da7a09a4..b39ecf2a8cf 100644 --- a/tests/st/ops/gpu/test_bce_with_logits_loss.py +++ b/tests/st/ops/gpu/test_bce_with_logits_loss.py @@ -32,7 +32,7 @@ class Net(nn.Cell): return self.loss(predict, target, weight, pos_weight) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reduction_none_testcases(): @@ -70,7 +70,7 @@ def test_reduction_none_testcases(): np.testing.assert_almost_equal(expected, output.asnumpy(), decimal=3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reduction_mean_testcases(): diff --git a/tests/st/ops/gpu/test_binary_cross_entropy_op.py b/tests/st/ops/gpu/test_binary_cross_entropy_op.py index 0d89a004483..9497b5c7b7a 100644 --- a/tests/st/ops/gpu/test_binary_cross_entropy_op.py +++ b/tests/st/ops/gpu/test_binary_cross_entropy_op.py @@ -34,7 +34,7 @@ class Net(nn.Cell): return self.BinaryCrossEntropy(x, y, weight) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_binary_cross_entropy_loss(): @@ -73,7 +73,7 @@ class Grad(nn.Cell): return gout -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_binary_cross_entropy_loss_grad(): diff --git a/tests/st/ops/gpu/test_boundingbox_decode_op.py b/tests/st/ops/gpu/test_boundingbox_decode_op.py index 8400ee02b9d..5f5d36d9750 100644 --- a/tests/st/ops/gpu/test_boundingbox_decode_op.py +++ b/tests/st/ops/gpu/test_boundingbox_decode_op.py @@ -32,7 +32,7 @@ class NetBoundingBoxDecode(nn.Cell): def construct(self, anchor, groundtruth): return self.decode(anchor, groundtruth) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_boundingbox_decode(): diff --git a/tests/st/ops/gpu/test_boundingbox_encode_op.py b/tests/st/ops/gpu/test_boundingbox_encode_op.py index c34e0e0e8e4..43b03a16aa0 100644 --- a/tests/st/ops/gpu/test_boundingbox_encode_op.py +++ b/tests/st/ops/gpu/test_boundingbox_encode_op.py @@ -53,7 +53,7 @@ def bbox2delta(proposals, gt, means, stds): return deltas -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_boundingbox_encode(): diff --git a/tests/st/ops/gpu/test_broadcast_op.py b/tests/st/ops/gpu/test_broadcast_op.py index eb9f0dac77f..e19404f7efb 100644 --- a/tests/st/ops/gpu/test_broadcast_op.py +++ b/tests/st/ops/gpu/test_broadcast_op.py @@ -21,7 +21,7 @@ from mindspore.common.tensor import Tensor from mindspore.ops import operations as P -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nobroadcast(): @@ -92,7 +92,7 @@ def test_nobroadcast(): assert np.allclose(output_ms.asnumpy(), output_np) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nobroadcast_fp16(): @@ -155,7 +155,7 @@ def test_nobroadcast_fp16(): assert np.allclose(output_ms.asnumpy(), output_np) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast(): @@ -226,7 +226,7 @@ def test_broadcast(): assert np.allclose(output_ms.asnumpy(), output_np) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_diff_dims(): @@ -357,7 +357,7 @@ def test_broadcast_diff_dims_float64(): assert np.allclose(output_ms.asnumpy(), output_np) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_fp16(): @@ -420,7 +420,7 @@ def test_broadcast_fp16(): assert np.allclose(output_ms.asnumpy(), output_np) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_divnonan_int8(): @@ -440,7 +440,7 @@ def test_divnonan_int8(): assert np.allclose(output_ms.asnumpy(), x2_np_zero) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_divnonan_uint8(): diff --git a/tests/st/ops/gpu/test_broadcast_to_ops.py b/tests/st/ops/gpu/test_broadcast_to_ops.py index 9e350c35df1..efa2b9e6159 100644 --- a/tests/st/ops/gpu/test_broadcast_to_ops.py +++ b/tests/st/ops/gpu/test_broadcast_to_ops.py @@ -21,7 +21,7 @@ from mindspore.common.tensor import Tensor from mindspore.ops import operations as P -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast(): @@ -45,7 +45,7 @@ def test_broadcast(): assert np.allclose(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_dyn_init(): diff --git a/tests/st/ops/gpu/test_cast_op.py b/tests/st/ops/gpu/test_cast_op.py index 527fc0d4568..576fd0735b5 100644 --- a/tests/st/ops/gpu/test_cast_op.py +++ b/tests/st/ops/gpu/test_cast_op.py @@ -53,7 +53,7 @@ class NetDynamic(Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_cast(): @@ -224,7 +224,7 @@ def test_cast9(): type1 = output[1].asnumpy().dtype assert type1 == 'float16' -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_cast10(): @@ -292,7 +292,7 @@ def test_cast13(): type1 = output[1].asnumpy().dtype assert type1 == 'float16' -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_cast14(): @@ -547,7 +547,7 @@ def test_cast28(): type1 = output[1].asnumpy().dtype assert type1 == 'int16' -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_cast29(): @@ -599,7 +599,7 @@ def test_cast31(): assert type1 == 'uint32' -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_cast32(): diff --git a/tests/st/ops/gpu/test_check_valid_op.py b/tests/st/ops/gpu/test_check_valid_op.py index fe90c8446b7..80c6319fb3d 100644 --- a/tests/st/ops/gpu/test_check_valid_op.py +++ b/tests/st/ops/gpu/test_check_valid_op.py @@ -47,13 +47,13 @@ def check_valid(nptype): output = boundingbox_decode(anchor_box, image_metas_box) assert np.array_equal(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_check_valid_float32(): check_valid(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_check_valid_float16(): diff --git a/tests/st/ops/gpu/test_concatv2_op.py b/tests/st/ops/gpu/test_concatv2_op.py index 20fc3ac7416..7ed506e4138 100644 --- a/tests/st/ops/gpu/test_concatv2_op.py +++ b/tests/st/ops/gpu/test_concatv2_op.py @@ -51,13 +51,13 @@ def axis32(nptype): [3., 6., 7.]]]).astype(nptype) assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_axis32_float64(): axis32(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_axis32_float32(): @@ -112,13 +112,13 @@ def axis43(nptype): [14., 15., 21., 22., 23.]]]]).astype(nptype) assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_axis43_float64(): axis43(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_axis43_float32(): @@ -130,7 +130,7 @@ def test_axis43_float32(): def test_axis43_int16(): axis43(np.int16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_axis43_uint8(): @@ -167,13 +167,13 @@ def axis21(nptype): [2., 3., 3., 4., 5.]]).astype(nptype) assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_axis21_float64(): axis21(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_axis21_float32(): @@ -185,7 +185,7 @@ def test_axis21_float32(): def test_axis21_int16(): axis21(np.int16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_axis21_uint8(): @@ -225,13 +225,13 @@ def concat_3i(nptype): diff = output_ms.asnumpy() - output_np assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_concat_3i_float64(): concat_3i(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_concat_3i_float32(): @@ -298,13 +298,13 @@ def concat_4i(nptype): diff = output_ms.asnumpy() - output_np assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_concat_4i_float64(): concat_4i(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_concat_4i_float32(): diff --git a/tests/st/ops/gpu/test_conv2d_backprop_filter_op.py b/tests/st/ops/gpu/test_conv2d_backprop_filter_op.py index 9171deeb04d..f5a2d14edf4 100644 --- a/tests/st/ops/gpu/test_conv2d_backprop_filter_op.py +++ b/tests/st/ops/gpu/test_conv2d_backprop_filter_op.py @@ -47,7 +47,7 @@ class Conv2dFilter(nn.Cell): return self.conv_filter(out, x, self.get_shape(w)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_conv2d_backprop_filter(): diff --git a/tests/st/ops/gpu/test_conv2d_backprop_input_op.py b/tests/st/ops/gpu/test_conv2d_backprop_input_op.py index f3cc2a920dd..c5c9c3a86ad 100644 --- a/tests/st/ops/gpu/test_conv2d_backprop_input_op.py +++ b/tests/st/ops/gpu/test_conv2d_backprop_input_op.py @@ -46,7 +46,7 @@ class Conv2dInput(nn.Cell): return self.conv_input(out, w, self.get_shape(x)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_conv2d_backprop_input(): diff --git a/tests/st/ops/gpu/test_conv2d_op.py b/tests/st/ops/gpu/test_conv2d_op.py index d5919c5b1c8..8a8482ba9c6 100644 --- a/tests/st/ops/gpu/test_conv2d_op.py +++ b/tests/st/ops/gpu/test_conv2d_op.py @@ -43,7 +43,7 @@ class NetConv2d(nn.Cell): return self.conv(x, w) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_conv2d(): @@ -86,7 +86,7 @@ class NetConv(nn.Cell): return self.conv(self.x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_conv(): @@ -190,7 +190,7 @@ class NetConv2dDynamic(nn.Cell): return x_conv -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_conv2d_dynamic(): @@ -251,7 +251,7 @@ class NetConvNHWC(nn.Cell): return self.conv(self.x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_conv_NHWC(): diff --git a/tests/st/ops/gpu/test_conv3dtranspose_op.py b/tests/st/ops/gpu/test_conv3dtranspose_op.py index 3c15387d686..cf892298db8 100644 --- a/tests/st/ops/gpu/test_conv3dtranspose_op.py +++ b/tests/st/ops/gpu/test_conv3dtranspose_op.py @@ -40,7 +40,7 @@ class NetConv3dTranspose(nn.Cell): return self.conv_trans(x, w) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_conv3d_transpose(): diff --git a/tests/st/ops/gpu/test_correction_mul_grad_op.py b/tests/st/ops/gpu/test_correction_mul_grad_op.py index efcd257ea17..fe1d9f97839 100644 --- a/tests/st/ops/gpu/test_correction_mul_grad_op.py +++ b/tests/st/ops/gpu/test_correction_mul_grad_op.py @@ -36,7 +36,7 @@ class Net(nn.Cell): return dx, d_batch_std -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_correction_mul_grad(): diff --git a/tests/st/ops/gpu/test_correction_mul_op.py b/tests/st/ops/gpu/test_correction_mul_op.py index 4767dee368d..335df895b1d 100644 --- a/tests/st/ops/gpu/test_correction_mul_op.py +++ b/tests/st/ops/gpu/test_correction_mul_op.py @@ -35,7 +35,7 @@ class Net(nn.Cell): return self.op(x, batch_var, moving_var) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_correction_mul(): diff --git a/tests/st/ops/gpu/test_cos_op.py b/tests/st/ops/gpu/test_cos_op.py index 0f512cf2d43..7bdba875a50 100644 --- a/tests/st/ops/gpu/test_cos_op.py +++ b/tests/st/ops/gpu/test_cos_op.py @@ -41,7 +41,7 @@ def test_cos_float16(): def test_cos_float32(): cos(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_cos_float64(): diff --git a/tests/st/ops/gpu/test_crop_and_resize.py b/tests/st/ops/gpu/test_crop_and_resize.py index 6be54e6545e..8309dbd59e9 100644 --- a/tests/st/ops/gpu/test_crop_and_resize.py +++ b/tests/st/ops/gpu/test_crop_and_resize.py @@ -190,7 +190,7 @@ def test_crop_and_resize_float16_nearest(datatype=np.float16): assert np.all(abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_crop_and_resize_float32_bilinear(datatype=np.float32): @@ -239,7 +239,7 @@ def test_crop_and_resize_float32_bilinear(datatype=np.float32): assert np.all(abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_crop_and_resize_float64_nearest(datatype=np.float64): @@ -284,7 +284,7 @@ def test_crop_and_resize_float64_nearest(datatype=np.float64): assert np.all(abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_crop_and_resize_int64_bilinearv2(datatype=np.int64): diff --git a/tests/st/ops/gpu/test_ctcloss_op.py b/tests/st/ops/gpu/test_ctcloss_op.py index ff7f5df5e6a..3b126b941a5 100644 --- a/tests/st/ops/gpu/test_ctcloss_op.py +++ b/tests/st/ops/gpu/test_ctcloss_op.py @@ -47,7 +47,7 @@ class GradData(nn.Cell): return self.grad(self.network)(probs, indices, labels, input_lengths) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ctcloss(): diff --git a/tests/st/ops/gpu/test_cudnn_inplace_fusion.py b/tests/st/ops/gpu/test_cudnn_inplace_fusion.py index c10f94a88b5..12244a2e1d9 100644 --- a/tests/st/ops/gpu/test_cudnn_inplace_fusion.py +++ b/tests/st/ops/gpu/test_cudnn_inplace_fusion.py @@ -47,7 +47,7 @@ class Conv2dBpropInputInplace(nn.Cell): return y -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_inplace_fusion1(): diff --git a/tests/st/ops/gpu/test_cumprod_op.py b/tests/st/ops/gpu/test_cumprod_op.py index 64b9e486f4d..96303f3abd4 100644 --- a/tests/st/ops/gpu/test_cumprod_op.py +++ b/tests/st/ops/gpu/test_cumprod_op.py @@ -126,7 +126,7 @@ def cum_prod(nptype): assert np.all(diff6 < error6) assert output[6].shape == expect6.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_cum_prod_uint8(): @@ -144,13 +144,13 @@ def test_cum_prod_int8(): def test_cum_prod_int32(): cum_prod(np.int32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_cum_prod_float16(): cum_prod(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_cum_prod_float32(): diff --git a/tests/st/ops/gpu/test_cumsum_op.py b/tests/st/ops/gpu/test_cumsum_op.py index 190975d6dda..5f270cb5a8e 100644 --- a/tests/st/ops/gpu/test_cumsum_op.py +++ b/tests/st/ops/gpu/test_cumsum_op.py @@ -132,7 +132,7 @@ def cum_sum(nptype): def test_cum_sum_uint8(): cum_sum(np.uint8) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_cum_sum_int8(): @@ -144,13 +144,13 @@ def test_cum_sum_int8(): def test_cum_sum_int32(): cum_sum(np.int32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_cum_sum_float16(): cum_sum(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_cum_sum_float32(): diff --git a/tests/st/ops/gpu/test_dense_op.py b/tests/st/ops/gpu/test_dense_op.py index a8cba4c1db2..cbb76b7e717 100644 --- a/tests/st/ops/gpu/test_dense_op.py +++ b/tests/st/ops/gpu/test_dense_op.py @@ -34,7 +34,7 @@ class BiasAdd(nn.Cell): return self.ba(x, b) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_biasadd(): @@ -98,7 +98,7 @@ class DenseNet(nn.Cell): return self.dense(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dx(): @@ -128,7 +128,7 @@ def test_dx(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dx_ND(): @@ -169,7 +169,7 @@ def test_dx_ND(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dw(): @@ -206,7 +206,7 @@ def test_dw(): assert np.all(-diff < db_error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dw_ND(): @@ -259,7 +259,7 @@ class Grad(nn.Cell): return self.grad(self.network)(input_, bias, dy) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_biasadd_3d(): @@ -326,7 +326,7 @@ def test_biasadd_3d(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_biasadd_4d(): @@ -455,7 +455,7 @@ class BiasAddDynamic(nn.Cell): output = self.ba(x, b) return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_bias_add_dynamic_two_inputs(): diff --git a/tests/st/ops/gpu/test_depthtospace_op.py b/tests/st/ops/gpu/test_depthtospace_op.py index 7079af26037..67d2dc8f408 100644 --- a/tests/st/ops/gpu/test_depthtospace_op.py +++ b/tests/st/ops/gpu/test_depthtospace_op.py @@ -75,13 +75,13 @@ def DepthToSpace_pynative(nptype, block_size=2, input_shape=(1, 12, 1, 1)): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_depthtospace_graph_float32(): DepthToSpace(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_depthtospace_graph_float16(): @@ -135,13 +135,13 @@ def test_depthtospace_graph_uint32(): def test_depthtospace_graph_uint64(): DepthToSpace(np.uint64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_depthtospace_pynative_float32(): DepthToSpace_pynative(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_depthtospace_pynative_float16(): diff --git a/tests/st/ops/gpu/test_div_op.py b/tests/st/ops/gpu/test_div_op.py index d07ed0d21e1..a7d6ccedc61 100644 --- a/tests/st/ops/gpu/test_div_op.py +++ b/tests/st/ops/gpu/test_div_op.py @@ -90,13 +90,13 @@ def div(nptype): assert np.all(diff4 < error4) assert output4.shape == expect4.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_div_float64(): div(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_div_float32(): diff --git a/tests/st/ops/gpu/test_dropout.py b/tests/st/ops/gpu/test_dropout.py index f00a0c091ff..b4d0b62139b 100644 --- a/tests/st/ops/gpu/test_dropout.py +++ b/tests/st/ops/gpu/test_dropout.py @@ -30,7 +30,7 @@ class Net(nn.Cell): return self.drop(x_) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dropout(): @@ -67,7 +67,7 @@ class DropoutDynamic(nn.Cell): return self.drop(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dropout_dynamic(): diff --git a/tests/st/ops/gpu/test_dtype_op.py b/tests/st/ops/gpu/test_dtype_op.py index bb316cea721..40d96fe9d71 100644 --- a/tests/st/ops/gpu/test_dtype_op.py +++ b/tests/st/ops/gpu/test_dtype_op.py @@ -116,14 +116,14 @@ def test_dtype_float16(): dtype_with_testcase(ms.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dtype_float32(): dtype_with_testcase(ms.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dtype_float64(): @@ -186,14 +186,14 @@ def test_dynamic_dtype_float16(): dtype_dynamic_with_testcase(ms.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dynamic_dtype_float32(): dtype_dynamic_with_testcase(ms.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dynamic_dtype_float64(): diff --git a/tests/st/ops/gpu/test_dynamic_shape_op.py b/tests/st/ops/gpu/test_dynamic_shape_op.py index e7929938d26..bb206210c48 100644 --- a/tests/st/ops/gpu/test_dynamic_shape_op.py +++ b/tests/st/ops/gpu/test_dynamic_shape_op.py @@ -98,13 +98,13 @@ def dynamic_shape(np_type): def test_dynamic_shape_int32(): dynamic_shape(np.int32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dynamic_shape_float16(): dynamic_shape(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dynamic_shape_float32(): diff --git a/tests/st/ops/gpu/test_elu_grad_op.py b/tests/st/ops/gpu/test_elu_grad_op.py index 8e21ed1e4fb..863aca95a4a 100644 --- a/tests/st/ops/gpu/test_elu_grad_op.py +++ b/tests/st/ops/gpu/test_elu_grad_op.py @@ -31,7 +31,7 @@ class NetEluGrad(nn.Cell): return self.eluGrad(dy, x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_elu_grad_fp16(): @@ -46,7 +46,7 @@ def test_elu_grad_fp16(): diff = output.asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_elu_grad_fp32(): diff --git a/tests/st/ops/gpu/test_elu_op.py b/tests/st/ops/gpu/test_elu_op.py index c0fbb792b8e..daa0fd9aee7 100644 --- a/tests/st/ops/gpu/test_elu_op.py +++ b/tests/st/ops/gpu/test_elu_op.py @@ -30,7 +30,7 @@ class NetElu(nn.Cell): return self.elu(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_elu_fp16(): @@ -50,7 +50,7 @@ def test_elu_fp16(): diff = output.asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_elu_fp32(): diff --git a/tests/st/ops/gpu/test_embedding_lookup_op.py b/tests/st/ops/gpu/test_embedding_lookup_op.py index d6cc0d93a2a..efd6c176564 100644 --- a/tests/st/ops/gpu/test_embedding_lookup_op.py +++ b/tests/st/ops/gpu/test_embedding_lookup_op.py @@ -49,14 +49,14 @@ def embeddinglookup_testcase(nptype): np.testing.assert_almost_equal(expect, output.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_embeddinglookup_float32(): embeddinglookup_testcase(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_embeddinglookup_float16(): diff --git a/tests/st/ops/gpu/test_equal_op.py b/tests/st/ops/gpu/test_equal_op.py index cb3e40378cc..bbd8179dfa4 100644 --- a/tests/st/ops/gpu/test_equal_op.py +++ b/tests/st/ops/gpu/test_equal_op.py @@ -136,7 +136,7 @@ def test_equal(): assert output.shape == expect[i].shape print('test [%d/%d] passed!' % (i, len(x))) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_notequal(): @@ -182,7 +182,7 @@ def test_notequal(): assert output.shape == expect[i].shape print('test [%d/%d] passed!' % (i, len(x))) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_greaterqual(): @@ -233,7 +233,7 @@ def test_greaterqual(): print('test [%d/%d] passed!' % (i, len(x))) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_equal_dynamic_shape(): diff --git a/tests/st/ops/gpu/test_equalcount_op.py b/tests/st/ops/gpu/test_equalcount_op.py index a3f629e7cc6..9a500af7ee6 100644 --- a/tests/st/ops/gpu/test_equalcount_op.py +++ b/tests/st/ops/gpu/test_equalcount_op.py @@ -31,7 +31,7 @@ class NetEqualCount(nn.Cell): return self.equalcount(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_equalcount(): diff --git a/tests/st/ops/gpu/test_erf_op.py b/tests/st/ops/gpu/test_erf_op.py index 93188bbc6cd..5fa3431b013 100644 --- a/tests/st/ops/gpu/test_erf_op.py +++ b/tests/st/ops/gpu/test_erf_op.py @@ -34,7 +34,7 @@ class NetErf(nn.Cell): return self.erf(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_erf_fp32(): @@ -45,7 +45,7 @@ def test_erf_fp32(): tol = 1e-6 assert (np.abs(output.asnumpy() - expect) < tol).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_erf_fp16(): diff --git a/tests/st/ops/gpu/test_erfc_op.py b/tests/st/ops/gpu/test_erfc_op.py index 01bdbbc0a4f..5da242fe051 100644 --- a/tests/st/ops/gpu/test_erfc_op.py +++ b/tests/st/ops/gpu/test_erfc_op.py @@ -34,7 +34,7 @@ class NetErfc(nn.Cell): return self.erfc(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_erfc_fp32(): @@ -45,7 +45,7 @@ def test_erfc_fp32(): tol = 1e-6 assert (np.abs(output.asnumpy() - expect) < tol).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_erfc_fp16(): diff --git a/tests/st/ops/gpu/test_exp_op.py b/tests/st/ops/gpu/test_exp_op.py index d84ba4a3b35..dc0726d4a2c 100644 --- a/tests/st/ops/gpu/test_exp_op.py +++ b/tests/st/ops/gpu/test_exp_op.py @@ -31,7 +31,7 @@ class NetExp(nn.Cell): return self.exp(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_exp(): diff --git a/tests/st/ops/gpu/test_expand_dims.py b/tests/st/ops/gpu/test_expand_dims.py index a67340c7edd..bdb5c6ac3b5 100644 --- a/tests/st/ops/gpu/test_expand_dims.py +++ b/tests/st/ops/gpu/test_expand_dims.py @@ -119,7 +119,7 @@ def test_net_float16(): assert np.all(output.asnumpy() == np.expand_dims(x, -1)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net_float32(): @@ -130,7 +130,7 @@ def test_net_float32(): assert np.all(output.asnumpy() == np.expand_dims(x, -1)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net_float64(): diff --git a/tests/st/ops/gpu/test_extract_image_patches_op.py b/tests/st/ops/gpu/test_extract_image_patches_op.py index e7bfdf07b64..cf2e41000b2 100644 --- a/tests/st/ops/gpu/test_extract_image_patches_op.py +++ b/tests/st/ops/gpu/test_extract_image_patches_op.py @@ -31,7 +31,7 @@ class Net(nn.Cell): return self.extractimagepatches(input_tensor) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_extract_image_patches_valid(): @@ -62,7 +62,7 @@ def test_extract_image_patches_valid(): assert np.all(output.asnumpy() == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_extract_image_patches_same(): diff --git a/tests/st/ops/gpu/test_fake_quant_perchannel.py b/tests/st/ops/gpu/test_fake_quant_perchannel.py index c4ddc285af7..26d8ac09083 100644 --- a/tests/st/ops/gpu/test_fake_quant_perchannel.py +++ b/tests/st/ops/gpu/test_fake_quant_perchannel.py @@ -36,7 +36,7 @@ class Net(nn.Cell): return self.op(x, minq, maxq) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_fake_quant_perchannel1(): @@ -77,7 +77,7 @@ def test_fake_quant_perchannel2(): assert np.all(np.abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_fake_quant_perchannel3(): @@ -387,7 +387,7 @@ def test_fake_quant_perchannel15(): assert np.all(np.abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_fake_quant_perchannel16(): @@ -408,7 +408,7 @@ def test_fake_quant_perchannel16(): assert np.all(np.abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_fake_quant_perchannel17(): @@ -571,7 +571,7 @@ def test_fake_quant_perchannel23(): assert np.all(np.abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_fake_quant_perchannel24(): @@ -598,7 +598,7 @@ def test_fake_quant_perchannel24(): assert np.all(np.abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_fake_quant_perchannel25(): diff --git a/tests/st/ops/gpu/test_fake_quant_perchannel_grad.py b/tests/st/ops/gpu/test_fake_quant_perchannel_grad.py index dde6279efa7..d50bac0b18b 100644 --- a/tests/st/ops/gpu/test_fake_quant_perchannel_grad.py +++ b/tests/st/ops/gpu/test_fake_quant_perchannel_grad.py @@ -309,7 +309,7 @@ def test_fake_quant_grad11(): assert np.all(np.abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_fake_quant_grad12(): @@ -341,7 +341,7 @@ def test_fake_quant_grad12(): assert np.all(np.abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_fake_quant_grad13(): diff --git a/tests/st/ops/gpu/test_fake_quant_perlayer.py b/tests/st/ops/gpu/test_fake_quant_perlayer.py index 49a24d076d8..1f3753bb1f3 100644 --- a/tests/st/ops/gpu/test_fake_quant_perlayer.py +++ b/tests/st/ops/gpu/test_fake_quant_perlayer.py @@ -42,7 +42,7 @@ class Net(nn.Cell): return self.fake_quant(x, minq, maxq) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_fake_quant1(): @@ -346,7 +346,7 @@ def test_fake_quant15(): assert np.all(np.abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_fake_quant16(): @@ -366,7 +366,7 @@ def test_fake_quant16(): assert np.all(np.abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_fake_quant17(): diff --git a/tests/st/ops/gpu/test_fake_quant_perlayer_grad.py b/tests/st/ops/gpu/test_fake_quant_perlayer_grad.py index 5e6a7ebf142..3d7b48cb5fd 100644 --- a/tests/st/ops/gpu/test_fake_quant_perlayer_grad.py +++ b/tests/st/ops/gpu/test_fake_quant_perlayer_grad.py @@ -179,7 +179,7 @@ def test_fake_quant_grad7(): assert np.all(np.abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_fake_quant_grad8(): @@ -200,7 +200,7 @@ def test_fake_quant_grad8(): assert np.all(np.abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_fake_quant_grad9(): diff --git a/tests/st/ops/gpu/test_flatten_grad_op.py b/tests/st/ops/gpu/test_flatten_grad_op.py index 0ee6c403189..05d44449299 100644 --- a/tests/st/ops/gpu/test_flatten_grad_op.py +++ b/tests/st/ops/gpu/test_flatten_grad_op.py @@ -32,7 +32,7 @@ class NetFlattenGrad(nn.Cell): return self.flattengrad(x, self.type) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_flatten_grad(): diff --git a/tests/st/ops/gpu/test_flatten_op.py b/tests/st/ops/gpu/test_flatten_op.py index 504d7c06b3c..4a887940b3a 100644 --- a/tests/st/ops/gpu/test_flatten_op.py +++ b/tests/st/ops/gpu/test_flatten_op.py @@ -74,7 +74,7 @@ class NetLastFlatten(nn.Cell): return x -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_flatten(): @@ -91,7 +91,7 @@ def test_flatten(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_all_flatten(): @@ -125,7 +125,7 @@ def test_first_flatten(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_last_flatten(): diff --git a/tests/st/ops/gpu/test_float_status_op.py b/tests/st/ops/gpu/test_float_status_op.py index 0edb9b92816..6fb37366e09 100644 --- a/tests/st/ops/gpu/test_float_status_op.py +++ b/tests/st/ops/gpu/test_float_status_op.py @@ -64,7 +64,7 @@ x2 = np.array([[np.inf, 1, 88.0, 0]]).astype(np.float32) x3 = np.array([[1, 2], [3, 4], [5.0, 88.0]]).astype(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float16, np.float32, np.float64]) @@ -88,7 +88,7 @@ def test_status(dtype): assert output3.asnumpy()[0] == expect3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float16, np.float32, np.float64]) @@ -112,7 +112,7 @@ def test_nan(dtype): assert (output3.asnumpy() == expect3).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float16, np.float32, np.float64]) @@ -136,7 +136,7 @@ def test_inf(dtype): assert (output3.asnumpy() == expect3).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float16, np.float32, np.float64]) diff --git a/tests/st/ops/gpu/test_floordiv_op.py b/tests/st/ops/gpu/test_floordiv_op.py index 26c32f92136..dd59623866d 100644 --- a/tests/st/ops/gpu/test_floordiv_op.py +++ b/tests/st/ops/gpu/test_floordiv_op.py @@ -29,7 +29,7 @@ class NetFloorDiv(nn.Cell): def construct(self, x, y): return self.floordiv(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_floor_div(): diff --git a/tests/st/ops/gpu/test_ftrl_op.py b/tests/st/ops/gpu/test_ftrl_op.py index e9518f7762f..fa3d331b611 100644 --- a/tests/st/ops/gpu/test_ftrl_op.py +++ b/tests/st/ops/gpu/test_ftrl_op.py @@ -41,7 +41,7 @@ class NetFtrl(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ftrl(): diff --git a/tests/st/ops/gpu/test_gatherV2_op.py b/tests/st/ops/gpu/test_gatherV2_op.py index 2102f8c9e6d..2424f270e06 100644 --- a/tests/st/ops/gpu/test_gatherV2_op.py +++ b/tests/st/ops/gpu/test_gatherV2_op.py @@ -32,7 +32,7 @@ class GatherNet(nn.Cell): return self.gather(x, indices, 1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather0(): @@ -856,7 +856,7 @@ class GatherNet1(nn.Cell): return self.gather(x, indices, -1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather1(): @@ -910,7 +910,7 @@ class GatherNet2(nn.Cell): return self.gather(x, indices, 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather2(): @@ -959,7 +959,7 @@ class GatherNetDynamic(nn.Cell): return self.gather(x, indices, self.axis) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gatherV2_dyn_ab(): @@ -990,7 +990,7 @@ def test_gatherV2_dyn_ab(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gatherV2_dyn_a(): @@ -1066,7 +1066,7 @@ def test_gatherV2_dyn_a(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gatherV2_dyn_b(): @@ -1137,7 +1137,7 @@ def test_gatherV2_dyn_b(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather1_float64(): @@ -1182,7 +1182,7 @@ def test_gather1_float64(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather1_int32(): diff --git a/tests/st/ops/gpu/test_gather_grad_op.py b/tests/st/ops/gpu/test_gather_grad_op.py index ebf6f29021f..c0a378283bf 100644 --- a/tests/st/ops/gpu/test_gather_grad_op.py +++ b/tests/st/ops/gpu/test_gather_grad_op.py @@ -33,7 +33,7 @@ class GatherDNet(nn.Cell): def construct(self, x, index): return self.gather_d(x, self.dim, index) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_grad_graph_int32_fp32(): @@ -52,7 +52,7 @@ def test_gather_grad_graph_int32_fp32(): diff = output[0].asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_grad_graph_int64_fp32(): @@ -71,7 +71,7 @@ def test_gather_grad_graph_int64_fp32(): diff = output[0].asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_grad_graph_int32_fp16(): @@ -90,7 +90,7 @@ def test_gather_grad_graph_int32_fp16(): diff = output[0].asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_grad_graph_int64_fp16(): @@ -109,7 +109,7 @@ def test_gather_grad_graph_int64_fp16(): diff = output[0].asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_grad_pynative_int32_fp32(): @@ -126,7 +126,7 @@ def test_gather_grad_pynative_int32_fp32(): diff = output.asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_grad_pynative_int64_fp32(): @@ -143,7 +143,7 @@ def test_gather_grad_pynative_int64_fp32(): diff = output.asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_grad_pynative_int32_fp16(): @@ -160,7 +160,7 @@ def test_gather_grad_pynative_int32_fp16(): diff = output.asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_grad_pynative_int64_fp16(): diff --git a/tests/st/ops/gpu/test_gather_op.py b/tests/st/ops/gpu/test_gather_op.py index 69a3cec1c8d..de76655465a 100644 --- a/tests/st/ops/gpu/test_gather_op.py +++ b/tests/st/ops/gpu/test_gather_op.py @@ -32,7 +32,7 @@ class GatherNet(nn.Cell): def construct(self, x, index): return self.gather(x, self.dim, index) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_pynative_fp32_int32(): @@ -46,7 +46,7 @@ def test_gather_pynative_fp32_int32(): diff = output.asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_pynative_fp32_int64(): @@ -60,7 +60,7 @@ def test_gather_pynative_fp32_int64(): diff = output.asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_pynative_fp16_int32(): @@ -74,7 +74,7 @@ def test_gather_pynative_fp16_int32(): diff = output.asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_pynative_fp16_int64(): diff --git a/tests/st/ops/gpu/test_gathernd_op.py b/tests/st/ops/gpu/test_gathernd_op.py index 96c9f1fe807..49f0a150bf3 100644 --- a/tests/st/ops/gpu/test_gathernd_op.py +++ b/tests/st/ops/gpu/test_gathernd_op.py @@ -40,7 +40,7 @@ def gathernd0(nptype): assert np.array_equal(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gathernd0_float64(): @@ -116,13 +116,13 @@ def gathernd1(nptype): assert np.array_equal(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gathernd1_float64(): gathernd1(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gathernd1_float32(): @@ -175,13 +175,13 @@ def gathernd2(nptype): assert np.array_equal(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gathernd2_float64(): gathernd2(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gathernd2_float32(): @@ -225,7 +225,7 @@ def test_gathernd_bool(): assert np.array_equal(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gathernd_indices_int64(): diff --git a/tests/st/ops/gpu/test_gelu_grad_op.py b/tests/st/ops/gpu/test_gelu_grad_op.py index 5eaecf4e089..003c379ae06 100644 --- a/tests/st/ops/gpu/test_gelu_grad_op.py +++ b/tests/st/ops/gpu/test_gelu_grad_op.py @@ -45,7 +45,7 @@ class Grad(nn.Cell): return gout -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gelugrad(): @@ -62,7 +62,7 @@ def test_gelugrad(): 0.34266686, 0.57757664, 0.04707306, 0.51536125] assert np.allclose(output[0].asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gelugrad_fp16(): diff --git a/tests/st/ops/gpu/test_gelu_op.py b/tests/st/ops/gpu/test_gelu_op.py index b6e449e081f..93cae5a2fbb 100644 --- a/tests/st/ops/gpu/test_gelu_op.py +++ b/tests/st/ops/gpu/test_gelu_op.py @@ -37,7 +37,7 @@ def GeluCompute(x): return 0.5 * x * (1.0 + np.tanh(np.sqrt(2 / np.pi) * (x + 0.044715 * x * x * x))) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gelu_1d(): @@ -51,7 +51,7 @@ def test_gelu_1d(): assert np.allclose(y_np, y_ms.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gelu_2d(): @@ -65,7 +65,7 @@ def test_gelu_2d(): assert np.allclose(y_np, y_ms.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gelu_4d(): @@ -79,7 +79,7 @@ def test_gelu_4d(): assert np.allclose(y_np, y_ms.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gelu_neg(): @@ -92,7 +92,7 @@ def test_gelu_neg(): assert np.allclose(y_np, y_ms.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gelu_4d_fp16(): diff --git a/tests/st/ops/gpu/test_gpu_convert_to_dynamic_shape_op.py b/tests/st/ops/gpu/test_gpu_convert_to_dynamic_shape_op.py index 765b5bc289d..cc99ddd0126 100644 --- a/tests/st/ops/gpu/test_gpu_convert_to_dynamic_shape_op.py +++ b/tests/st/ops/gpu/test_gpu_convert_to_dynamic_shape_op.py @@ -22,7 +22,7 @@ import mindspore.nn as nn import mindspore.context as context # test to make sure this op actually generates a dynamically shaped output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_convert_to_dyanamic_shape_confirm_dynamic(): @@ -103,13 +103,13 @@ def test_gpu_convert_to_dynamic_shape_bool(): def test_gpu_convert_to_dynamic_shape_float16(): gpu_convert_to_dynamic_shape_float(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_convert_to_dynamic_shape_float32(): gpu_convert_to_dynamic_shape_float(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_convert_to_dynamic_shape_float64(): diff --git a/tests/st/ops/gpu/test_hsigmoid_op.py b/tests/st/ops/gpu/test_hsigmoid_op.py index 09a772357da..0b9a02462a5 100644 --- a/tests/st/ops/gpu/test_hsigmoid_op.py +++ b/tests/st/ops/gpu/test_hsigmoid_op.py @@ -90,21 +90,21 @@ def generate_dynamic_testcase(nptype): np.testing.assert_almost_equal(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_hsigmoid_dynamic_float32(): generate_dynamic_testcase(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_hsigmoid_float32(): generate_testcases(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_hsigmoid_float16(): diff --git a/tests/st/ops/gpu/test_hswish_op.py b/tests/st/ops/gpu/test_hswish_op.py index fec7dc5a24b..7de5f6a7b39 100644 --- a/tests/st/ops/gpu/test_hswish_op.py +++ b/tests/st/ops/gpu/test_hswish_op.py @@ -71,7 +71,7 @@ def generate_test_cases(np_type, mode): judge_result_correct(output[0].asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_hswish_forward_and_backward(): diff --git a/tests/st/ops/gpu/test_identity_op.py b/tests/st/ops/gpu/test_identity_op.py index 127a2611eb5..a94c4c376d3 100644 --- a/tests/st/ops/gpu/test_identity_op.py +++ b/tests/st/ops/gpu/test_identity_op.py @@ -48,14 +48,14 @@ def generate_testcases(nptype): assert id(input_tensor) != id(output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_identity_float64(): generate_testcases(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_identity_float32(): diff --git a/tests/st/ops/gpu/test_in_top_k.py b/tests/st/ops/gpu/test_in_top_k.py index 9136904bd45..cf4d4c820c7 100644 --- a/tests/st/ops/gpu/test_in_top_k.py +++ b/tests/st/ops/gpu/test_in_top_k.py @@ -100,19 +100,19 @@ def in_top_k(nptype): expected_output = np.array([True, True, True]) np.testing.assert_array_equal(output.asnumpy(), expected_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_in_top_k_float16(): in_top_k(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_in_top_k_float32(): in_top_k(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_in_top_k_invalid_input(): diff --git a/tests/st/ops/gpu/test_index_add_op.py b/tests/st/ops/gpu/test_index_add_op.py index 7f5543005a8..3e3df144f83 100644 --- a/tests/st/ops/gpu/test_index_add_op.py +++ b/tests/st/ops/gpu/test_index_add_op.py @@ -35,7 +35,7 @@ class NetIndexAdd(nn.Cell): return z -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_index_add(): @@ -117,7 +117,7 @@ def index_add_forward(nptype): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_index_add_float16(): @@ -132,7 +132,7 @@ def test_index_add_float16(): index_add_forward(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_index_add_int32(): @@ -192,7 +192,7 @@ def test_index_add_float64(): index_add_forward(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_index_add_int16(): @@ -207,7 +207,7 @@ def test_index_add_int16(): index_add_forward(np.int16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_index_add_invalid_inputs(): @@ -288,7 +288,7 @@ def index_add_grad_with_type(nptype): np.testing.assert_array_equal(ygrad.asnumpy(), expect_ygrad) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_index_add_grad_float64(): @@ -303,7 +303,7 @@ def test_index_add_grad_float64(): index_add_grad_with_type(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_index_add_grad_float32(): @@ -348,7 +348,7 @@ def test_index_add_grad_int32(): index_add_grad_with_type(np.int32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_index_add_grad_int16(): diff --git a/tests/st/ops/gpu/test_instancenorm2d.py b/tests/st/ops/gpu/test_instancenorm2d.py index 1cf8c304de4..bdc6b8ff1a4 100644 --- a/tests/st/ops/gpu/test_instancenorm2d.py +++ b/tests/st/ops/gpu/test_instancenorm2d.py @@ -43,7 +43,7 @@ class Net(nn.Cell): shape = F.shape(x) return F.reshape(self.ops(F.reshape(x, (1, -1, shape[2], shape[3]))), shape) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_InstanceNorm2d_fp32(): diff --git a/tests/st/ops/gpu/test_iou_op.py b/tests/st/ops/gpu/test_iou_op.py index 17812f2d179..5ee4cc32e69 100644 --- a/tests/st/ops/gpu/test_iou_op.py +++ b/tests/st/ops/gpu/test_iou_op.py @@ -31,7 +31,7 @@ class NetIOU(nn.Cell): def construct(self, anchor, groundtruth): return self.encode(anchor, groundtruth) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_iou(): diff --git a/tests/st/ops/gpu/test_kl_div_op.py b/tests/st/ops/gpu/test_kl_div_op.py index 64c9845f25d..00a63c2aebe 100644 --- a/tests/st/ops/gpu/test_kl_div_op.py +++ b/tests/st/ops/gpu/test_kl_div_op.py @@ -34,7 +34,7 @@ class Net(nn.Cell): return self.KLDivLoss(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_binary_cross_entropy_loss(): @@ -61,7 +61,7 @@ class Grad(nn.Cell): return gout -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_binary_cross_entropy_loss_grad(): diff --git a/tests/st/ops/gpu/test_l2loss_op.py b/tests/st/ops/gpu/test_l2loss_op.py index 822d5b9ffc8..7477c48daa0 100644 --- a/tests/st/ops/gpu/test_l2loss_op.py +++ b/tests/st/ops/gpu/test_l2loss_op.py @@ -31,7 +31,7 @@ class L2LossNet(nn.Cell): def construct(self, x): return self.l2_loss(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_pynative_fp32_22(): @@ -43,7 +43,7 @@ def test_gather_pynative_fp32_22(): diff = output.asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_pynative_fp16_22(): @@ -55,7 +55,7 @@ def test_gather_pynative_fp16_22(): diff = output.asnumpy() - expect assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_pynative_fp32_14(): diff --git a/tests/st/ops/gpu/test_l2normalize_grad_op.py b/tests/st/ops/gpu/test_l2normalize_grad_op.py index 8171efb1462..05cee382c92 100644 --- a/tests/st/ops/gpu/test_l2normalize_grad_op.py +++ b/tests/st/ops/gpu/test_l2normalize_grad_op.py @@ -30,7 +30,7 @@ class Net(Cell): return self.norm_grad(x, out, dout) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_l2normalize_grad(): diff --git a/tests/st/ops/gpu/test_l2normalize_op.py b/tests/st/ops/gpu/test_l2normalize_op.py index 891eb863447..b6aeef52161 100644 --- a/tests/st/ops/gpu/test_l2normalize_op.py +++ b/tests/st/ops/gpu/test_l2normalize_op.py @@ -30,7 +30,7 @@ class Net(Cell): return self.norm(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_l2normalize(): diff --git a/tests/st/ops/gpu/test_layer_norm_grad_grad_op.py b/tests/st/ops/gpu/test_layer_norm_grad_grad_op.py index 1caad1b8c27..cedb3b08863 100644 --- a/tests/st/ops/gpu/test_layer_norm_grad_grad_op.py +++ b/tests/st/ops/gpu/test_layer_norm_grad_grad_op.py @@ -107,7 +107,7 @@ def LayerNormGradGradReference(x, dy, gamma, epsilon, grad_dx_np, grad_dg_np, gr return d_x, d_dy, d_gamma -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernormgradgrad0(): @@ -184,7 +184,7 @@ def test_layernormgradgrad1(): assert np.allclose(d_gamma_ms.asnumpy(), d_gamma_np, rtol=3e-3, atol=3e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernormgradgrad2(): @@ -298,7 +298,7 @@ def test_layernormgradgrad4(): assert np.allclose(d_gamma_ms.asnumpy(), d_gamma_np, rtol=3e-3, atol=3e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernormgradgrad5(): @@ -523,7 +523,7 @@ def test_layernormgradgrad10(): assert np.allclose(d_gamma_ms.asnumpy(), d_gamma_np, rtol=5e-3, atol=5e-1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernormgradgrad11(): diff --git a/tests/st/ops/gpu/test_layer_norm_grad_op.py b/tests/st/ops/gpu/test_layer_norm_grad_op.py index 4714f24ac34..f0a9e822e36 100644 --- a/tests/st/ops/gpu/test_layer_norm_grad_op.py +++ b/tests/st/ops/gpu/test_layer_norm_grad_op.py @@ -62,7 +62,7 @@ def LayerNormGradReference(x, dy, gamma, epsilon, begin_norm_axis, begin_params_ return dx, dg, db, mean, var -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernormgrad0(): @@ -116,7 +116,7 @@ def test_layernormgrad1(): assert np.allclose(db_ms.asnumpy(), db_np, rtol=1e-6, atol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernormgrad2(): @@ -195,7 +195,7 @@ def test_layernormgrad4(): assert np.allclose(db_ms.asnumpy(), db_np, rtol=1e-6, atol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernormgrad5(): diff --git a/tests/st/ops/gpu/test_layer_norm_op.py b/tests/st/ops/gpu/test_layer_norm_op.py index 6aa5f67d6e9..aff907e8f3b 100644 --- a/tests/st/ops/gpu/test_layer_norm_op.py +++ b/tests/st/ops/gpu/test_layer_norm_op.py @@ -47,7 +47,7 @@ def LayerNormReference(begin_norm_axis, begin_params_axis, x, gamma, beta): return y, mean, var -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernorm0(): @@ -70,7 +70,7 @@ def test_layernorm0(): assert np.allclose(var_ms.asnumpy(), var_np, atol=1e-6) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernorm1(): @@ -93,7 +93,7 @@ def test_layernorm1(): assert np.allclose(var_ms.asnumpy(), var_np, rtol=1e-6, atol=1e-6) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernorm3d_1(): @@ -116,7 +116,7 @@ def test_layernorm3d_1(): assert np.allclose(var_ms.asnumpy(), var_np, rtol=1e-6, atol=1e-6) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernorm3d_2(): @@ -139,7 +139,7 @@ def test_layernorm3d_2(): assert np.allclose(var_ms.asnumpy(), var_np, rtol=1e-6, atol=1e-6) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernorm2d_2(): @@ -161,7 +161,7 @@ def test_layernorm2d_2(): assert np.allclose(var_ms.asnumpy(), var_np, rtol=1e-6, atol=1e-6) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernorm2d_3(): @@ -183,7 +183,7 @@ def test_layernorm2d_3(): assert np.allclose(var_ms.asnumpy(), var_np, rtol=1e-6, atol=1e-6) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernorm2d_4(): diff --git a/tests/st/ops/gpu/test_lessequal_op.py b/tests/st/ops/gpu/test_lessequal_op.py index d3796f453ab..014a7df4777 100644 --- a/tests/st/ops/gpu/test_lessequal_op.py +++ b/tests/st/ops/gpu/test_lessequal_op.py @@ -31,7 +31,7 @@ class Net(Cell): return self.lessequal(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_lessequal(): diff --git a/tests/st/ops/gpu/test_lin_space.py b/tests/st/ops/gpu/test_lin_space.py index edd259ffbe0..4081d9f1a27 100644 --- a/tests/st/ops/gpu/test_lin_space.py +++ b/tests/st/ops/gpu/test_lin_space.py @@ -34,7 +34,7 @@ class LinSpaceNet(Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_lin_space_1(): @@ -51,7 +51,7 @@ def test_lin_space_1(): assert np.allclose(result_ms, result_np) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_lin_shape_2(): @@ -68,7 +68,7 @@ def test_lin_shape_2(): assert np.allclose(result_ms, result_np) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_lin_shape_3(): @@ -84,7 +84,7 @@ def test_lin_shape_3(): assert np.allclose(result_ms, result_np) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_lin_shape_4(): diff --git a/tests/st/ops/gpu/test_local_response_norm_op.py b/tests/st/ops/gpu/test_local_response_norm_op.py index ed164ece37c..026f558f123 100644 --- a/tests/st/ops/gpu/test_local_response_norm_op.py +++ b/tests/st/ops/gpu/test_local_response_norm_op.py @@ -47,7 +47,7 @@ class MSGradNet(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_lrn_ms(): @@ -73,7 +73,7 @@ def test_lrn_ms(): assert np.allclose(output.asnumpy(), y_exp) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_lrn_grad(): diff --git a/tests/st/ops/gpu/test_log1p_op.py b/tests/st/ops/gpu/test_log1p_op.py index e5e3583ff5e..ebd6c914d9b 100644 --- a/tests/st/ops/gpu/test_log1p_op.py +++ b/tests/st/ops/gpu/test_log1p_op.py @@ -33,7 +33,7 @@ class NetLog1p(nn.Cell): return self.log1p(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_log1p_fp32(): @@ -44,7 +44,7 @@ def test_log1p_fp32(): tol = 1e-6 assert (np.abs(output.asnumpy() - expect) < tol).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_log1p_fp16(): diff --git a/tests/st/ops/gpu/test_log_op.py b/tests/st/ops/gpu/test_log_op.py index 47af674f814..391821112bf 100644 --- a/tests/st/ops/gpu/test_log_op.py +++ b/tests/st/ops/gpu/test_log_op.py @@ -31,7 +31,7 @@ class NetLog(nn.Cell): return self.log(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_log(): diff --git a/tests/st/ops/gpu/test_logical_op.py b/tests/st/ops/gpu/test_logical_op.py index 1de7a7f8700..8706ab4a89a 100644 --- a/tests/st/ops/gpu/test_logical_op.py +++ b/tests/st/ops/gpu/test_logical_op.py @@ -68,7 +68,7 @@ def test_logicaland(): assert np.all(output.asnumpy() == np.logical_and(x, y)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_logicalor(): @@ -83,7 +83,7 @@ def test_logicalor(): assert np.all(output.asnumpy() == np.logical_or(x, y)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_logicalnot(): diff --git a/tests/st/ops/gpu/test_logsoftmax_op.py b/tests/st/ops/gpu/test_logsoftmax_op.py index 271be6367b4..fce422fcd84 100644 --- a/tests/st/ops/gpu/test_logsoftmax_op.py +++ b/tests/st/ops/gpu/test_logsoftmax_op.py @@ -23,7 +23,7 @@ from mindspore.ops import composite as C from mindspore.ops import operations as P -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_logsoftmax(): @@ -60,7 +60,7 @@ class Grad(nn.Cell): return gout -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_logsoftmaxgrad(): @@ -101,7 +101,7 @@ def test_logsoftmaxgrad(): assert np.allclose(dx[0].asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_logsoftmaxgrad1(): diff --git a/tests/st/ops/gpu/test_loss.py b/tests/st/ops/gpu/test_loss.py index dc6ffb58e60..9b64b468a1c 100644 --- a/tests/st/ops/gpu/test_loss.py +++ b/tests/st/ops/gpu/test_loss.py @@ -60,13 +60,13 @@ def weighted_loss(nptype): diff = test_data - output.asnumpy() assert np.all(abs(diff) < error_range) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_weighted_loss_float32(): weighted_loss(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_weighted_loss_float64(): @@ -101,13 +101,13 @@ def custom_loss(nptype): def test_custom_loss_float16(): custom_loss(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_custom_loss_float32(): custom_loss(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_custom_loss_float64(): diff --git a/tests/st/ops/gpu/test_matmul_op.py b/tests/st/ops/gpu/test_matmul_op.py index 30198d06dab..b645bc81adb 100644 --- a/tests/st/ops/gpu/test_matmul_op.py +++ b/tests/st/ops/gpu/test_matmul_op.py @@ -53,7 +53,7 @@ class MatMulComposite(nn.Cell): return self.matmul(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_MatMul_dynamic(): @@ -74,7 +74,7 @@ def test_MatMul_dynamic(): np.testing.assert_array_almost_equal(output2.asnumpy(), expect2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_matmul_float64(): @@ -88,7 +88,7 @@ def test_matmul_float64(): expect = np.matmul(x, y) np.testing.assert_array_almost_equal(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_matmul_composite(): diff --git a/tests/st/ops/gpu/test_matrix_inverse_op.py b/tests/st/ops/gpu/test_matrix_inverse_op.py index c11d22d445a..d362696bb0e 100644 --- a/tests/st/ops/gpu/test_matrix_inverse_op.py +++ b/tests/st/ops/gpu/test_matrix_inverse_op.py @@ -33,7 +33,7 @@ class NetMatrixInverse(nn.Cell): return self.matrix_inverse(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_matrix_inverse(): diff --git a/tests/st/ops/gpu/test_maximum_op.py b/tests/st/ops/gpu/test_maximum_op.py index 55874b52b5c..f8e096d8ce6 100644 --- a/tests/st/ops/gpu/test_maximum_op.py +++ b/tests/st/ops/gpu/test_maximum_op.py @@ -43,7 +43,7 @@ class Grad(Cell): return gout -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_maximum(): @@ -67,7 +67,7 @@ def test_maximum(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast(): @@ -191,7 +191,7 @@ def test_broadcast(): assert np.allclose(output_ms[1].asnumpy(), expect_dx2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_diff_dims(): diff --git a/tests/st/ops/gpu/test_maxpool_gpu_op.py b/tests/st/ops/gpu/test_maxpool_gpu_op.py index a2449032c78..5eb8103b75a 100644 --- a/tests/st/ops/gpu/test_maxpool_gpu_op.py +++ b/tests/st/ops/gpu/test_maxpool_gpu_op.py @@ -41,7 +41,7 @@ class Net_Pool2(nn.Cell): return self.maxpool_fun(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_maxpool2d(): @@ -81,7 +81,7 @@ def test_maxpool2d(): assert (output2.asnumpy() == expect_result2).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_max_pool3d_1(): @@ -108,7 +108,7 @@ def test_max_pool3d_1(): assert (output_ms.asnumpy() == expert_result).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_max_pool3d_2(): @@ -135,7 +135,7 @@ def test_max_pool3d_2(): assert (output_ms.asnumpy() == expert_result).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_max_pool3d_3(): @@ -156,7 +156,7 @@ def test_max_pool3d_3(): assert (output_ms.asnumpy() == expert_result).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_max_pool3d_4(): @@ -207,7 +207,7 @@ def test_max_pool3d_4(): assert (output_ms.asnumpy() == expert_result).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_max_pool3d_5(): diff --git a/tests/st/ops/gpu/test_maxpool_grad_gpu_op.py b/tests/st/ops/gpu/test_maxpool_grad_gpu_op.py index 80511d5a465..2d7db277e83 100644 --- a/tests/st/ops/gpu/test_maxpool_grad_gpu_op.py +++ b/tests/st/ops/gpu/test_maxpool_grad_gpu_op.py @@ -44,7 +44,7 @@ class Grad(nn.Cell): return self.grad(self.network)(x1, out, dout, grad) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_maxpool2d_grad(): diff --git a/tests/st/ops/gpu/test_maxpool_with_argmax_gpu_op.py b/tests/st/ops/gpu/test_maxpool_with_argmax_gpu_op.py index e634539942c..ab3916d173c 100644 --- a/tests/st/ops/gpu/test_maxpool_with_argmax_gpu_op.py +++ b/tests/st/ops/gpu/test_maxpool_with_argmax_gpu_op.py @@ -45,7 +45,7 @@ class Grad(Cell): return gout -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_train_forward_backward(): @@ -94,7 +94,7 @@ def test_train_forward_backward(): diff = dx - expect_dx assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_maxpool_with_argmax_2d(): @@ -147,7 +147,7 @@ def test_maxpool_with_argmax_2d(): assert (index.asnumpy() == expect_index_result).all() assert (index2.asnumpy() == expect__index_result2).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_maxpool_with_argmax_2d_fp16(): diff --git a/tests/st/ops/gpu/test_maxpool_with_argmax_grad_gpu_op.py b/tests/st/ops/gpu/test_maxpool_with_argmax_grad_gpu_op.py index 6d9a23702b7..62c53fd02c1 100644 --- a/tests/st/ops/gpu/test_maxpool_with_argmax_grad_gpu_op.py +++ b/tests/st/ops/gpu/test_maxpool_with_argmax_grad_gpu_op.py @@ -31,7 +31,7 @@ class Net_Pool_Grad(nn.Cell): return self.maxpool_grad_fun(x, dy, index) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_maxpool2d_grad(): @@ -73,7 +73,7 @@ def test_maxpool2d_grad(): assert np.allclose(expect_result, output.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_maxpool2d_grad_fp16(): diff --git a/tests/st/ops/gpu/test_minimum_op.py b/tests/st/ops/gpu/test_minimum_op.py index 17546cfc395..c7fdfd5e42c 100644 --- a/tests/st/ops/gpu/test_minimum_op.py +++ b/tests/st/ops/gpu/test_minimum_op.py @@ -44,7 +44,7 @@ class Grad(Cell): return gout -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nobroadcast(): @@ -62,7 +62,7 @@ def test_nobroadcast(): assert np.allclose(output_ms[1].asnumpy(), output1_np) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast(): @@ -187,7 +187,7 @@ def test_broadcast(): assert np.allclose(output_ms[1].asnumpy(), expect_dx2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_diff_dims(): @@ -220,7 +220,7 @@ def test_broadcast_diff_dims(): assert np.allclose(output_ms[1].asnumpy(), expect_dx2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_int32(): diff --git a/tests/st/ops/gpu/test_mirror_pad.py b/tests/st/ops/gpu/test_mirror_pad.py index a14bfa24f61..4d061443e68 100644 --- a/tests/st/ops/gpu/test_mirror_pad.py +++ b/tests/st/ops/gpu/test_mirror_pad.py @@ -23,7 +23,7 @@ import mindspore.context as context from mindspore import Tensor from mindspore.ops.composite import GradOperation -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_mirror_pad(): @@ -71,7 +71,7 @@ class Net(nn.Cell): return self.pad(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_mirror_pad_backprop(): @@ -87,7 +87,7 @@ def test_mirror_pad_backprop(): dx = dx[0].asnumpy() np.testing.assert_array_almost_equal(dx, expected_dx) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_mirror_pad_fwd_back_4d_int32_reflect(): @@ -128,7 +128,7 @@ def test_mirror_pad_fwd_back_4d_int32_reflect(): np.testing.assert_array_equal(dx_value_expected, dx_value_obtained) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_mirror_pad_fwd_back_4d_int32_symm(): diff --git a/tests/st/ops/gpu/test_momentum_fusion.py b/tests/st/ops/gpu/test_momentum_fusion.py index 61d4c0f4082..b8d249d48f0 100644 --- a/tests/st/ops/gpu/test_momentum_fusion.py +++ b/tests/st/ops/gpu/test_momentum_fusion.py @@ -40,7 +40,7 @@ class MomentumFusionNet(nn.Cell): return self.op(self.var, self.accum, self.lr, g, self.moment) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_momentum_fusion(): diff --git a/tests/st/ops/gpu/test_mul_op.py b/tests/st/ops/gpu/test_mul_op.py index 98ce628195c..599daa9a0dc 100644 --- a/tests/st/ops/gpu/test_mul_op.py +++ b/tests/st/ops/gpu/test_mul_op.py @@ -128,13 +128,13 @@ def mul(nptype): assert np.all(diff4 < error4) assert output4.shape == expect4.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_mul_float64(): mul(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_mul_float32(): @@ -199,13 +199,13 @@ def mul_dynamic(nptype): assert np.all(diff2 < error2) assert output2.shape == expect2.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_mul_dynamic_float64(): mul_dynamic(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_mul_dynamic_float32(): diff --git a/tests/st/ops/gpu/test_multinomial_op.py b/tests/st/ops/gpu/test_multinomial_op.py index 322a8fb85b5..8fae62d1c37 100644 --- a/tests/st/ops/gpu/test_multinomial_op.py +++ b/tests/st/ops/gpu/test_multinomial_op.py @@ -32,7 +32,7 @@ class Net(nn.Cell): def construct(self, x): return C.multinomial(x, self.sample, self.replacement, self.seed) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_multinomial(): diff --git a/tests/st/ops/gpu/test_neg_op.py b/tests/st/ops/gpu/test_neg_op.py index 0a73c0dda63..439456070d3 100644 --- a/tests/st/ops/gpu/test_neg_op.py +++ b/tests/st/ops/gpu/test_neg_op.py @@ -69,13 +69,13 @@ def neg(nptype): def test_neg_float16(): neg(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_neg_float32(): neg(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_neg_float64(): diff --git a/tests/st/ops/gpu/test_nll_loss.py b/tests/st/ops/gpu/test_nll_loss.py index e4454a1fd2d..a637442ef0e 100644 --- a/tests/st/ops/gpu/test_nll_loss.py +++ b/tests/st/ops/gpu/test_nll_loss.py @@ -123,7 +123,7 @@ def nll_loss_grad_template(nptype_input, nptype_weight, reduction): np.testing.assert_allclose(dx_np, dx_expected, ertol_loss) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nll_loss_no_reduction(): @@ -134,7 +134,7 @@ def test_nll_loss_no_reduction(): nll_loss_template(np.float16, np.float16, "none") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nll_loss_mean_reduction(): @@ -145,7 +145,7 @@ def test_nll_loss_mean_reduction(): nll_loss_template(np.float16, np.float16, "mean") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nll_loss_sum_reduction(): @@ -156,7 +156,7 @@ def test_nll_loss_sum_reduction(): nll_loss_template(np.float16, np.float16, "sum") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nll_loss_grad_mean_reduction(): @@ -167,7 +167,7 @@ def test_nll_loss_grad_mean_reduction(): nll_loss_grad_template(np.float16, np.float16, "mean") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nll_loss_grad_sum_reduction(): @@ -178,7 +178,7 @@ def test_nll_loss_grad_sum_reduction(): nll_loss_grad_template(np.float16, np.float16, "sum") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nll_loss_grad_no_reduction(): diff --git a/tests/st/ops/gpu/test_nms_with_mask_op.py b/tests/st/ops/gpu/test_nms_with_mask_op.py index 45764d7907e..446a7ffa28e 100644 --- a/tests/st/ops/gpu/test_nms_with_mask_op.py +++ b/tests/st/ops/gpu/test_nms_with_mask_op.py @@ -33,7 +33,7 @@ def runMSRun(op, bbox): return sel_rows, sel_score -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nms_with_mask_check_order(): @@ -54,7 +54,7 @@ def test_nms_with_mask_check_order(): ms_sorted_scores, np_sorted_scores) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nms_with_mask_edge_case_1(): @@ -75,7 +75,7 @@ def test_nms_with_mask_edge_case_1(): np.testing.assert_almost_equal(sel_score, expected_score) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nms_with_mask_edge_case_2(): @@ -92,7 +92,7 @@ def test_nms_with_mask_edge_case_2(): np.testing.assert_almost_equal(sel_score, expected_score) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nms_with_mask_edge_case_3(): diff --git a/tests/st/ops/gpu/test_one_hot_op.py b/tests/st/ops/gpu/test_one_hot_op.py index 49ccc7993ba..6a2115b3b1e 100644 --- a/tests/st/ops/gpu/test_one_hot_op.py +++ b/tests/st/ops/gpu/test_one_hot_op.py @@ -78,13 +78,13 @@ def one_hot(nptype): assert (output[2].asnumpy() == expect_2).all() assert (output[3].asnumpy() == expect_3).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_one_hot_int32(): one_hot(np.int32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_one_hot_int64(): diff --git a/tests/st/ops/gpu/test_oneslike_op.py b/tests/st/ops/gpu/test_oneslike_op.py index e721d157295..cbff539f28e 100644 --- a/tests/st/ops/gpu/test_oneslike_op.py +++ b/tests/st/ops/gpu/test_oneslike_op.py @@ -33,7 +33,7 @@ class NetOnesLike(nn.Cell): return self.ones_like(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_OnesLike(): diff --git a/tests/st/ops/gpu/test_pack_op.py b/tests/st/ops/gpu/test_pack_op.py index d59d4737040..79e730644f1 100644 --- a/tests/st/ops/gpu/test_pack_op.py +++ b/tests/st/ops/gpu/test_pack_op.py @@ -89,13 +89,13 @@ def stack_pynative(nptype): output = P.Stack(axis=2)((x1, x2)) assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_stack_graph_float32(): stack(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_stack_graph_float16(): @@ -125,13 +125,13 @@ def test_stack_graph_uint8(): def test_stack_graph_bool(): stack(np.bool) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_stack_pynative_float32(): stack_pynative(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_stack_pynative_float16(): diff --git a/tests/st/ops/gpu/test_pad.py b/tests/st/ops/gpu/test_pad.py index 4a756f61109..a6eaf54ddcc 100644 --- a/tests/st/ops/gpu/test_pad.py +++ b/tests/st/ops/gpu/test_pad.py @@ -24,7 +24,7 @@ from mindspore import Tensor from mindspore.ops.composite import GradOperation -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pad_basic(): @@ -52,7 +52,7 @@ def test_pad_basic(): np.testing.assert_array_equal(y_test, test_arr_expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pad_row(): @@ -83,7 +83,7 @@ def test_pad_row(): np.testing.assert_equal(y_test_2[:, :, 3:, :], test_arr_2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pad_column(): @@ -114,7 +114,7 @@ def test_pad_column(): np.testing.assert_equal(y_test_2[:, :, :, 6:-1], test_arr_2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pad_3d_pad(): @@ -172,7 +172,7 @@ class Net(nn.Cell): return self.pad(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pad_3d_backprop(): @@ -211,12 +211,12 @@ def test_pad_3d_backprop(): np.testing.assert_array_equal(dx, expected_dx) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pad_error_cases(): """ - Test against common errorneous inputs to trigger correct errors + Test against common erroneous inputs to trigger correct errors """ context.set_context(mode=context.GRAPH_MODE, device_target="GPU") diff --git a/tests/st/ops/gpu/test_prelu_op.py b/tests/st/ops/gpu/test_prelu_op.py index 4b50bd8aea9..0eb27d56e35 100644 --- a/tests/st/ops/gpu/test_prelu_op.py +++ b/tests/st/ops/gpu/test_prelu_op.py @@ -67,7 +67,7 @@ context.set_context(device_target="GPU", mode=context.GRAPH_MODE) dtypes = [mstype.float16, mstype.float32] -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_prelu_single_weight(): @@ -86,7 +86,7 @@ def test_prelu_single_weight(): test_prelu(x, weight, expect_forward, expect_dx, expect_dw) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_prelu_multiple_weight(): @@ -127,7 +127,7 @@ def test_prelu_multiple_weight(): test_prelu(x, weight, expect_forward, expect_dx, expect_dw) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_prelu_single_weight_0_D(): @@ -146,7 +146,7 @@ def test_prelu_single_weight_0_D(): test_prelu(x, weight, expect_forward, expect_dx, expect_dw) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_prelu_single_weight_1_D(): @@ -165,7 +165,7 @@ def test_prelu_single_weight_1_D(): test_prelu(x, weight, expect_forward, expect_dx, expect_dw) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_prelu_single_weight_2_D(): @@ -184,7 +184,7 @@ def test_prelu_single_weight_2_D(): test_prelu(x, weight, expect_forward, expect_dx, expect_dw) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_prelu_multiple_weight_2_D(): diff --git a/tests/st/ops/gpu/test_pyfunc_op.py b/tests/st/ops/gpu/test_pyfunc_op.py index b450764e134..4e5ec4c9010 100644 --- a/tests/st/ops/gpu/test_pyfunc_op.py +++ b/tests/st/ops/gpu/test_pyfunc_op.py @@ -58,7 +58,7 @@ def func_with_dtype(ms_dtype, np_dtype): assert np.allclose(x.asnumpy(), expect.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pyfunc_single_output(): @@ -70,7 +70,7 @@ def test_pyfunc_single_output(): func_with_dtype(ms.int64, np.int64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pyfunc_multi_output(): @@ -98,7 +98,7 @@ class PyFuncGraph(nn.Cell): def construct(self, x1, x2): return self.func((x1, x2)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pyfunc_no_output(): diff --git a/tests/st/ops/gpu/test_random_choice_with_mask.py b/tests/st/ops/gpu/test_random_choice_with_mask.py index 5944e15c3c2..6a442744484 100644 --- a/tests/st/ops/gpu/test_random_choice_with_mask.py +++ b/tests/st/ops/gpu/test_random_choice_with_mask.py @@ -58,7 +58,7 @@ class RCWM_1D(nn.Cell): return self.RCWM_1D(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_RCWM_3D(): @@ -72,7 +72,7 @@ def test_RCWM_3D(): assert output2.shape == expect2 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_RCWM_count_out(): @@ -87,7 +87,7 @@ def test_RCWM_count_out(): assert output2.shape == expect2 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_RCWM_count_in(): @@ -102,7 +102,7 @@ def test_RCWM_count_in(): assert output2.shape == expect2 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_RCWM_1D(): diff --git a/tests/st/ops/gpu/test_randperm.py b/tests/st/ops/gpu/test_randperm.py index ea78b95da95..7cea96fd7cb 100644 --- a/tests/st/ops/gpu/test_randperm.py +++ b/tests/st/ops/gpu/test_randperm.py @@ -87,19 +87,19 @@ def test_randperm_uint8(): def test_randperm_uint16(): randperm(8, 0, mindspore.uint16, 8) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_randperm_uint32(): randperm(4, 8, mindspore.uint32, 3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_randperm_uint64(): randperm(5, 4, mindspore.uint64, 5) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_randperm_n_too_large(): diff --git a/tests/st/ops/gpu/test_range_op.py b/tests/st/ops/gpu/test_range_op.py index 33eb0654664..b8020ed3c72 100644 --- a/tests/st/ops/gpu/test_range_op.py +++ b/tests/st/ops/gpu/test_range_op.py @@ -33,7 +33,7 @@ class RangeNet(nn.Cell): return self.range(start, limit, delta) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_range_precision_end_equals_last_element(): @@ -65,7 +65,7 @@ def test_range_precision_end_equals_last_element(): np_expected = np.arange(-12000, -12053, -1, dtype=np.float32) np.testing.assert_allclose(ms_out, np_expected, rtol=1e-5) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_range_int(): @@ -89,7 +89,7 @@ def test_range_int(): np_expected = np.array([3, -2, -7]) np.testing.assert_array_equal(ms_out, np_expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_range_float(): @@ -123,7 +123,7 @@ def test_range_invalid_max_output_length(): _ = P.Range(None) _ = P.Range('5') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_range_invalid_input(): diff --git a/tests/st/ops/gpu/test_realdiv_op.py b/tests/st/ops/gpu/test_realdiv_op.py index 29144ba9727..7c0265304f4 100644 --- a/tests/st/ops/gpu/test_realdiv_op.py +++ b/tests/st/ops/gpu/test_realdiv_op.py @@ -31,7 +31,7 @@ class NetRealDiv(nn.Cell): return self.divide(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_real_div(): diff --git a/tests/st/ops/gpu/test_reciprocal_grad_op.py b/tests/st/ops/gpu/test_reciprocal_grad_op.py index eb2f9cf5680..ddb30325d73 100644 --- a/tests/st/ops/gpu/test_reciprocal_grad_op.py +++ b/tests/st/ops/gpu/test_reciprocal_grad_op.py @@ -31,7 +31,7 @@ class NetReciprocalGrad(nn.Cell): return self.grad(y, dy) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reciprocal_grad_float32(): @@ -64,7 +64,7 @@ def test_reciprocal_grad_float32(): np.testing.assert_array_almost_equal(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reciprocal_grad_float16(): diff --git a/tests/st/ops/gpu/test_reciprocal_op.py b/tests/st/ops/gpu/test_reciprocal_op.py index 6fd20953530..ce85e7858a0 100644 --- a/tests/st/ops/gpu/test_reciprocal_op.py +++ b/tests/st/ops/gpu/test_reciprocal_op.py @@ -31,7 +31,7 @@ class NetReciprocal(nn.Cell): return self.reciprocal(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_Reciprocal(): diff --git a/tests/st/ops/gpu/test_reduce_all_op.py b/tests/st/ops/gpu/test_reduce_all_op.py index ed7dec41d78..4c1b7c24d3c 100644 --- a/tests/st/ops/gpu/test_reduce_all_op.py +++ b/tests/st/ops/gpu/test_reduce_all_op.py @@ -69,7 +69,7 @@ class ReduceAll(nn.Cell): P.ReduceAll(self.keep_dims3)(self.x3, self.axis3)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ReduceAll(): @@ -113,7 +113,7 @@ class ReduceAllDynamic(nn.Cell): return self.reduceall(dynamic_x, self.axis) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reduce_all_dynamic(): diff --git a/tests/st/ops/gpu/test_reduce_any_op.py b/tests/st/ops/gpu/test_reduce_any_op.py index 49f121c08cc..51e769f7fc2 100644 --- a/tests/st/ops/gpu/test_reduce_any_op.py +++ b/tests/st/ops/gpu/test_reduce_any_op.py @@ -69,7 +69,7 @@ class ReduceAny(nn.Cell): P.ReduceAny(self.keep_dims3)(self.x3, self.axis3)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ReduceAny(): @@ -113,7 +113,7 @@ class ReduceAnyDynamic(nn.Cell): return self.reduceany(dynamic_x, self.axis) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reduce_any_dynamic(): diff --git a/tests/st/ops/gpu/test_reduce_max_op.py b/tests/st/ops/gpu/test_reduce_max_op.py index bfdc240ff80..16519353946 100644 --- a/tests/st/ops/gpu/test_reduce_max_op.py +++ b/tests/st/ops/gpu/test_reduce_max_op.py @@ -117,7 +117,7 @@ class ReduceMax(nn.Cell): P.ReduceMax(self.keep_dims8)(self.x8, self.axis8)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ReduceMax(): @@ -196,7 +196,7 @@ class ReduceMaxDynamic(nn.Cell): dynamic_x = self.test_dynamic(self.x) return self.reducemax(dynamic_x, self.axis) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reduce_max_dynamic(): @@ -224,7 +224,7 @@ class ReduceMaxTypeNet(nn.Cell): def construct(self): return P.ReduceMax(self.keep_dims0)(self.x0, self.axis0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reduce_max_float64(): diff --git a/tests/st/ops/gpu/test_reduce_min_op.py b/tests/st/ops/gpu/test_reduce_min_op.py index 1b964984441..d6acd7d711e 100644 --- a/tests/st/ops/gpu/test_reduce_min_op.py +++ b/tests/st/ops/gpu/test_reduce_min_op.py @@ -32,7 +32,7 @@ class ReduceMin(nn.Cell): return self.reduce_min(x, axis) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float16, np.float32, np.float64]) @@ -74,7 +74,7 @@ class ReduceMinDynamic(nn.Cell): return self.reduce_min(dynamic_x, self.axis) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float32]) diff --git a/tests/st/ops/gpu/test_reduce_prod_op.py b/tests/st/ops/gpu/test_reduce_prod_op.py index d7e6a7c7fb3..d568deb718e 100644 --- a/tests/st/ops/gpu/test_reduce_prod_op.py +++ b/tests/st/ops/gpu/test_reduce_prod_op.py @@ -31,7 +31,7 @@ class ReduceProd(nn.Cell): return self.reduce_prod(x, axis) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @pytest.mark.parametrize('decimal, dtype', diff --git a/tests/st/ops/gpu/test_reduce_sum_op.py b/tests/st/ops/gpu/test_reduce_sum_op.py index 6ef0c80d509..c3eccecfc25 100644 --- a/tests/st/ops/gpu/test_reduce_sum_op.py +++ b/tests/st/ops/gpu/test_reduce_sum_op.py @@ -170,7 +170,7 @@ class ReduceSum(nn.Cell): P.ReduceSum(self.keep_dims14)(self.x14, self.axis14)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ReduceSum(): @@ -316,7 +316,7 @@ class ReduceSumTypeNet(nn.Cell): def construct(self): return P.ReduceSum(self.keep_dims0)(self.x0, self.axis0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reduce_sum_float64(): diff --git a/tests/st/ops/gpu/test_relu6_grad_op.py b/tests/st/ops/gpu/test_relu6_grad_op.py index 9be8ee6567e..bd3faadd15a 100644 --- a/tests/st/ops/gpu/test_relu6_grad_op.py +++ b/tests/st/ops/gpu/test_relu6_grad_op.py @@ -31,7 +31,7 @@ class NetReLU6Grad(nn.Cell): return self.relu6_grad(dy, x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_relu6_grad(): diff --git a/tests/st/ops/gpu/test_relu6_op.py b/tests/st/ops/gpu/test_relu6_op.py index 85d23d5f461..382def24c41 100644 --- a/tests/st/ops/gpu/test_relu6_op.py +++ b/tests/st/ops/gpu/test_relu6_op.py @@ -43,7 +43,7 @@ class NetRelu6Dynamic(nn.Cell): return self.relu6(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_relu6(): @@ -65,7 +65,7 @@ def test_relu6(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_relu6_dynamic(): diff --git a/tests/st/ops/gpu/test_relu_grad_grad_op.py b/tests/st/ops/gpu/test_relu_grad_grad_op.py index 68911198ad2..ded0cfbaefd 100644 --- a/tests/st/ops/gpu/test_relu_grad_grad_op.py +++ b/tests/st/ops/gpu/test_relu_grad_grad_op.py @@ -73,14 +73,14 @@ def test_relu_grad_grad_float16(): relu_grad_grad_base(np.float16, 1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_relu_grad_grad_float32(): relu_grad_grad_base(np.float32, 1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_relu_grad_grad_float64(): diff --git a/tests/st/ops/gpu/test_relu_op.py b/tests/st/ops/gpu/test_relu_op.py index b0dc21b0a68..4b9810ec783 100644 --- a/tests/st/ops/gpu/test_relu_op.py +++ b/tests/st/ops/gpu/test_relu_op.py @@ -43,7 +43,7 @@ class NetReluDynamic(nn.Cell): return self.relu(x_conv) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_relu_float32(): @@ -65,7 +65,7 @@ def test_relu_float32(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_relu_int8(): @@ -87,7 +87,7 @@ def test_relu_int8(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_relu_int32(): @@ -109,7 +109,7 @@ def test_relu_int32(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_relu_int64(): @@ -132,7 +132,7 @@ def test_relu_int64(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_relu_int64_dynamic_shape(): diff --git a/tests/st/ops/gpu/test_relu_v2.py b/tests/st/ops/gpu/test_relu_v2.py index f000531afa2..14db2e773b4 100644 --- a/tests/st/ops/gpu/test_relu_v2.py +++ b/tests/st/ops/gpu/test_relu_v2.py @@ -34,7 +34,7 @@ class ReluNet(nn.Cell): dx = self.relu_grad(dy, y) return y, dx -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ReluV2(): @@ -73,7 +73,7 @@ class AddReluNet(nn.Cell): return y, dx -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_AddRelu(): @@ -114,7 +114,7 @@ class AddReluGradNet(nn.Cell): return y, dx -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_AddReluGrad(): diff --git a/tests/st/ops/gpu/test_repeat_elements_op.py b/tests/st/ops/gpu/test_repeat_elements_op.py index e373f5c4ead..5b5006008e6 100644 --- a/tests/st/ops/gpu/test_repeat_elements_op.py +++ b/tests/st/ops/gpu/test_repeat_elements_op.py @@ -35,7 +35,7 @@ def repeat_elements(x, rep, axis): repeat_elements_net = RepeatElementsNet(rep, axis) return repeat_elements_net(Tensor(x.astype(np.int32))).asnumpy() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_1d_one_element_rep_1(): @@ -46,7 +46,7 @@ def test_repeat_elements_1d_one_element_rep_1(): np_out = a.repeat(1, 0) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_1d_one_element_rep_many(): @@ -61,7 +61,7 @@ def test_repeat_elements_1d_one_element_rep_many(): np_out = a.repeat(513, 0) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_1d_rep_1(): @@ -72,7 +72,7 @@ def test_repeat_elements_1d_rep_1(): np_out = a.repeat(1, 0) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_1d_rep_many(): @@ -83,7 +83,7 @@ def test_repeat_elements_1d_rep_many(): np_out = a.repeat(231, 0) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_2d_one_element_rep_1(): @@ -98,7 +98,7 @@ def test_repeat_elements_2d_one_element_rep_1(): np_out = a.repeat(1, 1) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_2d_one_element_rep_many(): @@ -113,7 +113,7 @@ def test_repeat_elements_2d_one_element_rep_many(): np_out = a.repeat(13, 1) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_2d_rep_1(): @@ -128,7 +128,7 @@ def test_repeat_elements_2d_rep_1(): np_out = a.repeat(1, 1) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_2d_rep_many(): @@ -143,7 +143,7 @@ def test_repeat_elements_2d_rep_many(): np_out = a.repeat(23, 1) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_3d_one_element_rep_1(): @@ -162,7 +162,7 @@ def test_repeat_elements_3d_one_element_rep_1(): np_out = a.repeat(1, 2) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_3d_one_element_rep_many(): @@ -181,7 +181,7 @@ def test_repeat_elements_3d_one_element_rep_many(): np_out = a.repeat(43, 2) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_3d_rep_1(): @@ -200,7 +200,7 @@ def test_repeat_elements_3d_rep_1(): np_out = a.repeat(1, 2) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_3d_rep_many(): @@ -219,7 +219,7 @@ def test_repeat_elements_3d_rep_many(): np_out = a.repeat(14, 2) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_4d_one_element_rep_1(): @@ -243,7 +243,7 @@ def test_repeat_elements_4d_one_element_rep_1(): np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_4d_one_element_rep_many(): @@ -266,7 +266,7 @@ def test_repeat_elements_4d_one_element_rep_many(): np_out = a.repeat(17, 3) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_4d_rep_1(): @@ -290,7 +290,7 @@ def test_repeat_elements_4d_rep_1(): np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_4d_rep_many(): @@ -313,7 +313,7 @@ def test_repeat_elements_4d_rep_many(): np_out = a.repeat(23, 3) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_5d_one_element_rep_1(): @@ -340,7 +340,7 @@ def test_repeat_elements_5d_one_element_rep_1(): np_out = a.repeat(1, 4) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_5d_one_element_rep_many(): @@ -367,7 +367,7 @@ def test_repeat_elements_5d_one_element_rep_many(): np_out = a.repeat(19, 4) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_5d_rep_1(): @@ -394,7 +394,7 @@ def test_repeat_elements_5d_rep_1(): np_out = a.repeat(1, 4) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_5d_rep_many(): @@ -421,7 +421,7 @@ def test_repeat_elements_5d_rep_many(): np_out = a.repeat(7, 4) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_large_one_element_rep_1(): @@ -460,7 +460,7 @@ def test_repeat_elements_large_one_element_rep_1(): np_out = a.repeat(1, 7) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_large_one_element_rep_many(): @@ -499,7 +499,7 @@ def test_repeat_elements_large_one_element_rep_many(): np_out = a.repeat(42, 7) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_large_rep_1(): @@ -538,7 +538,7 @@ def test_repeat_elements_large_rep_1(): np_out = a.repeat(1, 7) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_large_rep_many(): @@ -577,7 +577,7 @@ def test_repeat_elements_large_rep_many(): np_out = a.repeat(4, 7) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_half(): @@ -616,7 +616,7 @@ def test_repeat_elements_half(): np_out = a.repeat(4, 7) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_net_multi_use(): @@ -641,7 +641,7 @@ def test_repeat_elements_net_multi_use(): np_out = a.repeat(rep, axis) np.testing.assert_array_equal(np_out, ms_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_repeat_elements_invalid_input(): diff --git a/tests/st/ops/gpu/test_reshape_op.py b/tests/st/ops/gpu/test_reshape_op.py index acf3e74b0bb..681145b1d9a 100644 --- a/tests/st/ops/gpu/test_reshape_op.py +++ b/tests/st/ops/gpu/test_reshape_op.py @@ -95,13 +95,13 @@ def reshape_bool(): np.testing.assert_array_equal(output_tensor.asnumpy().flatten(), data) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reshape_float(): reshape(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reshape_float16(): diff --git a/tests/st/ops/gpu/test_resize_bilinear_grad_op.py b/tests/st/ops/gpu/test_resize_bilinear_grad_op.py index 6319e246372..41b0933010f 100644 --- a/tests/st/ops/gpu/test_resize_bilinear_grad_op.py +++ b/tests/st/ops/gpu/test_resize_bilinear_grad_op.py @@ -64,7 +64,7 @@ def test_resize_bilinear_grad_align_corners(): assert np.all(output.asnumpy() == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_resize_bilinear_grad(): diff --git a/tests/st/ops/gpu/test_resize_bilinear_op.py b/tests/st/ops/gpu/test_resize_bilinear_op.py index 0761e915de6..08067356982 100644 --- a/tests/st/ops/gpu/test_resize_bilinear_op.py +++ b/tests/st/ops/gpu/test_resize_bilinear_op.py @@ -31,7 +31,7 @@ class NetResizeBilinear(nn.Cell): return self.op(inputs) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_resize_nn_grayscale_integer_ratio_half(datatype=np.float16): @@ -141,7 +141,7 @@ def test_resize_nn_grayscale_integer_ratio_half(datatype=np.float16): assert np.all(abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_resize_nn_grayscale_integer_ratio_float(datatype=np.float32): @@ -353,7 +353,7 @@ def test_resize_nn_grayscale_not_integer_ratio_half(datatype=np.float16): assert np.all(abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_resize_nn_grayscale_not_integer_ratio_float(datatype=np.float32): @@ -465,7 +465,7 @@ def test_resize_nn_grayscale_not_integer_ratio_float(datatype=np.float32): assert np.all(abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_resize_nn_grayscale_multiple_images_half(datatype=np.float16): @@ -487,7 +487,7 @@ def test_resize_nn_grayscale_multiple_images_half(datatype=np.float16): assert np.all(abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_resize_nn_grayscale_multiple_images_float(datatype=np.float32): @@ -510,7 +510,7 @@ def test_resize_nn_grayscale_multiple_images_float(datatype=np.float32): assert np.all(abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_resize_nn_grayscale_align_corners_half(datatype=np.float16): @@ -538,7 +538,7 @@ def test_resize_nn_grayscale_align_corners_half(datatype=np.float16): assert np.all(abs(diff_align) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_resize_nn_grayscale_align_corners_float(datatype=np.float32): diff --git a/tests/st/ops/gpu/test_resize_nearest_neighbor_grad_op.py b/tests/st/ops/gpu/test_resize_nearest_neighbor_grad_op.py index 0203c872041..aaab763d907 100644 --- a/tests/st/ops/gpu/test_resize_nearest_neighbor_grad_op.py +++ b/tests/st/ops/gpu/test_resize_nearest_neighbor_grad_op.py @@ -38,7 +38,7 @@ class ResizeNearestNeighborGradAlignCornerF(nn.Cell): return self.ResizeNearestNeighborGradAlignCornerF(dy, size) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ResizeNearestNeighborGradAlignCornerT(): @@ -62,7 +62,7 @@ def test_ResizeNearestNeighborGradAlignCornerT(): output = rnn(Tensor(dy), size) assert np.all(output.asnumpy() == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ResizeNearestNeighborGradAlignCornerF(): diff --git a/tests/st/ops/gpu/test_reverse_sequence.py b/tests/st/ops/gpu/test_reverse_sequence.py index 16c41f1f3f8..7aada9efcd9 100644 --- a/tests/st/ops/gpu/test_reverse_sequence.py +++ b/tests/st/ops/gpu/test_reverse_sequence.py @@ -19,7 +19,7 @@ class Net(nn.Cell): return self.reverse_sequence(x, seq_lengths) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net_int8(): @@ -49,7 +49,7 @@ def test_net_int32(): assert np.array_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net_float32(): @@ -78,7 +78,7 @@ def test_net_float32(): assert np.array_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net_float64_0_dim(): diff --git a/tests/st/ops/gpu/test_reverse_v2_op.py b/tests/st/ops/gpu/test_reverse_v2_op.py index 20f8c7ec505..539712846ae 100644 --- a/tests/st/ops/gpu/test_reverse_v2_op.py +++ b/tests/st/ops/gpu/test_reverse_v2_op.py @@ -55,14 +55,14 @@ def reverse_v2_1d(nptype): reverse_v2(x_numpy, (0,)) reverse_v2(x_numpy, (-1,)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reverse_v2_float16(): reverse_v2_1d(np.float16) reverse_v2_3d(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reverse_v2_float32(): diff --git a/tests/st/ops/gpu/test_rint_op.py b/tests/st/ops/gpu/test_rint_op.py index e1e05c6e061..65bc2b3a351 100644 --- a/tests/st/ops/gpu/test_rint_op.py +++ b/tests/st/ops/gpu/test_rint_op.py @@ -46,14 +46,14 @@ def generate_testcases(nptype): np.testing.assert_almost_equal(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sign_float32(): generate_testcases(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sign_float16(): diff --git a/tests/st/ops/gpu/test_rl_buffer_net.py b/tests/st/ops/gpu/test_rl_buffer_net.py index 00d8a1c73a4..95eda4827bd 100644 --- a/tests/st/ops/gpu/test_rl_buffer_net.py +++ b/tests/st/ops/gpu/test_rl_buffer_net.py @@ -67,7 +67,7 @@ exp = [s, a, r, s_] exp1 = [s_, a, r, s] -@ pytest.mark.level0 +@ pytest.mark.level1 @ pytest.mark.platform_x86_gpu_training @ pytest.mark.env_onecard def test_Buffer(): diff --git a/tests/st/ops/gpu/test_rl_buffer_op.py b/tests/st/ops/gpu/test_rl_buffer_op.py index 44c5b34978d..3ed4bb789fd 100644 --- a/tests/st/ops/gpu/test_rl_buffer_op.py +++ b/tests/st/ops/gpu/test_rl_buffer_op.py @@ -112,7 +112,7 @@ def test_BufferGet(): np.testing.assert_almost_equal(ss_.asnumpy(), expect_s_) -@ pytest.mark.level0 +@ pytest.mark.level1 @ pytest.mark.platform_x86_gpu_training @ pytest.mark.env_onecard def test_BufferAppend(): diff --git a/tests/st/ops/gpu/test_rmsprop.py b/tests/st/ops/gpu/test_rmsprop.py index 4cad24ef539..1d29cd08d72 100644 --- a/tests/st/ops/gpu/test_rmsprop.py +++ b/tests/st/ops/gpu/test_rmsprop.py @@ -81,7 +81,7 @@ def rmspropcented_numpy(variable, gradients, mean_gradients, mean_square, moment return variable, gradients, mean_gradients, mean_square, moment -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_rmsprop(): @@ -142,7 +142,7 @@ def test_rmsprop(): assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_rmspropcenter(): diff --git a/tests/st/ops/gpu/test_roi_align_grad_op.py b/tests/st/ops/gpu/test_roi_align_grad_op.py index ddb4eb6591c..72eefbcff7d 100644 --- a/tests/st/ops/gpu/test_roi_align_grad_op.py +++ b/tests/st/ops/gpu/test_roi_align_grad_op.py @@ -38,7 +38,7 @@ class NetROIAlignGrad(nn.Cell): return self.roiAlignGrad(dy, rois) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_roi_align_grad(): diff --git a/tests/st/ops/gpu/test_round_op.py b/tests/st/ops/gpu/test_round_op.py index 28ca5bdafd7..643bfdcecd5 100644 --- a/tests/st/ops/gpu/test_round_op.py +++ b/tests/st/ops/gpu/test_round_op.py @@ -47,14 +47,14 @@ def generate_testcases(nptype): np.testing.assert_almost_equal(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_round_float32(): generate_testcases(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_round_float16(): diff --git a/tests/st/ops/gpu/test_rsqrt_grad_op.py b/tests/st/ops/gpu/test_rsqrt_grad_op.py index c84e7a92e76..643451dbeab 100644 --- a/tests/st/ops/gpu/test_rsqrt_grad_op.py +++ b/tests/st/ops/gpu/test_rsqrt_grad_op.py @@ -31,7 +31,7 @@ class NetRsqrtGrad(nn.Cell): return self.rsqrt_grad(x, dx) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_rsqrt_grad(): diff --git a/tests/st/ops/gpu/test_sampled_softmax_loss_op.py b/tests/st/ops/gpu/test_sampled_softmax_loss_op.py index fd2219b5289..f31e5cf7be4 100644 --- a/tests/st/ops/gpu/test_sampled_softmax_loss_op.py +++ b/tests/st/ops/gpu/test_sampled_softmax_loss_op.py @@ -36,7 +36,7 @@ def generate_test_data(num_classes, batch_size, sampled): return weights_s, biases_s, hidden_acts_s, sampled_values -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sampled_softmax_loss_assigned_sampler(): @@ -96,7 +96,7 @@ def test_sampled_softmax_loss_assigned_sampler(): case_remove_accidental_hits() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sampled_softmax_loss_none_sampler(): diff --git a/tests/st/ops/gpu/test_scatter_func_op.py b/tests/st/ops/gpu/test_scatter_func_op.py index d848af36080..f4c95bc0bdf 100644 --- a/tests/st/ops/gpu/test_scatter_func_op.py +++ b/tests/st/ops/gpu/test_scatter_func_op.py @@ -125,7 +125,7 @@ class VmapNet(nn.Cell): return vmap(self.net, self.in_axes, self.out_axes)(self.inputx, indices, updates) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_scatter_func_small_float32(): @@ -159,7 +159,7 @@ def test_scatter_func_small_float32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_scatter_func_input_updated(): @@ -199,7 +199,7 @@ def test_scatter_func_input_updated(): np.testing.assert_array_almost_equal(net.inputx.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_scatter_func_large_shape_float32(): @@ -370,7 +370,7 @@ def test_scatter_func_large_shape_float32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_scatter_func_small_float32_use_locking_false(): @@ -404,7 +404,7 @@ def test_scatter_func_small_float32_use_locking_false(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_scatter_func_input_less_than_1_float32(): @@ -469,7 +469,7 @@ def test_scatter_func_input_less_than_1_float32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_scatter_func_float16(): @@ -557,7 +557,7 @@ def test_scatter_func_large_float16(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_scatter_func_disordered_float16(): @@ -607,7 +607,7 @@ def test_scatter_func_disordered_float16(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_scatter_func_large_int32(): @@ -661,7 +661,7 @@ def test_scatter_func_large_int32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_scatter_func_disordered_int32(): @@ -711,7 +711,7 @@ def test_scatter_func_disordered_int32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_scatter_func_disordered_dynamic_int32(): @@ -757,7 +757,7 @@ def test_scatter_func_disordered_dynamic_int32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_scatter_func_disordered_dynamic_int8(): @@ -791,7 +791,7 @@ def test_scatter_func_disordered_dynamic_int8(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_scatter_func_disordered_dynamic_uint8(): @@ -821,7 +821,7 @@ def test_scatter_func_disordered_dynamic_uint8(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_scatter_func_input_less_than_1_dynamic_float32(): @@ -880,7 +880,7 @@ def test_scatter_func_input_less_than_1_dynamic_float32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_scatter_func_dynamic_two_inputs(): diff --git a/tests/st/ops/gpu/test_select_op.py b/tests/st/ops/gpu/test_select_op.py index c227e51318c..879f1d07c9b 100644 --- a/tests/st/ops/gpu/test_select_op.py +++ b/tests/st/ops/gpu/test_select_op.py @@ -33,7 +33,7 @@ class Net(nn.Cell): return self.select(cond_op, input_x, input_y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_select(): diff --git a/tests/st/ops/gpu/test_sequence_mask_op.py b/tests/st/ops/gpu/test_sequence_mask_op.py index d5394dd34fb..dfccc97d709 100644 --- a/tests/st/ops/gpu/test_sequence_mask_op.py +++ b/tests/st/ops/gpu/test_sequence_mask_op.py @@ -71,7 +71,7 @@ def test_sequence_mask_3d(): np.testing.assert_array_equal(expected_out.asnumpy(), ms_out.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sequence_mask_maxlen_1(): @@ -87,7 +87,7 @@ def test_sequence_mask_maxlen_1(): np.testing.assert_array_equal(expected_out.asnumpy(), ms_out.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sequence_mask_dynamic(): @@ -149,7 +149,7 @@ def sequence_mask_optional(x): return C.sequence_mask(Tensor(x.astype(np.int32))) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sequence_mask_optional_maxlen(): diff --git a/tests/st/ops/gpu/test_sgd_op.py b/tests/st/ops/gpu/test_sgd_op.py index f959d879cb2..b61e772f538 100644 --- a/tests/st/ops/gpu/test_sgd_op.py +++ b/tests/st/ops/gpu/test_sgd_op.py @@ -40,7 +40,7 @@ class NetSGD(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_SGD(): diff --git a/tests/st/ops/gpu/test_sigmoid_cross_entropy_with_logits_grad_op.py b/tests/st/ops/gpu/test_sigmoid_cross_entropy_with_logits_grad_op.py index 356f38ebab5..baa02dc8ddc 100644 --- a/tests/st/ops/gpu/test_sigmoid_cross_entropy_with_logits_grad_op.py +++ b/tests/st/ops/gpu/test_sigmoid_cross_entropy_with_logits_grad_op.py @@ -59,14 +59,14 @@ def sigmoid_cross_entropy_with_logits_grad(nptype): assert np.all(abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sigmoid_cross_entropy_with_logits_float32(): sigmoid_cross_entropy_with_logits_grad(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sigmoid_cross_entropy_with_logits_float64(): diff --git a/tests/st/ops/gpu/test_sigmoid_cross_entropy_with_logits_op.py b/tests/st/ops/gpu/test_sigmoid_cross_entropy_with_logits_op.py index bfc95c13d82..2110610cbb6 100644 --- a/tests/st/ops/gpu/test_sigmoid_cross_entropy_with_logits_op.py +++ b/tests/st/ops/gpu/test_sigmoid_cross_entropy_with_logits_op.py @@ -57,14 +57,14 @@ def sigmoid_cross_entropy_with_logits(nptype): assert np.all(abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sigmoid_cross_entropy_with_logits_float32(): sigmoid_cross_entropy_with_logits(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sigmoid_cross_entropy_with_logits_float64(): diff --git a/tests/st/ops/gpu/test_sigmoid_grad_grad_op.py b/tests/st/ops/gpu/test_sigmoid_grad_grad_op.py index 110230f49b1..fac8d5a85fc 100644 --- a/tests/st/ops/gpu/test_sigmoid_grad_grad_op.py +++ b/tests/st/ops/gpu/test_sigmoid_grad_grad_op.py @@ -42,7 +42,7 @@ class Grad(nn.Cell): return self.grad(self.network)(y, y_grad, dout) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sigmoid_grad_grad(): diff --git a/tests/st/ops/gpu/test_sigmoid_grad_op.py b/tests/st/ops/gpu/test_sigmoid_grad_op.py index 92d1d4d9f71..772fdb58549 100644 --- a/tests/st/ops/gpu/test_sigmoid_grad_op.py +++ b/tests/st/ops/gpu/test_sigmoid_grad_op.py @@ -31,7 +31,7 @@ class NetSigmoidGrad(nn.Cell): return self.sigmoid_grad(y, dy) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sigmoid_grad(): diff --git a/tests/st/ops/gpu/test_sigmoid_op.py b/tests/st/ops/gpu/test_sigmoid_op.py index f3d724a35b7..8f117b938b6 100644 --- a/tests/st/ops/gpu/test_sigmoid_op.py +++ b/tests/st/ops/gpu/test_sigmoid_op.py @@ -31,7 +31,7 @@ class NetSigmoid(nn.Cell): return self.sigmoid(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sigmoid(): diff --git a/tests/st/ops/gpu/test_sign_op.py b/tests/st/ops/gpu/test_sign_op.py index 2b7bf4f2016..60aba453d39 100644 --- a/tests/st/ops/gpu/test_sign_op.py +++ b/tests/st/ops/gpu/test_sign_op.py @@ -46,21 +46,21 @@ def generate_testcases(nptype): np.testing.assert_almost_equal(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sign_int32(): generate_testcases(np.int32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sign_float32(): generate_testcases(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sign_float16(): diff --git a/tests/st/ops/gpu/test_sin_op.py b/tests/st/ops/gpu/test_sin_op.py index 5b6a3c4b4fd..739be521c4e 100644 --- a/tests/st/ops/gpu/test_sin_op.py +++ b/tests/st/ops/gpu/test_sin_op.py @@ -35,13 +35,13 @@ def sin(nptype): def test_sin_float16(): sin(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sin_float32(): sin(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sin_float64(): diff --git a/tests/st/ops/gpu/test_slice.py b/tests/st/ops/gpu/test_slice.py index ebdce46a120..3b12591a633 100644 --- a/tests/st/ops/gpu/test_slice.py +++ b/tests/st/ops/gpu/test_slice.py @@ -31,7 +31,7 @@ class Slice(nn.Cell): return self.slice(x, (0, 1, 0), (2, 1, 3)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_slice(): @@ -55,7 +55,7 @@ class SliceNet(nn.Cell): return self.slice(x, (0, 11, 0, 0), (32, 7, 224, 224)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_slice_4d(): @@ -94,7 +94,7 @@ def test_slice_5d(): assert (output_ms.asnumpy() == output_np).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_slice_float64(): diff --git a/tests/st/ops/gpu/test_slice_grad.py b/tests/st/ops/gpu/test_slice_grad.py index 1b828006329..ab2cbaf45d2 100644 --- a/tests/st/ops/gpu/test_slice_grad.py +++ b/tests/st/ops/gpu/test_slice_grad.py @@ -35,7 +35,7 @@ class SliceGrad(nn.Cell): return self.slice_grad(dy, x, (0, 1, 0), (2, 1, 3)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_slice(): @@ -52,7 +52,7 @@ def test_slice(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_slice_float64(): diff --git a/tests/st/ops/gpu/test_smoothl1loss_op.py b/tests/st/ops/gpu/test_smoothl1loss_op.py index f4ef2851f5a..5b8edcadc2d 100644 --- a/tests/st/ops/gpu/test_smoothl1loss_op.py +++ b/tests/st/ops/gpu/test_smoothl1loss_op.py @@ -29,7 +29,7 @@ def smoothl1loss(beta): net = nn.SmoothL1Loss(beta) return net(Tensor(prediction), Tensor(target)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_smoothl1loss(): @@ -77,7 +77,7 @@ def smoothl1loss_grad(beta): grad = Grad(net) return grad(Tensor(prediction), Tensor(target), Tensor(sens)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_smoothl1loss_grad(): diff --git a/tests/st/ops/gpu/test_softmax_cross_entropy_with_logits_op.py b/tests/st/ops/gpu/test_softmax_cross_entropy_with_logits_op.py index f8d3ec0e8b3..4621446eea9 100644 --- a/tests/st/ops/gpu/test_softmax_cross_entropy_with_logits_op.py +++ b/tests/st/ops/gpu/test_softmax_cross_entropy_with_logits_op.py @@ -30,7 +30,7 @@ class NetSoftmaxCrossEntropyWithLogits(nn.Cell): return self.loss(logits, labels) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_softmax_cross_entropy_with_logits(): diff --git a/tests/st/ops/gpu/test_softmax_op.py b/tests/st/ops/gpu/test_softmax_op.py index 0f654e10bd9..9b7185ce191 100644 --- a/tests/st/ops/gpu/test_softmax_op.py +++ b/tests/st/ops/gpu/test_softmax_op.py @@ -34,7 +34,7 @@ class NetSoftmax(nn.Cell): return self.softmax1(x), self.softmax2(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_softmax(): @@ -87,7 +87,7 @@ class Grad(nn.Cell): return gout -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_softmax_4d(): diff --git a/tests/st/ops/gpu/test_softplus_grad_op.py b/tests/st/ops/gpu/test_softplus_grad_op.py index f3dda50877c..1394694857e 100644 --- a/tests/st/ops/gpu/test_softplus_grad_op.py +++ b/tests/st/ops/gpu/test_softplus_grad_op.py @@ -45,7 +45,7 @@ class Grad(nn.Cell): return gout -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_softplusgrad(): @@ -63,7 +63,7 @@ def test_softplusgrad(): expect = dy * np.exp(x) / (1 + np.exp(x)) assert np.allclose(output[0].asnumpy(), expect, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_softplusgrad_fp16(): diff --git a/tests/st/ops/gpu/test_softplus_op.py b/tests/st/ops/gpu/test_softplus_op.py index 1ee9d3c9c45..e08048f5a29 100644 --- a/tests/st/ops/gpu/test_softplus_op.py +++ b/tests/st/ops/gpu/test_softplus_op.py @@ -37,7 +37,7 @@ def SoftplusCompute(x): return np.log(1 + np.exp(x)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_softplus_1d(): @@ -51,7 +51,7 @@ def test_softplus_1d(): assert np.allclose(y_np, y_ms.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_softplus_2d(): @@ -65,7 +65,7 @@ def test_softplus_2d(): assert np.allclose(y_np, y_ms.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_softplus_4d(): @@ -79,7 +79,7 @@ def test_softplus_4d(): assert np.allclose(y_np, y_ms.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_softplus_neg(): @@ -92,7 +92,7 @@ def test_softplus_neg(): assert np.allclose(y_np, y_ms.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_softplus_4d_fp16(): diff --git a/tests/st/ops/gpu/test_sort.py b/tests/st/ops/gpu/test_sort.py index ac4e49ca015..635e99a1e55 100644 --- a/tests/st/ops/gpu/test_sort.py +++ b/tests/st/ops/gpu/test_sort.py @@ -111,49 +111,49 @@ def sort_3d(descending, nptype): np.testing.assert_array_equal(output.asnumpy(), expected_output) np.testing.assert_array_equal(indices.asnumpy(), expected_indices) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sort1d_float16(): sort_1d(False, np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sort1d_descending_float16(): sort_1d(True, np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sort1d_float32(): sort_1d(False, np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sort1d_descending_float32(): sort_1d(True, np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sort3d_float16(): sort_3d(False, np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sort3d_descending_float16(): sort_3d(True, np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sort3d_float32(): sort_3d(False, np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sort3d_descending_float32(): diff --git a/tests/st/ops/gpu/test_spacetodepth_op.py b/tests/st/ops/gpu/test_spacetodepth_op.py index 40bd927e64a..897c1f1ab54 100644 --- a/tests/st/ops/gpu/test_spacetodepth_op.py +++ b/tests/st/ops/gpu/test_spacetodepth_op.py @@ -64,13 +64,13 @@ def SpaceToDepth_pynative(nptype): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_spacetodepth_graph_float32(): SpaceToDepth(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_spacetodepth_graph_float16(): @@ -124,13 +124,13 @@ def test_spacetodepth_graph_uint32(): def test_spacetodepth_graph_uint64(): SpaceToDepth(np.uint64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_spacetodepth_pynative_float32(): SpaceToDepth_pynative(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_spacetodepth_pynative_float16(): @@ -148,13 +148,13 @@ def test_spacetodepth_pynative_int32(): def test_spacetodepth_pynative_int64(): SpaceToDepth_pynative(np.int64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_spacetodepth_pynative_int8(): SpaceToDepth_pynative(np.int8) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_spacetodepth_pynative_int16(): diff --git a/tests/st/ops/gpu/test_sparse_apply_ftrl_op.py b/tests/st/ops/gpu/test_sparse_apply_ftrl_op.py index fbda364d9bc..da83c41eb15 100644 --- a/tests/st/ops/gpu/test_sparse_apply_ftrl_op.py +++ b/tests/st/ops/gpu/test_sparse_apply_ftrl_op.py @@ -49,7 +49,7 @@ class Net_half(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ftrl(): @@ -74,7 +74,7 @@ def test_ftrl(): assert np.all(sparse_apply_ftrl.var.data.asnumpy() == expect_var) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ftrl_sparse_int64_ind(): @@ -99,7 +99,7 @@ def test_ftrl_sparse_int64_ind(): assert np.all(sparse_apply_ftrl.var.data.asnumpy() == expect_var) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ftrl_half(): @@ -124,7 +124,7 @@ def test_ftrl_half(): assert np.all(sparse_apply_ftrl.var.data.asnumpy() == expect_var) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ftrl_sparse_half_int64_ind(): @@ -148,7 +148,7 @@ def test_ftrl_sparse_half_int64_ind(): sparse_apply_ftrl(gradient, indices) assert np.all(sparse_apply_ftrl.var.data.asnumpy() == expect_var) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ftrl_half_return_output(): diff --git a/tests/st/ops/gpu/test_sparse_apply_proximal_adagrad_op.py b/tests/st/ops/gpu/test_sparse_apply_proximal_adagrad_op.py index 2d6f1cfe73a..5cb08c70e8b 100644 --- a/tests/st/ops/gpu/test_sparse_apply_proximal_adagrad_op.py +++ b/tests/st/ops/gpu/test_sparse_apply_proximal_adagrad_op.py @@ -62,7 +62,7 @@ def test_small_shape(): np.testing.assert_array_almost_equal(output1.asnumpy(), expect1) np.testing.assert_array_almost_equal(output2.asnumpy(), expect2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_small_shape_input_update(): @@ -84,7 +84,7 @@ def test_small_shape_input_update(): np.testing.assert_array_almost_equal(net.var.data.asnumpy(), expect1) np.testing.assert_array_almost_equal(net.accum.data.asnumpy(), expect2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_parameter_lr_l1_l2(): @@ -105,7 +105,7 @@ def test_parameter_lr_l1_l2(): np.testing.assert_array_almost_equal(output1.asnumpy(), expect1) np.testing.assert_array_almost_equal(output2.asnumpy(), expect2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_with_np_arange(): @@ -126,7 +126,7 @@ def test_with_np_arange(): np.testing.assert_array_almost_equal(output1.asnumpy(), expect1) np.testing.assert_array_almost_equal(output2.asnumpy(), expect2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_large_shape(): diff --git a/tests/st/ops/gpu/test_sparse_gather_v2_op.py b/tests/st/ops/gpu/test_sparse_gather_v2_op.py index 49c401b7ba4..3a34a6ed2a4 100644 --- a/tests/st/ops/gpu/test_sparse_gather_v2_op.py +++ b/tests/st/ops/gpu/test_sparse_gather_v2_op.py @@ -31,7 +31,7 @@ class SparseGatherNet(nn.Cell): return self.gather(x, indices, 1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather0(): @@ -696,7 +696,7 @@ class SparseGatherNet1(nn.Cell): return self.gather(x, indices, -1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather1(): @@ -745,7 +745,7 @@ class SparseGatherNet2(nn.Cell): return self.gather(x, indices, 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather2(): @@ -785,7 +785,7 @@ class SparseGatherDynamicNet(nn.Cell): return self.gather(out, indices, 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_d(): @@ -829,7 +829,7 @@ class SparseGatherDynamicNet2(nn.Cell): return (out_1, out_2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gather_d_two_inputs(): diff --git a/tests/st/ops/gpu/test_sparse_softmax_cross_entropy_with_logits_op.py b/tests/st/ops/gpu/test_sparse_softmax_cross_entropy_with_logits_op.py index c677d8c79f3..637148d22a3 100644 --- a/tests/st/ops/gpu/test_sparse_softmax_cross_entropy_with_logits_op.py +++ b/tests/st/ops/gpu/test_sparse_softmax_cross_entropy_with_logits_op.py @@ -29,7 +29,7 @@ class NetSparseSoftmaxCrossEntropyWithLogits(nn.Cell): return self.loss(logits, labels) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sparse_softmax_cross_entropy_with_logits(): diff --git a/tests/st/ops/gpu/test_split.py b/tests/st/ops/gpu/test_split.py index 70161e6660c..a0ef5c8ce33 100644 --- a/tests/st/ops/gpu/test_split.py +++ b/tests/st/ops/gpu/test_split.py @@ -64,14 +64,14 @@ def test_split_basic_float16(): split_basic(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_split_basic_float32(): split_basic(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_split_basic_float64(): @@ -106,7 +106,7 @@ def test_split_basic_bool(): split_basic(np.bool) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_split_4d(): @@ -120,7 +120,7 @@ def test_split_4d(): assert (out.asnumpy() == y[i]).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_split_dynamic(): @@ -134,7 +134,7 @@ def test_split_dynamic(): assert (out.asnumpy() == x[i]).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_split_dynamic_axis1(): @@ -149,7 +149,7 @@ def test_split_dynamic_axis1(): assert (out.asnumpy() == y[i]).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_split_dynamic_axis2(): @@ -164,7 +164,7 @@ def test_split_dynamic_axis2(): assert (out.asnumpy() == y[i]).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_split_invalid_input(): diff --git a/tests/st/ops/gpu/test_sqrt_grad_grad_op.py b/tests/st/ops/gpu/test_sqrt_grad_grad_op.py index 0f7363a555b..265ab79f0f1 100644 --- a/tests/st/ops/gpu/test_sqrt_grad_grad_op.py +++ b/tests/st/ops/gpu/test_sqrt_grad_grad_op.py @@ -73,14 +73,14 @@ def sqrt_grad_grad_base(dtype, loss): assert np.allclose(dgrad_ms.asnumpy(), dgrad_np, loss, loss) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sqrt_grad_grad_float16(): sqrt_grad_grad_base(np.float16, 1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sqrt_grad_grad_float32(): diff --git a/tests/st/ops/gpu/test_sqrt_grad_op.py b/tests/st/ops/gpu/test_sqrt_grad_op.py index df9b2893cca..dba8e8a8940 100644 --- a/tests/st/ops/gpu/test_sqrt_grad_op.py +++ b/tests/st/ops/gpu/test_sqrt_grad_op.py @@ -31,7 +31,7 @@ class NetSqrtGrad(nn.Cell): return self.sqrt_grad(x, dx) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sqrt_grad(): diff --git a/tests/st/ops/gpu/test_sqrt_op.py b/tests/st/ops/gpu/test_sqrt_op.py index 16844f5beed..385f180dd05 100644 --- a/tests/st/ops/gpu/test_sqrt_op.py +++ b/tests/st/ops/gpu/test_sqrt_op.py @@ -35,19 +35,19 @@ def sqrt(nptype): def test_sqrt_float16(): sqrt(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sqrt_float32(): sqrt(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sqrt_float64(): sqrt(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_rsqrt(): diff --git a/tests/st/ops/gpu/test_square_op.py b/tests/st/ops/gpu/test_square_op.py index 588f1ff62c8..45ea592cccb 100644 --- a/tests/st/ops/gpu/test_square_op.py +++ b/tests/st/ops/gpu/test_square_op.py @@ -23,7 +23,7 @@ from mindspore.ops.operations import _inner_ops as inner from mindspore.ops import operations as P -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_square_normal(): @@ -54,7 +54,7 @@ class SqaureNetDynamic(nn.Cell): return self.square(x_dyn) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_square_dynamic(): diff --git a/tests/st/ops/gpu/test_squared_difference_op.py b/tests/st/ops/gpu/test_squared_difference_op.py index d36826e1b3c..8b21d16829e 100644 --- a/tests/st/ops/gpu/test_squared_difference_op.py +++ b/tests/st/ops/gpu/test_squared_difference_op.py @@ -30,7 +30,7 @@ class SquaredDifference(nn.Cell): return self.squaredDiff(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nobroadcast_f16(): @@ -45,7 +45,7 @@ def test_nobroadcast_f16(): assert np.all(output == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nobroadcast_f32(): @@ -60,7 +60,7 @@ def test_nobroadcast_f32(): assert np.all(output == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nobroadcast_int32(): @@ -75,7 +75,7 @@ def test_nobroadcast_int32(): assert np.all(output == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_int32(): @@ -90,7 +90,7 @@ def test_broadcast_int32(): assert np.all(output == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_f32(): @@ -105,7 +105,7 @@ def test_broadcast_f32(): assert np.all(output == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_f16(): @@ -120,7 +120,7 @@ def test_broadcast_f16(): assert np.all(output == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_bool(): @@ -137,7 +137,7 @@ def test_broadcast_bool(): assert np.all(double_check < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nobroadcast_bool(): @@ -154,7 +154,7 @@ def test_nobroadcast_bool(): assert np.all(double_check < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_int32_f16(): @@ -171,7 +171,7 @@ def test_broadcast_int32_f16(): assert np.all(double_check < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_int32_f32(): @@ -188,7 +188,7 @@ def test_broadcast_int32_f32(): assert np.all(double_check < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nobroadcast_int32_f16(): @@ -205,7 +205,7 @@ def test_nobroadcast_int32_f16(): assert np.all(double_check < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_nobroadcast_int32_f32(): @@ -222,7 +222,7 @@ def test_nobroadcast_int32_f32(): assert np.all(double_check < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_f32_scalar_tensor(): @@ -237,7 +237,7 @@ def test_broadcast_f32_scalar_tensor(): assert np.all(output == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_f32_tensor_tensor(): @@ -252,7 +252,7 @@ def test_broadcast_f32_tensor_tensor(): assert np.all(output == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_f32_tensor_tensor_dim_over_7(): @@ -267,7 +267,7 @@ def test_broadcast_f32_tensor_tensor_dim_over_7(): assert True -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_f32_tensor_tensor_cannot_brocast(): @@ -282,7 +282,7 @@ def test_broadcast_f32_tensor_tensor_cannot_brocast(): assert True -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_int_f32_precision(): @@ -299,7 +299,7 @@ def test_broadcast_int_f32_precision(): assert np.all(double_thousand < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_broadcast_type_error(): diff --git a/tests/st/ops/gpu/test_squeeze_op.py b/tests/st/ops/gpu/test_squeeze_op.py index 259fd8fbd52..98099d7985e 100644 --- a/tests/st/ops/gpu/test_squeeze_op.py +++ b/tests/st/ops/gpu/test_squeeze_op.py @@ -92,13 +92,13 @@ def test_squeeze_int64(): def test_squeeze_float16(): squeeze(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_squeeze_float32(): squeeze(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_squeeze_float64(): diff --git a/tests/st/ops/gpu/test_standard_normal.py b/tests/st/ops/gpu/test_standard_normal.py index deb1d0c1a7d..68453b0ebb3 100644 --- a/tests/st/ops/gpu/test_standard_normal.py +++ b/tests/st/ops/gpu/test_standard_normal.py @@ -32,7 +32,7 @@ class Net(nn.Cell): def construct(self): return self.stdnormal(self.shape) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net(): diff --git a/tests/st/ops/gpu/test_stridedslice_grad_op.py b/tests/st/ops/gpu/test_stridedslice_grad_op.py index 91760aba18b..720bb726acb 100644 --- a/tests/st/ops/gpu/test_stridedslice_grad_op.py +++ b/tests/st/ops/gpu/test_stridedslice_grad_op.py @@ -239,73 +239,73 @@ def strided_slice_grad(nptype): [0., 0., 0., 0., 0.]]]]]]]).astype(nptype) assert np.allclose(dx[0].asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_grad_float64(): strided_slice_grad(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_grad_float32(): strided_slice_grad(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_grad_float16(): strided_slice_grad(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_grad_int64(): strided_slice_grad(np.int64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_grad_int32(): strided_slice_grad(np.int32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_grad_int16(): strided_slice_grad(np.int16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_grad_int8(): strided_slice_grad(np.int8) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_grad_uint64(): strided_slice_grad(np.uint64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_grad_uint32(): strided_slice_grad(np.uint32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_grad_uint16(): strided_slice_grad(np.uint16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_grad_uint8(): strided_slice_grad(np.uint8) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_grad_bool(): diff --git a/tests/st/ops/gpu/test_stridedslice_op.py b/tests/st/ops/gpu/test_stridedslice_op.py index 5b19f13bb46..cecae763711 100644 --- a/tests/st/ops/gpu/test_stridedslice_op.py +++ b/tests/st/ops/gpu/test_stridedslice_op.py @@ -102,13 +102,13 @@ def strided_slice(nptype): [[[2122.]]]]]]]).astype(nptype) assert np.allclose(y.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_float64(): strided_slice(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_float32(): @@ -138,7 +138,7 @@ def test_strided_slice_int32(): def test_strided_slice_int16(): strided_slice(np.int16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_strided_slice_int8(): diff --git a/tests/st/ops/gpu/test_sub_op.py b/tests/st/ops/gpu/test_sub_op.py index ab324682649..2c385fd7952 100644 --- a/tests/st/ops/gpu/test_sub_op.py +++ b/tests/st/ops/gpu/test_sub_op.py @@ -135,19 +135,19 @@ def sub(nptype): assert output0.shape == expect0.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sub_float64(): sub(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sub_float32(): sub(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sub_float16(): diff --git a/tests/st/ops/gpu/test_tanh_grad_grad_op.py b/tests/st/ops/gpu/test_tanh_grad_grad_op.py index f7028206768..a9f6fb313d1 100644 --- a/tests/st/ops/gpu/test_tanh_grad_grad_op.py +++ b/tests/st/ops/gpu/test_tanh_grad_grad_op.py @@ -71,14 +71,14 @@ def tanh_grad_grad_base(dtype, loss): assert np.allclose(dgrad_ms.asnumpy(), dgrad_np, loss, loss) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_tanh_grad_grad_float16(): tanh_grad_grad_base(np.float16, 1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_tanh_grad_grad_float32(): diff --git a/tests/st/ops/gpu/test_tanh_op.py b/tests/st/ops/gpu/test_tanh_op.py index b44c59570c7..d20667b141b 100644 --- a/tests/st/ops/gpu/test_tanh_op.py +++ b/tests/st/ops/gpu/test_tanh_op.py @@ -45,7 +45,7 @@ class Grad(nn.Cell): return gout -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_Tanh(): @@ -73,7 +73,7 @@ def test_Tanh(): assert np.allclose(output[0].asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_Tanh_fp16(): diff --git a/tests/st/ops/gpu/test_tensordot_op.py b/tests/st/ops/gpu/test_tensordot_op.py index c43b9eb7775..5f52c60c632 100644 --- a/tests/st/ops/gpu/test_tensordot_op.py +++ b/tests/st/ops/gpu/test_tensordot_op.py @@ -43,7 +43,7 @@ class GradNetwork(nn.Cell): return gout -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_tensor_dot_fp32(): @@ -167,7 +167,7 @@ def test_tensor_dot_fp16(): assert np.allclose(ms_result_np, np_result, rtol=1e-3, atol=6e0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_tensor_dot_outer(): @@ -188,7 +188,7 @@ def test_tensor_dot_outer(): np.testing.assert_array_almost_equal(ms_result_np, np_result) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_tensor_dot_reverse_axes(): @@ -265,7 +265,7 @@ def test_tensor_dot_reverse_axes(): np.testing.assert_array_almost_equal(ms_result_np, expected_result) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_tensor_dot_backprop(): diff --git a/tests/st/ops/gpu/test_time_distributed_op.py b/tests/st/ops/gpu/test_time_distributed_op.py index 6ff85747b2a..e171e218dd8 100644 --- a/tests/st/ops/gpu/test_time_distributed_op.py +++ b/tests/st/ops/gpu/test_time_distributed_op.py @@ -33,7 +33,7 @@ class TestTimeDistributed(nn.Cell): return self.time_distributed(inputs) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_time_distributed_conv2d(): @@ -48,7 +48,7 @@ def test_time_distributed_conv2d(): print("Conv2D layer wrapped successful") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_time_distributed_maxpool2d(): @@ -63,7 +63,7 @@ def test_time_distributed_maxpool2d(): print("MaxPooling2D layer wrapped successful") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_time_distributed_dense(): @@ -78,7 +78,7 @@ def test_time_distributed_dense(): print("Dense layer wrapped successful") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_time_distributed_dense_with_reshape_axis_not_first(): @@ -93,7 +93,7 @@ def test_time_distributed_dense_with_reshape_axis_not_first(): print("Dense layer wrapped successful") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_time_distributed_argmax(): @@ -108,7 +108,7 @@ def test_time_distributed_argmax(): print("Argmax op wrapped successful") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_time_distributed_flatten(): @@ -123,7 +123,7 @@ def test_time_distributed_flatten(): print("Flatten op wrapped successful") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_time_distributed_conv2d_no_reshape_axis(): @@ -138,7 +138,7 @@ def test_time_distributed_conv2d_no_reshape_axis(): print("Conv2D layer with no reshape axis wrapped successful") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_time_distributed_maxpool2d_no_reshape_axis(): @@ -153,7 +153,7 @@ def test_time_distributed_maxpool2d_no_reshape_axis(): print("MaxPooling2D layer with no reshape axis wrapped successful") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_time_distributed_dense_no_reshape_axis(): @@ -168,7 +168,7 @@ def test_time_distributed_dense_no_reshape_axis(): print("Dense layer with no reshape axis wrapped successful") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_time_distributed_argmax_no_reshape_axis(): @@ -183,7 +183,7 @@ def test_time_distributed_argmax_no_reshape_axis(): print("Argmax op with no reshape axis wrapped successful") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_time_distributed_flatten_no_reshape_axis(): diff --git a/tests/st/ops/gpu/test_topk_op.py b/tests/st/ops/gpu/test_topk_op.py index e938bd73b8f..68d5c61784a 100644 --- a/tests/st/ops/gpu/test_topk_op.py +++ b/tests/st/ops/gpu/test_topk_op.py @@ -40,7 +40,7 @@ def test_topk_small_2d(): assert np.allclose(ms_output[0].asnumpy(), np_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_topk_3d(): @@ -58,7 +58,7 @@ def test_topk_3d(): assert np.allclose(ms_output[0].asnumpy(), np_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_topk_big_2d(): @@ -90,7 +90,7 @@ def test_topk_big_2d(): assert np.allclose(ms_output[0].asnumpy(), np_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_topk_big_k(): @@ -102,7 +102,7 @@ def test_topk_big_k(): assert np.allclose(ms_output[0].asnumpy(), np_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_topk_1d(): diff --git a/tests/st/ops/gpu/test_transpose_op.py b/tests/st/ops/gpu/test_transpose_op.py index 5aa517fe904..f958c9dec8c 100644 --- a/tests/st/ops/gpu/test_transpose_op.py +++ b/tests/st/ops/gpu/test_transpose_op.py @@ -236,73 +236,73 @@ def transpose_d2(nptype): assert (output_1.asnumpy() == expect_1).all() assert (output_2.asnumpy() == expect_2).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_transpose_float32(): transpose1(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_transpose_float16(): transpose1(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_transpose_int32(): transpose1(np.int32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_transpose_int64(): transpose1(np.int64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_transpose_dynamic_int64(): transpose_d(np.int64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_transpose_dynamic_two_inputs_int64(): transpose_d2(np.int64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_transpose_dynamic_float32(): transpose_d(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_transpose_dynamic_float16(): transpose_d(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_transpose_dynamic_int32(): transpose_d(np.int32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_transpose_dynamic_two_inputs_float32(): transpose_d2(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_transpose_dynamic_two_inputs_float16(): transpose_d2(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_transpose_dynamic_two_inputs_int32(): diff --git a/tests/st/ops/gpu/test_uniform_candidate_sampler_op.py b/tests/st/ops/gpu/test_uniform_candidate_sampler_op.py index 38258e39e7e..49fa78cc8aa 100644 --- a/tests/st/ops/gpu/test_uniform_candidate_sampler_op.py +++ b/tests/st/ops/gpu/test_uniform_candidate_sampler_op.py @@ -71,7 +71,7 @@ def uniform_candidate_sampler_hit(x, num_true, num_sampled, unique, range_max, s return out1, out2, out3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_uniform_candidate_sampler_unique_1_true(): @@ -85,7 +85,7 @@ def test_uniform_candidate_sampler_unique_1_true(): np.testing.assert_array_equal(ms2, expected_2) np.testing.assert_array_equal(ms3, expected_3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_uniform_candidate_sampler_not_unique_1_true(): @@ -99,7 +99,7 @@ def test_uniform_candidate_sampler_not_unique_1_true(): np.testing.assert_array_equal(ms2, expected_2) np.testing.assert_array_equal(ms3, expected_3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_uniform_candidate_sampler_unique_2_true(): @@ -114,7 +114,7 @@ def test_uniform_candidate_sampler_unique_2_true(): np.testing.assert_array_equal(ms2, expected_2) np.testing.assert_array_equal(ms3, expected_3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_uniform_candidate_sampler_not_unique_2_true(): @@ -130,7 +130,7 @@ def test_uniform_candidate_sampler_not_unique_2_true(): np.testing.assert_array_equal(ms2, expected_2) np.testing.assert_array_equal(ms3, expected_3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_uniform_candidate_sampler_large(): @@ -149,7 +149,7 @@ def test_uniform_candidate_sampler_large(): np.testing.assert_array_equal(ms3, expected_3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_uniform_candidate_sampler_large_random(): @@ -163,7 +163,7 @@ def test_uniform_candidate_sampler_large_random(): np.testing.assert_array_equal(ms2, expected_2) np.testing.assert_array_equal(ms3, expected_3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_uniform_candidate_sampler_large_random_int64_input(): @@ -177,7 +177,7 @@ def test_uniform_candidate_sampler_large_random_int64_input(): np.testing.assert_array_equal(ms2, expected_2) np.testing.assert_array_equal(ms3, expected_3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_uniform_candidate_sampler_unique_1_true_hit(): @@ -187,7 +187,7 @@ def test_uniform_candidate_sampler_unique_1_true_hit(): np.testing.assert_array_equal(ms1.asnumpy(), expected_1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_uniform_candidate_sampler_unique_1_true_no_hit(): diff --git a/tests/st/ops/gpu/test_uniform_int.py b/tests/st/ops/gpu/test_uniform_int.py index c4ce3a138c0..87de908b8de 100644 --- a/tests/st/ops/gpu/test_uniform_int.py +++ b/tests/st/ops/gpu/test_uniform_int.py @@ -32,7 +32,7 @@ class Net(nn.Cell): def construct(self, a, b): return self.uniformint(self.shape, a, b) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net_1D(): diff --git a/tests/st/ops/gpu/test_uniform_real.py b/tests/st/ops/gpu/test_uniform_real.py index 56fe9325240..fe884738189 100644 --- a/tests/st/ops/gpu/test_uniform_real.py +++ b/tests/st/ops/gpu/test_uniform_real.py @@ -30,7 +30,7 @@ class Net(nn.Cell): def construct(self): return self.uniformreal(self.shape) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_net(): diff --git a/tests/st/ops/gpu/test_unique_op.py b/tests/st/ops/gpu/test_unique_op.py index 1ab835cd9af..d62f8c3cbf4 100644 --- a/tests/st/ops/gpu/test_unique_op.py +++ b/tests/st/ops/gpu/test_unique_op.py @@ -46,7 +46,7 @@ class NetUniqueDynamic(nn.Cell): return x_unique, x_idx, x_split -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_1d(): @@ -60,7 +60,7 @@ def test_unique_1d(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_1d_float(): @@ -74,7 +74,7 @@ def test_unique_1d_float(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_1d_sorted(): @@ -88,7 +88,7 @@ def test_unique_1d_sorted(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_zeros(): @@ -102,7 +102,7 @@ def test_unique_zeros(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_large(): @@ -120,7 +120,7 @@ def test_unique_large(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_1d_half(): @@ -134,7 +134,7 @@ def test_unique_1d_half(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_1d_sorted_half(): @@ -148,7 +148,7 @@ def test_unique_1d_sorted_half(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_zeros_half(): @@ -162,7 +162,7 @@ def test_unique_zeros_half(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_large_half(): @@ -180,7 +180,7 @@ def test_unique_large_half(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_1d_int32(): @@ -194,7 +194,7 @@ def test_unique_1d_int32(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_1d_sorted_int32(): @@ -208,7 +208,7 @@ def test_unique_1d_sorted_int32(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_zeros_int32(): @@ -222,7 +222,7 @@ def test_unique_zeros_int32(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_large_int32(): @@ -240,7 +240,7 @@ def test_unique_large_int32(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_dynamic(): @@ -269,7 +269,7 @@ def test_unique_dynamic(): assert (out.asnumpy() == expt_split2[i]).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_1d_int64(): @@ -285,7 +285,7 @@ def test_unique_1d_int64(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_1d_sorted_int64(): @@ -299,7 +299,7 @@ def test_unique_1d_sorted_int64(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_zeros_int64(): @@ -313,7 +313,7 @@ def test_unique_zeros_int64(): assert (x_idx.asnumpy() == exp_idx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unique_large_int64(): diff --git a/tests/st/ops/gpu/test_unpack_op.py b/tests/st/ops/gpu/test_unpack_op.py index cddd51a10f4..92064375eed 100644 --- a/tests/st/ops/gpu/test_unpack_op.py +++ b/tests/st/ops/gpu/test_unpack_op.py @@ -91,14 +91,14 @@ def unstack_pynative(nptype): assert (output[i].asnumpy() == exp).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unstack_graph_float32(): unstack(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unstack_graph_float16(): @@ -133,14 +133,14 @@ def test_unstack_graph_bool(): unstack(np.bool) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unstack_pynative_float32(): unstack_pynative(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_unstack_pynative_float16(): diff --git a/tests/st/ops/gpu/test_unsorted_segment_max.py b/tests/st/ops/gpu/test_unsorted_segment_max.py index b45b699e681..d49247703ba 100644 --- a/tests/st/ops/gpu/test_unsorted_segment_max.py +++ b/tests/st/ops/gpu/test_unsorted_segment_max.py @@ -35,7 +35,7 @@ class UnsortedSegmentMaxNet(nn.Cell): return self.unsorted_segment_max(data, ids, self.num_segments) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_1d_int32(): @@ -49,7 +49,7 @@ def test_1d_int32(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_2d_int32(): @@ -68,7 +68,7 @@ def test_2d_int32(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_3d_float16_int64(): @@ -107,7 +107,7 @@ def test_3d_float16_int64(): np.testing.assert_array_almost_equal(output, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_3d_float32_int64(): @@ -136,7 +136,7 @@ def test_3d_float32_int64(): np.testing.assert_array_almost_equal(output, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_3d_single_init(): @@ -223,7 +223,7 @@ class UnsortedSegmentMaxDynNet(nn.Cell): return self.unsorted_segment_max(data, ids, self.num_segments) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_3d_float32_dyn_ab(): @@ -261,7 +261,7 @@ def test_3d_float32_dyn_ab(): np.testing.assert_array_almost_equal(output, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_3d_single_init_dyn_a(): @@ -333,7 +333,7 @@ def test_3d_single_init_dyn_a(): np.testing.assert_array_almost_equal(output, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_3d_single_init_dyn_b(): diff --git a/tests/st/ops/gpu/test_unsorted_segment_min.py b/tests/st/ops/gpu/test_unsorted_segment_min.py index 9092d54fd64..15c1d922d12 100644 --- a/tests/st/ops/gpu/test_unsorted_segment_min.py +++ b/tests/st/ops/gpu/test_unsorted_segment_min.py @@ -35,7 +35,7 @@ class UnsortedSegmentMinNet(nn.Cell): return self.unsorted_segment_min(data, ids, self.num_segments) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_1d_int32(): @@ -49,7 +49,7 @@ def test_1d_int32(): np.testing.assert_array_almost_equal(output, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_2d_int32(): @@ -68,7 +68,7 @@ def test_2d_int32(): np.testing.assert_array_almost_equal(output, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_3d_float16(): @@ -107,7 +107,7 @@ def test_3d_float16(): np.testing.assert_array_almost_equal(output, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_3d_float32(): @@ -223,7 +223,7 @@ class UnsortedSegmentMinDynNet(nn.Cell): return self.unsorted_segment_min(data, ids, self.num_segments) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_3d_float32_ab_dyn(): @@ -255,7 +255,7 @@ def test_3d_float32_ab_dyn(): np.testing.assert_array_almost_equal(output, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_3d_float32_a_dyn(): @@ -327,7 +327,7 @@ def test_3d_float32_a_dyn(): np.testing.assert_array_almost_equal(output, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_3d_float32_b_dyn(): diff --git a/tests/st/ops/gpu/test_unsorted_segment_sum.py b/tests/st/ops/gpu/test_unsorted_segment_sum.py index 6a317b27998..9eb5cd05dd5 100644 --- a/tests/st/ops/gpu/test_unsorted_segment_sum.py +++ b/tests/st/ops/gpu/test_unsorted_segment_sum.py @@ -49,7 +49,7 @@ def test_1D(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_2D(): @@ -69,7 +69,7 @@ def test_2D(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_3D(): @@ -126,7 +126,7 @@ class UnsortedSegmentSumDynNet(nn.Cell): return self.unsorted_segment_sum(data, ids, self.num_segments) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dyn_ab(): @@ -180,7 +180,7 @@ def test_dyn_ab(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dyn_a(): @@ -246,7 +246,7 @@ def test_dyn_a(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dyn_b(): diff --git a/tests/st/ops/gpu/test_zeroslike_op.py b/tests/st/ops/gpu/test_zeroslike_op.py index 94d3b764b97..9d27b511bf0 100644 --- a/tests/st/ops/gpu/test_zeroslike_op.py +++ b/tests/st/ops/gpu/test_zeroslike_op.py @@ -31,7 +31,7 @@ class NetZerosLike(nn.Cell): return self.zeros_like(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_ZerosLike(): @@ -135,7 +135,7 @@ def test_zeros_like_dynamic_float16(): expected = np.zeros([3, 4, 1, 2, 5]) np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_zeros_like_dynamic_float32(): @@ -144,7 +144,7 @@ def test_zeros_like_dynamic_float32(): expected = np.zeros([3, 7, 3]) np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_zeros_like_dynamic_float64(): @@ -153,7 +153,7 @@ def test_zeros_like_dynamic_float64(): expected = np.zeros([2, 1, 1]) np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_zeros_like_dynamic_multiple_inputs(): diff --git a/tests/st/ops/graph_kernel/test_argmax.py b/tests/st/ops/graph_kernel/test_argmax.py index 458aebd45c2..47b38bb9f31 100644 --- a/tests/st/ops/graph_kernel/test_argmax.py +++ b/tests/st/ops/graph_kernel/test_argmax.py @@ -51,7 +51,7 @@ def test_argmax(): assert np.allclose(expect.asnumpy(), output.asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_argmax_gpu(): diff --git a/tests/st/ops/graph_kernel/test_atomic_add.py b/tests/st/ops/graph_kernel/test_atomic_add.py index d069d17f6b5..da2a57fab4f 100644 --- a/tests/st/ops/graph_kernel/test_atomic_add.py +++ b/tests/st/ops/graph_kernel/test_atomic_add.py @@ -100,7 +100,7 @@ def atomic_add_multi_output(): assert res -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_atomic_add_sum_output_gpu(): @@ -108,7 +108,7 @@ def test_atomic_add_sum_output_gpu(): atomic_add_sum_output() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_atomic_add_single_output_gpu(): diff --git a/tests/st/ops/graph_kernel/test_batchnorm.py b/tests/st/ops/graph_kernel/test_batchnorm.py index 57243340bc0..9f90f2c35ca 100644 --- a/tests/st/ops/graph_kernel/test_batchnorm.py +++ b/tests/st/ops/graph_kernel/test_batchnorm.py @@ -68,14 +68,14 @@ def test_bn_infer(): assert np.allclose(infer_expect[0][0].asnumpy(), infer_output[0][0].asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_bn_train_gpu(): context.set_context(mode=context.GRAPH_MODE, device_target="GPU") test_bn_train() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_bn_infer_gpu(): diff --git a/tests/st/ops/graph_kernel/test_batchnorm_grad.py b/tests/st/ops/graph_kernel/test_batchnorm_grad.py index 0abc3229279..ec93f7ec544 100644 --- a/tests/st/ops/graph_kernel/test_batchnorm_grad.py +++ b/tests/st/ops/graph_kernel/test_batchnorm_grad.py @@ -71,14 +71,14 @@ def test_bn_grad_infer(): assert np.allclose(expect[1].asnumpy(), output[1].asnumpy(), 0.0001, 0.0001) assert np.allclose(expect[2].asnumpy(), output[2].asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_bn_grad_train_gpu(): context.set_context(mode=context.GRAPH_MODE, device_target="GPU") test_bn_grad_train() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_bn_grad_infer_gpu(): diff --git a/tests/st/ops/graph_kernel/test_bias_add.py b/tests/st/ops/graph_kernel/test_bias_add.py index 0f87ac1badb..a89fda5e930 100644 --- a/tests/st/ops/graph_kernel/test_bias_add.py +++ b/tests/st/ops/graph_kernel/test_bias_add.py @@ -49,7 +49,7 @@ def test_bias_add(shape1, shape2, data_format, dtype): assert np.allclose(expect_np, output_np, 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_bias_add_gpu(): diff --git a/tests/st/ops/graph_kernel/test_compare.py b/tests/st/ops/graph_kernel/test_compare.py index feb6e9d6c21..d344e250e8c 100644 --- a/tests/st/ops/graph_kernel/test_compare.py +++ b/tests/st/ops/graph_kernel/test_compare.py @@ -185,7 +185,7 @@ def test_greater_equal_net(): assert out_gk_on_3.shape == out_gk_off_3.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_less_gpu(): diff --git a/tests/st/ops/graph_kernel/test_cse.py b/tests/st/ops/graph_kernel/test_cse.py index 757fa64b291..6a3f79833b7 100644 --- a/tests/st/ops/graph_kernel/test_cse.py +++ b/tests/st/ops/graph_kernel/test_cse.py @@ -46,7 +46,7 @@ def test_basic(): assert res -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_basic_gpu(): diff --git a/tests/st/ops/graph_kernel/test_dropoutgrad_reducesum_stitch.py b/tests/st/ops/graph_kernel/test_dropoutgrad_reducesum_stitch.py index e879b14afb7..3373366818c 100644 --- a/tests/st/ops/graph_kernel/test_dropoutgrad_reducesum_stitch.py +++ b/tests/st/ops/graph_kernel/test_dropoutgrad_reducesum_stitch.py @@ -82,7 +82,7 @@ def test_dropoutgrad_reducesum(shape, dtype): compare_result(expect, output, dtype) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_dropoutgrad_reducesum_gpu(): diff --git a/tests/st/ops/graph_kernel/test_equal_count.py b/tests/st/ops/graph_kernel/test_equal_count.py index 43984a24198..76ed8a3da69 100644 --- a/tests/st/ops/graph_kernel/test_equal_count.py +++ b/tests/st/ops/graph_kernel/test_equal_count.py @@ -46,7 +46,7 @@ def basic_test(datatype): assert np.allclose(expect_np, output_np, 1.e-4, 1.e-7) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_fp16(): @@ -59,7 +59,7 @@ def test_gpu_fp16(): basic_test(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_fp32(): diff --git a/tests/st/ops/graph_kernel/test_expand_dims.py b/tests/st/ops/graph_kernel/test_expand_dims.py index f1180018b96..4540e97946f 100644 --- a/tests/st/ops/graph_kernel/test_expand_dims.py +++ b/tests/st/ops/graph_kernel/test_expand_dims.py @@ -43,7 +43,7 @@ def test_expand_dims(shape, dtype, axis): assert np.allclose(expect.asnumpy(), output.asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_expand_dims_gpu(): diff --git a/tests/st/ops/graph_kernel/test_fused_adam.py b/tests/st/ops/graph_kernel/test_fused_adam.py index 2fe17dfd34d..de493f310f2 100644 --- a/tests/st/ops/graph_kernel/test_fused_adam.py +++ b/tests/st/ops/graph_kernel/test_fused_adam.py @@ -217,7 +217,7 @@ def test_adam_side_effect(): rtol=1.e-4, atol=1.e-8, equal_nan=True) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_adam_gpu(): @@ -232,7 +232,7 @@ def test_adam_ascend(): test_adam() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_adam_weight_decay_gpu(): @@ -247,7 +247,7 @@ def test_adam_weight_decay_ascend(): test_adam_weight_decay() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_adam_side_effect_gpu(): diff --git a/tests/st/ops/graph_kernel/test_gelu.py b/tests/st/ops/graph_kernel/test_gelu.py index a8434976799..86d89b78cd1 100644 --- a/tests/st/ops/graph_kernel/test_gelu.py +++ b/tests/st/ops/graph_kernel/test_gelu.py @@ -84,7 +84,7 @@ def test_gelu_grad(): assert res -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gelu_gpu(): @@ -100,7 +100,7 @@ def test_gelu_ascend(): test_gelu() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gelu_grad_gpu(): diff --git a/tests/st/ops/graph_kernel/test_identity.py b/tests/st/ops/graph_kernel/test_identity.py index 08365c682e2..ec935a6b105 100644 --- a/tests/st/ops/graph_kernel/test_identity.py +++ b/tests/st/ops/graph_kernel/test_identity.py @@ -41,7 +41,7 @@ def test_basic(dtype): output_np = output.asnumpy().copy() assert np.allclose(expect_np, output_np, 1.e-4, 1.e-7) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_1(): diff --git a/tests/st/ops/graph_kernel/test_layernorm.py b/tests/st/ops/graph_kernel/test_layernorm.py index 2021e128824..58c59cc7d2c 100644 --- a/tests/st/ops/graph_kernel/test_layernorm.py +++ b/tests/st/ops/graph_kernel/test_layernorm.py @@ -121,7 +121,7 @@ def test_layernorm_grad(shape, dtype, begin_norm_axis=-1, begin_params_axis=-1): compare_result(expect, output, dtype) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernorm_gpu(): @@ -139,7 +139,7 @@ def test_layernorm_ascend(): test_layernorm([4, 32, 32], np.float32, -1, -1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernorm_grad_gpu(): diff --git a/tests/st/ops/graph_kernel/test_layernorm_stitch.py b/tests/st/ops/graph_kernel/test_layernorm_stitch.py index d0c842dc5cc..7c46f958cee 100644 --- a/tests/st/ops/graph_kernel/test_layernorm_stitch.py +++ b/tests/st/ops/graph_kernel/test_layernorm_stitch.py @@ -79,7 +79,7 @@ def test_layernorm(shape1, shape2, dtype): compare_result(expect, output, dtype) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernorm_gpu(): diff --git a/tests/st/ops/graph_kernel/test_logsoftmax.py b/tests/st/ops/graph_kernel/test_logsoftmax.py index d2bf189d6a2..a9e88ad03c2 100644 --- a/tests/st/ops/graph_kernel/test_logsoftmax.py +++ b/tests/st/ops/graph_kernel/test_logsoftmax.py @@ -91,7 +91,7 @@ def test_logsoftmaxgrad(): assert np.allclose(dx[0].asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_logsoftmax_gpu(): @@ -99,7 +99,7 @@ def test_logsoftmax_gpu(): test_logsoftmax() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_logsoftmaxgrad_gpu(): diff --git a/tests/st/ops/graph_kernel/test_low_precision.py b/tests/st/ops/graph_kernel/test_low_precision.py index 75ecb0d3a1f..3893880d59a 100644 --- a/tests/st/ops/graph_kernel/test_low_precision.py +++ b/tests/st/ops/graph_kernel/test_low_precision.py @@ -85,7 +85,7 @@ def test_case_2(): output_np = output[0].asnumpy().copy() assert np.allclose(expect_np, output_np, 1.e-2, 1.e-2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_case_1(): @@ -93,7 +93,7 @@ def test_gpu_case_1(): context.set_context(graph_kernel_flags="--enable_low_precision=true --disable_pass=highlevelopt2.atomic_clean") test_case_1() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_case_2(): diff --git a/tests/st/ops/graph_kernel/test_matmul_to_mul.py b/tests/st/ops/graph_kernel/test_matmul_to_mul.py index 951ea112e4e..d7286119961 100644 --- a/tests/st/ops/graph_kernel/test_matmul_to_mul.py +++ b/tests/st/ops/graph_kernel/test_matmul_to_mul.py @@ -76,7 +76,7 @@ def test_batchmatmul_ascend(): context.set_context(mode=context.GRAPH_MODE, device_target="Ascend") test_batchmatmul() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_matmul_gpu(): diff --git a/tests/st/ops/graph_kernel/test_maximum_grad.py b/tests/st/ops/graph_kernel/test_maximum_grad.py index 286c5d1e735..9092483dad6 100644 --- a/tests/st/ops/graph_kernel/test_maximum_grad.py +++ b/tests/st/ops/graph_kernel/test_maximum_grad.py @@ -57,7 +57,7 @@ def test_maximum_grad(): -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_maximum_grad_gpu(): diff --git a/tests/st/ops/graph_kernel/test_minimum_grad.py b/tests/st/ops/graph_kernel/test_minimum_grad.py index 1a05b63f7f3..51a02ba7e8c 100644 --- a/tests/st/ops/graph_kernel/test_minimum_grad.py +++ b/tests/st/ops/graph_kernel/test_minimum_grad.py @@ -56,7 +56,7 @@ def test_minimum_grad(): assert np.allclose(result_on[1], result_off[1], rtol=1.e-4, atol=1.e-8, equal_nan=True) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_minimum_grad_gpu(): diff --git a/tests/st/ops/graph_kernel/test_oneslike.py b/tests/st/ops/graph_kernel/test_oneslike.py index 5f4f6e14baa..dd043770a08 100644 --- a/tests/st/ops/graph_kernel/test_oneslike.py +++ b/tests/st/ops/graph_kernel/test_oneslike.py @@ -41,14 +41,14 @@ def test_basic(shape, dtype, nptype): output = get_output(shape, dtype, nptype, True) assert np.allclose(expect.asnumpy(), output.asnumpy(), 1.e-4, 1.e-7) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_1(): context.set_context(mode=context.GRAPH_MODE, device_target="GPU") test_basic((2, 16), mstype.float16, np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_2(): diff --git a/tests/st/ops/graph_kernel/test_optimize_assign.py b/tests/st/ops/graph_kernel/test_optimize_assign.py index 026acd2b347..8d0abfead08 100644 --- a/tests/st/ops/graph_kernel/test_optimize_assign.py +++ b/tests/st/ops/graph_kernel/test_optimize_assign.py @@ -88,14 +88,14 @@ def test_opt_assign_output_2(): assert res -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_opt_assign_gpu_1(): test_opt_assign_output_1() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_opt_assign_gpu_2(): diff --git a/tests/st/ops/graph_kernel/test_recompute.py b/tests/st/ops/graph_kernel/test_recompute.py index 6d0bd3cbe1e..dd8c34194b8 100644 --- a/tests/st/ops/graph_kernel/test_recompute.py +++ b/tests/st/ops/graph_kernel/test_recompute.py @@ -161,21 +161,21 @@ def test_gpu_2(): context.set_context(mode=context.GRAPH_MODE, device_target="GPU") test_basic2(Net2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_3(): context.set_context(mode=context.GRAPH_MODE, device_target="GPU") test_basic2(Net3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_4(): context.set_context(mode=context.GRAPH_MODE, device_target="GPU") test_basic1(Net4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_5(): diff --git a/tests/st/ops/graph_kernel/test_reduce_max.py b/tests/st/ops/graph_kernel/test_reduce_max.py index d892ba238a9..958befad6f2 100644 --- a/tests/st/ops/graph_kernel/test_reduce_max.py +++ b/tests/st/ops/graph_kernel/test_reduce_max.py @@ -60,7 +60,7 @@ def test_reduce_max(): assert np.allclose(expect.asnumpy(), output.asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reduce_max_gpu(): diff --git a/tests/st/ops/graph_kernel/test_reduce_mean.py b/tests/st/ops/graph_kernel/test_reduce_mean.py index a7fb04e76d4..a27a334100a 100644 --- a/tests/st/ops/graph_kernel/test_reduce_mean.py +++ b/tests/st/ops/graph_kernel/test_reduce_mean.py @@ -40,7 +40,7 @@ def test_reduce_mean(): assert res -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reduce_mean_gpu(): diff --git a/tests/st/ops/graph_kernel/test_reduce_min.py b/tests/st/ops/graph_kernel/test_reduce_min.py index dfdda36d885..17184d263dd 100644 --- a/tests/st/ops/graph_kernel/test_reduce_min.py +++ b/tests/st/ops/graph_kernel/test_reduce_min.py @@ -60,7 +60,7 @@ def test_reduce_min(): assert np.allclose(expect.asnumpy(), output.asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_reduce_min_gpu(): diff --git a/tests/st/ops/graph_kernel/test_relu.py b/tests/st/ops/graph_kernel/test_relu.py index b9e4ad3169a..b7c76da4013 100644 --- a/tests/st/ops/graph_kernel/test_relu.py +++ b/tests/st/ops/graph_kernel/test_relu.py @@ -47,7 +47,7 @@ def test_relu(shape, dtype): assert np.allclose(expect_np, output_np, 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_relu_gpu(): diff --git a/tests/st/ops/graph_kernel/test_relu_grad.py b/tests/st/ops/graph_kernel/test_relu_grad.py index 1468771e4a5..9eefa06746c 100644 --- a/tests/st/ops/graph_kernel/test_relu_grad.py +++ b/tests/st/ops/graph_kernel/test_relu_grad.py @@ -48,7 +48,7 @@ def test_relu_grad(shape1, shape2, dtype): assert np.allclose(expect_np, output_np, 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_relu_grad_gpu(): diff --git a/tests/st/ops/graph_kernel/test_reorder_ops.py b/tests/st/ops/graph_kernel/test_reorder_ops.py index 391868c14d3..3d9fc21572c 100644 --- a/tests/st/ops/graph_kernel/test_reorder_ops.py +++ b/tests/st/ops/graph_kernel/test_reorder_ops.py @@ -81,7 +81,7 @@ def test_castdown(): assert np.allclose(expect_np, output_np, 1e-3, 1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_castup_gpu(): @@ -98,7 +98,7 @@ def test_castup_ascend(): test_castup() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_castdown_gpu(): diff --git a/tests/st/ops/graph_kernel/test_sigmoid.py b/tests/st/ops/graph_kernel/test_sigmoid.py index 4d72a393057..68b6af4e255 100644 --- a/tests/st/ops/graph_kernel/test_sigmoid.py +++ b/tests/st/ops/graph_kernel/test_sigmoid.py @@ -41,7 +41,7 @@ class NetSigmoidGrad(nn.Cell): return self.sigmoid_grad(y, dy) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sigmoid(): @@ -64,7 +64,7 @@ def test_sigmoid(): assert np.all(abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sigmoid_grad(): diff --git a/tests/st/ops/graph_kernel/test_sigmoid_cross_entropy_with_logits.py b/tests/st/ops/graph_kernel/test_sigmoid_cross_entropy_with_logits.py index 94a86bb5819..13391614a55 100644 --- a/tests/st/ops/graph_kernel/test_sigmoid_cross_entropy_with_logits.py +++ b/tests/st/ops/graph_kernel/test_sigmoid_cross_entropy_with_logits.py @@ -41,7 +41,7 @@ class NetSigmoidCrossEntropyWithLogitsGrad(nn.Cell): return self.sigmoid_cross_entropy_with_logits_grad(logits, labels, dout) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sigmoid_cross_entropy_with_logits(): @@ -67,7 +67,7 @@ def test_sigmoid_cross_entropy_with_logits(): assert np.all(abs(diff) < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_sigmoid_cross_entropy_with_logits_grad(): diff --git a/tests/st/ops/graph_kernel/test_simplify.py b/tests/st/ops/graph_kernel/test_simplify.py index d75d8076a17..41172c8b181 100644 --- a/tests/st/ops/graph_kernel/test_simplify.py +++ b/tests/st/ops/graph_kernel/test_simplify.py @@ -99,7 +99,7 @@ def test_empty_graph(): assert res -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_basic_gpu(): diff --git a/tests/st/ops/graph_kernel/test_slice.py b/tests/st/ops/graph_kernel/test_slice.py index 118675b46db..d7a13107302 100644 --- a/tests/st/ops/graph_kernel/test_slice.py +++ b/tests/st/ops/graph_kernel/test_slice.py @@ -47,7 +47,7 @@ def test_slice(): assert np.allclose(expect.asnumpy(), output.asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_slice_gpu(): diff --git a/tests/st/ops/graph_kernel/test_softmax.py b/tests/st/ops/graph_kernel/test_softmax.py index dbf6cdb5437..4f092247623 100644 --- a/tests/st/ops/graph_kernel/test_softmax.py +++ b/tests/st/ops/graph_kernel/test_softmax.py @@ -48,7 +48,7 @@ def test_softmax(shape, dtype): assert np.allclose(expect.asnumpy(), output.asnumpy(), rtol, atol, equal_nan=True) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_softmax_gpu(): diff --git a/tests/st/ops/graph_kernel/test_softmax_cross_entropy_with_logits.py b/tests/st/ops/graph_kernel/test_softmax_cross_entropy_with_logits.py index ec5f61056fb..29ca468c45a 100644 --- a/tests/st/ops/graph_kernel/test_softmax_cross_entropy_with_logits.py +++ b/tests/st/ops/graph_kernel/test_softmax_cross_entropy_with_logits.py @@ -31,7 +31,7 @@ class NetSoftmaxCrossEntropyWithLogits(nn.Cell): return self.loss(logits, labels) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_softmax_cross_entropy_with_logits(): diff --git a/tests/st/ops/graph_kernel/test_softmax_stitch.py b/tests/st/ops/graph_kernel/test_softmax_stitch.py index 6c8d7fcd3e4..86191d76f3c 100644 --- a/tests/st/ops/graph_kernel/test_softmax_stitch.py +++ b/tests/st/ops/graph_kernel/test_softmax_stitch.py @@ -85,7 +85,7 @@ def test_softmax(shape, dtype): compare_result(expect, output, dtype) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_softmax_gpu(): diff --git a/tests/st/ops/graph_kernel/test_square_sum_all.py b/tests/st/ops/graph_kernel/test_square_sum_all.py index 2a5c3559ad0..4ce06a6fbe5 100644 --- a/tests/st/ops/graph_kernel/test_square_sum_all.py +++ b/tests/st/ops/graph_kernel/test_square_sum_all.py @@ -46,14 +46,14 @@ def test_basic(datatype): assert np.allclose(expect_np0, output_np0, 1.e-4, 1.e-7) assert np.allclose(expect_np1, output_np1, 1.e-4, 1.e-7) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_1(): context.set_context(mode=context.GRAPH_MODE, device_target="GPU") test_basic(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_gpu_2(): diff --git a/tests/st/ops/graph_kernel/test_squared_difference.py b/tests/st/ops/graph_kernel/test_squared_difference.py index f2d49069219..a119a590f08 100644 --- a/tests/st/ops/graph_kernel/test_squared_difference.py +++ b/tests/st/ops/graph_kernel/test_squared_difference.py @@ -48,7 +48,7 @@ def test_squared_difference(shape1, shape2, dtype): assert np.allclose(expect_np, output_np, 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_squared_difference_gpu(): diff --git a/tests/st/ops/graph_kernel/test_tanh_grad.py b/tests/st/ops/graph_kernel/test_tanh_grad.py index 999ec9079fb..8700e881364 100644 --- a/tests/st/ops/graph_kernel/test_tanh_grad.py +++ b/tests/st/ops/graph_kernel/test_tanh_grad.py @@ -41,7 +41,7 @@ def test_tanh_grad(): assert res -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_tanh_grad_gpu(): diff --git a/tests/st/pynative/test_pynative_hook_grad.py b/tests/st/pynative/test_pynative_hook_grad.py index 768a531b683..4ae4d4e5e9d 100644 --- a/tests/st/pynative/test_pynative_hook_grad.py +++ b/tests/st/pynative/test_pynative_hook_grad.py @@ -414,7 +414,7 @@ def test_pynative_hook_diff_hook_ascend(): context.set_context(mode=context.PYNATIVE_MODE, device_target="Ascend") pynative_hook_diff_hook() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pynative_hook_diff_hook_gpu(): @@ -429,7 +429,7 @@ def test_pynative_hook_outermost_cell_not_change_grad_ascend(): context.set_context(mode=context.PYNATIVE_MODE, device_target="Ascend") pynative_hook_outermost_cell_not_change_grad() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pynative_hook_outermost_cell_not_change_grad_gpu(): @@ -444,7 +444,7 @@ def test_pynative_hook_all_cell_record_grad_ascend(): context.set_context(mode=context.PYNATIVE_MODE, device_target="Ascend") pynative_hook_all_cell_record_grad() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pynative_hook_all_cell_record_grad_gpu(): @@ -459,7 +459,7 @@ def test_pynative_hook_mul_change_input_grad_ascend(): context.set_context(mode=context.PYNATIVE_MODE, device_target="Ascend") pynative_hook_mul_change_input_grad() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pynative_hook_mul_change_input_grad_gpu(): @@ -474,7 +474,7 @@ def test_pynative_hook_mul2_change_input_grad_ascend(): context.set_context(mode=context.PYNATIVE_MODE, device_target="Ascend") pynative_hook_mul2_change_input_grad() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pynative_hook_mul2_change_input_grad_gpu(): @@ -489,7 +489,7 @@ def test_pynative_hook_outermost_cell_change_grad_ascend(): context.set_context(mode=context.PYNATIVE_MODE, device_target="Ascend") pynative_hook_outermost_cell_change_grad() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pynative_hook_outermost_cell_change_grad_gpu(): @@ -504,7 +504,7 @@ def test_pynative_hook_outermost_cell_record_grad_ascend(): context.set_context(mode=context.PYNATIVE_MODE, device_target="Ascend") pynative_hook_outermost_cell_record_grad() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pynative_hook_outermost_cell_record_grad_gpu(): @@ -519,7 +519,7 @@ def test_pynative_hook_bprop_outermost_cell_record_grad_ascend(): context.set_context(mode=context.PYNATIVE_MODE, device_target="Ascend") pynative_hook_bprop_outermost_cell_record_grad() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pynative_hook_bprop_outermost_cell_record_grad_gpu(): @@ -534,7 +534,7 @@ def test_pynative_hook_child_cell_record_grad_ascend(): context.set_context(mode=context.PYNATIVE_MODE, device_target="Ascend") pynative_hook_child_cell_record_grad() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_pynative_hook_child_cell_record_grad_gpu(): diff --git a/tests/st/pynative/test_pynative_layernorm_input_and_argmaxwithvalue.py b/tests/st/pynative/test_pynative_layernorm_input_and_argmaxwithvalue.py index 4bef68624cc..a412dbe9463 100644 --- a/tests/st/pynative/test_pynative_layernorm_input_and_argmaxwithvalue.py +++ b/tests/st/pynative/test_pynative_layernorm_input_and_argmaxwithvalue.py @@ -226,7 +226,7 @@ def test_layernorm_input_ascend(): context.set_context(device_target="Ascend") layernorm_input() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_layernorm_input_gpu(): diff --git a/tests/st/pynative/test_pynative_mixed_precision_cells.py b/tests/st/pynative/test_pynative_mixed_precision_cells.py index caac0cf5549..d903092bf80 100644 --- a/tests/st/pynative/test_pynative_mixed_precision_cells.py +++ b/tests/st/pynative/test_pynative_mixed_precision_cells.py @@ -154,7 +154,7 @@ def test_mixed_precision_multiples_cell_ascend_01(): context.set_context(device_target="Ascend") mixed_precision_multiples_cell_01() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_mixed_precision_multiples_cell_gpu_01(): @@ -169,7 +169,7 @@ def test_mixed_precision_multiples_cell_ascend_02(): context.set_context(device_target="Ascend") mixed_precision_multiples_cell_02() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_mixed_precision_multiples_cell_gpu_02(): @@ -184,7 +184,7 @@ def test_mixed_precision_multiples_cell_ascend_03(): context.set_context(device_target="Ascend") mixed_precision_multiples_cell_03() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_mixed_precision_multiples_cell_gpu_03(): diff --git a/tests/st/pynative/test_tensor_getitem.py b/tests/st/pynative/test_tensor_getitem.py index cef6c5f8515..deb63cd30b3 100644 --- a/tests/st/pynative/test_tensor_getitem.py +++ b/tests/st/pynative/test_tensor_getitem.py @@ -144,7 +144,7 @@ class NetWorkSliceStep(Cell): return ret1, ret2 -@pytest.mark.level0 +@pytest.mark.level1 # ascend op stridedslice has bug, and has not been fixed. @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard