From 12f2251ea3e73155f5e1c7fefe0d9c12f392563f Mon Sep 17 00:00:00 2001 From: liuruotao Date: Tue, 19 Dec 2023 11:13:10 +0800 Subject: [PATCH] takedown always success testcases --- .../auto_monad/test_auto_monad_expression.py | 6 +- tests/st/auto_parallel/test_comm_subgraph.py | 2 +- tests/st/container/test_CellDict.py | 26 +++---- tests/st/container/test_CellList.py | 4 +- .../st/dyn_shape_dev/test_abs_op_dyn_shape.py | 4 +- tests/st/dyn_shape_dev/test_add.py | 8 +-- tests/st/dyn_shape_dev/test_argmax.py | 6 +- tests/st/dyn_shape_dev/test_argmin.py | 6 +- tests/st/dyn_shape_dev/test_assign.py | 2 +- tests/st/dyn_shape_dev/test_avg_pool.py | 8 +-- tests/st/dyn_shape_dev/test_concat.py | 10 +-- tests/st/dyn_shape_dev/test_cosh.py | 6 +- tests/st/dyn_shape_dev/test_cummax.py | 6 +- tests/st/dyn_shape_dev/test_cumprod.py | 12 ++-- tests/st/dyn_shape_dev/test_cumsum.py | 12 ++-- tests/st/dyn_shape_dev/test_div.py | 2 +- .../dyn_shape_dev/test_framework_compile.py | 6 +- tests/st/dyn_shape_dev/test_gcd.py | 6 +- tests/st/dyn_shape_dev/test_gelu.py | 10 +-- tests/st/dyn_shape_dev/test_gelu_grad.py | 6 +- .../st/dyn_shape_dev/test_greater_equal_op.py | 6 +- tests/st/dyn_shape_dev/test_greater_op.py | 6 +- .../st/dyn_shape_dev/test_grid_sampler_2d.py | 4 +- .../test_grid_sampler_2d_grad.py | 2 +- .../st/dyn_shape_dev/test_grid_sampler_3d.py | 6 +- .../test_grid_sampler_3d_grad.py | 4 +- tests/st/dyn_shape_dev/test_hshrink.py | 10 +-- tests/st/dyn_shape_dev/test_hshrink_grad.py | 4 +- tests/st/dyn_shape_dev/test_hsigmoid.py | 10 +-- tests/st/dyn_shape_dev/test_hsigmoid_grad.py | 6 +- tests/st/dyn_shape_dev/test_hswish.py | 10 +-- tests/st/dyn_shape_dev/test_hswish_grad.py | 8 +-- tests/st/dyn_shape_dev/test_layer_norm.py | 8 +-- .../st/dyn_shape_dev/test_layer_norm_grad.py | 4 +- tests/st/dyn_shape_dev/test_less.py | 12 ++-- tests/st/dyn_shape_dev/test_less_equal.py | 12 ++-- tests/st/dyn_shape_dev/test_masked_fill.py | 6 +- tests/st/dyn_shape_dev/test_maximum.py | 8 +-- .../st/dyn_shape_dev/test_maximum_grad_op.py | 6 +- tests/st/dyn_shape_dev/test_minimum.py | 10 +-- tests/st/dyn_shape_dev/test_minimum_grad.py | 6 +- tests/st/dyn_shape_dev/test_mul.py | 10 +-- tests/st/dyn_shape_dev/test_neg.py | 6 +- tests/st/dyn_shape_dev/test_next_after.py | 6 +- tests/st/dyn_shape_dev/test_nllloss.py | 2 +- tests/st/dyn_shape_dev/test_nllloss_grad.py | 4 +- tests/st/dyn_shape_dev/test_nonzero.py | 4 +- tests/st/dyn_shape_dev/test_notequal.py | 8 +-- tests/st/dyn_shape_dev/test_onehot.py | 8 +-- tests/st/dyn_shape_dev/test_oneslike.py | 6 +- tests/st/dyn_shape_dev/test_ops_bias_add.py | 10 +-- .../dyn_shape_dev/test_ops_bias_add_grad.py | 6 +- tests/st/dyn_shape_dev/test_ops_cast.py | 6 +- tests/st/dyn_shape_dev/test_ops_ceil.py | 10 +-- tests/st/dyn_shape_dev/test_ops_celu.py | 10 +-- tests/st/dyn_shape_dev/test_ops_cholesky.py | 2 +- .../test_ops_cholesky_inverse.py | 2 +- tests/st/dyn_shape_dev/test_ops_conj.py | 2 +- tests/st/dyn_shape_dev/test_ops_cos.py | 10 +-- tests/st/dyn_shape_dev/test_ops_elu.py | 6 +- tests/st/dyn_shape_dev/test_ops_equal.py | 8 +-- tests/st/dyn_shape_dev/test_ops_erf.py | 6 +- tests/st/dyn_shape_dev/test_ops_erfc.py | 6 +- tests/st/dyn_shape_dev/test_ops_erfinv.py | 6 +- tests/st/dyn_shape_dev/test_ops_exp.py | 6 +- .../st/dyn_shape_dev/test_ops_expand_dims.py | 6 +- tests/st/dyn_shape_dev/test_ops_expm1.py | 6 +- tests/st/dyn_shape_dev/test_ops_flatten.py | 6 +- tests/st/dyn_shape_dev/test_ops_floor.py | 10 +-- tests/st/dyn_shape_dev/test_ops_floor_div.py | 6 +- tests/st/dyn_shape_dev/test_ops_floor_mod.py | 6 +- tests/st/dyn_shape_dev/test_ops_gather.py | 2 +- tests/st/dyn_shape_dev/test_ops_gather_d.py | 6 +- tests/st/dyn_shape_dev/test_ops_identity.py | 10 +-- tests/st/dyn_shape_dev/test_ops_lin_space.py | 4 +- .../test_ops_scalar_arithmetic.py | 24 +++---- .../test_ops_scalar_arithmetic_unary.py | 8 +-- tests/st/dyn_shape_dev/test_ops_split.py | 10 +-- tests/st/dyn_shape_dev/test_pow.py | 6 +- ...{test_prelu.py => test_prelu_dyn_shape.py} | 4 +- tests/st/dyn_shape_dev/test_prelu_grad.py | 2 +- tests/st/dyn_shape_dev/test_qr.py | 2 +- tests/st/dyn_shape_dev/test_range.py | 6 +- tests/st/dyn_shape_dev/test_rank.py | 4 +- tests/st/dyn_shape_dev/test_real.py | 2 +- tests/st/dyn_shape_dev/test_real_div.py | 6 +- tests/st/dyn_shape_dev/test_reciprocal.py | 6 +- .../st/dyn_shape_dev/test_reciprocal_grad.py | 6 +- tests/st/dyn_shape_dev/test_reduce_all.py | 4 +- tests/st/dyn_shape_dev/test_reduce_any.py | 4 +- tests/st/dyn_shape_dev/test_reduce_max.py | 4 +- tests/st/dyn_shape_dev/test_reduce_mean.py | 4 +- tests/st/dyn_shape_dev/test_reduce_min.py | 4 +- tests/st/dyn_shape_dev/test_reduce_prod.py | 4 +- tests/st/dyn_shape_dev/test_reduce_sum.py | 4 +- tests/st/dyn_shape_dev/test_relu.py | 4 +- tests/st/dyn_shape_dev/test_relu6.py | 4 +- tests/st/dyn_shape_dev/test_relu6_grad.py | 4 +- tests/st/dyn_shape_dev/test_relu_grad.py | 4 +- tests/st/dyn_shape_dev/test_reshape.py | 4 +- .../test_resize_nearest_neighbor.py | 4 +- .../test_resize_nearest_neighbor_grad.py | 2 +- .../test_resize_nearest_neighbor_v2.py | 2 +- tests/st/dyn_shape_dev/test_reverse_v2.py | 4 +- tests/st/dyn_shape_dev/test_right_shift.py | 2 +- tests/st/dyn_shape_dev/test_roll.py | 4 +- tests/st/dyn_shape_dev/test_round.py | 4 +- tests/st/dyn_shape_dev/test_rsqrt.py | 4 +- tests/st/dyn_shape_dev/test_rsqrt_grad.py | 4 +- tests/st/dyn_shape_dev/test_sigmoid.py | 2 +- tests/st/dyn_shape_dev/test_sigmoid_grad.py | 2 +- tests/st/dyn_shape_dev/test_sin.py | 6 +- tests/st/dyn_shape_dev/test_sinc.py | 4 +- tests/st/dyn_shape_dev/test_sinh.py | 6 +- tests/st/dyn_shape_dev/test_softmax.py | 10 +-- tests/st/dyn_shape_dev/test_sqrt.py | 10 +-- tests/st/dyn_shape_dev/test_sqrt_grad.py | 8 +-- tests/st/dyn_shape_dev/test_square.py | 8 +-- tests/st/dyn_shape_dev/test_tensor_shape.py | 6 +- tests/st/dyn_shape_dev/test_tile.py | 12 ++-- tests/st/dyn_shape_dev/test_transpose.py | 6 +- tests/st/dyn_shape_dev/test_zeros_like.py | 6 +- .../test_dynamic_stridedslice.py | 2 +- .../dynamic_shape/test_dynamic_transformer.py | 2 +- .../test_mindir_dyn_shape_and_dyn_seq.py | 2 +- tests/st/export_and_load/test_train_mindir.py | 4 +- .../test_graph_fallback_runtime_all_any.py | 4 +- .../test_graph_fallback_runtime_bool.py | 2 +- .../test_graph_fallback_runtime_dict.py | 2 +- .../test_graph_fallback_runtime_enumerate.py | 4 +- .../test_graph_fallback_runtime_filter.py | 2 +- .../test_graph_fallback_runtime_getattr.py | 2 +- .../test_graph_fallback_runtime_int_float.py | 2 +- .../test_graph_fallback_runtime_iter.py | 18 ++--- .../test_graph_fallback_runtime_list_tuple.py | 2 +- .../test_graph_fallback_runtime_max_min.py | 10 +-- .../test_graph_fallback_runtime_print.py | 6 +- .../test_graph_fallback_runtime_range.py | 4 +- .../test_graph_fallback_runtime_sum.py | 2 +- .../test_graph_fallback_runtime_zip.py | 6 +- .../test_graph_fallback_jit_syntax_level.py | 4 +- .../fallback/test_graph_fallback_runtime.py | 8 +-- .../test_graph_fallback_runtime_asnumpy.py | 2 +- .../test_graph_fallback_runtime_dict.py | 20 +++--- ...est_graph_fallback_runtime_dict_inplace.py | 18 ++--- ...raph_fallback_runtime_inplace_operation.py | 30 ++++---- ...est_graph_fallback_runtime_list_inplace.py | 4 +- .../test_graph_fallback_runtime_meta_fg.py | 20 +++--- ...test_graph_fallback_runtime_return_list.py | 2 +- .../test_graph_fallback_runtime_setattr.py | 18 ++--- .../test_graph_fallback_runtime_str_format.py | 24 +++---- .../test_graph_fallback_runtime_tuple_func.py | 32 ++++----- .../test_graph_fallback_side_effect.py | 6 +- .../fallback/test_graph_fallback_unsupport.py | 12 ++-- tests/st/ge/test_ge_call.py | 2 +- tests/st/ge/test_ge_dynamic_input.py | 2 +- tests/st/ge/test_ge_jit_level.py | 2 +- tests/st/ge/test_ge_nested_tuple.py | 2 +- tests/st/ge/test_if_by_if.py | 2 +- .../st/ge/test_while_by_cell_list_in_while.py | 2 +- tests/st/gradient/test_grad_graph.py | 2 +- tests/st/graph_syntax/dict/test_dict_comp.py | 2 +- .../operators/test_bitwise_operator.py | 2 +- .../python_builtin_functions/test_all_any.py | 2 +- .../test_dynamic_sequence_list_comp.py | 2 +- .../python_builtin_functions/test_len.py | 2 +- tests/st/graph_syntax/test_graph_joinedstr.py | 2 +- tests/st/mix_precision/test_mix_precision.py | 4 +- .../mix_precision/test_mix_precision_func.py | 2 +- .../resnet50/test_resnet50_cifar10.py | 2 +- tests/st/nccl/test_nccl_all.py | 4 +- .../bert_performance/test_bert_tdt_ge_run.py | 2 +- tests/st/nn/test_BiDense.py | 2 +- tests/st/nn/test_EmbeddingLookup.py | 2 +- tests/st/nn/test_GroupNorm.py | 2 +- tests/st/nn/test_LayerNorm.py | 2 +- tests/st/nn/test_MultiFieldEmbeddingLookup.py | 2 +- tests/st/nn/test_cell.py | 4 +- tests/st/nn/test_greater.py | 6 +- tests/st/nn/test_greater_equal.py | 6 +- tests/st/nn/test_grucell.py | 2 +- tests/st/nn/test_lstmcell.py | 2 +- tests/st/nn/test_rnn.py | 2 +- tests/st/nn/test_rnncell.py | 2 +- ...test_trainonestepcell_with_return_grads.py | 2 +- tests/st/nn/test_transformer.py | 6 +- tests/st/numpy_native/test_array_ops.py | 2 +- tests/st/ops/ascend/test_add.py | 2 +- tests/st/ops/ascend/test_addn.py | 2 +- tests/st/ops/ascend/test_equal_op.py | 2 +- tests/st/ops/ascend/test_exp_op.py | 2 +- .../ops/ascend/test_flash_attention_score.py | 2 +- tests/st/ops/ascend/test_ge_op.py | 2 +- tests/st/ops/ascend/test_less_equal_op.py | 2 +- tests/st/ops/ascend/test_maximum_op.py | 2 +- tests/st/ops/ascend/test_minimum_op.py | 2 +- tests/st/ops/ascend/test_mul_op.py | 2 +- .../ascend/test_tbe_ops/test_ReduceMean.py | 2 +- .../ops/ascend/test_tbe_ops/test_fast_gelu.py | 2 +- .../test_tbe_ops/test_fast_gelu_grad_sens.py | 2 +- .../ascend/test_tbe_ops/test_maximum_grad.py | 4 +- .../ascend/test_tbe_ops/test_minimum_grad.py | 4 +- .../st/ops/ascend/test_tbe_ops/test_select.py | 2 +- .../ascend/test_tbe_ops/test_stridedslice.py | 2 +- .../ascend/test_tbe_ops/test_tensor_move.py | 2 +- tests/st/ops/cpu/test_igamma_op.py | 6 +- tests/st/ops/cpu/test_im2col_op.py | 4 +- tests/st/ops/cpu/test_imag_op.py | 4 +- tests/st/ops/cpu/test_in_top_k_op.py | 6 +- tests/st/ops/cpu/test_index_add_op.py | 18 ++--- tests/st/ops/cpu/test_inplace_op.py | 6 +- tests/st/ops/cpu/test_matrix_band_part.py | 4 +- .../st/ops/cpu/test_matrix_diag_part_v3_op.py | 14 ++-- tests/st/ops/cpu/test_matrix_diag_v3_op.py | 6 +- tests/st/ops/cpu/test_matrix_logarithm_op.py | 2 +- tests/st/ops/cpu/test_matrix_solve.py | 8 +-- .../test_max_pool_grad_with_argmax_cpu_op.py | 6 +- ...est_max_pool_grad_with_argmax_v2_cpu_op.py | 8 +-- .../cpu/test_max_pool_with_argmax_cpu_op.py | 8 +-- .../test_max_pool_with_argmax_v2_cpu_op.py | 2 +- tests/st/ops/cpu/test_maximum_grad.py | 2 +- tests/st/ops/cpu/test_maximum_op.py | 16 ++--- .../cpu/test_maxpool_3d_grad_grad_cpu_op.py | 6 +- .../ops/cpu/test_maxpool_grad_grad_cpu_op.py | 6 +- ...st_maxpool_grad_grad_with_argmax_cpu_op.py | 6 +- tests/st/ops/cpu/test_maxpool_grad_op.py | 2 +- tests/st/ops/cpu/test_maxpool_op.py | 14 ++-- tests/st/ops/cpu/test_meshgrid_op.py | 8 +-- tests/st/ops/cpu/test_minimum_grad_op.py | 10 +-- tests/st/ops/cpu/test_minimum_op.py | 24 +++---- tests/st/ops/cpu/test_mirror_pad.py | 10 +-- tests/st/ops/cpu/test_mish_grad_op.py | 6 +- tests/st/ops/cpu/test_mish_op.py | 6 +- tests/st/ops/cpu/test_mul_op.py | 6 +- tests/st/ops/cpu/test_multinomial_op.py | 2 +- .../test_multinomial_with_replacement_op.py | 4 +- tests/st/ops/cpu/test_mvlgamma_op.py | 4 +- tests/st/ops/cpu/test_neg_op.py | 4 +- tests/st/ops/cpu/test_nextafter_op.py | 2 +- tests/st/ops/cpu/test_nllloss_op.py | 6 +- tests/st/ops/cpu/test_nms_with_mask_op.py | 8 +-- tests/st/ops/cpu/test_no_repeat_ngram.py | 4 +- tests/st/ops/cpu/test_norm_op.py | 2 +- tests/st/ops/cpu/test_notequal_op.py | 4 +- tests/st/ops/cpu/test_one_hot_op.py | 2 +- tests/st/ops/cpu/test_onednn_dfx.py | 2 +- tests/st/ops/cpu/test_oneslike_op.py | 4 +- tests/st/ops/cpu/test_pack_op.py | 4 +- tests/st/ops/cpu/test_pad.py | 12 ++-- tests/st/ops/cpu/test_pad_op.py | 2 +- tests/st/ops/cpu/test_pad_v3.py | 14 ++-- tests/st/ops/cpu/test_pad_v3_grad_op.py | 6 +- tests/st/ops/cpu/test_padding.py | 6 +- .../test_parameterized_truncated_normal_op.py | 4 +- tests/st/ops/cpu/test_pdist_op.py | 6 +- tests/st/ops/cpu/test_pinv_op.py | 2 +- tests/st/ops/cpu/test_population_count_op.py | 6 +- tests/st/ops/cpu/test_pow_op.py | 14 ++-- tests/st/ops/cpu/test_prelu_op.py | 12 ++-- tests/st/ops/cpu/test_primitive_cache.py | 6 +- tests/st/ops/cpu/test_print_op.py | 10 +-- .../cpu/test_priority_replay_buffer_test.py | 2 +- .../st/ops/cpu/test_ps_roi_pooling_grad_op.py | 48 ++++++------- tests/st/ops/cpu/test_ps_roi_pooling_v2_op.py | 26 +++---- .../st/ops/cpu/test_random_categorical_op.py | 26 +++---- .../cpu/test_random_choice_with_mask_op.py | 12 ++-- tests/st/ops/cpu/test_random_gamma_op.py | 6 +- tests/st/ops/cpu/test_random_poisson_op.py | 20 +++--- tests/st/ops/cpu/test_random_shuffle.py | 2 +- tests/st/ops/cpu/test_range_op.py | 16 ++--- tests/st/ops/cpu/test_rank_op.py | 10 +-- tests/st/ops/cpu/test_realdiv_op.py | 2 +- tests/st/ops/cpu/test_reduce_op.py | 6 +- tests/st/ops/cpu/test_reduce_std_op.py | 4 +- tests/st/ops/cpu/test_relu6_grad_op.py | 2 +- tests/st/ops/cpu/test_relu6_op.py | 4 +- tests/st/ops/cpu/test_relu_grad_op.py | 2 +- tests/st/ops/cpu/test_relu_grad_v2_op.py | 2 +- tests/st/ops/cpu/test_relu_op.py | 2 +- tests/st/ops/cpu/test_relu_v2_op.py | 2 +- tests/st/ops/cpu/test_renorm_op.py | 14 ++-- .../ops/cpu/test_resize_bilinear_grad_op.py | 2 +- .../ops/cpu/test_resize_linear_1d_grad_op.py | 6 +- tests/st/ops/cpu/test_resize_linear_1d_op.py | 6 +- tests/st/ops/cpu/test_rgb_to_hsv_op.py | 2 +- tests/st/ops/cpu/test_rrelu.py | 6 +- .../st/ops/cpu/test_scatter_arithmetic_op.py | 70 +++++++++---------- .../ops/cpu/test_scatter_nd_arithmetic_op.py | 20 +++--- tests/st/ops/cpu/test_scatter_nd_op.py | 30 ++++---- tests/st/ops/cpu/test_scatter_nd_update_op.py | 10 +-- tests/st/ops/cpu/test_searchsorted_op.py | 8 +-- tests/st/ops/cpu/test_select_op.py | 12 ++-- tests/st/ops/cpu/test_selu_op.py | 6 +- tests/st/ops/cpu/test_sgd_op.py | 2 +- tests/st/ops/cpu/test_shift_op.py | 8 +-- .../st/ops/cpu/test_sparse_matrix_softmax.py | 2 +- .../test_dynamic_sequence_stack.py | 2 +- .../ops/dynamic_shape/grad/test_besselj1.py | 2 +- .../ops/dynamic_shape/grad/test_index_fill.py | 2 +- .../ops/dynamic_shape/grad/test_reduce_sum.py | 2 +- .../ops/dynamic_shape/test_dynamic_getitem.py | 18 ++--- .../ops/dynamic_shape/test_dynamic_setitem.py | 20 +++--- tests/st/ops/test_eps_op.py | 2 +- ...t_func_binary_cross_entropy_with_logits.py | 2 +- tests/st/ops/test_func_dense.py | 2 +- tests/st/ops/test_func_diagonal.py | 4 +- tests/st/ops/test_func_eq.py | 2 +- tests/st/ops/test_func_inner.py | 2 +- tests/st/ops/test_func_interpolate.py | 2 +- tests/st/ops/test_func_nansum.py | 2 +- tests/st/ops/test_func_one_hot.py | 2 +- tests/st/ops/test_func_select_scatter.py | 2 +- tests/st/ops/test_func_slice_scatter.py | 4 +- tests/st/ops/test_func_squeeze.py | 2 +- tests/st/ops/test_func_sum.py | 2 +- tests/st/ops/test_func_vector_norm.py | 2 +- tests/st/ops/test_func_zeta.py | 2 +- tests/st/ops/test_gaussian_nll_loss.py | 4 +- tests/st/ops/test_lppool1d.py | 2 +- tests/st/ops/test_margin_ranking_loss.py | 2 +- tests/st/ops/test_ops_clip_func.py | 2 +- tests/st/ops/test_ops_isposinf.py | 2 +- tests/st/ops/test_split.py | 4 +- tests/st/optimizer/test_asgd.py | 2 +- tests/st/optimizer/test_sgd.py | 2 +- tests/st/optimizer_ex/test_adam_cmp.py | 6 +- tests/st/optimizer_ex/test_adamw_cmp.py | 6 +- tests/st/optimizer_ex/test_lr_scheduler.py | 22 +++--- .../data_parallel/test_pynative_hccl.py | 2 +- .../data_parallel/test_pynative_nccl_all.py | 2 +- tests/st/pynative/hook/test_hook_function.py | 2 +- .../hook/test_pynative_backward_hook.py | 2 +- .../hook/test_pynative_forward_hook.py | 4 +- .../jit/test_outmost_dyn_shape_jit.py | 2 +- tests/st/pynative/jit/test_pynative_jit.py | 2 +- tests/st/pynative/ops/test_ops.py | 2 +- tests/st/pynative/test_tensor_getitem.py | 2 +- tests/st/pynative/test_tensor_setitem.py | 2 +- tests/st/rl/test_cell_list_getattr.py | 6 +- tests/st/scipy_st/test_grad.py | 2 +- tests/st/sparse/test_control_flow.py | 6 +- tests/st/sparse/test_coo.py | 6 +- tests/st/sparse/test_csr.py | 10 +-- tests/st/sparse/test_sparse_dynshape.py | 2 +- .../syntax/simple_expression/test_assign.py | 2 +- .../simple_expression/test_assignment_ops.py | 2 +- .../syntax/simple_expression/test_compare.py | 8 +-- .../test_sequence_getitem.py | 18 ++--- .../simple_expression/test_zip_operation.py | 2 +- tests/st/tensor/test_abs.py | 4 +- tests/st/tensor/test_cholesky_solve.py | 2 +- tests/st/tensor/test_diagonal.py | 4 +- tests/st/tensor/test_diagonal_scatter.py | 2 +- tests/st/tensor/test_eq.py | 4 +- tests/st/tensor/test_floor_divide.py | 2 +- tests/st/tensor/test_inner.py | 2 +- tests/st/tensor/test_item.py | 2 +- tests/st/tensor/test_movedim.py | 2 +- tests/st/tensor/test_permute.py | 2 +- tests/st/tensor/test_squeeze.py | 2 +- tests/st/tensor/test_sum.py | 2 +- tests/st/tensor/test_tolist.py | 2 +- tests/st/tensor/test_type.py | 2 +- tests/st/tensor/test_type_as.py | 2 +- tests/st/tensor/test_var.py | 2 +- tests/st/vmap/test_vmap_graph.py | 2 +- 366 files changed, 1096 insertions(+), 1096 deletions(-) rename tests/st/dyn_shape_dev/{test_prelu.py => test_prelu_dyn_shape.py} (98%) diff --git a/tests/st/auto_monad/test_auto_monad_expression.py b/tests/st/auto_monad/test_auto_monad_expression.py index bb2275a2d53..afc6ba810a6 100644 --- a/tests/st/auto_monad/test_auto_monad_expression.py +++ b/tests/st/auto_monad/test_auto_monad_expression.py @@ -201,7 +201,7 @@ class ForwardNet2(Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -364,7 +364,7 @@ def test_parameter_tuple_assign_addn_inner_net(): assert out == (12, 15) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -454,7 +454,7 @@ def test_parameter_value_control_flow_ascend(): assert graph_out == (9, 4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/auto_parallel/test_comm_subgraph.py b/tests/st/auto_parallel/test_comm_subgraph.py index 4af381d03e5..578096f3900 100644 --- a/tests/st/auto_parallel/test_comm_subgraph.py +++ b/tests/st/auto_parallel/test_comm_subgraph.py @@ -63,7 +63,7 @@ def compute_process(q, device_id, device_num, enable_comm_subgraph): q.put(res.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_single diff --git a/tests/st/container/test_CellDict.py b/tests/st/container/test_CellDict.py index 00aa1065f81..25bf154f02d 100644 --- a/tests/st/container/test_CellDict.py +++ b/tests/st/container/test_CellDict.py @@ -33,7 +33,7 @@ class TestGetitemMethodNet(nn.Cell): return self.cell_dict['conv'] -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -69,7 +69,7 @@ class TestSetitemMethodNet(nn.Cell): return self.cell_dict['conv'] -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -105,7 +105,7 @@ class TestSetitemMethodErrCaseNet(nn.Cell): return self.cell_dict[key] -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -165,7 +165,7 @@ class TestDelitemMethodNet(nn.Cell): return len(self.cell_dict) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -199,7 +199,7 @@ class TestContainsMethodNet(nn.Cell): return ret1, ret2 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -234,7 +234,7 @@ class TestClearMethodNet(nn.Cell): return len(self.cell_dict) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -268,7 +268,7 @@ class TestPopMethodNet(nn.Cell): return op, cell_dict_len -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -305,7 +305,7 @@ class TestKeysMethodNet(nn.Cell): return self.cell_dict.keys() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -338,7 +338,7 @@ class TestValuesMethodNet(nn.Cell): return self.cell_dict.values() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -380,7 +380,7 @@ class TestItemsMethodNet(nn.Cell): return self.cell_dict.items() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -457,7 +457,7 @@ class TestUpdateMethodNet(nn.Cell): return output1, output2, output3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -505,7 +505,7 @@ class TestUpdateMethodEmbeddedNet(nn.Cell): def construct(self, object_list): self.cell_dict.update(object_list) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -567,7 +567,7 @@ class DupParaNameNet2(nn.Cell): return a + b -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/container/test_CellList.py b/tests/st/container/test_CellList.py index fff7e615830..d139602ace9 100644 --- a/tests/st/container/test_CellList.py +++ b/tests/st/container/test_CellList.py @@ -33,7 +33,7 @@ class TestCellListInsertNet(nn.Cell): return len(self.cell_list) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -69,7 +69,7 @@ class EmbeddedCellDictNet(nn.Cell): return x -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_abs_op_dyn_shape.py b/tests/st/dyn_shape_dev/test_abs_op_dyn_shape.py index 86dd2fe18ce..fc6013e0831 100644 --- a/tests/st/dyn_shape_dev/test_abs_op_dyn_shape.py +++ b/tests/st/dyn_shape_dev/test_abs_op_dyn_shape.py @@ -72,7 +72,7 @@ def abs_infervalue_func2(): return ops.auto_generate.abs(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -132,7 +132,7 @@ def test_abs_vmap(mode): eltwise_case_vmap(abs_func, mode) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training diff --git a/tests/st/dyn_shape_dev/test_add.py b/tests/st/dyn_shape_dev/test_add.py index f4ad06805ea..41dbf4b8f49 100644 --- a/tests/st/dyn_shape_dev/test_add.py +++ b/tests/st/dyn_shape_dev/test_add.py @@ -38,7 +38,7 @@ def add_infervalue_func2(): y = ms.Tensor(np.array([3, 5, 1]).astype(np.float32)) return ops.auto_generate.add(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -64,7 +64,7 @@ def test_add(mode): compare(output_grads, expect_grads) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -88,7 +88,7 @@ def test_add_vmap(mode): assert np.allclose(output_vmap.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -121,7 +121,7 @@ def test_add_dyn(mode): compare(output_grads, expect_grads) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_argmax.py b/tests/st/dyn_shape_dev/test_argmax.py index 52b5ace6467..fb55f1d29a1 100644 --- a/tests/st/dyn_shape_dev/test_argmax.py +++ b/tests/st/dyn_shape_dev/test_argmax.py @@ -20,7 +20,7 @@ from mindspore.ops import auto_generate as P from test_argmin import argmin_argmax_case, argmin_argmax_case_dyn, argmin_argmax_case_vmap -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -36,7 +36,7 @@ def test_argmax(mode): argmin_argmax_case(P.argmax_, np.argmax) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -52,7 +52,7 @@ def test_argmax_vmap(mode): argmin_argmax_case_vmap(P.argmax_) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training diff --git a/tests/st/dyn_shape_dev/test_argmin.py b/tests/st/dyn_shape_dev/test_argmin.py index 251f97a6dad..a8aba7de319 100644 --- a/tests/st/dyn_shape_dev/test_argmin.py +++ b/tests/st/dyn_shape_dev/test_argmin.py @@ -69,7 +69,7 @@ def argmin_argmax_case_dyn(op_func, np_func): assert np.allclose(output2.asnumpy(), expect2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -85,7 +85,7 @@ def test_argmin(mode): argmin_argmax_case(P.argmin_, np.argmin) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -101,7 +101,7 @@ def test_argmin_vmap(mode): argmin_argmax_case_vmap(P.argmin_) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training diff --git a/tests/st/dyn_shape_dev/test_assign.py b/tests/st/dyn_shape_dev/test_assign.py index 28e6a2e0e00..e00069ea47b 100644 --- a/tests/st/dyn_shape_dev/test_assign.py +++ b/tests/st/dyn_shape_dev/test_assign.py @@ -66,7 +66,7 @@ def test_assign_forward_ascend(mode): output = assign_forward_func(variable, value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_avg_pool.py b/tests/st/dyn_shape_dev/test_avg_pool.py index 697d8fe1463..083098028fd 100644 --- a/tests/st/dyn_shape_dev/test_avg_pool.py +++ b/tests/st/dyn_shape_dev/test_avg_pool.py @@ -35,7 +35,7 @@ def avg_pool_dyn_shape_func(x): return ops.auto_generate.avg_pool(x, kernel_size=2, strides=2, pad_mode="VALID", data_format="NCHW") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -56,7 +56,7 @@ def test_avg_pool_forward(mode): assert np.allclose(out.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -75,7 +75,7 @@ def test_avg_pool_backward(mode): grads = avg_pool_backward_func(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -100,7 +100,7 @@ def test_avg_pool_vmap(mode): assert np.allclose(out.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_concat.py b/tests/st/dyn_shape_dev/test_concat.py index 737461c5f4e..17833f97f07 100644 --- a/tests/st/dyn_shape_dev/test_concat.py +++ b/tests/st/dyn_shape_dev/test_concat.py @@ -61,7 +61,7 @@ def forward_datas_prepare(shape, num=2, axis=0, diff_shapes=False, need_expect=T return tuple(tensor_inputs), np_expect -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -79,7 +79,7 @@ def test_concat_forward(mode): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -128,7 +128,7 @@ def test_concat_vmap(mode): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -164,7 +164,7 @@ def test_concat_forward_dynamic(mode, dyn_mode): assert np.allclose(out3.asnumpy(), expect3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -199,7 +199,7 @@ def test_concat_backward_dynamic(mode, dyn_mode): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_cosh.py b/tests/st/dyn_shape_dev/test_cosh.py index 2f467289c28..dd4c6063a99 100644 --- a/tests/st/dyn_shape_dev/test_cosh.py +++ b/tests/st/dyn_shape_dev/test_cosh.py @@ -33,7 +33,7 @@ def cosh_backward_func(x): return ops.grad(cosh_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -54,7 +54,7 @@ def test_cosh_forward(mode): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -75,7 +75,7 @@ def test_cosh_backward(mode): assert np.allclose(grads.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_cummax.py b/tests/st/dyn_shape_dev/test_cummax.py index 1269064b4af..447b3b9daeb 100644 --- a/tests/st/dyn_shape_dev/test_cummax.py +++ b/tests/st/dyn_shape_dev/test_cummax.py @@ -58,7 +58,7 @@ def test_cummax_forward(context_mode, dtype): assert (indices.asnumpy() == expect_indices).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -84,7 +84,7 @@ def test_cummax_vmap(context_mode, dtype): assert (indices.asnumpy() == expect_indices).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -118,7 +118,7 @@ def test_cummax_dynamic(context_mode): assert np.allclose(indices2.asnumpy(), expect_indices2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_cumprod.py b/tests/st/dyn_shape_dev/test_cumprod.py index ad379e9c6c0..89bf85b704e 100644 --- a/tests/st/dyn_shape_dev/test_cumprod.py +++ b/tests/st/dyn_shape_dev/test_cumprod.py @@ -37,7 +37,7 @@ def cumprod_dyn_shape_func(x, axis, exclusive, reverse): return ops.auto_generate.cum_prod(x, axis, exclusive, reverse) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -63,7 +63,7 @@ def test_cumprod_forward(context_mode, dtype): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -88,7 +88,7 @@ def test_cumprod_backward(context_mode, dtype): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -120,7 +120,7 @@ def test_cumprod_dynamic(context_mode): assert np.allclose(out2.asnumpy(), expect2, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -152,7 +152,7 @@ def test_cumprod_dynamic_rank(context_mode): assert np.allclose(out2.asnumpy(), expect2, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -184,7 +184,7 @@ def test_cumprod_backward_dynamic(context_mode): assert np.allclose(out2.asnumpy(), expect2, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_cumsum.py b/tests/st/dyn_shape_dev/test_cumsum.py index 909fec4dba6..b31314c89d0 100644 --- a/tests/st/dyn_shape_dev/test_cumsum.py +++ b/tests/st/dyn_shape_dev/test_cumsum.py @@ -37,7 +37,7 @@ def cumsum_dyn_shape_func(x, axis, exclusive, reverse): return ops.auto_generate.cum_sum(x, axis, exclusive, reverse) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -62,7 +62,7 @@ def test_cumsum_forward(context_mode, dtype): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -88,7 +88,7 @@ def test_cumsum_backward(context_mode, dtype): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -120,7 +120,7 @@ def test_cumsum_dynamic(context_mode): assert np.allclose(out2.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -152,7 +152,7 @@ def test_cumsum_dynamic_rank(context_mode): assert np.allclose(out2.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -184,7 +184,7 @@ def test_cumsum_backward_dynamic(context_mode): assert np.allclose(out2.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_div.py b/tests/st/dyn_shape_dev/test_div.py index d5217e26c20..8deb9c58947 100644 --- a/tests/st/dyn_shape_dev/test_div.py +++ b/tests/st/dyn_shape_dev/test_div.py @@ -51,7 +51,7 @@ def div_infer_value1(): return div_forward_func(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_framework_compile.py b/tests/st/dyn_shape_dev/test_framework_compile.py index b2f886630ac..c24ef2d9e2a 100644 --- a/tests/st/dyn_shape_dev/test_framework_compile.py +++ b/tests/st/dyn_shape_dev/test_framework_compile.py @@ -41,7 +41,7 @@ class AvgPoolCreateInstanceNet(nn.Cell): return op(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -59,7 +59,7 @@ def test_avg_pool(): out = net(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -77,7 +77,7 @@ def test_avg_pool_create_instance_const_args(): out = net(x, 1, 1, "VALID", "NCHW") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_gcd.py b/tests/st/dyn_shape_dev/test_gcd.py index ada2c91d251..d0956b8c7b7 100644 --- a/tests/st/dyn_shape_dev/test_gcd.py +++ b/tests/st/dyn_shape_dev/test_gcd.py @@ -31,7 +31,7 @@ def gcd_backward_func(x1, x2): return ops.grad(gcd_forward_func, (0,))(x1, x2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -52,7 +52,7 @@ def test_gcd_forward(mode): assert np.allclose(out.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -72,7 +72,7 @@ def test_gcd_backward(mode): gcd_backward_func(x1, x2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_gelu.py b/tests/st/dyn_shape_dev/test_gelu.py index e21b87a7013..30c0e4db5c1 100644 --- a/tests/st/dyn_shape_dev/test_gelu.py +++ b/tests/st/dyn_shape_dev/test_gelu.py @@ -38,7 +38,7 @@ def gelu_dyn_shape_func(x): return ops.auto_generate.gelu_(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -59,7 +59,7 @@ def test_gelu_forward(mode): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -80,7 +80,7 @@ def test_gelu_backward(mode): assert np.allclose(grads.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -103,7 +103,7 @@ def test_gelu_vmap(mode): assert np.allclose(out.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -133,7 +133,7 @@ def test_gelu_dynamic(mode): assert np.allclose(output1.asnumpy(), expect1, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_gelu_grad.py b/tests/st/dyn_shape_dev/test_gelu_grad.py index fd07104b933..0ac249beb42 100644 --- a/tests/st/dyn_shape_dev/test_gelu_grad.py +++ b/tests/st/dyn_shape_dev/test_gelu_grad.py @@ -37,7 +37,7 @@ def gelu_grad_dyn_shape_func(dy, x, y): return ops.auto_generate.gelu_grad(dy, x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -59,7 +59,7 @@ def test_gelu_grad_forward(mode): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -93,7 +93,7 @@ def test_gelu_grad_dynamic(mode): assert np.allclose(output1.asnumpy(), expect1, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_greater_equal_op.py b/tests/st/dyn_shape_dev/test_greater_equal_op.py index 34b9aebdeb5..7feb0e6c4a5 100644 --- a/tests/st/dyn_shape_dev/test_greater_equal_op.py +++ b/tests/st/dyn_shape_dev/test_greater_equal_op.py @@ -31,7 +31,7 @@ def greater_equal_backward_func(x, y): return ops.grad(greater_equal_forward_func, (0,))(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -52,7 +52,7 @@ def test_greater_equal_forward(mode): assert np.allclose(out.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -73,7 +73,7 @@ def test_greater_equal_backward(mode): assert np.allclose(grads.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_greater_op.py b/tests/st/dyn_shape_dev/test_greater_op.py index 72e743560af..88bbe5f1591 100644 --- a/tests/st/dyn_shape_dev/test_greater_op.py +++ b/tests/st/dyn_shape_dev/test_greater_op.py @@ -31,7 +31,7 @@ def greater_backward_func(x, y): return ops.grad(greater_forward_func, (0,))(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -52,7 +52,7 @@ def test_greater_forward(mode): assert np.allclose(out.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -73,7 +73,7 @@ def test_greater_backward(mode): assert np.allclose(grads.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_grid_sampler_2d.py b/tests/st/dyn_shape_dev/test_grid_sampler_2d.py index eec92e0295d..2c4b3dc1a1e 100644 --- a/tests/st/dyn_shape_dev/test_grid_sampler_2d.py +++ b/tests/st/dyn_shape_dev/test_grid_sampler_2d.py @@ -32,7 +32,7 @@ def grid_sampler_2d_backward_func(input_x, grid): return ops.grad(grid_sampler_2d_forward_func, (0, 1))(input_x, grid) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -90,7 +90,7 @@ def test_grid_sampler_2d_forward(mode): assert np.allclose(out.asnumpy(), expect_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_grid_sampler_2d_grad.py b/tests/st/dyn_shape_dev/test_grid_sampler_2d_grad.py index f87d21bf7c1..b97cb5ba816 100644 --- a/tests/st/dyn_shape_dev/test_grid_sampler_2d_grad.py +++ b/tests/st/dyn_shape_dev/test_grid_sampler_2d_grad.py @@ -27,7 +27,7 @@ def grid_sampler_2d_grad_forward_func(grad, input_x, grid): return grid_sampler_2d_grad -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_grid_sampler_3d.py b/tests/st/dyn_shape_dev/test_grid_sampler_3d.py index ca4c981a57d..a23128e9c3e 100644 --- a/tests/st/dyn_shape_dev/test_grid_sampler_3d.py +++ b/tests/st/dyn_shape_dev/test_grid_sampler_3d.py @@ -33,7 +33,7 @@ def grid_sampler_3d_backward_func(input_x, grid): return ops.grad(grid_sampler_3d_forward_func, (0, 1))(input_x, grid) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -80,7 +80,7 @@ def test_grid_sampler_3d_forward(mode): assert np.allclose(out.asnumpy(), except_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -139,7 +139,7 @@ def test_grid_sampler_3d_backward(mode): assert np.allclose(grads[1].asnumpy(), expect_out2, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_grid_sampler_3d_grad.py b/tests/st/dyn_shape_dev/test_grid_sampler_3d_grad.py index 23982fc3af7..df7e8632c01 100644 --- a/tests/st/dyn_shape_dev/test_grid_sampler_3d_grad.py +++ b/tests/st/dyn_shape_dev/test_grid_sampler_3d_grad.py @@ -28,7 +28,7 @@ def grid_sampler_3d_grad_forward_func(grad, input_x, grid): return grid_sampler_3d_grad -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -96,7 +96,7 @@ def test_grid_sampler_3d_grad_forward(mode): assert np.allclose(out[1].asnumpy(), expect_out2, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_hshrink.py b/tests/st/dyn_shape_dev/test_hshrink.py index 6378c142ca5..3aaeec2d71b 100644 --- a/tests/st/dyn_shape_dev/test_hshrink.py +++ b/tests/st/dyn_shape_dev/test_hshrink.py @@ -37,7 +37,7 @@ def hshrink_dyn_shape_func(input_x, lambd): return ops.auto_generate.hshrink(input_x, lambd) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -59,7 +59,7 @@ def test_hshrink_forward(context_mode): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -81,7 +81,7 @@ def test_hshrink_backward(context_mode): assert np.allclose(grads.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -106,7 +106,7 @@ def test_hshrink_vmap(context_mode): assert np.allclose(out.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -140,7 +140,7 @@ def test_hshrink_dynamic(mode): assert np.allclose(out1.asnumpy(), expect1, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_hshrink_grad.py b/tests/st/dyn_shape_dev/test_hshrink_grad.py index 803377a7b9f..1a8c5923769 100644 --- a/tests/st/dyn_shape_dev/test_hshrink_grad.py +++ b/tests/st/dyn_shape_dev/test_hshrink_grad.py @@ -32,7 +32,7 @@ def hshrink_grad_dyn_shape_func(gradients, features, lambd): return ops.auto_generate.hshrink_grad(gradients, features, lambd) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -56,7 +56,7 @@ def test_hshrink_grad_forward(context_mode): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_hsigmoid.py b/tests/st/dyn_shape_dev/test_hsigmoid.py index b2088149175..bb923b6f6ba 100644 --- a/tests/st/dyn_shape_dev/test_hsigmoid.py +++ b/tests/st/dyn_shape_dev/test_hsigmoid.py @@ -38,7 +38,7 @@ def hsigmoid_dyn_shape_func(x): return ops.auto_generate.hsigmoid(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -59,7 +59,7 @@ def test_hsigmoid_forward(mode): assert np.allclose(out.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -80,7 +80,7 @@ def test_hsigmoid_backward(mode): assert np.allclose(grads.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -103,7 +103,7 @@ def test_hsigmoid_vmap(mode): assert np.allclose(out.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -130,7 +130,7 @@ def test_hsigmoid_dynamic(mode): assert np.allclose(out1.asnumpy(), expect1, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_hsigmoid_grad.py b/tests/st/dyn_shape_dev/test_hsigmoid_grad.py index 345c0f3aa1d..a79839c11e1 100644 --- a/tests/st/dyn_shape_dev/test_hsigmoid_grad.py +++ b/tests/st/dyn_shape_dev/test_hsigmoid_grad.py @@ -32,7 +32,7 @@ def hsigmoid_grad_dyn_shape_func(grads, x): return ops.auto_generate.hsigmoid_grad(grads, x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -79,7 +79,7 @@ def test_hsigmoid_grad_vmap(mode): assert np.allclose(out.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -111,7 +111,7 @@ def test_hsigmoid_grad_dynamic(mode): assert np.allclose(out1.asnumpy(), expect1, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_hswish.py b/tests/st/dyn_shape_dev/test_hswish.py index 098ccaf4474..c15dfe44d7a 100644 --- a/tests/st/dyn_shape_dev/test_hswish.py +++ b/tests/st/dyn_shape_dev/test_hswish.py @@ -36,7 +36,7 @@ def hswish_dyn_shape_func(x): return ops.auto_generate.hardswish_(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -55,7 +55,7 @@ def test_hswish_forward(mode): assert np.allclose(out.asnumpy(), expect_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -74,7 +74,7 @@ def test_hswish_backward(mode): assert np.allclose(grads.asnumpy(), expect_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -97,7 +97,7 @@ def test_hswish_vmap(mode): assert np.allclose(out.asnumpy(), expect_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -124,7 +124,7 @@ def test_hswish_dynamic(mode): assert np.allclose(out1.asnumpy(), expect1, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_hswish_grad.py b/tests/st/dyn_shape_dev/test_hswish_grad.py index a2986cbe3eb..3f349af9d56 100644 --- a/tests/st/dyn_shape_dev/test_hswish_grad.py +++ b/tests/st/dyn_shape_dev/test_hswish_grad.py @@ -31,7 +31,7 @@ def hswish_grad_dyn_shape_func(y_grad, x): return ops.auto_generate.hswish_grad(y_grad, x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -52,7 +52,7 @@ def test_hswish_grad_forward(mode): assert np.allclose(out.asnumpy(), expect_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -77,7 +77,7 @@ def test_hswish_grad_vmap(mode): assert np.allclose(out.asnumpy(), expect_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -109,7 +109,7 @@ def test_hswish_grad_dynamic(mode): assert np.allclose(out1.asnumpy(), expect1, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_layer_norm.py b/tests/st/dyn_shape_dev/test_layer_norm.py index 4aac1295299..2cfed710999 100644 --- a/tests/st/dyn_shape_dev/test_layer_norm.py +++ b/tests/st/dyn_shape_dev/test_layer_norm.py @@ -45,7 +45,7 @@ def layer_norm_dyn_shape_func(input_x, gamma, beta): epsilon=1e-7) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -74,7 +74,7 @@ def test_layer_norm_forward(mode): atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -147,7 +147,7 @@ def test_layer_norm_vmap(mode): atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -204,7 +204,7 @@ def test_layer_norm_dynamic_shape(mode): atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_layer_norm_grad.py b/tests/st/dyn_shape_dev/test_layer_norm_grad.py index bddb8781a6c..19356a5a516 100644 --- a/tests/st/dyn_shape_dev/test_layer_norm_grad.py +++ b/tests/st/dyn_shape_dev/test_layer_norm_grad.py @@ -42,7 +42,7 @@ def layer_norm_grad_dyn_shape_func(x, dy, variance, mean, gamma): begin_params_axis=1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -125,7 +125,7 @@ def test_layer_norm_grad_vmap(mode): atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_less.py b/tests/st/dyn_shape_dev/test_less.py index 136b9af8696..cbaec06abef 100644 --- a/tests/st/dyn_shape_dev/test_less.py +++ b/tests/st/dyn_shape_dev/test_less.py @@ -49,7 +49,7 @@ def less_infervalue_func2(): return ops.auto_generate.less(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -70,7 +70,7 @@ def test_less_forward(mode): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -93,7 +93,7 @@ def test_less_backward(mode): assert np.allclose(grad_y.asnumpy(), expect_grad_y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -116,7 +116,7 @@ def test_less_vmap(mode): assert np.allclose(out.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -150,7 +150,7 @@ def test_less_dynamic_shape(mode): assert np.allclose(out_2.asnumpy(), expect_2, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -184,7 +184,7 @@ def test_less_dynamic_rank(mode): assert np.allclose(out_2.asnumpy(), expect_2, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_less_equal.py b/tests/st/dyn_shape_dev/test_less_equal.py index 15f4bf11c83..e48d0f572b8 100644 --- a/tests/st/dyn_shape_dev/test_less_equal.py +++ b/tests/st/dyn_shape_dev/test_less_equal.py @@ -49,7 +49,7 @@ def less_equal_infervalue_func2(): return ops.auto_generate.less_equal(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -70,7 +70,7 @@ def test_less_equal_forward(mode): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -93,7 +93,7 @@ def test_less_equal_backward(mode): assert np.allclose(grad_y.asnumpy(), expect_grad_y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -116,7 +116,7 @@ def test_less_equal_vmap(mode): assert np.allclose(out.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -150,7 +150,7 @@ def test_less_equal_dynamic_shape(mode): assert np.allclose(out_2.asnumpy(), expect_2, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -184,7 +184,7 @@ def test_less_equal_dynamic_rank(mode): assert np.allclose(out_2.asnumpy(), expect_2, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_masked_fill.py b/tests/st/dyn_shape_dev/test_masked_fill.py index bb627fcdf11..5fea86a6a69 100644 --- a/tests/st/dyn_shape_dev/test_masked_fill.py +++ b/tests/st/dyn_shape_dev/test_masked_fill.py @@ -30,7 +30,7 @@ def masked_fill_backward_func(input_x, mask, value): return ops.grad(masked_fill_forward_func, (0, 1))(input_x, mask, value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -51,7 +51,7 @@ def test_masked_fill_forward(mode): np.testing.assert_array_almost_equal(output.asnumpy(), expect_output, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -74,7 +74,7 @@ def test_masked_fill_backward(mode): np.testing.assert_array_almost_equal(mask_output.asnumpy(), expect_mask_output, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_maximum.py b/tests/st/dyn_shape_dev/test_maximum.py index 5ebdce5d464..09bed5dce64 100644 --- a/tests/st/dyn_shape_dev/test_maximum.py +++ b/tests/st/dyn_shape_dev/test_maximum.py @@ -36,7 +36,7 @@ def maximum_vmap_func(x, y): return ops.vmap(maximum_forward_func, in_axes=0, out_axes=0)(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -57,7 +57,7 @@ def test_maximum_op_forward(context_mode, data_type): np.testing.assert_allclose(out.asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -79,7 +79,7 @@ def test_maximum_op_backward(context_mode, data_type): np.testing.assert_allclose(grads[1].asnumpy(), expect_out[1], rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -100,7 +100,7 @@ def test_maximum_op_vmap(context_mode, data_type): np.testing.assert_allclose(out.asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_maximum_grad_op.py b/tests/st/dyn_shape_dev/test_maximum_grad_op.py index 20e35658f33..7dcacb29f3c 100644 --- a/tests/st/dyn_shape_dev/test_maximum_grad_op.py +++ b/tests/st/dyn_shape_dev/test_maximum_grad_op.py @@ -31,7 +31,7 @@ def maximum_grad_vmap_func(x1, x2, grads): return ops.vmap(maximum_grad_forward_func, in_axes=0, out_axes=0)(x1, x2, grads) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -55,7 +55,7 @@ def test_maximum_grad_op_forward(context_mode, data_type): -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -78,7 +78,7 @@ def test_maximum_grad_op_vmap(context_mode, data_type): np.testing.assert_allclose(out[1].asnumpy(), expect_out[1], rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_minimum.py b/tests/st/dyn_shape_dev/test_minimum.py index 45d5c9afaca..7bc1fdcda09 100644 --- a/tests/st/dyn_shape_dev/test_minimum.py +++ b/tests/st/dyn_shape_dev/test_minimum.py @@ -50,7 +50,7 @@ def minimum_infervalue_func2(): return ops.auto_generate.minimum(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -72,7 +72,7 @@ def test_minimum_op_forward(context_mode, data_type): np.testing.assert_allclose(out.asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -95,7 +95,7 @@ def test_minimum_op_backward(context_mode, data_type): np.testing.assert_allclose(grads[1].asnumpy(), expect_out[1], rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -117,7 +117,7 @@ def test_minimum_op_vmap(context_mode, data_type): np.testing.assert_allclose(out.asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -139,7 +139,7 @@ def test_minimum_op_infervalue(context_mode): np.testing.assert_array_equal(out_2.asnumpy(), expect_out_2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_minimum_grad.py b/tests/st/dyn_shape_dev/test_minimum_grad.py index d8d0f20d2a8..8271c54a0b4 100644 --- a/tests/st/dyn_shape_dev/test_minimum_grad.py +++ b/tests/st/dyn_shape_dev/test_minimum_grad.py @@ -31,7 +31,7 @@ def minimum_grad_vmap_func(x1, x2, grads): return ops.vmap(minimum_grad_forward_func, in_axes=0, out_axes=0)(x1, x2, grads) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -56,7 +56,7 @@ def test_minimum_grad_op_forward(context_mode, data_type): -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -80,7 +80,7 @@ def test_minimum_grad_op_vmap(context_mode, data_type): np.testing.assert_allclose(out[1].asnumpy(), expect_out[1], rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_mul.py b/tests/st/dyn_shape_dev/test_mul.py index e8f022f8518..e37c702d3b0 100644 --- a/tests/st/dyn_shape_dev/test_mul.py +++ b/tests/st/dyn_shape_dev/test_mul.py @@ -50,7 +50,7 @@ def mul_infervalue_func2(): return ops.auto_generate.mul(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -72,7 +72,7 @@ def test_mul_op_forward(context_mode, data_type): np.testing.assert_allclose(out.asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -95,7 +95,7 @@ def test_mul_op_backward(context_mode, data_type): np.testing.assert_allclose(grads[1].asnumpy(), expect_out[1], rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -117,7 +117,7 @@ def test_mul_op_vmap(context_mode, data_type): np.testing.assert_allclose(out.asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -139,7 +139,7 @@ def test_mul_op_infervalue(context_mode): np.testing.assert_array_equal(out_2.asnumpy(), expect_out_2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_neg.py b/tests/st/dyn_shape_dev/test_neg.py index ac6a395cb38..2eab7baad7a 100644 --- a/tests/st/dyn_shape_dev/test_neg.py +++ b/tests/st/dyn_shape_dev/test_neg.py @@ -36,7 +36,7 @@ def neg_vmap_func(x): return ops.vmap(neg_forward_func, in_axes=0, out_axes=0)(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -57,7 +57,7 @@ def test_neg_op_forward(context_mode, data_type): np.testing.assert_allclose(out.asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -78,7 +78,7 @@ def test_neg_op_backward(context_mode, data_type): np.testing.assert_allclose(grads.asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_next_after.py b/tests/st/dyn_shape_dev/test_next_after.py index 5c803108346..f7143a6a614 100644 --- a/tests/st/dyn_shape_dev/test_next_after.py +++ b/tests/st/dyn_shape_dev/test_next_after.py @@ -31,7 +31,7 @@ def next_after_backward_func(x, other): return ops.grad(next_after_forward_func, 0)(x, other) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -53,7 +53,7 @@ def test_next_after_op_forward(context_mode, data_type): @pytest.mark.skip(reason="dynamic shape not support now") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize("context_mode", [ms.GRAPH_MODE, ms.PYNATIVE_MODE]) @@ -72,7 +72,7 @@ def test_next_after_op_forward_cpu(context_mode, data_type): np.testing.assert_allclose(out.asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_nllloss.py b/tests/st/dyn_shape_dev/test_nllloss.py index 614b9d06b22..b29f42bedce 100644 --- a/tests/st/dyn_shape_dev/test_nllloss.py +++ b/tests/st/dyn_shape_dev/test_nllloss.py @@ -80,7 +80,7 @@ def test_nllloss_forward_ascend(mode, data_type): assert np.allclose(actual_output[1].asnumpy(), expect_total_weight) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_nllloss_grad.py b/tests/st/dyn_shape_dev/test_nllloss_grad.py index 146b2e56978..514999a8a08 100644 --- a/tests/st/dyn_shape_dev/test_nllloss_grad.py +++ b/tests/st/dyn_shape_dev/test_nllloss_grad.py @@ -30,7 +30,7 @@ def nllloss_grad_vmap_func(logits, loss_grad, labels, weight, total_weight): weight, total_weight) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -87,7 +87,7 @@ def get_grad_inputs_and_output(nptype_input, nptype_weight, reduction, input_typ return inputs, outputs -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_nonzero.py b/tests/st/dyn_shape_dev/test_nonzero.py index c316eeec0d6..7507f0a36af 100644 --- a/tests/st/dyn_shape_dev/test_nonzero.py +++ b/tests/st/dyn_shape_dev/test_nonzero.py @@ -31,7 +31,7 @@ def nonzero_backward_func(x): return ops.grad(nonzero_forward_func, 0)(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -52,7 +52,7 @@ def test_nonzero_op_forward(context_mode, data_type): np.testing.assert_array_equal(out.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_notequal.py b/tests/st/dyn_shape_dev/test_notequal.py index 561314f867e..9ffa79e7338 100644 --- a/tests/st/dyn_shape_dev/test_notequal.py +++ b/tests/st/dyn_shape_dev/test_notequal.py @@ -50,7 +50,7 @@ def notequal_vmap_func(x, y): return ops.vmap(notequal_forward_func, in_axes=0, out_axes=0)(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -72,7 +72,7 @@ def test_notequal_op_forward(context_mode, data_type): np.testing.assert_array_equal(out.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -94,7 +94,7 @@ def test_notequal_op_infervalue(context_mode): np.testing.assert_array_equal(out.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -116,7 +116,7 @@ def test_notequal_op_backward(context_mode, data_type): np.testing.assert_allclose(grads[0].asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_onehot.py b/tests/st/dyn_shape_dev/test_onehot.py index 1211a15c7ce..6f8a1bbaae6 100644 --- a/tests/st/dyn_shape_dev/test_onehot.py +++ b/tests/st/dyn_shape_dev/test_onehot.py @@ -37,7 +37,7 @@ def onehot_vmap_func(indices, depth, on_value, off_value, axis): return ops.vmap(onehot_forward_func, in_axes=in_axis, out_axes=0)(indices, depth, on_value, off_value, axis) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -61,7 +61,7 @@ def test_onehot_op_forward(context_mode, data_type): np.testing.assert_allclose(out.asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -84,7 +84,7 @@ def test_onehot_op_forward_depth_tensor(context_mode): np.testing.assert_allclose(out.asnumpy(), expect_out, rtol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -108,7 +108,7 @@ def test_onehot_op_backward(context_mode, data_type): np.testing.assert_allclose(grads[0].asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_oneslike.py b/tests/st/dyn_shape_dev/test_oneslike.py index 8bbb6d5bed9..9b74b56a49d 100644 --- a/tests/st/dyn_shape_dev/test_oneslike.py +++ b/tests/st/dyn_shape_dev/test_oneslike.py @@ -36,7 +36,7 @@ def oneslike_vmap_func(x): return ops.vmap(oneslike_forward_func, in_axes=0, out_axes=0)(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -57,7 +57,7 @@ def test_oneslike_op_forward(context_mode, data_type): np.testing.assert_array_equal(out.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -78,7 +78,7 @@ def test_oneslike_op_backward(context_mode, data_type): np.testing.assert_allclose(grads.asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_bias_add.py b/tests/st/dyn_shape_dev/test_ops_bias_add.py index d7b43b0f34d..503d7af2a39 100644 --- a/tests/st/dyn_shape_dev/test_ops_bias_add.py +++ b/tests/st/dyn_shape_dev/test_ops_bias_add.py @@ -49,7 +49,7 @@ def test_bias_add_4d(mode): assert np.all(output.asnumpy() == expect_output), "bias_add execute failed, please check current code commit" -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -77,7 +77,7 @@ def test_bias_add_2d(mode): assert np.all(output.asnumpy() == expect_output), "bias_add execute failed, please check current code commit" -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -105,7 +105,7 @@ def test_bias_add_3d(mode): assert np.all(output.asnumpy() == expect_output), "bias_add execute failed, please check current code commit" -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -133,7 +133,7 @@ def test_bias_add_5d(mode): assert np.all(output.asnumpy() == expect_output), "bias_add execute failed, please check current code commit" -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -157,7 +157,7 @@ def test_bias_add_backward(mode): assert np.all(output.asnumpy() == expect_output), "bias_add_grad execute failed, please check current code commit" -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training diff --git a/tests/st/dyn_shape_dev/test_ops_bias_add_grad.py b/tests/st/dyn_shape_dev/test_ops_bias_add_grad.py index 655da7466d4..441bdcbb274 100644 --- a/tests/st/dyn_shape_dev/test_ops_bias_add_grad.py +++ b/tests/st/dyn_shape_dev/test_ops_bias_add_grad.py @@ -20,7 +20,7 @@ from mindspore import ops import test_utils -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -45,7 +45,7 @@ def test_bias_add_grad_2d(data_type, mode): assert np.all(output.asnumpy() == expect_output), "bias_add_grad execute failed, please check current code commit" -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -71,7 +71,7 @@ def test_bias_add_grad_4d(data_type, mode): assert np.all(output.asnumpy() == expect_output), "bias_add_grad execute failed, please check current code commit" -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training diff --git a/tests/st/dyn_shape_dev/test_ops_cast.py b/tests/st/dyn_shape_dev/test_ops_cast.py index c1b4d0a91e7..4c0f0687439 100644 --- a/tests/st/dyn_shape_dev/test_ops_cast.py +++ b/tests/st/dyn_shape_dev/test_ops_cast.py @@ -31,7 +31,7 @@ class Net(ms.nn.Cell): return self.Cast(x, self.dtype) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -53,7 +53,7 @@ def test_cast_forward(mode): assert output.asnumpy().dtype == 'float64' -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -74,7 +74,7 @@ def test_cast_backward(mode): assert np.allclose(grads.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_ceil.py b/tests/st/dyn_shape_dev/test_ops_ceil.py index 6e57b96586f..83116b1f68c 100644 --- a/tests/st/dyn_shape_dev/test_ops_ceil.py +++ b/tests/st/dyn_shape_dev/test_ops_ceil.py @@ -35,7 +35,7 @@ def ceil_dyn_shape_func(x): return ops.auto_generate.ceil(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -56,7 +56,7 @@ def test_ceil_forward(mode): assert np.allclose(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -77,7 +77,7 @@ def test_ceil_backward(mode): assert np.allclose(grads.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -103,7 +103,7 @@ def test_ceil_vmap(mode): assert np.allclose(vmap_out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -132,7 +132,7 @@ def test_ceil_dynamic(mode): assert np.allclose(output1.asnumpy(), expect1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_celu.py b/tests/st/dyn_shape_dev/test_ops_celu.py index 6fa334292ac..4807943c2fb 100644 --- a/tests/st/dyn_shape_dev/test_ops_celu.py +++ b/tests/st/dyn_shape_dev/test_ops_celu.py @@ -35,7 +35,7 @@ def celu_dyn_shape_func(x, alpha=1.0): return ops.auto_generate.celu_(x, alpha) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -57,7 +57,7 @@ def test_celu_forward(mode): np.testing.assert_allclose(output.asnumpy(), expect, rtol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -79,7 +79,7 @@ def test_celu_backward(mode): np.testing.assert_allclose(grads.asnumpy(), expect, rtol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -102,7 +102,7 @@ def test_celu_vmap(mode): np.testing.assert_allclose(vmap_out.asnumpy(), expect, rtol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -131,7 +131,7 @@ def test_celu_dynamic(mode): np.testing.assert_allclose(output1.asnumpy(), expect1, rtol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_cholesky.py b/tests/st/dyn_shape_dev/test_ops_cholesky.py index 5e4fe2f540e..3983c37a7df 100644 --- a/tests/st/dyn_shape_dev/test_ops_cholesky.py +++ b/tests/st/dyn_shape_dev/test_ops_cholesky.py @@ -26,7 +26,7 @@ def cholesky_forward_func(x, upper): return P.Cholesky(upper)(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_cholesky_inverse.py b/tests/st/dyn_shape_dev/test_ops_cholesky_inverse.py index 33906954a82..3365d51fac1 100644 --- a/tests/st/dyn_shape_dev/test_ops_cholesky_inverse.py +++ b/tests/st/dyn_shape_dev/test_ops_cholesky_inverse.py @@ -30,7 +30,7 @@ class CholeskyInverseNet(nn.Cell): return self.cholesky_inverse(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_conj.py b/tests/st/dyn_shape_dev/test_ops_conj.py index 67b7262cc21..0e7aca172ac 100644 --- a/tests/st/dyn_shape_dev/test_ops_conj.py +++ b/tests/st/dyn_shape_dev/test_ops_conj.py @@ -31,7 +31,7 @@ def conj_backward_func(x): return ops.grad(conj_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_cos.py b/tests/st/dyn_shape_dev/test_ops_cos.py index f2bd77c7d73..b9297737910 100644 --- a/tests/st/dyn_shape_dev/test_ops_cos.py +++ b/tests/st/dyn_shape_dev/test_ops_cos.py @@ -36,7 +36,7 @@ def cos_dyn_shape_func(x): return ops.auto_generate.cos(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -57,7 +57,7 @@ def test_cos_forward(mode): assert np.allclose(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -78,7 +78,7 @@ def test_cos_backward(mode): assert np.allclose(grads.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -104,7 +104,7 @@ def test_cos_vmap(mode): assert np.allclose(vmap_out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -133,7 +133,7 @@ def test_cos_dynamic(mode): assert np.allclose(output1.asnumpy(), expect1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_elu.py b/tests/st/dyn_shape_dev/test_ops_elu.py index 4eb4b522da6..ad9b4b3eb05 100644 --- a/tests/st/dyn_shape_dev/test_ops_elu.py +++ b/tests/st/dyn_shape_dev/test_ops_elu.py @@ -30,7 +30,7 @@ def elu_backward_func(x): return ops.grad(elu_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -50,7 +50,7 @@ def test_elu_forward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-3, atol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -70,7 +70,7 @@ def test_elu_backward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_equal.py b/tests/st/dyn_shape_dev/test_ops_equal.py index 8984210667f..073e93b8a1e 100644 --- a/tests/st/dyn_shape_dev/test_ops_equal.py +++ b/tests/st/dyn_shape_dev/test_ops_equal.py @@ -44,7 +44,7 @@ def equal_infervalue_func2(): return ops.auto_generate.equal(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -65,7 +65,7 @@ def test_equal_forward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -86,7 +86,7 @@ def test_equal_backward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -108,7 +108,7 @@ def test_equal_vmap(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_erf.py b/tests/st/dyn_shape_dev/test_ops_erf.py index 3e53a48df95..08f3a78276f 100644 --- a/tests/st/dyn_shape_dev/test_ops_erf.py +++ b/tests/st/dyn_shape_dev/test_ops_erf.py @@ -30,7 +30,7 @@ def erf_backward_func(x): return ops.grad(erf_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -50,7 +50,7 @@ def test_erf_forward(mode): assert np.allclose(output.asnumpy(), expect, rtol=0.001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -70,7 +70,7 @@ def test_erf_backward(mode): assert np.allclose(output.asnumpy(), expect, rtol=0.001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_erfc.py b/tests/st/dyn_shape_dev/test_ops_erfc.py index b3ee42e041d..e31ea3075a4 100644 --- a/tests/st/dyn_shape_dev/test_ops_erfc.py +++ b/tests/st/dyn_shape_dev/test_ops_erfc.py @@ -30,7 +30,7 @@ def erfc_backward_func(x): return ops.grad(erfc_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -50,7 +50,7 @@ def test_erfc_forward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -70,7 +70,7 @@ def test_erfc_backward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_erfinv.py b/tests/st/dyn_shape_dev/test_ops_erfinv.py index f9474b37507..4ae1dd18b0f 100644 --- a/tests/st/dyn_shape_dev/test_ops_erfinv.py +++ b/tests/st/dyn_shape_dev/test_ops_erfinv.py @@ -30,7 +30,7 @@ def erfinv_backward_func(x): return ops.grad(erfinv_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -50,7 +50,7 @@ def test_erfinv_forward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -70,7 +70,7 @@ def test_erfinv_backward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_exp.py b/tests/st/dyn_shape_dev/test_ops_exp.py index 6870a6b6953..c076776f9a2 100644 --- a/tests/st/dyn_shape_dev/test_ops_exp.py +++ b/tests/st/dyn_shape_dev/test_ops_exp.py @@ -30,7 +30,7 @@ def exp_backward_func(x): return ops.grad(exp_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -50,7 +50,7 @@ def test_exp_forward(mode): assert np.allclose(output.asnumpy(), expect, rtol=0.001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -70,7 +70,7 @@ def test_exp_backward(mode): assert np.allclose(output.asnumpy(), expect, rtol=0.001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_expand_dims.py b/tests/st/dyn_shape_dev/test_ops_expand_dims.py index 3eb5f1f9c62..b98bfaeac90 100644 --- a/tests/st/dyn_shape_dev/test_ops_expand_dims.py +++ b/tests/st/dyn_shape_dev/test_ops_expand_dims.py @@ -30,7 +30,7 @@ def expand_dims_backward_func(x, axis): return ops.grad(expand_dims_forward_func, (0,))(x, axis) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -51,7 +51,7 @@ def test_expand_dims_forward(mode): assert np.allclose(output.asnumpy(), expect, rtol=0.001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -72,7 +72,7 @@ def test_expand_dims_backward(mode): assert np.allclose(output.asnumpy(), expect, rtol=0.001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_expm1.py b/tests/st/dyn_shape_dev/test_ops_expm1.py index 0595d3685c2..6b7c55e04fd 100644 --- a/tests/st/dyn_shape_dev/test_ops_expm1.py +++ b/tests/st/dyn_shape_dev/test_ops_expm1.py @@ -30,7 +30,7 @@ def expm1_backward_func(x): return ops.grad(expm1_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -50,7 +50,7 @@ def test_expm1_forward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -70,7 +70,7 @@ def test_expm1_backward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_flatten.py b/tests/st/dyn_shape_dev/test_ops_flatten.py index 07b57e472eb..5a63cffd65e 100644 --- a/tests/st/dyn_shape_dev/test_ops_flatten.py +++ b/tests/st/dyn_shape_dev/test_ops_flatten.py @@ -30,7 +30,7 @@ def flatten_backward_func(x): return ops.grad(flatten_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -50,7 +50,7 @@ def test_flatten_forward(mode): assert np.allclose(output.asnumpy(), expect, rtol=0.001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -70,7 +70,7 @@ def test_flatten_backward(mode): assert np.allclose(output.asnumpy(), expect, rtol=0.001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_floor.py b/tests/st/dyn_shape_dev/test_ops_floor.py index 914fb431575..ae81a7d94c6 100644 --- a/tests/st/dyn_shape_dev/test_ops_floor.py +++ b/tests/st/dyn_shape_dev/test_ops_floor.py @@ -34,7 +34,7 @@ def floor_dyn_shape_func(x): return ops.auto_generate.floor(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -54,7 +54,7 @@ def test_floor_forward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -74,7 +74,7 @@ def test_floor_backward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -95,7 +95,7 @@ def test_floor_vmap(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -124,7 +124,7 @@ def test_floor_dynamic(mode): assert np.allclose(output1.asnumpy(), expect1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_floor_div.py b/tests/st/dyn_shape_dev/test_ops_floor_div.py index e35b311b24f..290d2afbd7d 100644 --- a/tests/st/dyn_shape_dev/test_ops_floor_div.py +++ b/tests/st/dyn_shape_dev/test_ops_floor_div.py @@ -30,7 +30,7 @@ def floor_div_backward_func(x, y): return ops.grad(floor_div_forward_func, (0,))(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -51,7 +51,7 @@ def test_floor_div_forward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -72,7 +72,7 @@ def test_floor_div_backward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_floor_mod.py b/tests/st/dyn_shape_dev/test_ops_floor_mod.py index cc829fc0971..fb123f96840 100644 --- a/tests/st/dyn_shape_dev/test_ops_floor_mod.py +++ b/tests/st/dyn_shape_dev/test_ops_floor_mod.py @@ -30,7 +30,7 @@ def floor_mod_backward_func(x, y): return ops.grad(floor_mod_forward_func, (0,))(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -51,7 +51,7 @@ def test_floor_mod_forward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -72,7 +72,7 @@ def test_floor_mod_backward(mode): assert np.allclose(output.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_gather.py b/tests/st/dyn_shape_dev/test_ops_gather.py index f21d06aac23..c84d6e51812 100644 --- a/tests/st/dyn_shape_dev/test_ops_gather.py +++ b/tests/st/dyn_shape_dev/test_ops_gather.py @@ -30,7 +30,7 @@ def gather_backward_func(input_params, input_indices, axis, batch_dims=0): return ops.grad(gather_forward_func, (0,))(input_params, input_indices, axis, batch_dims) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_gather_d.py b/tests/st/dyn_shape_dev/test_ops_gather_d.py index 02b1c0e2a49..82b18505d50 100644 --- a/tests/st/dyn_shape_dev/test_ops_gather_d.py +++ b/tests/st/dyn_shape_dev/test_ops_gather_d.py @@ -31,7 +31,7 @@ def gather_d_backward_func(x, dim, index): return ops.grad(gather_d_forward_func, (0,))(x, dim, index) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -53,7 +53,7 @@ def test_gather_d_forward(mode): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -74,7 +74,7 @@ def test_gather_d_backward(mode): assert np.allclose(grads.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_identity.py b/tests/st/dyn_shape_dev/test_ops_identity.py index d1981fcb170..3265359edc5 100644 --- a/tests/st/dyn_shape_dev/test_ops_identity.py +++ b/tests/st/dyn_shape_dev/test_ops_identity.py @@ -35,7 +35,7 @@ def identity_dyn_shape_func(x): return ops.auto_generate.deepcopy(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -55,7 +55,7 @@ def test_identity_forward(mode): assert np.allclose(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -75,7 +75,7 @@ def test_identity_backward(mode): assert np.allclose(grads.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -100,7 +100,7 @@ def test_identity_vmap(mode): assert np.allclose(vmap_out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -128,7 +128,7 @@ def test_identity_dynamic(mode): assert np.allclose(output1.asnumpy(), expect1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_lin_space.py b/tests/st/dyn_shape_dev/test_ops_lin_space.py index 675aacc5a07..c27f3d3fc14 100644 --- a/tests/st/dyn_shape_dev/test_ops_lin_space.py +++ b/tests/st/dyn_shape_dev/test_ops_lin_space.py @@ -35,7 +35,7 @@ def lin_sapce_dyn_shape_func(start, stop, num=5): return ops.auto_generate.lin_space_(start, stop, num) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -55,7 +55,7 @@ def test_lin_space_forward(mode): assert np.allclose(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_ops_scalar_arithmetic.py b/tests/st/dyn_shape_dev/test_ops_scalar_arithmetic.py index 4cf1b29f6c7..8e6aa9787dc 100644 --- a/tests/st/dyn_shape_dev/test_ops_scalar_arithmetic.py +++ b/tests/st/dyn_shape_dev/test_ops_scalar_arithmetic.py @@ -21,7 +21,7 @@ import mindspore.ops.operations.manually_defined as F import test_utils -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -61,7 +61,7 @@ def test_scalar_add(mode): assert np.allclose(mutable_grad_output, expect_grad_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -98,7 +98,7 @@ def test_scalar_sub(mode): assert np.allclose(mutable_grad_output, expect_grad_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -134,7 +134,7 @@ def test_scalar_mul(mode): assert np.allclose(mutable_grad_output, expect_grad_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -170,7 +170,7 @@ def test_scalar_div(mode): assert np.allclose(mutable_grad_output, expect_grad_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -206,7 +206,7 @@ def test_scalar_mod(mode): assert np.allclose(mutable_grad_output, expect_grad_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -242,7 +242,7 @@ def test_scalar_floordiv(mode): assert np.allclose(mutable_grad_output, expect_grad_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -278,7 +278,7 @@ def test_scalar_eq(mode): assert np.allclose(mutable_grad_output, expect_grad_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -314,7 +314,7 @@ def test_scalar_ge(mode): assert np.allclose(mutable_grad_output, expect_grad_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -350,7 +350,7 @@ def test_scalar_gt(mode): assert np.allclose(mutable_grad_output, expect_grad_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -386,7 +386,7 @@ def test_scalar_le(mode): assert np.allclose(mutable_grad_output, expect_grad_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -421,7 +421,7 @@ def test_scalar_lt(mode): mutable_grad_output = scalar_lt_backward_func(ms.mutable(input_x), ms.mutable(input_y)) assert np.allclose(mutable_grad_output, expect_grad_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training diff --git a/tests/st/dyn_shape_dev/test_ops_scalar_arithmetic_unary.py b/tests/st/dyn_shape_dev/test_ops_scalar_arithmetic_unary.py index 98ced81abea..4ac90eaf2fc 100644 --- a/tests/st/dyn_shape_dev/test_ops_scalar_arithmetic_unary.py +++ b/tests/st/dyn_shape_dev/test_ops_scalar_arithmetic_unary.py @@ -20,7 +20,7 @@ from mindspore import ops import mindspore.ops.operations.manually_defined as F import test_utils -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -54,7 +54,7 @@ def test_scalar_bool_mutable(mode): assert np.allclose(mutable_grad_output, expect_grad_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -88,7 +88,7 @@ def test_scalar_uadd_mutable(mode): assert np.allclose(mutable_grad_output, expect_grad_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -122,7 +122,7 @@ def test_scalar_usub_mutable(mode): assert np.allclose(mutable_grad_output, expect_grad_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training diff --git a/tests/st/dyn_shape_dev/test_ops_split.py b/tests/st/dyn_shape_dev/test_ops_split.py index 8c329c9ee2d..eb8f8ffa710 100644 --- a/tests/st/dyn_shape_dev/test_ops_split.py +++ b/tests/st/dyn_shape_dev/test_ops_split.py @@ -35,7 +35,7 @@ def split_dyn_shape_func(x, axis=0, output_num=2): return ops.auto_generate.split_(x, axis, output_num) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -57,7 +57,7 @@ def test_split_forward(mode): assert np.allclose(res.asnumpy(), exp) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -77,7 +77,7 @@ def test_split_backward(mode): assert np.allclose(grads.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -100,7 +100,7 @@ def test_split_vmap(mode): assert np.allclose(res.asnumpy(), exp) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -130,7 +130,7 @@ def test_split_dynamic(mode): assert np.allclose(res.asnumpy(), exp) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_pow.py b/tests/st/dyn_shape_dev/test_pow.py index 8bb70bf44ed..14512ce3ba2 100644 --- a/tests/st/dyn_shape_dev/test_pow.py +++ b/tests/st/dyn_shape_dev/test_pow.py @@ -36,7 +36,7 @@ def pow_vmap_func(x, y): return ops.vmap(pow_forward_func, in_axes=0, out_axes=0)(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -58,7 +58,7 @@ def test_pow_op_forward(context_mode, data_type): np.testing.assert_allclose(out.asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -81,7 +81,7 @@ def test_pow_op_backward(context_mode, data_type): np.testing.assert_allclose(grads[1].asnumpy(), expect_out[1], rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_prelu.py b/tests/st/dyn_shape_dev/test_prelu_dyn_shape.py similarity index 98% rename from tests/st/dyn_shape_dev/test_prelu.py rename to tests/st/dyn_shape_dev/test_prelu_dyn_shape.py index 23741385685..480f582e4c4 100644 --- a/tests/st/dyn_shape_dev/test_prelu.py +++ b/tests/st/dyn_shape_dev/test_prelu_dyn_shape.py @@ -31,7 +31,7 @@ def prelu_backward_func(x, weight): return ops.grad(prelu_forward_func, (0, 1))(x, weight) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -57,7 +57,7 @@ def test_prelu_forward(mode): np.testing.assert_array_almost_equal(output.asnumpy(), expect_output, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_prelu_grad.py b/tests/st/dyn_shape_dev/test_prelu_grad.py index c9c0e2793b9..f1d715b9b4e 100644 --- a/tests/st/dyn_shape_dev/test_prelu_grad.py +++ b/tests/st/dyn_shape_dev/test_prelu_grad.py @@ -26,7 +26,7 @@ def prelu_grad_func(y, x, weight): return ops.auto_generate.prelu_grad(y, x, weight) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_qr.py b/tests/st/dyn_shape_dev/test_qr.py index 098332b93d5..92463223d96 100644 --- a/tests/st/dyn_shape_dev/test_qr.py +++ b/tests/st/dyn_shape_dev/test_qr.py @@ -30,7 +30,7 @@ def qr_backward_func(x, full_matrices): return ops.grad(qr_forward_func, (0, 1))(x, full_matrices) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_range.py b/tests/st/dyn_shape_dev/test_range.py index 953581c38d4..e7503c5ee47 100644 --- a/tests/st/dyn_shape_dev/test_range.py +++ b/tests/st/dyn_shape_dev/test_range.py @@ -31,7 +31,7 @@ def range_backward_func(start, limit, delta): return ops.grad(range_forward_func, (0, 1, 2))(start, limit, delta) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize('mode', [ms.GRAPH_MODE, ms.PYNATIVE_MODE]) @@ -52,7 +52,7 @@ def test_range_forward_tensor_input(mode): np.testing.assert_array_equal(output.asnumpy(), expect_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -75,7 +75,7 @@ def test_range_forward(mode): np.testing.assert_array_equal(output.asnumpy(), expect_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_rank.py b/tests/st/dyn_shape_dev/test_rank.py index f61039ad3b5..03dddfd431a 100644 --- a/tests/st/dyn_shape_dev/test_rank.py +++ b/tests/st/dyn_shape_dev/test_rank.py @@ -30,7 +30,7 @@ def rank_backward_func(x): return ops.grad(rank_forward_func, (0))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -50,7 +50,7 @@ def test_rank_forward(mode): np.testing.assert_equal(output, expect_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_real.py b/tests/st/dyn_shape_dev/test_real.py index ff32426f5de..a5287c3f0ef 100644 --- a/tests/st/dyn_shape_dev/test_real.py +++ b/tests/st/dyn_shape_dev/test_real.py @@ -30,7 +30,7 @@ def real_backward_func(x): return ops.grad(real_forward_func, (0))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_real_div.py b/tests/st/dyn_shape_dev/test_real_div.py index 5b048d28087..ea68e5a5b6d 100644 --- a/tests/st/dyn_shape_dev/test_real_div.py +++ b/tests/st/dyn_shape_dev/test_real_div.py @@ -31,7 +31,7 @@ def real_div_backward_func(x, y): return ops.grad(real_div_forward_func, (0, 1))(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -52,7 +52,7 @@ def test_real_div_forward(mode): np.testing.assert_allclose(output.asnumpy(), expect_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -75,7 +75,7 @@ def test_real_div_backward(mode): np.testing.assert_array_almost_equal(dy.asnumpy(), except_dy, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_reciprocal.py b/tests/st/dyn_shape_dev/test_reciprocal.py index 221767c0dc8..8247ff285ad 100644 --- a/tests/st/dyn_shape_dev/test_reciprocal.py +++ b/tests/st/dyn_shape_dev/test_reciprocal.py @@ -30,7 +30,7 @@ def reciprocal_backward_func(x): return ops.grad(reciprocal_forward_func, (0))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -50,7 +50,7 @@ def test_reciprocal_forward(mode): np.testing.assert_array_almost_equal(output.asnumpy(), expect_output, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -70,7 +70,7 @@ def test_reciprocal_backward(mode): np.testing.assert_array_almost_equal(output.asnumpy(), expect_output, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_reciprocal_grad.py b/tests/st/dyn_shape_dev/test_reciprocal_grad.py index 8eacb402b5f..be198c4d820 100644 --- a/tests/st/dyn_shape_dev/test_reciprocal_grad.py +++ b/tests/st/dyn_shape_dev/test_reciprocal_grad.py @@ -25,7 +25,7 @@ def reciprocal_grad_func(y, dy): return ops.auto_generate.reciprocal_grad(y, dy) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -46,7 +46,7 @@ def test_reciprocal_grad(mode): np.testing.assert_array_almost_equal(output.asnumpy(), expect_output, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -68,7 +68,7 @@ def test_reciprocal_grad_vmap(mode): np.testing.assert_equal(output.asnumpy(), expect_out.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_reduce_all.py b/tests/st/dyn_shape_dev/test_reduce_all.py index b00f8d3de8b..e0f452c5092 100644 --- a/tests/st/dyn_shape_dev/test_reduce_all.py +++ b/tests/st/dyn_shape_dev/test_reduce_all.py @@ -31,7 +31,7 @@ def reduce_all_backward_func(x): return ops.grad(reduce_all_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -56,7 +56,7 @@ def test_reduce_all(mode): assert (grad.asnumpy() == expect_grad).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_reduce_any.py b/tests/st/dyn_shape_dev/test_reduce_any.py index 24448cdab3d..1bf0d713a41 100644 --- a/tests/st/dyn_shape_dev/test_reduce_any.py +++ b/tests/st/dyn_shape_dev/test_reduce_any.py @@ -31,7 +31,7 @@ def reduce_any_backward_func(x): return ops.grad(reduce_any_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -56,7 +56,7 @@ def test_reduce_any(mode): assert (grad.asnumpy() == expect_grad).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_reduce_max.py b/tests/st/dyn_shape_dev/test_reduce_max.py index 27033d79076..d3c348c79f2 100644 --- a/tests/st/dyn_shape_dev/test_reduce_max.py +++ b/tests/st/dyn_shape_dev/test_reduce_max.py @@ -32,7 +32,7 @@ def reduce_max_backward_func(x): return ops.grad(reduce_max_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -58,7 +58,7 @@ def test_reduce_max(mode): assert (grad.asnumpy() == expect_grad).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_reduce_mean.py b/tests/st/dyn_shape_dev/test_reduce_mean.py index be821ac5463..1acd3fb5d44 100644 --- a/tests/st/dyn_shape_dev/test_reduce_mean.py +++ b/tests/st/dyn_shape_dev/test_reduce_mean.py @@ -32,7 +32,7 @@ def reduce_mean_backward_func(x): return ops.grad(reduce_mean_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -58,7 +58,7 @@ def test_reduce_mean(mode): assert np.allclose(grad.asnumpy(), expect_grad, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_reduce_min.py b/tests/st/dyn_shape_dev/test_reduce_min.py index c6f4db64e67..8eed4548733 100644 --- a/tests/st/dyn_shape_dev/test_reduce_min.py +++ b/tests/st/dyn_shape_dev/test_reduce_min.py @@ -32,7 +32,7 @@ def reduce_min_backward_func(x): return ops.grad(reduce_min_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -58,7 +58,7 @@ def test_reduce_min(mode): assert (grad.asnumpy() == expect_grad).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_reduce_prod.py b/tests/st/dyn_shape_dev/test_reduce_prod.py index 2dd3697fd9a..4feaa4c9e8e 100644 --- a/tests/st/dyn_shape_dev/test_reduce_prod.py +++ b/tests/st/dyn_shape_dev/test_reduce_prod.py @@ -31,7 +31,7 @@ def reduce_prod_backward_func(x): return ops.grad(reduce_prod_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -57,7 +57,7 @@ def test_reduce_prod(mode): assert np.allclose(grad.asnumpy(), expect_grad, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_reduce_sum.py b/tests/st/dyn_shape_dev/test_reduce_sum.py index 781d1445df4..fca0076da7e 100644 --- a/tests/st/dyn_shape_dev/test_reduce_sum.py +++ b/tests/st/dyn_shape_dev/test_reduce_sum.py @@ -32,7 +32,7 @@ def reduce_sum_backward_func(x): return ops.grad(reduce_sum_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -58,7 +58,7 @@ def test_reduce_sum(mode): assert np.allclose(grad.asnumpy(), expect_grad, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_relu.py b/tests/st/dyn_shape_dev/test_relu.py index 01b9a488bfe..ee17afdddf2 100644 --- a/tests/st/dyn_shape_dev/test_relu.py +++ b/tests/st/dyn_shape_dev/test_relu.py @@ -31,7 +31,7 @@ def relu_backward_func(x): return ops.grad(relu_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -61,7 +61,7 @@ def test_relu(mode): assert (grad.asnumpy() == expect_grad).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_relu6.py b/tests/st/dyn_shape_dev/test_relu6.py index d27a6165870..c6efd0176e7 100644 --- a/tests/st/dyn_shape_dev/test_relu6.py +++ b/tests/st/dyn_shape_dev/test_relu6.py @@ -31,7 +31,7 @@ def relu6_backward_func(x): return ops.grad(relu6_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -61,7 +61,7 @@ def test_relu6(mode): assert (grad.asnumpy() == expect_grad).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_relu6_grad.py b/tests/st/dyn_shape_dev/test_relu6_grad.py index 928c88f228f..b2bb67b90bf 100644 --- a/tests/st/dyn_shape_dev/test_relu6_grad.py +++ b/tests/st/dyn_shape_dev/test_relu6_grad.py @@ -27,7 +27,7 @@ def relu6_grad_func(dy, x): return _relu6_grad(dy, x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -54,7 +54,7 @@ def test_relu6_grad(mode): assert (out.asnumpy() == expect_out).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_relu_grad.py b/tests/st/dyn_shape_dev/test_relu_grad.py index 99fa06abe6a..8c7b244cc68 100644 --- a/tests/st/dyn_shape_dev/test_relu_grad.py +++ b/tests/st/dyn_shape_dev/test_relu_grad.py @@ -27,7 +27,7 @@ def relu_grad_func(dy, x): return _relu_grad(dy, x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -54,7 +54,7 @@ def test_relu_grad(mode): assert (out.asnumpy() == expect_out).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_reshape.py b/tests/st/dyn_shape_dev/test_reshape.py index d881279789a..9c0d2534198 100644 --- a/tests/st/dyn_shape_dev/test_reshape.py +++ b/tests/st/dyn_shape_dev/test_reshape.py @@ -30,7 +30,7 @@ def reshape_backward_func(x, shape): return ops.grad(reshape_forward_func, (0,))(x, shape) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -56,7 +56,7 @@ def test_reshape_op(mode): assert (grad.asnumpy() == expect_grad).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_resize_nearest_neighbor.py b/tests/st/dyn_shape_dev/test_resize_nearest_neighbor.py index ede17f0afb3..efab07fa777 100644 --- a/tests/st/dyn_shape_dev/test_resize_nearest_neighbor.py +++ b/tests/st/dyn_shape_dev/test_resize_nearest_neighbor.py @@ -34,7 +34,7 @@ def resize_nearest_neighbor_backward_func(x, size, align_corners): def resize_nearest_neighbor_dyn_shape_func(x, size, align_corners): return ops.auto_generate.resize_nearest_neighbor(x, size, align_corners) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -56,7 +56,7 @@ def test_resize_nearest_neighbor_op_forward(context_mode, data_type): np.testing.assert_allclose(out.asnumpy(), expect_out, rtol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_resize_nearest_neighbor_grad.py b/tests/st/dyn_shape_dev/test_resize_nearest_neighbor_grad.py index 86607164f48..f3fc7c63403 100644 --- a/tests/st/dyn_shape_dev/test_resize_nearest_neighbor_grad.py +++ b/tests/st/dyn_shape_dev/test_resize_nearest_neighbor_grad.py @@ -30,7 +30,7 @@ def resize_nearest_neighbor_grad_dyn_shape_func(x, size, align_corners): return ops.auto_generate.resize_nearest_neighbor_grad(x, size, align_corners) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_resize_nearest_neighbor_v2.py b/tests/st/dyn_shape_dev/test_resize_nearest_neighbor_v2.py index c67fe28aa80..9d04e4b3942 100644 --- a/tests/st/dyn_shape_dev/test_resize_nearest_neighbor_v2.py +++ b/tests/st/dyn_shape_dev/test_resize_nearest_neighbor_v2.py @@ -35,7 +35,7 @@ def resize_nearest_neighbor_v2_dyn_shape_func(x, size, align_corners, half_pixel return ops.auto_generate.resize_nearest_neighbor_v2(x, size, align_corners, half_pixel_centers) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_reverse_v2.py b/tests/st/dyn_shape_dev/test_reverse_v2.py index 33603012f43..552c8e5e1be 100644 --- a/tests/st/dyn_shape_dev/test_reverse_v2.py +++ b/tests/st/dyn_shape_dev/test_reverse_v2.py @@ -30,7 +30,7 @@ def reverse_v2_backward_func(x): return ops.grad(reverse_v2_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -55,7 +55,7 @@ def test_reverse_v2(mode): assert (grad.asnumpy() == expect_grad).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_right_shift.py b/tests/st/dyn_shape_dev/test_right_shift.py index bfe297f2ad9..a6c34fe946f 100644 --- a/tests/st/dyn_shape_dev/test_right_shift.py +++ b/tests/st/dyn_shape_dev/test_right_shift.py @@ -30,7 +30,7 @@ def right_shift_backward_func(x, y): return ops.grad(right_shift_forward_func, (0,))(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_roll.py b/tests/st/dyn_shape_dev/test_roll.py index cc941731d1e..d8c362175d8 100644 --- a/tests/st/dyn_shape_dev/test_roll.py +++ b/tests/st/dyn_shape_dev/test_roll.py @@ -30,7 +30,7 @@ def roll_backward_func(x): return ops.grad(roll_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -53,7 +53,7 @@ def test_roll(mode): assert (grad.asnumpy() == expect_grad).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training diff --git a/tests/st/dyn_shape_dev/test_round.py b/tests/st/dyn_shape_dev/test_round.py index 995663884bc..ab4115b48b5 100644 --- a/tests/st/dyn_shape_dev/test_round.py +++ b/tests/st/dyn_shape_dev/test_round.py @@ -31,7 +31,7 @@ def round_backward_func(x): return ops.grad(round_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -56,7 +56,7 @@ def test_round(mode): assert (grad.asnumpy() == expect_grad).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_rsqrt.py b/tests/st/dyn_shape_dev/test_rsqrt.py index 43c877c04ce..d59aa58b687 100644 --- a/tests/st/dyn_shape_dev/test_rsqrt.py +++ b/tests/st/dyn_shape_dev/test_rsqrt.py @@ -31,7 +31,7 @@ def rsqrt_backward_func(x): return ops.grad(rsqrt_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -55,7 +55,7 @@ def test_rsqrt(mode): np.testing.assert_allclose(grad.asnumpy(), expect_grad, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_rsqrt_grad.py b/tests/st/dyn_shape_dev/test_rsqrt_grad.py index 3688e8db226..3f4c1ec1dfb 100644 --- a/tests/st/dyn_shape_dev/test_rsqrt_grad.py +++ b/tests/st/dyn_shape_dev/test_rsqrt_grad.py @@ -27,7 +27,7 @@ def rsqrt_grad_func(dy, x): return _rsqrt_grad(dy, x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -54,7 +54,7 @@ def test_rsqrt_grad(mode): np.testing.assert_allclose(out.asnumpy(), expect_out, rtol=1e-3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_sigmoid.py b/tests/st/dyn_shape_dev/test_sigmoid.py index 9105fe3803b..8616ad7e89e 100644 --- a/tests/st/dyn_shape_dev/test_sigmoid.py +++ b/tests/st/dyn_shape_dev/test_sigmoid.py @@ -69,7 +69,7 @@ def test_sigmoid_backward(mode): grads = sigmoid_backward_func(x) assert np.allclose(grads.numpy(), expect_out.numpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_sigmoid_grad.py b/tests/st/dyn_shape_dev/test_sigmoid_grad.py index e83e5b2ec56..f8f726704dd 100644 --- a/tests/st/dyn_shape_dev/test_sigmoid_grad.py +++ b/tests/st/dyn_shape_dev/test_sigmoid_grad.py @@ -71,7 +71,7 @@ def test_sigmoid_grad_backward(mode): ddy, _ = sigmoid_grad_backward_func(y, dy) assert np.allclose(ddy.numpy(), expect_out.numpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_sin.py b/tests/st/dyn_shape_dev/test_sin.py index 6b07f536b0b..fcf75b59703 100644 --- a/tests/st/dyn_shape_dev/test_sin.py +++ b/tests/st/dyn_shape_dev/test_sin.py @@ -31,7 +31,7 @@ def sin_backward_func(x): return ops.grad(sin_forward_func, (0))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -51,7 +51,7 @@ def test_sin_forward(mode): np.testing.assert_array_almost_equal(output.asnumpy(), expect_output, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -71,7 +71,7 @@ def test_sin_backward(mode): np.testing.assert_array_almost_equal(output.asnumpy(), expect_output, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_sinc.py b/tests/st/dyn_shape_dev/test_sinc.py index 43a721a8eee..b7fe1afff62 100644 --- a/tests/st/dyn_shape_dev/test_sinc.py +++ b/tests/st/dyn_shape_dev/test_sinc.py @@ -30,7 +30,7 @@ def sinc_backward_func(x): return ops.grad(sinc_forward_func, (0))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -50,7 +50,7 @@ def test_sinc_forward(mode): np.testing.assert_array_almost_equal(output.asnumpy(), expect_output, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_sinh.py b/tests/st/dyn_shape_dev/test_sinh.py index cfbd1a89a97..ed6794dfd26 100644 --- a/tests/st/dyn_shape_dev/test_sinh.py +++ b/tests/st/dyn_shape_dev/test_sinh.py @@ -31,7 +31,7 @@ def sinh_backward_func(x): return ops.grad(sinh_forward_func, (0))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -51,7 +51,7 @@ def test_sinh_forward(mode): np.testing.assert_array_almost_equal(output.asnumpy(), expect_output, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -71,7 +71,7 @@ def test_sinh_backward(mode): np.testing.assert_array_almost_equal(output.asnumpy(), expect_output, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_softmax.py b/tests/st/dyn_shape_dev/test_softmax.py index 196370bae5a..ddc074ea42d 100644 --- a/tests/st/dyn_shape_dev/test_softmax.py +++ b/tests/st/dyn_shape_dev/test_softmax.py @@ -31,7 +31,7 @@ def softmax_backward_func(x): return ops.grad(softmax_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -54,7 +54,7 @@ def test_softmax_op(mode): assert np.allclose(output.asnumpy(), expect_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -73,7 +73,7 @@ def test_softmax_backward(mode): assert grads.asnumpy().shape == expect_shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -98,7 +98,7 @@ def test_sofmax_vmap(mode): assert out.asnumpy().shape == expect_shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -180,7 +180,7 @@ def test_softmax_dynamic_backward_shape(mode): # 反向动态shape有公共问题,待解决后再放开用例 @pytest.mark.skip -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_sqrt.py b/tests/st/dyn_shape_dev/test_sqrt.py index d123c8ff16e..178e922fe80 100644 --- a/tests/st/dyn_shape_dev/test_sqrt.py +++ b/tests/st/dyn_shape_dev/test_sqrt.py @@ -35,7 +35,7 @@ def sqrt_dyn_shape_func(x): return ops.auto_generate.sqrt_(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -55,7 +55,7 @@ def test_sqrt_forward(mode): assert np.allclose(out.asnumpy(), expect_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -75,7 +75,7 @@ def test_sqrt_backward(mode): assert np.allclose(grads.asnumpy(), expect_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -98,7 +98,7 @@ def test_sqrt_vmap(mode): assert np.allclose(out.asnumpy(), expect_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -126,7 +126,7 @@ def test_sqrt_dynamic(mode): assert np.allclose(output1.asnumpy(), expect_out1, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_sqrt_grad.py b/tests/st/dyn_shape_dev/test_sqrt_grad.py index 28b9d1a7c93..d05d86b0678 100644 --- a/tests/st/dyn_shape_dev/test_sqrt_grad.py +++ b/tests/st/dyn_shape_dev/test_sqrt_grad.py @@ -30,7 +30,7 @@ def sqrt_grad_dyn_shape_func(dy, x): return ops.auto_generate.sqrt_grad(dy, x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -52,7 +52,7 @@ def test_sqrt_grad_forward(mode): assert np.allclose(out.asnumpy(), expect_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -77,7 +77,7 @@ def test_sqrt_grad_vmap(mode): assert np.allclose(out.asnumpy(), expect_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -109,7 +109,7 @@ def test_sqrt_grad_dynamic(mode): assert np.allclose(output1.asnumpy(), expect_out1, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_square.py b/tests/st/dyn_shape_dev/test_square.py index f4733e3956a..778669939a8 100644 --- a/tests/st/dyn_shape_dev/test_square.py +++ b/tests/st/dyn_shape_dev/test_square.py @@ -35,7 +35,7 @@ def square_dyn_shape_func(x): return ops.auto_generate.square_(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -72,7 +72,7 @@ def test_square_backward(mode): assert np.allclose(grads.asnumpy(), expect_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -94,7 +94,7 @@ def test_square_vmap(mode): assert np.allclose(out.asnumpy(), expect_out, 1e-04, 1e-04) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -122,7 +122,7 @@ def test_square_dynamic(mode): assert np.allclose(output1.asnumpy(), expect_out1, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_tensor_shape.py b/tests/st/dyn_shape_dev/test_tensor_shape.py index 11ecfc456b0..b1e37ad4e3c 100644 --- a/tests/st/dyn_shape_dev/test_tensor_shape.py +++ b/tests/st/dyn_shape_dev/test_tensor_shape.py @@ -30,7 +30,7 @@ def tensor_shape_backward_func(x): return ops.grad(tensor_shape_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -50,7 +50,7 @@ def test_tensor_shape_forward(mode): assert (out.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -70,7 +70,7 @@ def test_tensor_shape_backward(mode): assert (grad.asnumpy() == expect_grad).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_tile.py b/tests/st/dyn_shape_dev/test_tile.py index 3eb8283d1b2..fa35796dda5 100644 --- a/tests/st/dyn_shape_dev/test_tile.py +++ b/tests/st/dyn_shape_dev/test_tile.py @@ -45,7 +45,7 @@ def tile_infer_value(): return tile_func(x, (1, 2)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -64,7 +64,7 @@ def test_tile_infer_value(): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -85,7 +85,7 @@ def test_tile_forward(mode): assert np.allclose(out.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -105,7 +105,7 @@ def test_tile_backward(mode): assert np.allclose(grads.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -128,7 +128,7 @@ def test_tile_vmap(mode): assert np.allclose(out.asnumpy(), expect, rtol=1e-4, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -159,7 +159,7 @@ def test_tile_forward_dyn(mode): assert np.allclose(out2.asnumpy(), expect2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_transpose.py b/tests/st/dyn_shape_dev/test_transpose.py index a2d0978ac76..8ad8740ff27 100644 --- a/tests/st/dyn_shape_dev/test_transpose.py +++ b/tests/st/dyn_shape_dev/test_transpose.py @@ -30,7 +30,7 @@ def transpose_backward_func(x): return ops.grad(transpose_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -55,7 +55,7 @@ def test_transpose_forward(mode): assert (out.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -75,7 +75,7 @@ def test_transpose_backward(mode): assert (grad.asnumpy() == expect_grad).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dyn_shape_dev/test_zeros_like.py b/tests/st/dyn_shape_dev/test_zeros_like.py index 4b81cf59873..8af075be5c3 100644 --- a/tests/st/dyn_shape_dev/test_zeros_like.py +++ b/tests/st/dyn_shape_dev/test_zeros_like.py @@ -32,7 +32,7 @@ def zeros_like_backward_func(x): return ops.grad(zeros_like_forward_func, (0,))(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -53,7 +53,7 @@ def test_zeros_like_forward(mode): assert (out.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -74,7 +74,7 @@ def test_zeros_like_backward(mode): assert (grad.asnumpy() == expect_grad).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/dynamic_shape/test_dynamic_stridedslice.py b/tests/st/dynamic_shape/test_dynamic_stridedslice.py index debc845c378..dac4e86bda3 100644 --- a/tests/st/dynamic_shape/test_dynamic_stridedslice.py +++ b/tests/st/dynamic_shape/test_dynamic_stridedslice.py @@ -67,7 +67,7 @@ def test_stridedslice_dynamic_shape(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard diff --git a/tests/st/dynamic_shape/test_dynamic_transformer.py b/tests/st/dynamic_shape/test_dynamic_transformer.py index 222cc07f8e3..871d2b2c070 100644 --- a/tests/st/dynamic_shape/test_dynamic_transformer.py +++ b/tests/st/dynamic_shape/test_dynamic_transformer.py @@ -197,7 +197,7 @@ def get_train_loss(is_graph_mode, device_target, device_id=0): return loss_list -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_train_graph_mode_gpu(): diff --git a/tests/st/export_and_load/test_mindir_dyn_shape_and_dyn_seq.py b/tests/st/export_and_load/test_mindir_dyn_shape_and_dyn_seq.py index 6fae5ec6353..da445b04869 100644 --- a/tests/st/export_and_load/test_mindir_dyn_shape_and_dyn_seq.py +++ b/tests/st/export_and_load/test_mindir_dyn_shape_and_dyn_seq.py @@ -133,7 +133,7 @@ def test_mutable_dynamic_tuple(): os.remove(verify_name) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/export_and_load/test_train_mindir.py b/tests/st/export_and_load/test_train_mindir.py index 1bae8aca1a6..670e3d2868f 100644 --- a/tests/st/export_and_load/test_train_mindir.py +++ b/tests/st/export_and_load/test_train_mindir.py @@ -112,7 +112,7 @@ def decrypt_func(cipher_file, key): return plain_data -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -134,7 +134,7 @@ def test_export_lenet_grad_mindir(): assert os.path.exists(verify_name) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_all_any.py b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_all_any.py index 1aa5eb98763..87c6bda2d5d 100644 --- a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_all_any.py +++ b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_all_any.py @@ -70,7 +70,7 @@ def test_fallback_any_dict(): @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -94,7 +94,7 @@ def test_fallback_all_asnumpy(): @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_bool.py b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_bool.py index 47682373332..e5f0e576591 100644 --- a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_bool.py +++ b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_bool.py @@ -41,7 +41,7 @@ def test_fallback_bool_tensor_asnumpy(): @pytest.mark.skip(reason="RebuildKernelSelectBackoffOp Unsupported op[Shape].") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_dict.py b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_dict.py index 6142f93a8fd..9659a9d8805 100644 --- a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_dict.py +++ b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_dict.py @@ -20,7 +20,7 @@ context.set_context(mode=context.GRAPH_MODE) @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_enumerate.py b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_enumerate.py index ae0fca45d5c..58997c5f9f9 100644 --- a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_enumerate.py +++ b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_enumerate.py @@ -21,7 +21,7 @@ from mindspore import Tensor, jit, context context.set_context(mode=context.GRAPH_MODE) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -46,7 +46,7 @@ def test_enumerate_asnumpy(): assert ret == (6, 110) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_filter.py b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_filter.py index 2ba4c98f3bf..900021a8b08 100644 --- a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_filter.py +++ b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_filter.py @@ -25,7 +25,7 @@ def filter_fn(x): @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_getattr.py b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_getattr.py index 8214ef5e21e..9f545140e6f 100644 --- a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_getattr.py +++ b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_getattr.py @@ -100,7 +100,7 @@ def test_getattr_numpy_array_2(): assert (out == np.array([0, 1, 2, 3, 4])).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_int_float.py b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_int_float.py index 875b794e331..639a14c2be9 100644 --- a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_int_float.py +++ b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_int_float.py @@ -59,7 +59,7 @@ def test_fallback_float(): @pytest.mark.skip(reason="ScalarToRawMemory memcpy failed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_iter.py b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_iter.py index ae8dabe565b..53f6ed53d6a 100644 --- a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_iter.py +++ b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_iter.py @@ -24,7 +24,7 @@ context.set_context(mode=context.GRAPH_MODE) @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -52,7 +52,7 @@ def test_list_tuple_string_dict_tensor_iter(): @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -75,7 +75,7 @@ def test_tensor_iter(): @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -102,7 +102,7 @@ def test_list_iter_in_for(): @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -128,7 +128,7 @@ def test_iter_asnumpy(): @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -156,7 +156,7 @@ def test_tuple_iter_next(): @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -200,7 +200,7 @@ class MyIterator(nn.Cell): @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -226,7 +226,7 @@ def test_custom_self_iterator_1(): @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -253,7 +253,7 @@ def test_custom_self_iterator_2(): @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_list_tuple.py b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_list_tuple.py index c3cde1cb147..cec6e4fce01 100644 --- a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_list_tuple.py +++ b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_list_tuple.py @@ -20,7 +20,7 @@ context.set_context(mode=context.GRAPH_MODE) @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_max_min.py b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_max_min.py index 20743d3bada..105f5a2e5c5 100644 --- a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_max_min.py +++ b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_max_min.py @@ -19,7 +19,7 @@ from mindspore import Tensor, jit, context context.set_context(mode=context.GRAPH_MODE) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -39,7 +39,7 @@ def test_fallback_runtime_max(): assert out == 3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -60,7 +60,7 @@ def test_fallback_runtime_min(): @pytest.mark.skip(reason="Invalid call node") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -82,7 +82,7 @@ def test_fallback_runtime_max_min(): assert out[0] == 3, out[1] == 1 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -102,7 +102,7 @@ def test_fallback_runtime_max_min_scalar_tensor(): assert out == (2, 3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_fallback_runtime_max_scalar_asnumpy(): diff --git a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_print.py b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_print.py index 39a6d40ea86..4fe7514ead0 100644 --- a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_print.py +++ b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_print.py @@ -127,7 +127,7 @@ def test_fallback_print_asnumpy_custom_class(): @pytest.mark.skip(reason="print mutalble is not supported") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -188,7 +188,7 @@ def test_fallback_print_pyinterpret(): check_output(cap.output, patterns) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -220,7 +220,7 @@ def test_np_init(): check_output(cap.output, patterns) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_range.py b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_range.py index fb141ff4c78..d90227b5b4c 100644 --- a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_range.py +++ b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_range.py @@ -19,7 +19,7 @@ from mindspore import Tensor, jit, context context.set_context(mode=context.GRAPH_MODE) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -43,7 +43,7 @@ def test_fallback_runtime_range(): assert out == 6 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_sum.py b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_sum.py index e0077258d9c..56b28399668 100644 --- a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_sum.py +++ b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_sum.py @@ -20,7 +20,7 @@ from mindspore import Tensor, jit, context context.set_context(mode=context.GRAPH_MODE) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_zip.py b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_zip.py index 17de6fc8638..a4f19843c98 100644 --- a/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_zip.py +++ b/tests/st/fallback/python_builtin_funtions/test_graph_fallback_runtime_zip.py @@ -67,7 +67,7 @@ def test_fallback_runtime_zip_asnumpy(): assert str(type(out)) == "" -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -90,7 +90,7 @@ def test_fallback_runtime_zip_asnumpy_tuple(): assert out == ((10, 10), (20, 20)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -114,7 +114,7 @@ def test_fallback_runtime_zip_tensor(): assert out == 30 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/test_graph_fallback_jit_syntax_level.py b/tests/st/fallback/test_graph_fallback_jit_syntax_level.py index 263fc97e1b6..76769e6be2b 100644 --- a/tests/st/fallback/test_graph_fallback_jit_syntax_level.py +++ b/tests/st/fallback/test_graph_fallback_jit_syntax_level.py @@ -27,7 +27,7 @@ jit_config_compatible = ms.JitConfig(jit_syntax_level="COMPATIBLE") jit_config_lax = ms.JitConfig(jit_syntax_level="LAX") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -54,7 +54,7 @@ def test_dict_in_cell(): assert isinstance(out2, dict) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/test_graph_fallback_runtime.py b/tests/st/fallback/test_graph_fallback_runtime.py index d8644f14b2c..2d5be1271e2 100644 --- a/tests/st/fallback/test_graph_fallback_runtime.py +++ b/tests/st/fallback/test_graph_fallback_runtime.py @@ -49,7 +49,7 @@ def test_getattr_tensor_with_wrong_attr(): assert "object has no attribute" in str(err.value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -280,7 +280,7 @@ def test_pyexecute_with_func_graph_input(): @pytest.mark.skip('backend not support different type in value tuple') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -346,7 +346,7 @@ class CreateNotDynTensor(nn.Cell): @pytest.mark.skip('ops.shape(x) is constant, not mutable.') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -442,7 +442,7 @@ def test_make_tensor_with_dynamic_shape_dtype(): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/fallback/test_graph_fallback_runtime_asnumpy.py b/tests/st/fallback/test_graph_fallback_runtime_asnumpy.py index 64d7787555b..d0e88f1616a 100644 --- a/tests/st/fallback/test_graph_fallback_runtime_asnumpy.py +++ b/tests/st/fallback/test_graph_fallback_runtime_asnumpy.py @@ -195,7 +195,7 @@ def test_tensor_asnumpy_2(): @pytest.mark.skip(reason="No support yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/test_graph_fallback_runtime_dict.py b/tests/st/fallback/test_graph_fallback_runtime_dict.py index b352e7b8486..c09930d9ee3 100644 --- a/tests/st/fallback/test_graph_fallback_runtime_dict.py +++ b/tests/st/fallback/test_graph_fallback_runtime_dict.py @@ -253,7 +253,7 @@ def fc_with_initialize(input_channels, out_channels): @pytest.mark.skip(reason="Pyexecute output is not any and type is wrong.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -305,7 +305,7 @@ def test_net_dict_1(): @pytest.mark.skip(reason="Pyexecute output is not any and type is wrong.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -584,7 +584,7 @@ def test_return_nested_dict_with_inputs(): assert out == {'a': [x, y], 'b': (1, 2, {'a': 1})} -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -888,7 +888,7 @@ def test_return_dict_with_dict_items(): assert out == {'a': (x, y)} -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -908,7 +908,7 @@ def test_return_dict_with_empty_shape_tensor(): assert out['a'].shape == () -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -960,7 +960,7 @@ def test_return_different_size_dict_in_if_else(): assert "Cannot join the return values of different branches, perhaps you need to make them equal." in str(e) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -982,7 +982,7 @@ def test_get_item_with_string_input(): assert out == x -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -1004,7 +1004,7 @@ def test_get_item_with_string_input_grad(): assert out == Tensor(1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -1025,7 +1025,7 @@ def test_return_dict_with_string_input(): assert out == {"a": x} -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -1046,7 +1046,7 @@ def test_return_dict_with_string_input_grad(): assert out == Tensor(1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/test_graph_fallback_runtime_dict_inplace.py b/tests/st/fallback/test_graph_fallback_runtime_dict_inplace.py index 9ca3b094d28..34882f08dcc 100644 --- a/tests/st/fallback/test_graph_fallback_runtime_dict_inplace.py +++ b/tests/st/fallback/test_graph_fallback_runtime_dict_inplace.py @@ -26,7 +26,7 @@ context.set_context(mode=context.GRAPH_MODE) global_dict_1 = {"1": 1} -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -95,7 +95,7 @@ def test_global_dict_used_in_graph_3(): global_input_dict_1 = {"1": 1} -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -252,7 +252,7 @@ def test_dict_inplace_setitem_2(): assert x == {"a": 3, "b": 2} -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -277,7 +277,7 @@ def test_dict_inplace_setitem_3(): assert id(y) == id(res["b"]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -307,7 +307,7 @@ def test_dict_inplace_setitem_with_attribute(): assert id(x) == id(ret) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -343,7 +343,7 @@ def test_dict_inplace_setitem_with_attribute_2(): @pytest.mark.skip(reason="setitem with abstract any do not convert to pyexecute yet") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -377,7 +377,7 @@ def test_dict_inplace_setitem_with_attribute_3(): assert id(x) == id(ret) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @@ -407,7 +407,7 @@ def test_dict_getitem_after_setitem(): global_dict_for_update = {'Name': 'a', 'Age': 7} -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @@ -439,7 +439,7 @@ def test_dict_getitem_after_setitem_2(): assert ret3 == {'Name': 'c', 'Age': 18, 'Grade': 'college'} -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training diff --git a/tests/st/fallback/test_graph_fallback_runtime_inplace_operation.py b/tests/st/fallback/test_graph_fallback_runtime_inplace_operation.py index 9a99d4013f8..361ab4f4c94 100644 --- a/tests/st/fallback/test_graph_fallback_runtime_inplace_operation.py +++ b/tests/st/fallback/test_graph_fallback_runtime_inplace_operation.py @@ -120,7 +120,7 @@ def test_change_net_list_2(): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -150,7 +150,7 @@ class ListNet2(Cell): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -168,7 +168,7 @@ def test_change_net_list_4(): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -196,7 +196,7 @@ class TensorNet(Cell): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -214,7 +214,7 @@ def test_change_net_tensor(): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -245,7 +245,7 @@ class ExternalNet(Cell): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -263,7 +263,7 @@ def test_change_net_external_object(): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -291,7 +291,7 @@ class ChangeNet(Cell): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -309,7 +309,7 @@ def test_change_net_attr_type_number_to_list(): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -327,7 +327,7 @@ def test_change_net_attr_type_tuple_to_number(): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -345,7 +345,7 @@ def test_change_net_attr_type_tuple_to_tensor(): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -374,7 +374,7 @@ class ChangeNet2(Cell): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -392,7 +392,7 @@ def test_change_net_attr_type_tensor_to_list_with_cal(): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -429,7 +429,7 @@ class HybridNet(Cell): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -450,7 +450,7 @@ def test_change_net_attr(): @pytest.mark.skip(reason="Not support to change attribute of cell object") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/test_graph_fallback_runtime_list_inplace.py b/tests/st/fallback/test_graph_fallback_runtime_list_inplace.py index 9b38595aaba..398ae9520ff 100644 --- a/tests/st/fallback/test_graph_fallback_runtime_list_inplace.py +++ b/tests/st/fallback/test_graph_fallback_runtime_list_inplace.py @@ -290,7 +290,7 @@ def test_return_local_list_4(): assert ret == [1.0, 2, 3.0, Tensor([1]), 2] -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -316,7 +316,7 @@ def test_return_local_list_5(): @pytest.mark.skip(reason="No need to convert to PyExecute node. SequenceMul execute fail in Ascend.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/test_graph_fallback_runtime_meta_fg.py b/tests/st/fallback/test_graph_fallback_runtime_meta_fg.py index a700ee2fa16..6b8d4327a3e 100644 --- a/tests/st/fallback/test_graph_fallback_runtime_meta_fg.py +++ b/tests/st/fallback/test_graph_fallback_runtime_meta_fg.py @@ -392,7 +392,7 @@ def test_fallback_getitem_meta_2(): assert ret == (2, 3) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -572,7 +572,7 @@ def test_fallback_meta_fg_not_support_type_bitwise_xor(): assert "unsupported operand type" in str(err.value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -824,7 +824,7 @@ def test_fallback_meta_fg_not_support_type_pow(): assert "unsupported operand type" in str(err.value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -959,7 +959,7 @@ def test_fallback_meta_fg_not_support_type_not_equal(): @pytest.mark.skip(reason="do not support inplace operation yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -989,7 +989,7 @@ def test_fallback_setitem_meta(): @pytest.mark.skip(reason="do not support inplace operation yet.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -1045,7 +1045,7 @@ def test_shift_operator_error_list_input(): assert "unsupported operand type" in str(err.value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -1077,7 +1077,7 @@ def test_user_made_meta_fg(): assert res == 3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -1109,7 +1109,7 @@ def test_user_made_meta_fg_with_error(): assert "cannot find fn match given args." in str(err.value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -1141,7 +1141,7 @@ def test_user_made_meta_fg_with_hyper_map(): assert res == 3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -1165,7 +1165,7 @@ def test_getitem_with_no_script(): assert res == (1, 2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/test_graph_fallback_runtime_return_list.py b/tests/st/fallback/test_graph_fallback_runtime_return_list.py index 0538ff99f70..c3695a936ae 100644 --- a/tests/st/fallback/test_graph_fallback_runtime_return_list.py +++ b/tests/st/fallback/test_graph_fallback_runtime_return_list.py @@ -485,7 +485,7 @@ def test_return_list_from_third_party_3(): assert res == [0, 2, 4, 6, 8] -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/test_graph_fallback_runtime_setattr.py b/tests/st/fallback/test_graph_fallback_runtime_setattr.py index 9643c9fb61f..79a6ef2cf26 100644 --- a/tests/st/fallback/test_graph_fallback_runtime_setattr.py +++ b/tests/st/fallback/test_graph_fallback_runtime_setattr.py @@ -376,7 +376,7 @@ def test_setattr_global_obj_nested_attr2(): @pytest.mark.skip(reason="Return value of x.shape is not changed because of the way InterpretNode generate.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -665,7 +665,7 @@ def test_global_getattr_after_setattr_2(): assert obj2.x == 5 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -697,7 +697,7 @@ def test_global_getattr_after_setattr_3(): assert obj.x == 3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -729,7 +729,7 @@ def test_global_getattr_before_setattr(): assert obj.x == 3 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -756,7 +756,7 @@ def test_global_setattr_in_control_flow(): assert ret == -1 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -785,7 +785,7 @@ def test_global_setattr_in_control_flow_2(): assert ret == 4 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -807,7 +807,7 @@ def test_setattr_for_attribute_no_exist(): assert net.x == 4 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -833,7 +833,7 @@ def test_setattr_for_attribute_no_exist_2(): assert obj.y == 2 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -925,7 +925,7 @@ def test_getattr_assign(class_type_choice): assert res2 == 2 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/test_graph_fallback_runtime_str_format.py b/tests/st/fallback/test_graph_fallback_runtime_str_format.py index 34a2d488ddb..8fef204fb29 100644 --- a/tests/st/fallback/test_graph_fallback_runtime_str_format.py +++ b/tests/st/fallback/test_graph_fallback_runtime_str_format.py @@ -44,7 +44,7 @@ def test_str_format_single_input(): assert foo(x) == "string is {}".format(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -65,7 +65,7 @@ def test_str_format_mutiple_input(): assert foo(x) == "{} is {}".format("string", x + x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -89,7 +89,7 @@ def test_fallback_str_format_input(): assert ms_str2 == "[1] " -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -113,7 +113,7 @@ def test_format_with_number_placeholder_input(): assert ms_str == "{1} {0} {1}".format(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -139,7 +139,7 @@ def test_format_with_key_input(): assert result_st == "hello {name2},It's me, {name1}".format(name2=x, name1=y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -164,7 +164,7 @@ def test_format_with_list_index(): assert result_st == "hello {0[1]},It's me {0[0]}".format([x, y]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -190,7 +190,7 @@ def test_format_with_tuple_index(): @pytest.mark.skip("need make dict do not eliminate before opt A.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -220,7 +220,7 @@ def test_format_with_map(): assert "tuple indices must be integers or slices, not str." in str(err.value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -246,7 +246,7 @@ def test_format_as_function(): assert result_st == func([x, y]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -283,7 +283,7 @@ def test_format_number(): assert result_str == correct_str -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -315,7 +315,7 @@ def test_format_padding(): assert result_str == correct_str -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -346,7 +346,7 @@ def test_str_format_using_cell(): assert "Unsupported parameter type for python primitive, the parameter value is DeadNode" in str(err.value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/test_graph_fallback_runtime_tuple_func.py b/tests/st/fallback/test_graph_fallback_runtime_tuple_func.py index d7e3357000f..fdf217f11c7 100644 --- a/tests/st/fallback/test_graph_fallback_runtime_tuple_func.py +++ b/tests/st/fallback/test_graph_fallback_runtime_tuple_func.py @@ -24,7 +24,7 @@ from mindspore.ops.composite import GradOperation context.set_context(mode=context.GRAPH_MODE) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -82,7 +82,7 @@ def test_sequence_in_with_irregular_sequence_3(): assert not foo(Tensor([1, 1, 1])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -140,7 +140,7 @@ def test_sequence_in_with_nested_sequence_3(): assert foo(Tensor([1, 2, 3])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -209,7 +209,7 @@ def test_sequence_mul_with_irregular_sequence_3(): @pytest.mark.skip(reason="dynamic length sequence output format failed") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -231,7 +231,7 @@ def test_sequence_mul_with_nested_sequence(): @pytest.mark.skip(reason="dynamic length sequence output format failed") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -253,7 +253,7 @@ def test_sequence_mul_with_nested_sequence_2(): @pytest.mark.skip(reason="Tuple output with AbstractSequence can not be used in operator") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -275,7 +275,7 @@ def test_sequence_mul_used_in_operator(): assert ret == Tensor([5]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_ascend_training @@ -301,7 +301,7 @@ def test_ops_with_sequence_of_any_input(): assert ret == Tensor([1]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_ascend_training @@ -448,7 +448,7 @@ def test_sequence_compare_with_operation_5(): @pytest.mark.skip(reason="dynamic length sequence output format failed") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -468,7 +468,7 @@ def test_sequence_len_with_operation(): assert ret == 2 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -588,7 +588,7 @@ def test_sequence_count_with_operation_6(): assert ret == 2 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -776,7 +776,7 @@ def test_sequence_getitem_with_tensor_index_3(): assert ret2 == Tensor([2]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -836,7 +836,7 @@ def test_sequence_getitem_with_index_2(): assert np.all(ret[3] == np.array([1, 2, 3, 4])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -887,7 +887,7 @@ def test_sequence_getitem_with_slice_2(): assert np.all(ret[1][1].asnumpy() == np.array([5, 4, 3, 2])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -949,7 +949,7 @@ def test_sequence_ops_with_grad_3(): @pytest.mark.skip(reason="PyExecuteGradient with AbstractAny extra input") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -973,7 +973,7 @@ def test_sequence_ops_with_grad_4(): assert grad1 == grad2 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/test_graph_fallback_side_effect.py b/tests/st/fallback/test_graph_fallback_side_effect.py index 0dd4ff44c7d..c2fcf1f9822 100644 --- a/tests/st/fallback/test_graph_fallback_side_effect.py +++ b/tests/st/fallback/test_graph_fallback_side_effect.py @@ -406,7 +406,7 @@ def test_if_after_for_in_if_numpy(): assert (output.asnumpy() == [3, 4]).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -446,7 +446,7 @@ def test_fallback_save_load(): pass -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -500,7 +500,7 @@ class TransformerNet(nn.Cell): return dense_output.asnumpy(), Tensor(np.array([1, 2])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/fallback/test_graph_fallback_unsupport.py b/tests/st/fallback/test_graph_fallback_unsupport.py index bceca50240a..bfe1ddd50f8 100644 --- a/tests/st/fallback/test_graph_fallback_unsupport.py +++ b/tests/st/fallback/test_graph_fallback_unsupport.py @@ -153,7 +153,7 @@ def test_sequence_in_sequence(): @pytest.mark.skip(reason="not support now") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -175,7 +175,7 @@ def test_all_with_variable(): @pytest.mark.skip(reason="not support now") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -197,7 +197,7 @@ def test_slice_with_variable(): @pytest.mark.skip(reason="not support now") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -260,7 +260,7 @@ def test_math_ceil_with_variable(): @pytest.mark.skip(reason="not support now") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -382,7 +382,7 @@ def test_numpy_prod_with_variable_axis(): @pytest.mark.skip(reason="not support now") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -406,7 +406,7 @@ def test_for_with_interpret_object(): @pytest.mark.skip(reason="not support now") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/ge/test_ge_call.py b/tests/st/ge/test_ge_call.py index d55aa16b7e3..d6b5c53ca1e 100644 --- a/tests/st/ge/test_ge_call.py +++ b/tests/st/ge/test_ge_call.py @@ -15,7 +15,7 @@ import pytest from ge_test_utils import run_testcase -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend910b_training diff --git a/tests/st/ge/test_ge_dynamic_input.py b/tests/st/ge/test_ge_dynamic_input.py index 955ad0f6c3a..4a8c1036c32 100644 --- a/tests/st/ge/test_ge_dynamic_input.py +++ b/tests/st/ge/test_ge_dynamic_input.py @@ -16,7 +16,7 @@ import pytest from ge_test_utils import run_testcase -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend910b_training diff --git a/tests/st/ge/test_ge_jit_level.py b/tests/st/ge/test_ge_jit_level.py index 8db6e4fd578..28567c7e2ce 100644 --- a/tests/st/ge/test_ge_jit_level.py +++ b/tests/st/ge/test_ge_jit_level.py @@ -15,7 +15,7 @@ import pytest from ge_test_utils import run_testcase -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend910b_training diff --git a/tests/st/ge/test_ge_nested_tuple.py b/tests/st/ge/test_ge_nested_tuple.py index 71213b27c40..cd96d68ff3f 100644 --- a/tests/st/ge/test_ge_nested_tuple.py +++ b/tests/st/ge/test_ge_nested_tuple.py @@ -15,7 +15,7 @@ import pytest from ge_test_utils import run_testcase -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend910b_training diff --git a/tests/st/ge/test_if_by_if.py b/tests/st/ge/test_if_by_if.py index dda3b3ce7e3..c0149f88662 100644 --- a/tests/st/ge/test_if_by_if.py +++ b/tests/st/ge/test_if_by_if.py @@ -16,7 +16,7 @@ import os import pytest -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_arm_ascend910b_training diff --git a/tests/st/ge/test_while_by_cell_list_in_while.py b/tests/st/ge/test_while_by_cell_list_in_while.py index 715b15b917a..f09e8238e1f 100644 --- a/tests/st/ge/test_while_by_cell_list_in_while.py +++ b/tests/st/ge/test_while_by_cell_list_in_while.py @@ -16,7 +16,7 @@ import os import pytest -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_arm_ascend910b_training diff --git a/tests/st/gradient/test_grad_graph.py b/tests/st/gradient/test_grad_graph.py index aa8b5efae92..9403533368d 100644 --- a/tests/st/gradient/test_grad_graph.py +++ b/tests/st/gradient/test_grad_graph.py @@ -722,7 +722,7 @@ def test_value_and_grad_nest_with_weights_graph(): assert np.allclose(gradient[1][1].asnumpy(), expect_grad_weight2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/graph_syntax/dict/test_dict_comp.py b/tests/st/graph_syntax/dict/test_dict_comp.py index 5e58b05db85..ab9ae203c5b 100644 --- a/tests/st/graph_syntax/dict/test_dict_comp.py +++ b/tests/st/graph_syntax/dict/test_dict_comp.py @@ -20,7 +20,7 @@ from mindspore import context, jit context.set_context(mode=context.GRAPH_MODE) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/graph_syntax/operators/test_bitwise_operator.py b/tests/st/graph_syntax/operators/test_bitwise_operator.py index 1985fb39da4..e91f5f9fc03 100644 --- a/tests/st/graph_syntax/operators/test_bitwise_operator.py +++ b/tests/st/graph_syntax/operators/test_bitwise_operator.py @@ -154,7 +154,7 @@ def test_bitwise_or_3(): assert np.allclose(result.asnumpy(), np.array([1, 3, -3])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard diff --git a/tests/st/graph_syntax/python_builtin_functions/test_all_any.py b/tests/st/graph_syntax/python_builtin_functions/test_all_any.py index f6ce857155f..8af34a952f9 100644 --- a/tests/st/graph_syntax/python_builtin_functions/test_all_any.py +++ b/tests/st/graph_syntax/python_builtin_functions/test_all_any.py @@ -43,7 +43,7 @@ def test_fallback_all_tensor(): assert (not out1) and out2 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/graph_syntax/python_builtin_functions/test_dynamic_sequence_list_comp.py b/tests/st/graph_syntax/python_builtin_functions/test_dynamic_sequence_list_comp.py index 5eb1b905d32..b7b2b045422 100644 --- a/tests/st/graph_syntax/python_builtin_functions/test_dynamic_sequence_list_comp.py +++ b/tests/st/graph_syntax/python_builtin_functions/test_dynamic_sequence_list_comp.py @@ -22,7 +22,7 @@ context.set_context(mode=context.GRAPH_MODE) @pytest.mark.skip(reason='temporarily skip this case to pass ci') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/graph_syntax/python_builtin_functions/test_len.py b/tests/st/graph_syntax/python_builtin_functions/test_len.py index 7e1a171ecf5..c7b3d7deff2 100644 --- a/tests/st/graph_syntax/python_builtin_functions/test_len.py +++ b/tests/st/graph_syntax/python_builtin_functions/test_len.py @@ -62,7 +62,7 @@ def test_len_list_with_variable(): assert out == 14 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/graph_syntax/test_graph_joinedstr.py b/tests/st/graph_syntax/test_graph_joinedstr.py index 814cc9de6d7..d1c093390bd 100644 --- a/tests/st/graph_syntax/test_graph_joinedstr.py +++ b/tests/st/graph_syntax/test_graph_joinedstr.py @@ -116,7 +116,7 @@ def test_joinedstr_out_tensor(): @pytest.mark.skip("the print output when I6WM9U has been resolved." "otherwise the print may cause sync stream error because the data has not been sync to device") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/mix_precision/test_mix_precision.py b/tests/st/mix_precision/test_mix_precision.py index 8b1273419ca..2ea311a933c 100644 --- a/tests/st/mix_precision/test_mix_precision.py +++ b/tests/st/mix_precision/test_mix_precision.py @@ -152,7 +152,7 @@ def test_sit_auto_mix_precision_model_o0(): clean_all_ir_files('./test_amp_o0/') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -240,7 +240,7 @@ def test_sit_auto_mix_precision_model_o1(): allclose_nparray(out_graph.asnumpy(), out_pynative.asnumpy(), 0.001, 0.001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/mix_precision/test_mix_precision_func.py b/tests/st/mix_precision/test_mix_precision_func.py index c40ccaf8a1f..4a75d328714 100644 --- a/tests/st/mix_precision/test_mix_precision_func.py +++ b/tests/st/mix_precision/test_mix_precision_func.py @@ -48,7 +48,7 @@ def test_static_loss_scaler(mode): assert (unscaled_grads[1] == grads[1] / 1024.).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/model_zoo_tests/resnet50/test_resnet50_cifar10.py b/tests/st/model_zoo_tests/resnet50/test_resnet50_cifar10.py index 26f1ba06512..debbfbc261c 100644 --- a/tests/st/model_zoo_tests/resnet50/test_resnet50_cifar10.py +++ b/tests/st/model_zoo_tests/resnet50/test_resnet50_cifar10.py @@ -131,7 +131,7 @@ def test_resnet50_cifar10_gpu_accuracy(): print("loss_list is", loss_list) assert sum(loss_list) / len(loss_list) < 0.70 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard def test_resnet50_cifar10_gpu_performance(): diff --git a/tests/st/nccl/test_nccl_all.py b/tests/st/nccl/test_nccl_all.py index 1eaa7df4332..f64f786b573 100644 --- a/tests/st/nccl/test_nccl_all.py +++ b/tests/st/nccl/test_nccl_all.py @@ -54,14 +54,14 @@ def test_nccl_broadcast_op(): return_code = os.system("mpirun -n 8 pytest -s test_nccl_broadcast_op.py") assert return_code == 0 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_single def test_nccl_send_recv_op(): return_code = os.system("mpirun -n 8 pytest -s test_nccl_send_recv_op.py") assert return_code == 0 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_single def test_nccl_all_to_all_op(): diff --git a/tests/st/networks/models/bert/bert_performance/test_bert_tdt_ge_run.py b/tests/st/networks/models/bert/bert_performance/test_bert_tdt_ge_run.py index 3a11ce869f5..ede8eaab9f3 100644 --- a/tests/st/networks/models/bert/bert_performance/test_bert_tdt_ge_run.py +++ b/tests/st/networks/models/bert/bert_performance/test_bert_tdt_ge_run.py @@ -19,7 +19,7 @@ import pytest import numpy as np -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend910b_training diff --git a/tests/st/nn/test_BiDense.py b/tests/st/nn/test_BiDense.py index 0e710a2d696..a429dd69f26 100644 --- a/tests/st/nn/test_BiDense.py +++ b/tests/st/nn/test_BiDense.py @@ -31,7 +31,7 @@ class Net(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/nn/test_EmbeddingLookup.py b/tests/st/nn/test_EmbeddingLookup.py index 00e5de88977..7bd6af27707 100644 --- a/tests/st/nn/test_EmbeddingLookup.py +++ b/tests/st/nn/test_EmbeddingLookup.py @@ -31,7 +31,7 @@ class Net(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/nn/test_GroupNorm.py b/tests/st/nn/test_GroupNorm.py index 3c5804d76fe..b4c4fd71ff0 100644 --- a/tests/st/nn/test_GroupNorm.py +++ b/tests/st/nn/test_GroupNorm.py @@ -31,7 +31,7 @@ class Net(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/nn/test_LayerNorm.py b/tests/st/nn/test_LayerNorm.py index ecbcc74b3e3..fe6225b9cd3 100644 --- a/tests/st/nn/test_LayerNorm.py +++ b/tests/st/nn/test_LayerNorm.py @@ -33,7 +33,7 @@ class Net(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/nn/test_MultiFieldEmbeddingLookup.py b/tests/st/nn/test_MultiFieldEmbeddingLookup.py index 18398011c8e..cb124b7676d 100644 --- a/tests/st/nn/test_MultiFieldEmbeddingLookup.py +++ b/tests/st/nn/test_MultiFieldEmbeddingLookup.py @@ -32,7 +32,7 @@ class Net(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/nn/test_cell.py b/tests/st/nn/test_cell.py index 920b8f120ee..bf54775ff4a 100644 --- a/tests/st/nn/test_cell.py +++ b/tests/st/nn/test_cell.py @@ -28,7 +28,7 @@ class Net(nn.Cell): self.b = pt -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -57,7 +57,7 @@ class AttrNet(nn.Cell): self.a = ms.Tensor(2.) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/nn/test_greater.py b/tests/st/nn/test_greater.py index c7ddf4183dc..a285a751f9d 100644 --- a/tests/st/nn/test_greater.py +++ b/tests/st/nn/test_greater.py @@ -30,7 +30,7 @@ def greater_backward_func(x, y): return ops.grad(greater_forward_func, (0,))(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -48,7 +48,7 @@ def test_greater_forward(): assert np.allclose(out.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -66,7 +66,7 @@ def test_greater_backward(): assert np.allclose(grads.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/nn/test_greater_equal.py b/tests/st/nn/test_greater_equal.py index b6601f78936..48ab7d9711d 100644 --- a/tests/st/nn/test_greater_equal.py +++ b/tests/st/nn/test_greater_equal.py @@ -30,7 +30,7 @@ def greater_equal_backward_func(x, y): return ops.grad(greater_equal_forward_func, (0,))(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -48,7 +48,7 @@ def test_greater_equal_forward(): assert np.allclose(out.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @@ -66,7 +66,7 @@ def test_greater_equal_backward(): assert np.allclose(grads.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/nn/test_grucell.py b/tests/st/nn/test_grucell.py index 9679b8430f2..123ede677a2 100644 --- a/tests/st/nn/test_grucell.py +++ b/tests/st/nn/test_grucell.py @@ -30,7 +30,7 @@ class Net(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/nn/test_lstmcell.py b/tests/st/nn/test_lstmcell.py index cd6e0683a45..77968209a56 100644 --- a/tests/st/nn/test_lstmcell.py +++ b/tests/st/nn/test_lstmcell.py @@ -30,7 +30,7 @@ class Net(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/nn/test_rnn.py b/tests/st/nn/test_rnn.py index e10272674d6..edb1c3611c3 100644 --- a/tests/st/nn/test_rnn.py +++ b/tests/st/nn/test_rnn.py @@ -30,7 +30,7 @@ class Net(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/nn/test_rnncell.py b/tests/st/nn/test_rnncell.py index 88684040194..546e0f731e4 100644 --- a/tests/st/nn/test_rnncell.py +++ b/tests/st/nn/test_rnncell.py @@ -30,7 +30,7 @@ class Net(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/nn/test_trainonestepcell_with_return_grads.py b/tests/st/nn/test_trainonestepcell_with_return_grads.py index 931dfde3b15..73d034f6d1f 100644 --- a/tests/st/nn/test_trainonestepcell_with_return_grads.py +++ b/tests/st/nn/test_trainonestepcell_with_return_grads.py @@ -34,7 +34,7 @@ class NetAdam(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/nn/test_transformer.py b/tests/st/nn/test_transformer.py index a9ff943b778..0398b472084 100644 --- a/tests/st/nn/test_transformer.py +++ b/tests/st/nn/test_transformer.py @@ -187,7 +187,7 @@ def test_transformerencoder_square_input_ascend(dtype, training, jit): np.allclose(result.asnumpy(), ref_output.asnumpy(), rtol=1e-7, atol=1e-5) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [ms.float16, ms.float32]) @@ -236,7 +236,7 @@ def test_transformerencoder_square_input_gpu(dtype, training, jit): np.allclose(result.asnumpy(), ref_output.asnumpy(), rtol=1e-7, atol=1e-5) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [ms.float16, ms.float32]) @@ -370,7 +370,7 @@ def test_transformer_cpu(dtype, training, jit): assert result.shape == tgt.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [ms.float16, ms.float32]) diff --git a/tests/st/numpy_native/test_array_ops.py b/tests/st/numpy_native/test_array_ops.py index 13f1932e012..fe6e163d6dc 100644 --- a/tests/st/numpy_native/test_array_ops.py +++ b/tests/st/numpy_native/test_array_ops.py @@ -1376,7 +1376,7 @@ def test_tensor_reshape(): _pynative_executor.sync() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/ascend/test_add.py b/tests/st/ops/ascend/test_add.py index 3dcdc3297fe..5f8ec85ced2 100644 --- a/tests/st/ops/ascend/test_add.py +++ b/tests/st/ops/ascend/test_add.py @@ -82,7 +82,7 @@ def test_add_float32_tensor_api(): -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_add_bfloat16_tensor_api(): diff --git a/tests/st/ops/ascend/test_addn.py b/tests/st/ops/ascend/test_addn.py index 29592c08682..62a8b54ccea 100644 --- a/tests/st/ops/ascend/test_addn.py +++ b/tests/st/ops/ascend/test_addn.py @@ -58,7 +58,7 @@ def test_grad_addn_with_list(): grad_op(AddN())(inp, inp) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.PYNATIVE_MODE, context.GRAPH_MODE]) diff --git a/tests/st/ops/ascend/test_equal_op.py b/tests/st/ops/ascend/test_equal_op.py index 7a5c1343174..584bd8603f1 100644 --- a/tests/st/ops/ascend/test_equal_op.py +++ b/tests/st/ops/ascend/test_equal_op.py @@ -61,7 +61,7 @@ def test_equal_tensor_modes(): test_equal_tensor_api() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_f_equal_api_modes(): diff --git a/tests/st/ops/ascend/test_exp_op.py b/tests/st/ops/ascend/test_exp_op.py index 632e20650e6..a880a24d976 100644 --- a/tests/st/ops/ascend/test_exp_op.py +++ b/tests/st/ops/ascend/test_exp_op.py @@ -31,7 +31,7 @@ def test_exp_tensor_api(ms_type): assert np.allclose(output, expected, rtol=0.004, atol=0.004) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_exp_tensor_bf16_api(): diff --git a/tests/st/ops/ascend/test_flash_attention_score.py b/tests/st/ops/ascend/test_flash_attention_score.py index 89654f82db9..829f64b0971 100644 --- a/tests/st/ops/ascend/test_flash_attention_score.py +++ b/tests/st/ops/ascend/test_flash_attention_score.py @@ -71,7 +71,7 @@ def test_flash_attention_score_fwd(): assert softmax_sum.shape == (B, N, S, 8) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_flash_attention_score_fwd_bwd(): diff --git a/tests/st/ops/ascend/test_ge_op.py b/tests/st/ops/ascend/test_ge_op.py index aadd5d49ef6..6fa0983f6d3 100644 --- a/tests/st/ops/ascend/test_ge_op.py +++ b/tests/st/ops/ascend/test_ge_op.py @@ -31,7 +31,7 @@ def test_f_ge_api(): np.testing.assert_array_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_f_ge_api_modes(): diff --git a/tests/st/ops/ascend/test_less_equal_op.py b/tests/st/ops/ascend/test_less_equal_op.py index 8fba64e1733..294dd76980b 100644 --- a/tests/st/ops/ascend/test_less_equal_op.py +++ b/tests/st/ops/ascend/test_less_equal_op.py @@ -77,7 +77,7 @@ def test_less_equal_functional_tensor_modes(): test_less_equal_tensor_api() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_less_equal_functional_tensor_modes_910b(): diff --git a/tests/st/ops/ascend/test_maximum_op.py b/tests/st/ops/ascend/test_maximum_op.py index 31f240a7c44..22a1b4a9216 100644 --- a/tests/st/ops/ascend/test_maximum_op.py +++ b/tests/st/ops/ascend/test_maximum_op.py @@ -41,7 +41,7 @@ def test_maximum_tensor_api_modes(mode): np.testing.assert_array_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_maximum_tensor_with_bfloat16(): diff --git a/tests/st/ops/ascend/test_minimum_op.py b/tests/st/ops/ascend/test_minimum_op.py index fb10feced1f..54fbe9874f1 100644 --- a/tests/st/ops/ascend/test_minimum_op.py +++ b/tests/st/ops/ascend/test_minimum_op.py @@ -57,7 +57,7 @@ def test_minimum_two_tensors_tensor_dynamic(): assert np.all(output.asnumpy() == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_min_tensor_with_bfloat16(): diff --git a/tests/st/ops/ascend/test_mul_op.py b/tests/st/ops/ascend/test_mul_op.py index 6072f566948..20275b92e35 100644 --- a/tests/st/ops/ascend/test_mul_op.py +++ b/tests/st/ops/ascend/test_mul_op.py @@ -40,7 +40,7 @@ def test_mul_tensor_api_modes(mode): np.testing.assert_array_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) diff --git a/tests/st/ops/ascend/test_tbe_ops/test_ReduceMean.py b/tests/st/ops/ascend/test_tbe_ops/test_ReduceMean.py index f6d81c06a35..b9540615a3b 100644 --- a/tests/st/ops/ascend/test_tbe_ops/test_ReduceMean.py +++ b/tests/st/ops/ascend/test_tbe_ops/test_ReduceMean.py @@ -65,7 +65,7 @@ def test_net(): print(output.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_net_bf16(): diff --git a/tests/st/ops/ascend/test_tbe_ops/test_fast_gelu.py b/tests/st/ops/ascend/test_tbe_ops/test_fast_gelu.py index 734c496c62a..ce7197eed8e 100644 --- a/tests/st/ops/ascend/test_tbe_ops/test_fast_gelu.py +++ b/tests/st/ops/ascend/test_tbe_ops/test_fast_gelu.py @@ -126,7 +126,7 @@ def test_fast_gelu_input_dim_20480_1024(): input_shape = [20480, 1024] fast_gelu_forward_cmp(input_shape) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_fast_gelu_input_dim_20480_1024_bf16(): diff --git a/tests/st/ops/ascend/test_tbe_ops/test_fast_gelu_grad_sens.py b/tests/st/ops/ascend/test_tbe_ops/test_fast_gelu_grad_sens.py index d97715b82b2..7009f972576 100755 --- a/tests/st/ops/ascend/test_tbe_ops/test_fast_gelu_grad_sens.py +++ b/tests/st/ops/ascend/test_tbe_ops/test_fast_gelu_grad_sens.py @@ -95,7 +95,7 @@ def test_grad_fast_gelu_input_10240_1024(): input_shape = [10240, 1024] fast_gelu_backward_cmp(input_shape) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_grad_fast_gelu_input_10240_1024_bf16(): diff --git a/tests/st/ops/ascend/test_tbe_ops/test_maximum_grad.py b/tests/st/ops/ascend/test_tbe_ops/test_maximum_grad.py index a7843e3f222..03623268509 100644 --- a/tests/st/ops/ascend/test_tbe_ops/test_maximum_grad.py +++ b/tests/st/ops/ascend/test_tbe_ops/test_maximum_grad.py @@ -104,7 +104,7 @@ def test_max_tensor_grad_with_same_input(): gen_data(inputA_np, inputB_np, grad_, ms_type=ms.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_max_tensor_grad_with_bfloat16(): @@ -119,7 +119,7 @@ def test_max_tensor_grad_with_bfloat16(): gen_data(inputA_np, inputB_np, grad_, ms_type=ms.bfloat16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_max_tensor_grad_with_bf16(): diff --git a/tests/st/ops/ascend/test_tbe_ops/test_minimum_grad.py b/tests/st/ops/ascend/test_tbe_ops/test_minimum_grad.py index f33e439bd88..ebb31c1642b 100644 --- a/tests/st/ops/ascend/test_tbe_ops/test_minimum_grad.py +++ b/tests/st/ops/ascend/test_tbe_ops/test_minimum_grad.py @@ -103,7 +103,7 @@ def test_min_tensor_grad_with_same_input(): gen_data(inputA_np, inputB_np, grad_, ms_type=ms.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_min_tensor_grad_with_bfloat16(): @@ -118,7 +118,7 @@ def test_min_tensor_grad_with_bfloat16(): gen_data(inputA_np, inputB_np, grad_, ms_type=ms.bfloat16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_min_tensor_grad_with_bf16(): diff --git a/tests/st/ops/ascend/test_tbe_ops/test_select.py b/tests/st/ops/ascend/test_tbe_ops/test_select.py index 13c7539cd61..6d88ab85116 100644 --- a/tests/st/ops/ascend/test_tbe_ops/test_select.py +++ b/tests/st/ops/ascend/test_tbe_ops/test_select.py @@ -100,7 +100,7 @@ def test_functional_select_scalar(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_functional_select_bf16(): diff --git a/tests/st/ops/ascend/test_tbe_ops/test_stridedslice.py b/tests/st/ops/ascend/test_tbe_ops/test_stridedslice.py index d4eb786b79e..e66fadf83ec 100644 --- a/tests/st/ops/ascend/test_tbe_ops/test_stridedslice.py +++ b/tests/st/ops/ascend/test_tbe_ops/test_stridedslice.py @@ -69,7 +69,7 @@ def test_stridedslice_input_3d(): stride = (1, 1, 1) me_stridedslice(input_, begin, end, stride) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_stridedslice_input_3d_bf16(): diff --git a/tests/st/ops/ascend/test_tbe_ops/test_tensor_move.py b/tests/st/ops/ascend/test_tbe_ops/test_tensor_move.py index ec7b01fe9ee..3c32c0b9adc 100644 --- a/tests/st/ops/ascend/test_tbe_ops/test_tensor_move.py +++ b/tests/st/ops/ascend/test_tbe_ops/test_tensor_move.py @@ -21,7 +21,7 @@ import pytest tensor_move = Primitive('TensorMove') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend910b_training @pytest.mark.env_onecard def test_tensor_move(): diff --git a/tests/st/ops/cpu/test_igamma_op.py b/tests/st/ops/cpu/test_igamma_op.py index db4fd8c5a0f..a48f8342f3c 100644 --- a/tests/st/ops/cpu/test_igamma_op.py +++ b/tests/st/ops/cpu/test_igamma_op.py @@ -40,7 +40,7 @@ def test_igamma_functional_api_modes(mode): np.testing.assert_array_almost_equal(output.asnumpy(), expected, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -58,7 +58,7 @@ def test_igamma_tensor_api_modes(mode): np.testing.assert_array_almost_equal(output.asnumpy(), expected, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -76,7 +76,7 @@ def test_igammac_functional_api_modes(mode): np.testing.assert_array_almost_equal(output.asnumpy(), expected, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) diff --git a/tests/st/ops/cpu/test_im2col_op.py b/tests/st/ops/cpu/test_im2col_op.py index fe30e2d4fc3..7fab9deacf6 100644 --- a/tests/st/ops/cpu/test_im2col_op.py +++ b/tests/st/ops/cpu/test_im2col_op.py @@ -26,7 +26,7 @@ from mindspore.ops.functional import vmap context.set_context(mode=context.GRAPH_MODE, device_target="CPU") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_im2col_vmap(): @@ -69,7 +69,7 @@ class NetIm2Col(nn.Cell): return self.im2col(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_im2col_cpu_dynamic_shape(): diff --git a/tests/st/ops/cpu/test_imag_op.py b/tests/st/ops/cpu/test_imag_op.py index 78abe0828f0..0763e0dfe39 100644 --- a/tests/st/ops/cpu/test_imag_op.py +++ b/tests/st/ops/cpu/test_imag_op.py @@ -20,7 +20,7 @@ from mindspore.ops import functional as F from mindspore.common import dtype as mstype -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -37,7 +37,7 @@ def test_imag_functional_api_modes(mode): np.testing.assert_array_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) diff --git a/tests/st/ops/cpu/test_in_top_k_op.py b/tests/st/ops/cpu/test_in_top_k_op.py index 1d91e1ab06d..50d1818593b 100644 --- a/tests/st/ops/cpu/test_in_top_k_op.py +++ b/tests/st/ops/cpu/test_in_top_k_op.py @@ -82,7 +82,7 @@ def in_top_k(nptype): np.testing.assert_array_equal(output.asnumpy(), expected_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_in_top_k_float16(): @@ -94,7 +94,7 @@ def test_in_top_k_float16(): in_top_k(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_in_top_k_float32(): @@ -106,7 +106,7 @@ def test_in_top_k_float32(): in_top_k(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_in_top_k_invalid_input(): diff --git a/tests/st/ops/cpu/test_index_add_op.py b/tests/st/ops/cpu/test_index_add_op.py index c20e96e78e1..8f8a60d5f56 100644 --- a/tests/st/ops/cpu/test_index_add_op.py +++ b/tests/st/ops/cpu/test_index_add_op.py @@ -46,7 +46,7 @@ def index_add_forward(nptype): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_index_add_float64(): @@ -61,7 +61,7 @@ def test_index_add_float64(): index_add_forward(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_index_add_float16(): @@ -76,7 +76,7 @@ def test_index_add_float16(): index_add_forward(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_index_add_int32(): @@ -91,7 +91,7 @@ def test_index_add_int32(): index_add_forward(np.int32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_index_add_int16(): @@ -176,7 +176,7 @@ def index_add_grad_with_type(nptype): np.testing.assert_array_equal(ygrad.asnumpy(), expect_ygrad) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_index_add_grad_float64(): @@ -191,7 +191,7 @@ def test_index_add_grad_float64(): index_add_grad_with_type(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_index_add_grad_float32(): @@ -236,7 +236,7 @@ def test_index_add_grad_int32(): index_add_grad_with_type(np.int32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_index_add_grad_int16(): @@ -281,7 +281,7 @@ def test_index_add_grad_uint8(): index_add_grad_with_type(np.uint8) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_index_add_function(): @@ -299,7 +299,7 @@ def test_index_add_function(): np.testing.assert_array_equal(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_index_add_dynamic(): diff --git a/tests/st/ops/cpu/test_inplace_op.py b/tests/st/ops/cpu/test_inplace_op.py index 5fa3bf0d2be..b2c9e57f1bb 100644 --- a/tests/st/ops/cpu/test_inplace_op.py +++ b/tests/st/ops/cpu/test_inplace_op.py @@ -62,7 +62,7 @@ def inplace_op_np(op, x: np.ndarray, v: np.ndarray, indices): return result -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('shape, indice_len', [((10, 4, 3, 2), 4), ((5, 2, 4, 6), 3)]) @@ -85,7 +85,7 @@ def test_inplace_add(shape, indice_len, dtype): np.allclose(result.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('shape', [(10, 4, 3, 2), (5, 2, 4, 6)]) @@ -107,7 +107,7 @@ def test_inplace_add_same_indice(shape, dtype): np.allclose(result.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('shape, indice', [((10, 4, 3, 2), 4), ((5, 2, 4, 6), 3)]) diff --git a/tests/st/ops/cpu/test_matrix_band_part.py b/tests/st/ops/cpu/test_matrix_band_part.py index 16507de1cd8..0ee1ee518f3 100644 --- a/tests/st/ops/cpu/test_matrix_band_part.py +++ b/tests/st/ops/cpu/test_matrix_band_part.py @@ -61,7 +61,7 @@ def test_matrix_band_part(mode, dtype, batch_shape, rows, cols): np.testing.assert_array_almost_equal(ms_output.asnumpy(), np_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -164,7 +164,7 @@ def test_matrix_band_part_vmap(mode): np.testing.assert_almost_equal(output.asnumpy(), expect_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) diff --git a/tests/st/ops/cpu/test_matrix_diag_part_v3_op.py b/tests/st/ops/cpu/test_matrix_diag_part_v3_op.py index bf8e9116371..3ebf894151d 100644 --- a/tests/st/ops/cpu/test_matrix_diag_part_v3_op.py +++ b/tests/st/ops/cpu/test_matrix_diag_part_v3_op.py @@ -30,7 +30,7 @@ class MatrixDiagPartV3Net(nn.Cell): return self.matrix_diag_dart_v3(x, k, padding_value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_matrix_diag_part_v3_function(): @@ -63,7 +63,7 @@ def test_matrix_diag_part_v3_function(): np.testing.assert_allclose(result, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_matrix_diag_part_v3_function_int_k(): @@ -89,7 +89,7 @@ def test_matrix_diag_part_v3_function_int_k(): np.testing.assert_allclose(result, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_matrix_diag_part_v3_pynative(): @@ -122,7 +122,7 @@ def test_matrix_diag_part_v3_pynative(): np.testing.assert_allclose(result, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_matrix_diag_part_v3_graph(): @@ -155,7 +155,7 @@ def test_matrix_diag_part_v3_graph(): np.testing.assert_allclose(result, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_matrix_diag_part_v3_primitive_negative_k(): @@ -183,7 +183,7 @@ def test_matrix_diag_part_v3_primitive_negative_k(): np.testing.assert_allclose(result, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_matrix_diag_part_v3_vmap(): @@ -216,7 +216,7 @@ def test_matrix_diag_part_v3_vmap(): np.testing.assert_allclose(result, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_matrix_diag_part_v3_dynamic_shape(): diff --git a/tests/st/ops/cpu/test_matrix_diag_v3_op.py b/tests/st/ops/cpu/test_matrix_diag_v3_op.py index 492abb7a3c4..d313b08ad10 100644 --- a/tests/st/ops/cpu/test_matrix_diag_v3_op.py +++ b/tests/st/ops/cpu/test_matrix_diag_v3_op.py @@ -51,7 +51,7 @@ def benchmark(diagonal, expect, align="RIGHT_LEFT", k=None, num_rows=None, num_c np.testing.assert_allclose(output.asnumpy(), expect, rtol=error, atol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize("data_shape", [(8,)]) @@ -68,7 +68,7 @@ def test_matrix_diag_v1(data_shape, data_type): benchmark(diagonal, expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize("data_type", [np.uint32, np.uint64, np.int8, np.int16]) @@ -121,7 +121,7 @@ def test_matrix_diag_v3_1(data_type): benchmark(diagonal, expect, k=k) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize("data_type", [np.uint8, np.uint16, np.float16]) diff --git a/tests/st/ops/cpu/test_matrix_logarithm_op.py b/tests/st/ops/cpu/test_matrix_logarithm_op.py index b7a620887a8..b08823002c3 100644 --- a/tests/st/ops/cpu/test_matrix_logarithm_op.py +++ b/tests/st/ops/cpu/test_matrix_logarithm_op.py @@ -42,7 +42,7 @@ def dyn_case(): assert y.asnumpy().shape == (2, 2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_matrix_logarithm_dyn(): diff --git a/tests/st/ops/cpu/test_matrix_solve.py b/tests/st/ops/cpu/test_matrix_solve.py index cbc4c0764e9..bbe95c975ca 100644 --- a/tests/st/ops/cpu/test_matrix_solve.py +++ b/tests/st/ops/cpu/test_matrix_solve.py @@ -30,7 +30,7 @@ class MatrixSolveNet(nn.Cell): return ops.matrix_solve(matrix, rhs, self.adjoint) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('adjoint', [True, False]) @@ -62,7 +62,7 @@ def test_matrix_solve(adjoint, rhs_shape, dtype, error): assert np.allclose(result, expected, atol=error, rtol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('adjoint', [True, False]) @@ -92,7 +92,7 @@ def test_matrix_solve_complex(adjoint, m, k, dtype, error): assert np.allclose(result, expected, atol=error, rtol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('adjoint', [True, False]) @@ -119,7 +119,7 @@ def test_matrix_solve_vmap(adjoint, dtype, error): assert np.allclose(result, expected, atol=error, rtol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_matrix_solve_dynamic_shape(): diff --git a/tests/st/ops/cpu/test_max_pool_grad_with_argmax_cpu_op.py b/tests/st/ops/cpu/test_max_pool_grad_with_argmax_cpu_op.py index eecaad568f9..a74a573cc8a 100644 --- a/tests/st/ops/cpu/test_max_pool_grad_with_argmax_cpu_op.py +++ b/tests/st/ops/cpu/test_max_pool_grad_with_argmax_cpu_op.py @@ -34,7 +34,7 @@ class MaxPoolGradWithArgmax(nn.Cell): return self.grad(x, grad, argmax) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool_grad_with_argmax(): @@ -73,7 +73,7 @@ def test_maxpool_grad_with_argmax(): assert (actual_output.asnumpy() == expect_result).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool_grad_with_argmax_fp16(): @@ -113,7 +113,7 @@ def test_maxpool_grad_with_argmax_fp16(): assert (actual_output.asnumpy() == expect_result).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool_grad_with_argmax_x_dynamic_shape(): diff --git a/tests/st/ops/cpu/test_max_pool_grad_with_argmax_v2_cpu_op.py b/tests/st/ops/cpu/test_max_pool_grad_with_argmax_v2_cpu_op.py index 3d43a1d5475..ad5953c2cf8 100644 --- a/tests/st/ops/cpu/test_max_pool_grad_with_argmax_v2_cpu_op.py +++ b/tests/st/ops/cpu/test_max_pool_grad_with_argmax_v2_cpu_op.py @@ -36,7 +36,7 @@ class MaxPoolGradWithArgmaxV2(nn.Cell): return self.grad(x, grad, argmax) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool_grad_with_argmax_v2(): @@ -75,7 +75,7 @@ def test_maxpool_grad_with_argmax_v2(): assert (actual_output.asnumpy() == expect_result).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool_grad_with_argmax_v2_fp16(): @@ -115,7 +115,7 @@ def test_maxpool_grad_with_argmax_v2_fp16(): assert (actual_output.asnumpy() == expect_result).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool_grad_with_argmax_v2_vmap(): @@ -141,7 +141,7 @@ def test_maxpool_grad_with_argmax_v2_vmap(): assert actual_output.shape == (1, 1, 1, 6, 6) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool_grad_with_argmax_v2_x_dynamic_shape(): diff --git a/tests/st/ops/cpu/test_max_pool_with_argmax_cpu_op.py b/tests/st/ops/cpu/test_max_pool_with_argmax_cpu_op.py index 2b1a566fb65..12fd78f1b0a 100644 --- a/tests/st/ops/cpu/test_max_pool_with_argmax_cpu_op.py +++ b/tests/st/ops/cpu/test_max_pool_with_argmax_cpu_op.py @@ -36,7 +36,7 @@ class MaxPoolWithArgmaxOp(nn.Cell): return self.max_pool_op(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool_with_argmax_valid(): @@ -63,7 +63,7 @@ def test_maxpool_with_argmax_valid(): assert (argmax.asnumpy() == expect_argmax).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool_with_argmax_same(): @@ -90,7 +90,7 @@ def test_maxpool_with_argmax_same(): assert (argmax.asnumpy() == expect_argmax).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool_with_argmax_tensorflow(): @@ -134,7 +134,7 @@ def test_maxpool_with_argmax_tensorflow(): assert np.allclose(actual_output.asnumpy(), expect_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool_with_argmax_dynamic_shape(): diff --git a/tests/st/ops/cpu/test_max_pool_with_argmax_v2_cpu_op.py b/tests/st/ops/cpu/test_max_pool_with_argmax_v2_cpu_op.py index ca48b6b4418..dc662039305 100644 --- a/tests/st/ops/cpu/test_max_pool_with_argmax_v2_cpu_op.py +++ b/tests/st/ops/cpu/test_max_pool_with_argmax_v2_cpu_op.py @@ -37,7 +37,7 @@ class MaxPoolWithArgmaxV2Net(Cell): return output, argmax -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool_with_argmax_v2_vmap(): diff --git a/tests/st/ops/cpu/test_maximum_grad.py b/tests/st/ops/cpu/test_maximum_grad.py index 91c64d6ca82..920154502a5 100644 --- a/tests/st/ops/cpu/test_maximum_grad.py +++ b/tests/st/ops/cpu/test_maximum_grad.py @@ -43,7 +43,7 @@ class MaxmumGradNet(Cell): return self.maximum_grad(x, y, dy) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maximum_grad_random(): diff --git a/tests/st/ops/cpu/test_maximum_op.py b/tests/st/ops/cpu/test_maximum_op.py index faeb41dae14..f221f66e412 100644 --- a/tests/st/ops/cpu/test_maximum_op.py +++ b/tests/st/ops/cpu/test_maximum_op.py @@ -42,7 +42,7 @@ class TwoTensorsMaximum(Cell): return self.max(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maximum_const_scalar_tensor_int(): @@ -63,7 +63,7 @@ def test_maximum_const_scalar_tensor_int(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maximum_two_tensors_not_broadcast_int(): @@ -85,7 +85,7 @@ def test_maximum_two_tensors_not_broadcast_int(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maximum_two_tensors_broadcast_int(): @@ -107,7 +107,7 @@ def test_maximum_two_tensors_broadcast_int(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maximum_two_tensors_broadcast_one_dimension_int(): @@ -129,7 +129,7 @@ def test_maximum_two_tensors_broadcast_one_dimension_int(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maximum_two_tensors_not_broadcast_all_one_dimension_int(): @@ -151,7 +151,7 @@ def test_maximum_two_tensors_not_broadcast_all_one_dimension_int(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maximum_two_tensors_not_broadcast_float32(): @@ -173,7 +173,7 @@ def test_maximum_two_tensors_not_broadcast_float32(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maximum_two_tensors_not_broadcast_float64(): @@ -195,7 +195,7 @@ def test_maximum_two_tensors_not_broadcast_float64(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) diff --git a/tests/st/ops/cpu/test_maxpool_3d_grad_grad_cpu_op.py b/tests/st/ops/cpu/test_maxpool_3d_grad_grad_cpu_op.py index 28e14132a60..117dbcc7ecb 100644 --- a/tests/st/ops/cpu/test_maxpool_3d_grad_grad_cpu_op.py +++ b/tests/st/ops/cpu/test_maxpool_3d_grad_grad_cpu_op.py @@ -36,7 +36,7 @@ class NetMaxPool3DGradGrad(nn.Cell): return self.maxpool_grad_grad_fun(x, out, grad) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool3d_grad_grad_fp16(): @@ -67,7 +67,7 @@ def test_maxpool3d_grad_grad_fp16(): assert np.allclose(output.asnumpy(), expect_result) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool3d_grad_grad_fp32(): @@ -125,7 +125,7 @@ def test_maxpool3d_grad_grad_fp32(): assert np.allclose(output.asnumpy(), expect_result) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('axis', [2]) diff --git a/tests/st/ops/cpu/test_maxpool_grad_grad_cpu_op.py b/tests/st/ops/cpu/test_maxpool_grad_grad_cpu_op.py index 438803af39d..ac94034ab04 100644 --- a/tests/st/ops/cpu/test_maxpool_grad_grad_cpu_op.py +++ b/tests/st/ops/cpu/test_maxpool_grad_grad_cpu_op.py @@ -36,7 +36,7 @@ class NetMaxPoolGradGrad(nn.Cell): return self.maxpool_grad_grad_fun(x, out, grad) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool2d_grad_grad_fp16(): @@ -66,7 +66,7 @@ def test_maxpool2d_grad_grad_fp16(): assert np.allclose(output.asnumpy(), expect_result) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool2d_grad_grad_fp32(): @@ -105,7 +105,7 @@ def test_maxpool2d_grad_grad_fp32(): assert np.allclose(output.asnumpy(), expect_result) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('axis', [2]) diff --git a/tests/st/ops/cpu/test_maxpool_grad_grad_with_argmax_cpu_op.py b/tests/st/ops/cpu/test_maxpool_grad_grad_with_argmax_cpu_op.py index 2bcf77b7814..85557bba836 100644 --- a/tests/st/ops/cpu/test_maxpool_grad_grad_with_argmax_cpu_op.py +++ b/tests/st/ops/cpu/test_maxpool_grad_grad_with_argmax_cpu_op.py @@ -36,7 +36,7 @@ class NetMaxPoolGradGradWithArgmax(nn.Cell): return self.maxpool_grad_grad_argmax_fun(x, grad, argmax) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize("argmax_type", [np.int32, np.int64]) @@ -67,7 +67,7 @@ def test_maxpool_grad_grad_argmax_fp16(argmax_type): assert np.allclose(output.asnumpy(), expect_result) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize("argmax_type", [np.int32, np.int64]) @@ -108,7 +108,7 @@ def test_maxpool_grad_grad_argmax_fp32(argmax_type): assert np.allclose(output.asnumpy(), expect_result) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('axis', [2]) diff --git a/tests/st/ops/cpu/test_maxpool_grad_op.py b/tests/st/ops/cpu/test_maxpool_grad_op.py index a3af5d99b61..5996f7fba9e 100644 --- a/tests/st/ops/cpu/test_maxpool_grad_op.py +++ b/tests/st/ops/cpu/test_maxpool_grad_op.py @@ -45,7 +45,7 @@ class MaxPoolGrad(nn.Cell): return self.grad(self.forward)(x, sens) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool_grad_vmap(): diff --git a/tests/st/ops/cpu/test_maxpool_op.py b/tests/st/ops/cpu/test_maxpool_op.py index d111261e9f3..f327a1d5486 100644 --- a/tests/st/ops/cpu/test_maxpool_op.py +++ b/tests/st/ops/cpu/test_maxpool_op.py @@ -50,7 +50,7 @@ class MaxPoolGrad(nn.Cell): return self.grad(self.forward)(x, sens) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool2d_valid(): @@ -84,7 +84,7 @@ def test_maxpool2d_valid(): assert (actual_grad[0].asnumpy() == expect_dx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_maxpool2d_same(): @@ -118,7 +118,7 @@ def test_maxpool2d_same(): assert (actual_grad[0].asnumpy() == expect_dx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize("dtype", [np.float32, np.float16, np.float64]) @@ -164,7 +164,7 @@ def test_maxpool3d_1(dtype): assert (actual_grad[0].asnumpy() == expect_dx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize("dtype", [np.float32, np.float16, np.float64]) @@ -210,7 +210,7 @@ def test_maxpool3d_2(dtype): assert (actual_grad[0].asnumpy() == expect_dx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize("dtype", [np.float32, np.float16, np.float64]) @@ -253,7 +253,7 @@ def test_maxpool3d_3(dtype): assert (actual_grad[0].asnumpy() == expect_dx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize("dtype", [np.float32, np.float16, np.float64]) @@ -311,7 +311,7 @@ def test_maxpool3d_4(dtype): assert (actual_grad[0].asnumpy() == expect_dx).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_max_pool2d_vmap(): diff --git a/tests/st/ops/cpu/test_meshgrid_op.py b/tests/st/ops/cpu/test_meshgrid_op.py index e04bad047ca..3d8615daabc 100644 --- a/tests/st/ops/cpu/test_meshgrid_op.py +++ b/tests/st/ops/cpu/test_meshgrid_op.py @@ -37,7 +37,7 @@ class NetMeshgrid(nn.Cell): return self.meshgrid(inputs) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_meshgrid_dshape(): @@ -59,7 +59,7 @@ def test_meshgrid_dshape(): assert output[1].asnumpy().shape == expect_shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', @@ -99,7 +99,7 @@ def test_meshgrid(dtype, indexing): assert np.array_equal(output[2].asnumpy(), np_output[2]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('axis', [2]) @@ -134,7 +134,7 @@ def test_meshgrid_vmap_cpu(axis): assert np.array_equal(output_vmap1.asnumpy(), output_manually1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('axis0', [2]) diff --git a/tests/st/ops/cpu/test_minimum_grad_op.py b/tests/st/ops/cpu/test_minimum_grad_op.py index c675f582e3a..5d4f7203db2 100644 --- a/tests/st/ops/cpu/test_minimum_grad_op.py +++ b/tests/st/ops/cpu/test_minimum_grad_op.py @@ -102,7 +102,7 @@ def test_min_tensor_grad_with_same_input(): assert np.allclose(output[1].asnumpy(), expect1, rtol=1e-6, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_min_tensor_grad_result(): @@ -203,7 +203,7 @@ def test_min_tensor_grad_result(): assert np.all(np.abs(output[1].asnumpy() - expect1) < error1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_min_grad_vmap(): @@ -230,7 +230,7 @@ def test_min_grad_vmap(): assert np.allclose(dy.asnumpy(), expect1, rtol=1e-6, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_min_grad_vmap_none(): @@ -258,7 +258,7 @@ def test_min_grad_vmap_none(): assert np.allclose(dy.asnumpy(), expect1, rtol=1e-6, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_min_grad_vmap_scalar(): @@ -282,7 +282,7 @@ def test_min_grad_vmap_scalar(): assert np.allclose(dy.asnumpy(), expect1, rtol=1e-6, atol=1e-4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_min_grad_vmap_special(): diff --git a/tests/st/ops/cpu/test_minimum_op.py b/tests/st/ops/cpu/test_minimum_op.py index 05ecf9a87cf..816b71b239a 100644 --- a/tests/st/ops/cpu/test_minimum_op.py +++ b/tests/st/ops/cpu/test_minimum_op.py @@ -41,7 +41,7 @@ class TwoTensorsMinimum(Cell): return self.min(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_minimum_constScalar_tensor_int(): @@ -54,7 +54,7 @@ def test_minimum_constScalar_tensor_int(): assert np.all(output.asnumpy() == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_minimum_two_tensors_Not_Broadcast_int(): @@ -69,7 +69,7 @@ def test_minimum_two_tensors_Not_Broadcast_int(): assert np.all(output.asnumpy() == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_minimum_two_tensors_Broadcast_int(): @@ -84,7 +84,7 @@ def test_minimum_two_tensors_Broadcast_int(): assert np.all(output.asnumpy() == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_minimum_two_tensors_Broadcast_oneDimension_int(): @@ -99,7 +99,7 @@ def test_minimum_two_tensors_Broadcast_oneDimension_int(): assert np.all(output.asnumpy() == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_minimum_two_tensors_notBroadcast_all_oneDimension_int(): @@ -113,7 +113,7 @@ def test_minimum_two_tensors_notBroadcast_all_oneDimension_int(): assert np.all(output.asnumpy() == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_minimum_two_tensors_notBroadcast_float32(): @@ -131,7 +131,7 @@ def test_minimum_two_tensors_notBroadcast_float32(): assert output.shape == expect.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_minimum_two_tensors_notBroadcast_float16(): @@ -149,7 +149,7 @@ def test_minimum_two_tensors_notBroadcast_float16(): assert output.shape == expect.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_minimum_two_tensors_Broadcast_float16(): @@ -167,7 +167,7 @@ def test_minimum_two_tensors_Broadcast_float16(): assert output.shape == expect.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_minimum_two_tensors_notBroadcast_float64(): @@ -185,7 +185,7 @@ def test_minimum_two_tensors_notBroadcast_float64(): assert output.shape == expect.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_minimum_two_tensors_functional_int8(): @@ -209,7 +209,7 @@ class MinimumTensorNet(Cell): return x.minimum(y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_minimum_two_tensors_tensor_graph_uint16(): @@ -229,7 +229,7 @@ def test_minimum_two_tensors_tensor_graph_uint16(): assert np.all(output.asnumpy() == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_minimum_two_tensors_tensor_pynative_uint8(): diff --git a/tests/st/ops/cpu/test_mirror_pad.py b/tests/st/ops/cpu/test_mirror_pad.py index 362d38115cb..492fd668c84 100644 --- a/tests/st/ops/cpu/test_mirror_pad.py +++ b/tests/st/ops/cpu/test_mirror_pad.py @@ -24,7 +24,7 @@ from mindspore import Tensor from mindspore.ops.composite import GradOperation -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_mirror_pad(): @@ -75,7 +75,7 @@ class Net(nn.Cell): return self.pad(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_mirror_pad_backprop(): @@ -92,7 +92,7 @@ def test_mirror_pad_backprop(): np.testing.assert_array_almost_equal(dx, expected_dx) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_mirror_pad_fwd_back_4d_int32_reflect(): @@ -132,7 +132,7 @@ def test_mirror_pad_fwd_back_4d_int32_reflect(): np.testing.assert_array_equal(dx_value_expected, dx_value_obtained) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_mirror_pad_fwd_back_4d_int32_symm(): @@ -172,7 +172,7 @@ def test_mirror_pad_fwd_back_4d_int32_symm(): np.testing.assert_array_equal(dx_value_expected, dx_value_obtained) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_mirror_pad_grad_dynamic_shape(): diff --git a/tests/st/ops/cpu/test_mish_grad_op.py b/tests/st/ops/cpu/test_mish_grad_op.py index 93d4ba7f27f..6d8cd1d4666 100644 --- a/tests/st/ops/cpu/test_mish_grad_op.py +++ b/tests/st/ops/cpu/test_mish_grad_op.py @@ -60,7 +60,7 @@ class MishGradDynamicShapeNet(nn.Cell): return self.grad(self.network)(x_unique, dy_unique) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -102,7 +102,7 @@ def test_mish_grad(mode, dtype, tol): assert np.allclose(output[0].asnumpy(), expect, atol=tol, rtol=tol, equal_nan=True) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -166,7 +166,7 @@ def test_mish_grad_vmap(mode): assert np.allclose(output[0].asnumpy(), expect_output, atol=1e-4, rtol=1e-4, equal_nan=True) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) diff --git a/tests/st/ops/cpu/test_mish_op.py b/tests/st/ops/cpu/test_mish_op.py index d095cbefced..6f22c2e1e37 100644 --- a/tests/st/ops/cpu/test_mish_op.py +++ b/tests/st/ops/cpu/test_mish_op.py @@ -56,7 +56,7 @@ def mish_np_bencmark(x): return result -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize("data_shape", [(4,), (3, 4), (4, 5, 7)]) @@ -81,7 +81,7 @@ def test_mish(data_shape, data_type): np.testing.assert_allclose(output.asnumpy(), benchmark_output, rtol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_mish_vmap(): @@ -103,7 +103,7 @@ def test_mish_vmap(): np.testing.assert_allclose(output.asnumpy(), benchmark_output, rtol=error, atol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_mish_dy_shape(): diff --git a/tests/st/ops/cpu/test_mul_op.py b/tests/st/ops/cpu/test_mul_op.py index a4d60a145e7..a8c6dc9dacf 100644 --- a/tests/st/ops/cpu/test_mul_op.py +++ b/tests/st/ops/cpu/test_mul_op.py @@ -35,7 +35,7 @@ class Net(nn.Cell): return self.mul(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_mul(): @@ -86,7 +86,7 @@ def test_mul(): assert out.shape == exp.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_mul_int32(): @@ -137,7 +137,7 @@ def test_mul_int32(): assert out.shape == exp.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) diff --git a/tests/st/ops/cpu/test_multinomial_op.py b/tests/st/ops/cpu/test_multinomial_op.py index 991d83f7bd3..63aacb3072c 100644 --- a/tests/st/ops/cpu/test_multinomial_op.py +++ b/tests/st/ops/cpu/test_multinomial_op.py @@ -37,7 +37,7 @@ class Net(nn.Cell): return C.multinomial(x, self.sample, self.replacement, self.seed) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_multinomial_net(): diff --git a/tests/st/ops/cpu/test_multinomial_with_replacement_op.py b/tests/st/ops/cpu/test_multinomial_with_replacement_op.py index b2149de850d..254d03766e6 100644 --- a/tests/st/ops/cpu/test_multinomial_with_replacement_op.py +++ b/tests/st/ops/cpu/test_multinomial_with_replacement_op.py @@ -34,7 +34,7 @@ class Net(nn.Cell): return self.multinomialwithreplacement(x, seed, offset) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_multinomial_with_replacement_net(): @@ -57,7 +57,7 @@ def test_multinomial_with_replacement_net(): assert out2.asnumpy().shape == (10, 15) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_multinomial_with_replacement_functional(): diff --git a/tests/st/ops/cpu/test_mvlgamma_op.py b/tests/st/ops/cpu/test_mvlgamma_op.py index cfa92965dec..23f40bdbc20 100644 --- a/tests/st/ops/cpu/test_mvlgamma_op.py +++ b/tests/st/ops/cpu/test_mvlgamma_op.py @@ -22,7 +22,7 @@ from mindspore.ops import functional as F from mindspore.common import dtype as mstype -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -39,7 +39,7 @@ def test_mvlgamma_functional_api_modes(mode): np.testing.assert_array_almost_equal(output.asnumpy(), expected, decimal=4) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) diff --git a/tests/st/ops/cpu/test_neg_op.py b/tests/st/ops/cpu/test_neg_op.py index 00afd09e372..54434fda685 100644 --- a/tests/st/ops/cpu/test_neg_op.py +++ b/tests/st/ops/cpu/test_neg_op.py @@ -45,7 +45,7 @@ class Net(nn.Cell): return self.ops(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_net(): @@ -61,7 +61,7 @@ def test_net(): print(output[0].asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) diff --git a/tests/st/ops/cpu/test_nextafter_op.py b/tests/st/ops/cpu/test_nextafter_op.py index 90324adedf9..fe6a9b8026a 100644 --- a/tests/st/ops/cpu/test_nextafter_op.py +++ b/tests/st/ops/cpu/test_nextafter_op.py @@ -64,7 +64,7 @@ def test_nextafter_float64_pynative(): assert out_msp.asnumpy() == out_tf -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_nextafter_cpu_dynamic_shape(): diff --git a/tests/st/ops/cpu/test_nllloss_op.py b/tests/st/ops/cpu/test_nllloss_op.py index 4d1ad719242..d219d03e9ca 100644 --- a/tests/st/ops/cpu/test_nllloss_op.py +++ b/tests/st/ops/cpu/test_nllloss_op.py @@ -56,7 +56,7 @@ target = Tensor(np.array([1, 0, 4]).astype(np.int32)) weight = Tensor(np.array([0.2, 0.3, 0.1, 0.15, 0.25]).astype(np_type)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_NLLLoss_none(): @@ -84,7 +84,7 @@ def test_NLLLoss_none(): assert np.allclose(net(logits, target, weight)[0].asnumpy(), expect_dx) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_NLLLoss_sum(): @@ -112,7 +112,7 @@ def test_NLLLoss_sum(): assert np.allclose(net(logits, target, weight)[0].asnumpy(), expect_dx) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_NLLLoss_mean(): diff --git a/tests/st/ops/cpu/test_nms_with_mask_op.py b/tests/st/ops/cpu/test_nms_with_mask_op.py index 845d657c40b..4a5fde256a8 100644 --- a/tests/st/ops/cpu/test_nms_with_mask_op.py +++ b/tests/st/ops/cpu/test_nms_with_mask_op.py @@ -33,7 +33,7 @@ def runMSRun(op, bbox): return sel_rows, sel_score -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_nms_with_mask_check_order(): @@ -54,7 +54,7 @@ def test_nms_with_mask_check_order(): ms_sorted_scores, np_sorted_scores) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_nms_with_mask_edge_case_1(): @@ -75,7 +75,7 @@ def test_nms_with_mask_edge_case_1(): np.testing.assert_almost_equal(sel_score, expected_score) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_nms_with_mask_edge_case_2(): @@ -92,7 +92,7 @@ def test_nms_with_mask_edge_case_2(): np.testing.assert_almost_equal(sel_score, expected_score) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_nms_with_mask_edge_case_3(): diff --git a/tests/st/ops/cpu/test_no_repeat_ngram.py b/tests/st/ops/cpu/test_no_repeat_ngram.py index 2256b9806ce..7189c55e519 100644 --- a/tests/st/ops/cpu/test_no_repeat_ngram.py +++ b/tests/st/ops/cpu/test_no_repeat_ngram.py @@ -165,7 +165,7 @@ def vmap_nested_case(): np.testing.assert_allclose(output.asnumpy(), fornet_output.asnumpy(), rtol=1e-6) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_vmap_cpu(): @@ -178,7 +178,7 @@ def test_vmap_cpu(): vmap_case() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_vmap_cpu_nested(): diff --git a/tests/st/ops/cpu/test_norm_op.py b/tests/st/ops/cpu/test_norm_op.py index 26c668bb8b9..061922de395 100644 --- a/tests/st/ops/cpu/test_norm_op.py +++ b/tests/st/ops/cpu/test_norm_op.py @@ -40,7 +40,7 @@ class NetNorm(nn.Cell): self.norm_3(indices), self.norm_4(indices)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_norm(): diff --git a/tests/st/ops/cpu/test_notequal_op.py b/tests/st/ops/cpu/test_notequal_op.py index 6ff2a7e7bc4..d6c7b3e6c9c 100644 --- a/tests/st/ops/cpu/test_notequal_op.py +++ b/tests/st/ops/cpu/test_notequal_op.py @@ -33,7 +33,7 @@ class OpNetWrapper(nn.Cell): return self.op(*inputs) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_notequal_int(): @@ -48,7 +48,7 @@ def test_notequal_int(): assert np.allclose(outputs.asnumpy(), (True, False, True)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_notequal_float(): diff --git a/tests/st/ops/cpu/test_one_hot_op.py b/tests/st/ops/cpu/test_one_hot_op.py index 6218836a3c9..80b6d348ebe 100644 --- a/tests/st/ops/cpu/test_one_hot_op.py +++ b/tests/st/ops/cpu/test_one_hot_op.py @@ -44,7 +44,7 @@ class NetOneHot(nn.Cell): self.one_hot_3(indices3), self.one_hot_4(indices4)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_one_hot(): diff --git a/tests/st/ops/cpu/test_onednn_dfx.py b/tests/st/ops/cpu/test_onednn_dfx.py index db9594bff3d..1c47726669e 100644 --- a/tests/st/ops/cpu/test_onednn_dfx.py +++ b/tests/st/ops/cpu/test_onednn_dfx.py @@ -23,7 +23,7 @@ import mindspore.context as context context.set_context(mode=context.GRAPH_MODE, device_target='CPU') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_onednn_dfx_log(): diff --git a/tests/st/ops/cpu/test_oneslike_op.py b/tests/st/ops/cpu/test_oneslike_op.py index eceea590bea..7b0d5a3f85a 100644 --- a/tests/st/ops/cpu/test_oneslike_op.py +++ b/tests/st/ops/cpu/test_oneslike_op.py @@ -32,7 +32,7 @@ class NetOnesLike(nn.Cell): return self.ones_like(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.int32, np.float32, np.float64]) @@ -65,7 +65,7 @@ def test_OnesLike(dtype): assert output1.shape == expect1.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_oneslike_cpu_dynamic_shape(): diff --git a/tests/st/ops/cpu/test_pack_op.py b/tests/st/ops/cpu/test_pack_op.py index bb5f8b4848b..6a3a4614630 100644 --- a/tests/st/ops/cpu/test_pack_op.py +++ b/tests/st/ops/cpu/test_pack_op.py @@ -35,7 +35,7 @@ class PackNet(nn.Cell): return self.stack((x1, x2)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float16, np.float32, np.int32, np.int16, np.uint8, np.bool]) @@ -70,7 +70,7 @@ def test_pack_graph(dtype): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_pack_graph_float32_dynamic_shape(): diff --git a/tests/st/ops/cpu/test_pad.py b/tests/st/ops/cpu/test_pad.py index 6de4a513113..feba5c7eaa6 100644 --- a/tests/st/ops/cpu/test_pad.py +++ b/tests/st/ops/cpu/test_pad.py @@ -23,7 +23,7 @@ from mindspore import Tensor from mindspore.ops.composite import GradOperation -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_pad_basic(): @@ -51,7 +51,7 @@ def test_pad_basic(): np.testing.assert_array_equal(y_test, test_arr_expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_pad_row(): @@ -82,7 +82,7 @@ def test_pad_row(): np.testing.assert_equal(y_test_2[:, :, 3:, :], test_arr_2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_pad_column(): @@ -113,7 +113,7 @@ def test_pad_column(): np.testing.assert_equal(y_test_2[:, :, :, 6:-1], test_arr_2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_pad_3d_pad(): @@ -171,7 +171,7 @@ class Net(nn.Cell): return self.pad(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_pad_3d_backprop(): @@ -210,7 +210,7 @@ def test_pad_3d_backprop(): np.testing.assert_array_equal(dx, expected_dx) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_pad_error_cases(): diff --git a/tests/st/ops/cpu/test_pad_op.py b/tests/st/ops/cpu/test_pad_op.py index 1edba212065..a4c1f39256a 100644 --- a/tests/st/ops/cpu/test_pad_op.py +++ b/tests/st/ops/cpu/test_pad_op.py @@ -43,7 +43,7 @@ def vmap_case(): assert np.allclose(expect1, out_ms1.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_pad_vmap_cpu(): diff --git a/tests/st/ops/cpu/test_pad_v3.py b/tests/st/ops/cpu/test_pad_v3.py index 5a2ef3dbe3e..10c463bb4b7 100644 --- a/tests/st/ops/cpu/test_pad_v3.py +++ b/tests/st/ops/cpu/test_pad_v3.py @@ -38,7 +38,7 @@ class NetDynamic(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_padv3_circular_dynamic_shape_3d(): @@ -64,7 +64,7 @@ def test_padv3_circular_dynamic_shape_3d(): np.testing.assert_almost_equal(expect, out.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_padv3_circular_dynamic_shape_4d(): @@ -89,7 +89,7 @@ def test_padv3_circular_dynamic_shape_4d(): np.testing.assert_almost_equal(expect, out.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_padv3_circular_dynamic_shape_5d(): @@ -115,7 +115,7 @@ def test_padv3_circular_dynamic_shape_5d(): np.testing.assert_almost_equal(expect, out.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_padv3_x_padding_dynamic_shape(): @@ -147,7 +147,7 @@ def test_padv3_x_padding_dynamic_shape(): np.testing.assert_almost_equal(expect, out.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_padv3_x_dynamic_shape(): @@ -178,7 +178,7 @@ def test_padv3_x_dynamic_shape(): np.testing.assert_almost_equal(expect, out.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_padv3_padding_dynamic_shape(): @@ -209,7 +209,7 @@ def test_padv3_padding_dynamic_shape(): np.testing.assert_almost_equal(expect, out.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_padv3_padding_list(): diff --git a/tests/st/ops/cpu/test_pad_v3_grad_op.py b/tests/st/ops/cpu/test_pad_v3_grad_op.py index eda2aafafe3..9114015bd0c 100644 --- a/tests/st/ops/cpu/test_pad_v3_grad_op.py +++ b/tests/st/ops/cpu/test_pad_v3_grad_op.py @@ -31,7 +31,7 @@ class PadV3GradNet(nn.Cell): return self.op(x, paddings) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_padv3grad_circular_3d(): @@ -51,7 +51,7 @@ def test_padv3grad_circular_3d(): np.testing.assert_almost_equal(expect, output.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_padv3grad_circular_4d(): @@ -71,7 +71,7 @@ def test_padv3grad_circular_4d(): np.testing.assert_almost_equal(expect, output.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_padv3grad_circular_5d(): diff --git a/tests/st/ops/cpu/test_padding.py b/tests/st/ops/cpu/test_padding.py index 1779c0595ee..bbf67bd3a49 100644 --- a/tests/st/ops/cpu/test_padding.py +++ b/tests/st/ops/cpu/test_padding.py @@ -43,7 +43,7 @@ class PaddingDynamicShapeNet(nn.Cell): return F.padding(x_unique, pad_dim_size) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -67,7 +67,7 @@ def test_padding(mode, shape, dtype, pad_dim_size): np.testing.assert_array_almost_equal(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -123,7 +123,7 @@ def test_padding_vmap(mode): np.testing.assert_almost_equal(output.asnumpy(), expect_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) diff --git a/tests/st/ops/cpu/test_parameterized_truncated_normal_op.py b/tests/st/ops/cpu/test_parameterized_truncated_normal_op.py index 13004db1eb0..5b874977383 100644 --- a/tests/st/ops/cpu/test_parameterized_truncated_normal_op.py +++ b/tests/st/ops/cpu/test_parameterized_truncated_normal_op.py @@ -31,7 +31,7 @@ class ParameterizedTruncatedNormalTEST(nn.Cell): return self.parameterized_truncated_normal(shape, mean, stdevs, minvals, maxvals) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_parameterized_truncated_normal_op_case1(): @@ -62,7 +62,7 @@ def test_parameterized_truncated_normal_op_case1(): assert not (output0.asnumpy() == output1.asnumpy()).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_parameterized_truncated_normal_op_case2(): diff --git a/tests/st/ops/cpu/test_pdist_op.py b/tests/st/ops/cpu/test_pdist_op.py index 7ebc57c8750..581488a9c0c 100644 --- a/tests/st/ops/cpu/test_pdist_op.py +++ b/tests/st/ops/cpu/test_pdist_op.py @@ -23,7 +23,7 @@ import mindspore.context as context context.set_context(mode=context.GRAPH_MODE, device_target='CPU') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype, eps', [(np.float16, 1.0e-3), (np.float32, 1.0e-6), (np.float64, 1.0e-6)]) @@ -41,7 +41,7 @@ def test_pdist_normal(dtype, eps): assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype, eps', [(np.float16, 1.0e-3), (np.float32, 1.0e-6), (np.float64, 1.0e-6)]) @@ -59,7 +59,7 @@ def test_pdist_zero(dtype, eps): assert np.all(diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype, eps', [(np.float16, 1.0e-3), (np.float32, 1.0e-6), (np.float64, 1.0e-6)]) diff --git a/tests/st/ops/cpu/test_pinv_op.py b/tests/st/ops/cpu/test_pinv_op.py index 7829e2eb081..18d816dabd1 100644 --- a/tests/st/ops/cpu/test_pinv_op.py +++ b/tests/st/ops/cpu/test_pinv_op.py @@ -24,7 +24,7 @@ from mindspore import Tensor context.set_context(mode=context.GRAPH_MODE, device_target='CPU') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float32, np.float64]) diff --git a/tests/st/ops/cpu/test_population_count_op.py b/tests/st/ops/cpu/test_population_count_op.py index 20d754f7a52..1c5717c65e4 100644 --- a/tests/st/ops/cpu/test_population_count_op.py +++ b/tests/st/ops/cpu/test_population_count_op.py @@ -34,7 +34,7 @@ class Net(nn.Cell): return self.population_count(x0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test16_net(): @@ -46,7 +46,7 @@ def test16_net(): assert (output.asnumpy() == expect_x_result).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test8_net(): @@ -58,7 +58,7 @@ def test8_net(): assert (output.asnumpy() == expect_x_result).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_vmap_population_count(): diff --git a/tests/st/ops/cpu/test_pow_op.py b/tests/st/ops/cpu/test_pow_op.py index 0888d8701df..6b31319e621 100644 --- a/tests/st/ops/cpu/test_pow_op.py +++ b/tests/st/ops/cpu/test_pow_op.py @@ -32,7 +32,7 @@ class Net(nn.Cell): return self.ops(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_net(): @@ -82,7 +82,7 @@ def test_net(): assert out.shape == expect.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_tensor_pow_pynative(): @@ -103,7 +103,7 @@ def tensor_pow_func(x, y): return x.pow(y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_tensor_pow_graph(): @@ -118,7 +118,7 @@ def test_tensor_pow_graph(): assert np.all(output.asnumpy() == np.array([1, 4, 27])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_pow_functional(): @@ -133,7 +133,7 @@ def test_pow_functional(): assert np.all(output.asnumpy() == np.array([1, 4, 27])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_pow_vmap(): @@ -149,7 +149,7 @@ def test_pow_vmap(): assert np.all(output.asnumpy() == np.array([[1, 4, 27], [27, 4, 1]])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_pow_dynamic_shape(): @@ -170,7 +170,7 @@ def test_pow_dynamic_shape(): assert output.asnumpy().shape == except_shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_pow_complex(): diff --git a/tests/st/ops/cpu/test_prelu_op.py b/tests/st/ops/cpu/test_prelu_op.py index eff99a0b94c..33b59279039 100644 --- a/tests/st/ops/cpu/test_prelu_op.py +++ b/tests/st/ops/cpu/test_prelu_op.py @@ -63,7 +63,7 @@ context.set_context(device_target="CPU", mode=context.GRAPH_MODE) dtypes = [mstype.float16, mstype.float32] -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_prelu_single_weight(): @@ -87,7 +87,7 @@ def test_prelu_single_weight(): prelu_test(x, weight, expect_forward, expect_dx, expect_dw) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_prelu_multiple_weight(): @@ -133,7 +133,7 @@ def test_prelu_multiple_weight(): prelu_test(x, weight, expect_forward, expect_dx, expect_dw) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_prelu_single_weight_scalar(): @@ -157,7 +157,7 @@ def test_prelu_single_weight_scalar(): prelu_test(x, weight, expect_forward, expect_dx, expect_dw) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_prelu_single_weight_1d(): @@ -181,7 +181,7 @@ def test_prelu_single_weight_1d(): prelu_test(x, weight, expect_forward, expect_dx, expect_dw) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_prelu_single_weight_2d(): @@ -205,7 +205,7 @@ def test_prelu_single_weight_2d(): prelu_test(x, weight, expect_forward, expect_dx, expect_dw) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_prelu_multiple_weight_2d(): diff --git a/tests/st/ops/cpu/test_primitive_cache.py b/tests/st/ops/cpu/test_primitive_cache.py index f3b4fd4b685..2170f7d0932 100644 --- a/tests/st/ops/cpu/test_primitive_cache.py +++ b/tests/st/ops/cpu/test_primitive_cache.py @@ -28,7 +28,7 @@ from mindspore.ops._primitive_cache import _get_cache_prim # pylint: disable=W0235 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_jit_function_run_in_pynative(): @@ -58,7 +58,7 @@ def test_jit_function_run_in_pynative(): np.testing.assert_almost_equal(output.asnumpy(), expect_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_run_pynative_and_then_run_graph(): @@ -88,7 +88,7 @@ def test_run_pynative_and_then_run_graph(): np.testing.assert_almost_equal(output1.asnumpy(), output2.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_continuous_cache(): diff --git a/tests/st/ops/cpu/test_print_op.py b/tests/st/ops/cpu/test_print_op.py index 26ca1ee6b76..2b8eebe0e02 100644 --- a/tests/st/ops/cpu/test_print_op.py +++ b/tests/st/ops/cpu/test_print_op.py @@ -38,7 +38,7 @@ class PrintFunc(nn.Cell): return x -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -57,7 +57,7 @@ def test_print_op_dtype(mode, dtype): net(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -76,7 +76,7 @@ def test_print_op_dynamic_shape(mode): net(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -93,7 +93,7 @@ def test_print_op_functional(mode): net(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_print_op_tuple(): @@ -114,7 +114,7 @@ def test_print_op_tuple(): net(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu diff --git a/tests/st/ops/cpu/test_priority_replay_buffer_test.py b/tests/st/ops/cpu/test_priority_replay_buffer_test.py index 85ccbe6e537..0f56fc2d077 100644 --- a/tests/st/ops/cpu/test_priority_replay_buffer_test.py +++ b/tests/st/ops/cpu/test_priority_replay_buffer_test.py @@ -47,7 +47,7 @@ class PriorityReplayBuffer(nn.Cell): return self.destroy_op() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_priority_replay_buffer_ops(): diff --git a/tests/st/ops/cpu/test_ps_roi_pooling_grad_op.py b/tests/st/ops/cpu/test_ps_roi_pooling_grad_op.py index fc02d1ab2b3..36f68aee0a9 100644 --- a/tests/st/ops/cpu/test_ps_roi_pooling_grad_op.py +++ b/tests/st/ops/cpu/test_ps_roi_pooling_grad_op.py @@ -167,7 +167,7 @@ def _ps_roi_pooling_grad_case(data_type, mode, y_size_adjust=None, dynamic_shape atol=ALL_CLOSE_CRITERION, rtol=ALL_CLOSE_CRITERION) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_mind_ir(): @@ -216,7 +216,7 @@ def test_ps_roi_pooling_grad_mind_ir(): atol=ALL_CLOSE_CRITERION, rtol=ALL_CLOSE_CRITERION) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_graph_mode(): @@ -230,7 +230,7 @@ def test_ps_roi_pooling_grad_graph_mode(): mode=CTX_MODE) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_y_0_shape_wrong(): @@ -246,7 +246,7 @@ def test_ps_roi_pooling_grad_y_0_shape_wrong(): ) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_y_1_shape_wrong(): @@ -262,7 +262,7 @@ def test_ps_roi_pooling_grad_y_1_shape_wrong(): ) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_y_2_shape_wrong(): @@ -278,7 +278,7 @@ def test_ps_roi_pooling_grad_y_2_shape_wrong(): ) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_y_3_shape_wrong(): @@ -294,7 +294,7 @@ def test_ps_roi_pooling_grad_y_3_shape_wrong(): ) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_pynative_mode(): @@ -308,7 +308,7 @@ def test_ps_roi_pooling_grad_pynative_mode(): mode=ms.context.PYNATIVE_MODE) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_input_size_attr_type_wrong(): @@ -324,7 +324,7 @@ def test_ps_roi_pooling_grad_input_size_attr_type_wrong(): _check_attr_validation(arg_name, arg_value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_input_size_attr_type_wrong2(): @@ -340,7 +340,7 @@ def test_ps_roi_pooling_grad_input_size_attr_type_wrong2(): _check_attr_validation(arg_name, arg_value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_input_size_attr_type_wrong3(): @@ -356,7 +356,7 @@ def test_ps_roi_pooling_grad_input_size_attr_type_wrong3(): _check_attr_validation(arg_name, arg_value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_input_size_attr_range_wrong(): @@ -368,7 +368,7 @@ def test_ps_roi_pooling_grad_input_size_attr_range_wrong(): _check_attr_validation(arg_name="input_size", arg_value=-1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_input_size_attr_range_wrong2(): @@ -380,7 +380,7 @@ def test_ps_roi_pooling_grad_input_size_attr_range_wrong2(): _check_attr_validation(arg_name="input_size", arg_value=0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_input_size_attr_range_wrong3(): @@ -392,7 +392,7 @@ def test_ps_roi_pooling_grad_input_size_attr_range_wrong3(): _check_attr_validation(arg_name="input_size", arg_value=(100, -1)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_input_size_attr_range_wrong4(): @@ -404,7 +404,7 @@ def test_ps_roi_pooling_grad_input_size_attr_range_wrong4(): _check_attr_validation(arg_name="input_size", arg_value=(0, 100)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_spatial_scale_attr_type_wrong(): @@ -416,7 +416,7 @@ def test_ps_roi_pooling_grad_spatial_scale_attr_type_wrong(): _check_attr_validation(arg_name="spatial_scale", arg_value=object()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_spatial_scale_attr_range_wrong(): @@ -428,7 +428,7 @@ def test_ps_roi_pooling_grad_spatial_scale_attr_range_wrong(): _check_attr_validation(arg_name="spatial_scale", arg_value=-0.15) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_group_size_attr_type_wrong(): @@ -440,7 +440,7 @@ def test_ps_roi_pooling_grad_group_size_attr_type_wrong(): _check_attr_validation(arg_name="group_size", arg_value=7.1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_group_size_attr_range_wrong(): @@ -452,7 +452,7 @@ def test_ps_roi_pooling_grad_group_size_attr_range_wrong(): _check_attr_validation(arg_name="group_size", arg_value=-1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_output_dim_attr_type_wrong(): @@ -464,7 +464,7 @@ def test_ps_roi_pooling_output_dim_attr_type_wrong(): _check_attr_validation(arg_name="output_dim", arg_value=7.1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_output_dim_attr_range_wrong(): @@ -496,7 +496,7 @@ def _check_attr_validation(arg_name, arg_value): assert False -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_input_args_num(): @@ -526,7 +526,7 @@ def test_ps_roi_pooling_grad_input_args_num(): assert False, "Expected ValueError to occur, but it does not occur." -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_input_type_unsupported(): @@ -556,7 +556,7 @@ def test_ps_roi_pooling_grad_input_type_unsupported(): assert False, "Expected TypeError to occur, but it does not occur." -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_input_type_unsupported2(): @@ -586,7 +586,7 @@ def test_ps_roi_pooling_grad_input_type_unsupported2(): assert False, "Expected TypeError to occur, but it does not occur." -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_grad_input_type_unsupported3(): diff --git a/tests/st/ops/cpu/test_ps_roi_pooling_v2_op.py b/tests/st/ops/cpu/test_ps_roi_pooling_v2_op.py index 35b3e792338..80fd5ebe63f 100644 --- a/tests/st/ops/cpu/test_ps_roi_pooling_v2_op.py +++ b/tests/st/ops/cpu/test_ps_roi_pooling_v2_op.py @@ -127,7 +127,7 @@ def test_ps_roi_pooling_dynamic_shape(): _ps_roi_pooling_case(np.float32, CTX_MODE, dynamic_shape=True) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_mind_ir(): @@ -177,7 +177,7 @@ def test_ps_roi_pooling_mind_ir(): atol=ALL_CLOSE_CRITERION, rtol=ALL_CLOSE_CRITERION) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_graph_mode(): @@ -191,7 +191,7 @@ def test_ps_roi_pooling_graph_mode(): mode=CTX_MODE) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_x_1_shape_wrong(): @@ -207,7 +207,7 @@ def test_ps_roi_pooling_x_1_shape_wrong(): ) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_pynative_mode(): @@ -221,7 +221,7 @@ def test_ps_roi_pooling_pynative_mode(): mode=ms.context.PYNATIVE_MODE) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_spatial_scale_attr_type_wrong(): @@ -233,7 +233,7 @@ def test_ps_roi_pooling_spatial_scale_attr_type_wrong(): _check_attr_validation(arg_name="spatial_scale", arg_value=object()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_spatial_scale_attr_range_wrong(): @@ -245,7 +245,7 @@ def test_ps_roi_pooling_spatial_scale_attr_range_wrong(): _check_attr_validation(arg_name="spatial_scale", arg_value=-0.15) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_group_size_attr_type_wrong(): @@ -257,7 +257,7 @@ def test_ps_roi_pooling_group_size_attr_type_wrong(): _check_attr_validation(arg_name="group_size", arg_value=7.1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_group_size_attr_range_wrong(): @@ -269,7 +269,7 @@ def test_ps_roi_pooling_group_size_attr_range_wrong(): _check_attr_validation(arg_name="group_size", arg_value=-1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_output_dim_attr_type_wrong(): @@ -281,7 +281,7 @@ def test_ps_roi_pooling_output_dim_attr_type_wrong(): _check_attr_validation(arg_name="output_dim", arg_value=7.1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_output_dim_attr_range_wrong(): @@ -312,7 +312,7 @@ def _check_attr_validation(arg_name, arg_value): assert False -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_input_args_num(): @@ -354,7 +354,7 @@ def test_ps_roi_pooling_input_args_num(): assert False, "Expected ValueError to occur, but it does not occur." -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_input_type_unsupported1(): @@ -392,7 +392,7 @@ def test_ps_roi_pooling_input_type_unsupported1(): assert False, "Expected TypeError to occur, but it does not occur." -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_ps_roi_pooling_input_type_unsupported2(): diff --git a/tests/st/ops/cpu/test_random_categorical_op.py b/tests/st/ops/cpu/test_random_categorical_op.py index 701848c6ab3..650cc14a990 100644 --- a/tests/st/ops/cpu/test_random_categorical_op.py +++ b/tests/st/ops/cpu/test_random_categorical_op.py @@ -35,7 +35,7 @@ class RCnet(nn.Cell): TARGET = "CPU" -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_rc_graph_fp16_int64(): @@ -67,7 +67,7 @@ def test_rc_graph_fp16_int64(): assert np.all(diff == 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_rc_graph_fp32_int64(): @@ -98,7 +98,7 @@ def test_rc_graph_fp32_int64(): assert np.all(diff == 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_rc_graph_fp64_int64(): @@ -129,7 +129,7 @@ def test_rc_graph_fp64_int64(): assert np.all(diff == 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_rc_graph_fp16_int16(): @@ -160,7 +160,7 @@ def test_rc_graph_fp16_int16(): assert np.all(diff == 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_rc_graph_fp16_int32(): @@ -191,7 +191,7 @@ def test_rc_graph_fp16_int32(): assert np.all(diff == 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_rc_pynative_fp16_int64(): @@ -222,7 +222,7 @@ def test_rc_pynative_fp16_int64(): assert np.all(diff == 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_rc_pynative_fp32_int64(): @@ -253,7 +253,7 @@ def test_rc_pynative_fp32_int64(): assert np.all(diff == 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_rc_pynative_fp64_int64(): @@ -284,7 +284,7 @@ def test_rc_pynative_fp64_int64(): assert np.all(diff == 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_rc_pynative_fp16_int16(): @@ -315,7 +315,7 @@ def test_rc_pynative_fp16_int16(): assert np.all(diff == 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_rc_pynative_fp16_int32(): @@ -346,7 +346,7 @@ def test_rc_pynative_fp16_int32(): assert np.all(diff == 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_rc_pynative_fp16_int32_result_random(): @@ -370,7 +370,7 @@ def test_rc_pynative_fp16_int32_result_random(): assert np.any(diff != 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_rc_graph_fp16_int32_result_random(): @@ -394,7 +394,7 @@ def test_rc_graph_fp16_int32_result_random(): assert np.any(diff != 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_rc_pynative_fp16_int32_dynamic_shape(): diff --git a/tests/st/ops/cpu/test_random_choice_with_mask_op.py b/tests/st/ops/cpu/test_random_choice_with_mask_op.py index 51bcdc35ac2..08c262b6fbf 100644 --- a/tests/st/ops/cpu/test_random_choice_with_mask_op.py +++ b/tests/st/ops/cpu/test_random_choice_with_mask_op.py @@ -69,7 +69,7 @@ class RCWM_max_count(nn.Cell): return self.rcwm_max_count(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_RCWM_3D(): @@ -88,7 +88,7 @@ def test_RCWM_3D(): assert output2.shape == expect2 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_RCWM_count_out(): @@ -108,7 +108,7 @@ def test_RCWM_count_out(): assert output2.shape == expect2 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_RCWM_count_in(): @@ -128,7 +128,7 @@ def test_RCWM_count_in(): assert output2.shape == expect2 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_RCWM_max_count_graph(): @@ -144,7 +144,7 @@ def test_RCWM_max_count_graph(): rcwm(input_tensor) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_RCWM_max_count_pynative(): @@ -160,7 +160,7 @@ def test_RCWM_max_count_pynative(): rcwm(input_tensor) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_RCWM_1D(): diff --git a/tests/st/ops/cpu/test_random_gamma_op.py b/tests/st/ops/cpu/test_random_gamma_op.py index 5ca9b9818da..a27293fbd90 100644 --- a/tests/st/ops/cpu/test_random_gamma_op.py +++ b/tests/st/ops/cpu/test_random_gamma_op.py @@ -52,7 +52,7 @@ class RandomGammaDR(nn.Cell): return F.random_gamma(outshape, outalpha, seed), outshape, outalpha -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_dynamic_rank(): @@ -75,7 +75,7 @@ def test_dynamic_rank(): assert (output.shape == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize("dtype", [ms.float64, ms.float32, ms.float16]) @@ -110,7 +110,7 @@ def test_random_gamma_op(dtype): print(output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize("dtype", [ms.float64, ms.float32, ms.float16]) diff --git a/tests/st/ops/cpu/test_random_poisson_op.py b/tests/st/ops/cpu/test_random_poisson_op.py index c049c6ccbf8..2c5b1fd8a0d 100644 --- a/tests/st/ops/cpu/test_random_poisson_op.py +++ b/tests/st/ops/cpu/test_random_poisson_op.py @@ -22,7 +22,7 @@ from mindspore import Tensor from mindspore.common.api import _pynative_executor -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize("dtype", [ms.float64, ms.float32, ms.float16, ms.int64, ms.int32]) @@ -50,7 +50,7 @@ def test_poisson_function(dtype, shape_dtype, rate_dtype): assert output.dtype == dtype -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_poisson_function_shape_type_error(): @@ -70,7 +70,7 @@ def test_poisson_function_shape_type_error(): assert False -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_poisson_function_shape_dim_error(): @@ -90,7 +90,7 @@ def test_poisson_function_shape_dim_error(): assert False -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_poisson_function_shape_dtype_error(): @@ -110,7 +110,7 @@ def test_poisson_function_shape_dtype_error(): assert False -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_poisson_function_shape_value_error(): @@ -130,7 +130,7 @@ def test_poisson_function_shape_value_error(): assert False -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_poisson_function_rate_type_error(): @@ -150,7 +150,7 @@ def test_poisson_function_rate_type_error(): assert False -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_poisson_function_rate_dtype_error(): @@ -170,7 +170,7 @@ def test_poisson_function_rate_dtype_error(): assert False -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_poisson_function_seed_type_error(): @@ -190,7 +190,7 @@ def test_poisson_function_seed_type_error(): assert False -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_poisson_function_out_dtype_error(): @@ -262,7 +262,7 @@ class PoissonDSFactory: return output.shape, output.dtype -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize("max_dims", [2, 3, 4, 5, 6]) diff --git a/tests/st/ops/cpu/test_random_shuffle.py b/tests/st/ops/cpu/test_random_shuffle.py index 9dd3dccd280..81edb0d9487 100644 --- a/tests/st/ops/cpu/test_random_shuffle.py +++ b/tests/st/ops/cpu/test_random_shuffle.py @@ -34,7 +34,7 @@ class RandomShuffleNet(nn.Cell): return self.random_shuffle(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) diff --git a/tests/st/ops/cpu/test_range_op.py b/tests/st/ops/cpu/test_range_op.py index 15216d371df..fc08b87eac2 100644 --- a/tests/st/ops/cpu/test_range_op.py +++ b/tests/st/ops/cpu/test_range_op.py @@ -32,7 +32,7 @@ class OpNetWrapper(nn.Cell): return self.op(*inputs) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_int(): @@ -45,7 +45,7 @@ def test_int(): assert np.allclose(outputs.asnumpy(), range(0, 100, 10)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_float(): @@ -58,7 +58,7 @@ def test_float(): assert np.allclose(outputs.asnumpy(), [10., 30., 50., 70., 90.]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_range_op_int(): @@ -73,7 +73,7 @@ def test_range_op_int(): assert np.array_equal(result.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_range_op_float(): @@ -88,7 +88,7 @@ def test_range_op_float(): assert np.array_equal(result.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_range_op_int64(): @@ -103,7 +103,7 @@ def test_range_op_int64(): assert np.array_equal(result.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_range_op_float64(): @@ -118,7 +118,7 @@ def test_range_op_float64(): assert np.array_equal(result.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_range_op_int_reserve(): @@ -133,7 +133,7 @@ def test_range_op_int_reserve(): assert np.array_equal(result.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_range_op_float_reserve(): diff --git a/tests/st/ops/cpu/test_rank_op.py b/tests/st/ops/cpu/test_rank_op.py index 455a7b05573..b37a6c510a4 100644 --- a/tests/st/ops/cpu/test_rank_op.py +++ b/tests/st/ops/cpu/test_rank_op.py @@ -32,7 +32,7 @@ class RankNet(nn.Cell): return self.rank(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -50,7 +50,7 @@ def test_rank_scalar(mode): assert output == expect -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -68,7 +68,7 @@ def test_rank_3d_tensor(mode): assert output == expect -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -86,7 +86,7 @@ def test_rank_4d_tensor(mode): assert output == expect -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) @@ -108,7 +108,7 @@ def test_rank_dynamic_shape(mode): assert output == expect -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) diff --git a/tests/st/ops/cpu/test_realdiv_op.py b/tests/st/ops/cpu/test_realdiv_op.py index c0f7c7a9023..5375e693460 100644 --- a/tests/st/ops/cpu/test_realdiv_op.py +++ b/tests/st/ops/cpu/test_realdiv_op.py @@ -31,7 +31,7 @@ class NetRealDiv(nn.Cell): return self.divide(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float32, np.float64, np.float16, np.uint32, np.uint16]) diff --git a/tests/st/ops/cpu/test_reduce_op.py b/tests/st/ops/cpu/test_reduce_op.py index 234b9f52795..bc367465c35 100644 --- a/tests/st/ops/cpu/test_reduce_op.py +++ b/tests/st/ops/cpu/test_reduce_op.py @@ -184,7 +184,7 @@ def test_reduce_all_onnx(): os.remove(file) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_reduce(): @@ -253,7 +253,7 @@ def test_reduce(): assert (output[17].asnumpy() == 0.0).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_reduce_logic(): @@ -290,7 +290,7 @@ def test_reduce_logic(): assert (output[7].asnumpy() == expect_any_4).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_reduce_prod(): diff --git a/tests/st/ops/cpu/test_reduce_std_op.py b/tests/st/ops/cpu/test_reduce_std_op.py index 478ad8f8b33..7ba4a2682b2 100644 --- a/tests/st/ops/cpu/test_reduce_std_op.py +++ b/tests/st/ops/cpu/test_reduce_std_op.py @@ -38,7 +38,7 @@ class NetReduceStd(nn.Cell): return F.std_mean(indice, axis=self._axis, ddof=False, keepdims=self._keep_dims) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('axis', [None, 0, 1, -1, (-1, 0, 1), (0, 1, 2)]) @@ -76,7 +76,7 @@ class ReduceStdDynamicShapeNet(nn.Cell): return F.std_mean(x_unique, ddof=False, keepdims=False) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) diff --git a/tests/st/ops/cpu/test_relu6_grad_op.py b/tests/st/ops/cpu/test_relu6_grad_op.py index b5e2725319a..390b0d49704 100644 --- a/tests/st/ops/cpu/test_relu6_grad_op.py +++ b/tests/st/ops/cpu/test_relu6_grad_op.py @@ -41,7 +41,7 @@ class NetReluGrad(nn.Cell): return self.relu6_grad(self.dy, self.x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_relu_grad(): diff --git a/tests/st/ops/cpu/test_relu6_op.py b/tests/st/ops/cpu/test_relu6_op.py index e2f8b0cb4b2..d0d5723c990 100644 --- a/tests/st/ops/cpu/test_relu6_op.py +++ b/tests/st/ops/cpu/test_relu6_op.py @@ -40,7 +40,7 @@ class NetReLU6Grad(nn.Cell): def construct(self, x, dy): return self.relu6_grad(dy, x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_relu6(): @@ -55,7 +55,7 @@ def test_relu6(): output = relu6(x) assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_relu6_grad(): diff --git a/tests/st/ops/cpu/test_relu_grad_op.py b/tests/st/ops/cpu/test_relu_grad_op.py index 82c821351c6..7ad3c8e5620 100644 --- a/tests/st/ops/cpu/test_relu_grad_op.py +++ b/tests/st/ops/cpu/test_relu_grad_op.py @@ -41,7 +41,7 @@ class NetReluGrad(nn.Cell): return self.relu_grad(self.dy, self.x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_relu_grad(): diff --git a/tests/st/ops/cpu/test_relu_grad_v2_op.py b/tests/st/ops/cpu/test_relu_grad_v2_op.py index 45e356abbe3..3207645fc0c 100644 --- a/tests/st/ops/cpu/test_relu_grad_v2_op.py +++ b/tests/st/ops/cpu/test_relu_grad_v2_op.py @@ -25,7 +25,7 @@ from mindspore.ops.operations import _grad_ops as G context.set_context(mode=context.GRAPH_MODE, device_target='CPU') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float16, np.float32, np.float64, np.int32]) diff --git a/tests/st/ops/cpu/test_relu_op.py b/tests/st/ops/cpu/test_relu_op.py index 03680fe9cc7..e0edd0546e6 100644 --- a/tests/st/ops/cpu/test_relu_op.py +++ b/tests/st/ops/cpu/test_relu_op.py @@ -38,7 +38,7 @@ class NetRelu(nn.Cell): return self.relu(self.x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_relu(): diff --git a/tests/st/ops/cpu/test_relu_v2_op.py b/tests/st/ops/cpu/test_relu_v2_op.py index 1369c72b536..896376dcc39 100644 --- a/tests/st/ops/cpu/test_relu_v2_op.py +++ b/tests/st/ops/cpu/test_relu_v2_op.py @@ -25,7 +25,7 @@ from mindspore.ops import operations as P context.set_context(mode=context.GRAPH_MODE, device_target='CPU') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float16, np.float32, np.float64, np.int32]) diff --git a/tests/st/ops/cpu/test_renorm_op.py b/tests/st/ops/cpu/test_renorm_op.py index c501406781c..fe5c6f05d96 100644 --- a/tests/st/ops/cpu/test_renorm_op.py +++ b/tests/st/ops/cpu/test_renorm_op.py @@ -33,7 +33,7 @@ class ReNormNet(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_renorm_op_float32(data_type=np.float32): @@ -59,7 +59,7 @@ def test_renorm_op_float32(data_type=np.float32): np.testing.assert_allclose(output.asnumpy(), benchmark_output, rtol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_renorm_op_float16(data_type=np.float16): @@ -85,7 +85,7 @@ def test_renorm_op_float16(data_type=np.float16): np.testing.assert_allclose(output.asnumpy(), benchmark_output, rtol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_renorm_op1_float32(data_type=np.float32): @@ -111,7 +111,7 @@ def test_renorm_op1_float32(data_type=np.float32): np.testing.assert_allclose(output.asnumpy(), benchmark_output, rtol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_renorm_op2_float16(data_type=np.float16): @@ -137,7 +137,7 @@ def test_renorm_op2_float16(data_type=np.float16): np.testing.assert_allclose(output.asnumpy(), benchmark_output, rtol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_renorm_op2_float64(data_type=np.float64): @@ -283,7 +283,7 @@ def vmap_nested_case(): np.testing.assert_allclose(output.asnumpy(), fornet_output.asnumpy(), rtol=1e-6) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_renorm_vmap_cpu(): @@ -296,7 +296,7 @@ def test_renorm_vmap_cpu(): vmap_case() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_renorm_vmap_cpu_nested(): diff --git a/tests/st/ops/cpu/test_resize_bilinear_grad_op.py b/tests/st/ops/cpu/test_resize_bilinear_grad_op.py index 7d61709405e..2bbd67d0a71 100644 --- a/tests/st/ops/cpu/test_resize_bilinear_grad_op.py +++ b/tests/st/ops/cpu/test_resize_bilinear_grad_op.py @@ -100,7 +100,7 @@ def test_resize_bilinear_grad_align_corner_false(): assert np.all(output.asnumpy() == expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize('mode', [context.GRAPH_MODE, context.PYNATIVE_MODE]) diff --git a/tests/st/ops/cpu/test_resize_linear_1d_grad_op.py b/tests/st/ops/cpu/test_resize_linear_1d_grad_op.py index b653b8edffc..8b0ae2663de 100644 --- a/tests/st/ops/cpu/test_resize_linear_1d_grad_op.py +++ b/tests/st/ops/cpu/test_resize_linear_1d_grad_op.py @@ -54,7 +54,7 @@ class ResizeLinear1DGradNet(nn.Cell): return gout -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float32, np.float64]) @@ -78,7 +78,7 @@ def test_resize_linear_1d_grad_align_corners(dtype): assert np.allclose(output[0].asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float32, np.float64]) @@ -102,7 +102,7 @@ def test_resize_linear_1d_grad_half_pixel(dtype): assert np.allclose(output[0].asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float16, np.float32, np.float64]) diff --git a/tests/st/ops/cpu/test_resize_linear_1d_op.py b/tests/st/ops/cpu/test_resize_linear_1d_op.py index d7818768241..9634579a6f4 100644 --- a/tests/st/ops/cpu/test_resize_linear_1d_op.py +++ b/tests/st/ops/cpu/test_resize_linear_1d_op.py @@ -23,7 +23,7 @@ from mindspore.ops.operations import ResizeLinear1D context.set_context(mode=context.GRAPH_MODE, device_target='CPU') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float16, np.float32, np.float64]) @@ -42,7 +42,7 @@ def test_resize_linear_1d_align_corners(dtype): assert np.allclose(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float16, np.float32, np.float64]) @@ -62,7 +62,7 @@ def test_resize_linear_1d_half_pixel(dtype): assert np.allclose(output.asnumpy(), expect) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float16, np.float32, np.float64]) diff --git a/tests/st/ops/cpu/test_rgb_to_hsv_op.py b/tests/st/ops/cpu/test_rgb_to_hsv_op.py index 42469ce9b42..823076f2b0c 100644 --- a/tests/st/ops/cpu/test_rgb_to_hsv_op.py +++ b/tests/st/ops/cpu/test_rgb_to_hsv_op.py @@ -29,7 +29,7 @@ class Net(nn.Cell): return self.op(images) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_rgb_to_csv_dyn(): diff --git a/tests/st/ops/cpu/test_rrelu.py b/tests/st/ops/cpu/test_rrelu.py index 8d79263cdaf..99dbad108bf 100644 --- a/tests/st/ops/cpu/test_rrelu.py +++ b/tests/st/ops/cpu/test_rrelu.py @@ -27,7 +27,7 @@ class Net(nn.Cell): return self.rrelu(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.env_onecard @@ -44,7 +44,7 @@ def test_rrelu_normal(): assert np.array_equal(output, expected_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.env_onecard @@ -61,7 +61,7 @@ def test_rrelu_negative_lu(): assert np.array_equal(output, expected_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.env_onecard diff --git a/tests/st/ops/cpu/test_scatter_arithmetic_op.py b/tests/st/ops/cpu/test_scatter_arithmetic_op.py index 7c3be5b9a4a..ad79f8cc1f7 100644 --- a/tests/st/ops/cpu/test_scatter_arithmetic_op.py +++ b/tests/st/ops/cpu/test_scatter_arithmetic_op.py @@ -50,7 +50,7 @@ def scatter_add_use_locking_false_net(inputx, indices, updates): return net() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_add_small_float32(): @@ -63,7 +63,7 @@ def test_scatter_add_small_float32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_add_input_updated(): @@ -78,7 +78,7 @@ def test_scatter_add_input_updated(): np.testing.assert_array_almost_equal(net.inputx.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_add_large_shape_float32(): @@ -113,7 +113,7 @@ def test_scatter_add_large_shape_float32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_add_small_float32_use_locking_false(): @@ -126,7 +126,7 @@ def test_scatter_add_small_float32_use_locking_false(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_add_input_less_than_1_float32(): @@ -145,7 +145,7 @@ def test_scatter_add_input_less_than_1_float32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_add_float16(): @@ -158,7 +158,7 @@ def test_scatter_add_float16(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_add_large_float16(): @@ -175,7 +175,7 @@ def test_scatter_add_large_float16(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_add_disordered_float16(): @@ -192,7 +192,7 @@ def test_scatter_add_disordered_float16(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_add_large_int32(): @@ -209,7 +209,7 @@ def test_scatter_add_large_int32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_add_disordered_int32(): @@ -226,7 +226,7 @@ def test_scatter_add_disordered_int32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_add_function(): @@ -273,7 +273,7 @@ def scatter_sub_use_locking_false_net(inputx, indices, updates): return net() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_sub_input_updated(): @@ -288,7 +288,7 @@ def test_scatter_sub_input_updated(): np.testing.assert_array_almost_equal(net.inputx.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_sub_large_shape_float32(): @@ -324,7 +324,7 @@ def test_scatter_sub_large_shape_float32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_sub_small_float32_use_locking_false(): @@ -362,7 +362,7 @@ def scatter_mul_use_locking_false_net(inputx, indices, updates): return net() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_mul_input_updated(): @@ -377,7 +377,7 @@ def test_scatter_mul_input_updated(): np.testing.assert_array_almost_equal(net.inputx.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_mul_output_updated_float32(): @@ -390,7 +390,7 @@ def test_scatter_mul_output_updated_float32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_mul_small_float32_use_locking_false(): @@ -428,7 +428,7 @@ def scatter_div_use_locking_false_net(inputx, indices, updates): return net() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_div_input_updated(): @@ -443,7 +443,7 @@ def test_scatter_div_input_updated(): np.testing.assert_array_almost_equal(net.inputx.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_div_output_updated_float32(): @@ -456,7 +456,7 @@ def test_scatter_div_output_updated_float32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_div_small_float32_use_locking_false(): @@ -469,7 +469,7 @@ def test_scatter_div_small_float32_use_locking_false(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_div_output_int16(): @@ -485,7 +485,7 @@ def test_scatter_div_output_int16(): assert np.allclose(output.asnumpy(), input_x.asnumpy(), 0.0001, 0.0001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_div_output_float64(): @@ -526,7 +526,7 @@ def scatter_max_use_locking_false_net(inputx, indices, updates): return net() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_max_input_updated(): @@ -541,7 +541,7 @@ def test_scatter_max_input_updated(): np.testing.assert_array_almost_equal(net.inputx.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_max_output_updated_float32(): @@ -554,7 +554,7 @@ def test_scatter_max_output_updated_float32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_max_small_float32_use_locking_false(): @@ -592,7 +592,7 @@ def scatter_min_use_locking_false_net(inputx, indices, updates): return net() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_min_input_updated(): @@ -607,7 +607,7 @@ def test_scatter_min_input_updated(): np.testing.assert_array_almost_equal(net.inputx.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_min_output_updated_float32(): @@ -620,7 +620,7 @@ def test_scatter_min_output_updated_float32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_min_small_float32_use_locking_false(): @@ -658,7 +658,7 @@ def scatter_update_use_locking_false_net(inputx, indices, updates): return net() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_update_input_updated(): @@ -673,7 +673,7 @@ def test_scatter_update_input_updated(): np.testing.assert_array_almost_equal(net.inputx.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_update_output_updated_float32(): @@ -686,7 +686,7 @@ def test_scatter_update_output_updated_float32(): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_update_output_updated_huge_tensor_float32(): @@ -706,7 +706,7 @@ def test_scatter_update_output_updated_huge_tensor_float32(): np.testing.assert_array_almost_equal(output.asnumpy()[0:2], expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_update_small_float32_use_locking_false(): @@ -729,7 +729,7 @@ class TestScatterAddNetDynamic(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_add_dynamic_shape(): @@ -760,7 +760,7 @@ class TestScatterSubNetDynamic(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_sub_dynamic_shape(): @@ -792,7 +792,7 @@ class TestScatterUpdateNetDynamic(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_update_dynamic_shape(): diff --git a/tests/st/ops/cpu/test_scatter_nd_arithmetic_op.py b/tests/st/ops/cpu/test_scatter_nd_arithmetic_op.py index dfd2e476aa6..75b4a604da3 100644 --- a/tests/st/ops/cpu/test_scatter_nd_arithmetic_op.py +++ b/tests/st/ops/cpu/test_scatter_nd_arithmetic_op.py @@ -118,7 +118,7 @@ def compare_with_numpy(func, lock, input_x, indices, updates): np.testing.assert_array_almost_equal(pynative_output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('lock', [True, False]) @@ -138,7 +138,7 @@ def test_scatter_nd_small_float(lock, func, data_type, index_type): compare_with_numpy(func, lock, input_x, indices, updates) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('lock', [True, False]) @@ -158,7 +158,7 @@ def test_scatter_nd_small_int(lock, func, data_type, index_type): compare_with_numpy(func, lock, input_x, indices, updates) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('lock', [True, False]) @@ -186,7 +186,7 @@ def test_scatter_nd_multi_dims(lock, func, data_type, index_type): compare_with_numpy(func, lock, input_x, indices, updates) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('lock', [True, False]) @@ -220,7 +220,7 @@ def test_scatter_nd_indices_out_of_range(lock, func, data_type, index_type): _ = TestScatterNdNet(func, lock, input_x, indices, updates)() _pynative_executor.sync() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('lock', [True, False]) @@ -240,7 +240,7 @@ def test_scatter_nd_one_value(lock, func, data_type, index_type): compare_with_numpy(func, lock, input_x, indices, updates) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('lock', [True]) @@ -260,7 +260,7 @@ def test_scatter_nd_lock(lock, func, data_type, index_type): compare_with_numpy(func, lock, input_x, indices, updates) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize("func_name", ["add", "sub"]) @@ -289,7 +289,7 @@ def test_scatter_nd_dy_shape(func_name): np.testing.assert_allclose(np_result, ms_result.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize('data_type', [mstype.int64]) @@ -309,7 +309,7 @@ def test_scatter_nd_div_division_by_zero(data_type, index_type): compare_with_numpy('div', False, input_x, indices, updates) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('func', ['add', 'sub', 'div', 'mul']) @@ -336,7 +336,7 @@ def test_scatter_func_indices_vmap(func): np.testing.assert_array_almost_equal(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('func', ['add', 'sub', 'div', 'mul']) diff --git a/tests/st/ops/cpu/test_scatter_nd_op.py b/tests/st/ops/cpu/test_scatter_nd_op.py index df9c30e7429..25bec8c1600 100644 --- a/tests/st/ops/cpu/test_scatter_nd_op.py +++ b/tests/st/ops/cpu/test_scatter_nd_op.py @@ -127,7 +127,7 @@ def scatternd_indices_out_of_range(nptype): _pynative_executor.sync() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_float64(): @@ -140,7 +140,7 @@ def test_scatternd_float64(): scatternd_negative(np.float64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_float32(): @@ -153,7 +153,7 @@ def test_scatternd_float32(): scatternd_negative(np.float32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_float16(): @@ -166,7 +166,7 @@ def test_scatternd_float16(): scatternd_negative(np.float16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_int64(): @@ -179,7 +179,7 @@ def test_scatternd_int64(): scatternd_negative(np.int64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_int32(): @@ -192,7 +192,7 @@ def test_scatternd_int32(): scatternd_negative(np.int32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_int16(): @@ -205,7 +205,7 @@ def test_scatternd_int16(): scatternd_negative(np.int16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_int8(): @@ -218,7 +218,7 @@ def test_scatternd_int8(): scatternd_negative(np.int8) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_uint64(): @@ -230,7 +230,7 @@ def test_scatternd_uint64(): scatternd_positive_uint(np.uint64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_uint32(): @@ -242,7 +242,7 @@ def test_scatternd_uint32(): scatternd_positive_uint(np.uint32) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_uint16(): @@ -254,7 +254,7 @@ def test_scatternd_uint16(): scatternd_positive_uint(np.uint16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_indices_out_of_range(): @@ -266,7 +266,7 @@ def test_scatternd_indices_out_of_range(): scatternd_indices_out_of_range(np.int16) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_uint8(): @@ -307,7 +307,7 @@ def vmap_1_batch(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_vmap(): @@ -328,7 +328,7 @@ class FunctionalNet(nn.Cell): return ops.scatter_nd(indices, update, self.shape) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_functional(): @@ -351,7 +351,7 @@ def test_scatternd_functional(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatternd_functional_pynative(): diff --git a/tests/st/ops/cpu/test_scatter_nd_update_op.py b/tests/st/ops/cpu/test_scatter_nd_update_op.py index 59f1015f985..7c1096ca4fb 100644 --- a/tests/st/ops/cpu/test_scatter_nd_update_op.py +++ b/tests/st/ops/cpu/test_scatter_nd_update_op.py @@ -26,7 +26,7 @@ from mindspore.ops import operations as P context.set_context(mode=context.GRAPH_MODE, device_target='CPU') -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float32, np.float64]) @@ -58,7 +58,7 @@ def test_op1(dtype): np.array(expect, dtype=dtype)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float32, np.float64, np.int32, np.int64]) @@ -90,7 +90,7 @@ def test_op2(dtype): np.array(expect, dtype=dtype)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float32, np.float64, np.int32, np.int64]) @@ -126,7 +126,7 @@ def test_op3(dtype): assert np.allclose(output.asnumpy(), np.array(expect, dtype=dtype)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype', [np.float32, np.float64]) @@ -184,7 +184,7 @@ def test_op5(dtype): scatter_nd_update(indices, update) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_scatter_nd_update_dyn_shape(): diff --git a/tests/st/ops/cpu/test_searchsorted_op.py b/tests/st/ops/cpu/test_searchsorted_op.py index 94fa793d844..3fb51bb7b7a 100644 --- a/tests/st/ops/cpu/test_searchsorted_op.py +++ b/tests/st/ops/cpu/test_searchsorted_op.py @@ -33,7 +33,7 @@ class SearchSortedNet(nn.Cell): def construct(self, sequence, values): return self.searchsorted(sequence, values) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_right_out32(): @@ -55,7 +55,7 @@ def test_right_out32(): assert output.dtype == mstype.int32 assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_out32(): @@ -77,7 +77,7 @@ def test_out32(): assert output.dtype == mstype.int32 assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_right_out64(): @@ -96,7 +96,7 @@ def test_right_out64(): assert (output.asnumpy() == expect).all() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_out64(): diff --git a/tests/st/ops/cpu/test_select_op.py b/tests/st/ops/cpu/test_select_op.py index f5a613bcdd2..7fb2a793faa 100644 --- a/tests/st/ops/cpu/test_select_op.py +++ b/tests/st/ops/cpu/test_select_op.py @@ -39,7 +39,7 @@ class Net(nn.Cell): context.set_context(mode=context.GRAPH_MODE, device_target="CPU") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_select_float32(): @@ -56,7 +56,7 @@ def test_select_float32(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_select_float16(): @@ -73,7 +73,7 @@ def test_select_float16(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_select_int32(): @@ -90,7 +90,7 @@ def test_select_int32(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_functional_select_scalar(): @@ -111,7 +111,7 @@ def test_functional_select_scalar(): assert np.all(-diff < error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_functional_select_broadcast(): @@ -130,7 +130,7 @@ def test_functional_select_broadcast(): assert ret.shape == (5, 5, 65, 54, 12, 5, 2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_functional_select_type_error(): diff --git a/tests/st/ops/cpu/test_selu_op.py b/tests/st/ops/cpu/test_selu_op.py index c01f60c3b31..a877004aa10 100644 --- a/tests/st/ops/cpu/test_selu_op.py +++ b/tests/st/ops/cpu/test_selu_op.py @@ -62,7 +62,7 @@ def selu_np_bencmark(input_x): return result -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.parametrize("data_shape", [(4,), (3, 4), (4, 5, 7)]) @@ -87,7 +87,7 @@ def test_selu(data_shape, data_type): np.testing.assert_allclose(output.asnumpy(), benchmark_output, rtol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_selu_vmap_cpu(): @@ -109,7 +109,7 @@ def test_selu_vmap_cpu(): assert np.allclose(output.asnumpy(), benchmark_output, rtol=error, atol=error) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu def test_selu_dy_shape(): diff --git a/tests/st/ops/cpu/test_sgd_op.py b/tests/st/ops/cpu/test_sgd_op.py index 37eb6e42f24..aa25980ae7c 100644 --- a/tests/st/ops/cpu/test_sgd_op.py +++ b/tests/st/ops/cpu/test_sgd_op.py @@ -40,7 +40,7 @@ class NetSGD(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard def test_SGD(): diff --git a/tests/st/ops/cpu/test_shift_op.py b/tests/st/ops/cpu/test_shift_op.py index 5e7093207e1..a0f7ef3e259 100644 --- a/tests/st/ops/cpu/test_shift_op.py +++ b/tests/st/ops/cpu/test_shift_op.py @@ -102,7 +102,7 @@ def compare(arr: np.ndarray, periods: int, axis: int, fill_value=np.nan): assert np.allclose(numpy_result, mindspore_result, equal_nan=True) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype, fill_value', @@ -117,7 +117,7 @@ def test_no_shift(fill_value, dtype, axis): compare(arr, axis=axis, periods=0, fill_value=fill_value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype, fill_value', @@ -138,7 +138,7 @@ def test_fancy_1d(fill_value, dtype, periods): compare(arr, axis=3, periods=periods, fill_value=fill_value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype, fill_value', @@ -154,7 +154,7 @@ def test_2d(fill_value, dtype, axis, periods): compare(arr, axis=axis, periods=periods, fill_value=fill_value) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @pytest.mark.parametrize('dtype, fill_value', diff --git a/tests/st/ops/cpu/test_sparse_matrix_softmax.py b/tests/st/ops/cpu/test_sparse_matrix_softmax.py index 0bea475b1ab..0aab70b05a3 100644 --- a/tests/st/ops/cpu/test_sparse_matrix_softmax.py +++ b/tests/st/ops/cpu/test_sparse_matrix_softmax.py @@ -111,7 +111,7 @@ def test_ops_sparse_matrix_softmax_vs_nn_softmax_int64(): assert np.all(weights - c[4] < eps) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard diff --git a/tests/st/ops/dynamic_sequence/test_dynamic_sequence_stack.py b/tests/st/ops/dynamic_sequence/test_dynamic_sequence_stack.py index bd6d554fde8..c24edd48cba 100644 --- a/tests/st/ops/dynamic_sequence/test_dynamic_sequence_stack.py +++ b/tests/st/ops/dynamic_sequence/test_dynamic_sequence_stack.py @@ -35,7 +35,7 @@ class NetSequenceStack(nn.Cell): return self.op(seq) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/ops/dynamic_shape/grad/test_besselj1.py b/tests/st/ops/dynamic_shape/grad/test_besselj1.py index f767e279259..3c825dadcb2 100644 --- a/tests/st/ops/dynamic_shape/grad/test_besselj1.py +++ b/tests/st/ops/dynamic_shape/grad/test_besselj1.py @@ -77,7 +77,7 @@ def test_gpu_grad_dynamic_shape_2(): grad_dyn_case(False) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard diff --git a/tests/st/ops/dynamic_shape/grad/test_index_fill.py b/tests/st/ops/dynamic_shape/grad/test_index_fill.py index 82a9ad6f98f..ffb9581f08b 100644 --- a/tests/st/ops/dynamic_shape/grad/test_index_fill.py +++ b/tests/st/ops/dynamic_shape/grad/test_index_fill.py @@ -54,7 +54,7 @@ def test_dynamic_shape_index_fill(): [Tensor(x_np), dim, Tensor(index_np), value], False) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.env_onecard @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/dynamic_shape/grad/test_reduce_sum.py b/tests/st/ops/dynamic_shape/grad/test_reduce_sum.py index c72c94c079e..3ed1030d22a 100644 --- a/tests/st/ops/dynamic_shape/grad/test_reduce_sum.py +++ b/tests/st/ops/dynamic_shape/grad/test_reduce_sum.py @@ -73,7 +73,7 @@ def test_dynamic_reducesum_rank(): test_dynamic.test_dynamic_grad_net(x, True) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard diff --git a/tests/st/ops/dynamic_shape/test_dynamic_getitem.py b/tests/st/ops/dynamic_shape/test_dynamic_getitem.py index dbf960481a7..e3b2183810b 100644 --- a/tests/st/ops/dynamic_shape/test_dynamic_getitem.py +++ b/tests/st/ops/dynamic_shape/test_dynamic_getitem.py @@ -262,7 +262,7 @@ def test_dynamic_getitem_tensor(): fact.grad_impl() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -381,7 +381,7 @@ def test_dynamic_rank_getitem_slice(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -422,7 +422,7 @@ def test_dynamic_rank_getitem_with_single_basic_index(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -460,7 +460,7 @@ def test_dynamic_rank_getitem_tuple_with_basic_index(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -497,7 +497,7 @@ def test_dynamic_rank_getitem_with_tensor_index(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -535,7 +535,7 @@ def test_dynamic_rank_getitem_tuple_with_multi_tensor_index(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -570,7 +570,7 @@ def test_dynamic_rank_getitem_with_list_index(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -725,7 +725,7 @@ def test_dynamic_getitem_int_001(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -911,7 +911,7 @@ def test_dynamic_getitem_tuple_002(): fact.grad_impl() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard diff --git a/tests/st/ops/dynamic_shape/test_dynamic_setitem.py b/tests/st/ops/dynamic_shape/test_dynamic_setitem.py index 3ea2ee70a67..80e23a57000 100644 --- a/tests/st/ops/dynamic_shape/test_dynamic_setitem.py +++ b/tests/st/ops/dynamic_shape/test_dynamic_setitem.py @@ -154,7 +154,7 @@ class TensorSetItem(nn.Cell): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -194,7 +194,7 @@ def test_dynamic_setitem_int_tensor(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -383,7 +383,7 @@ def test_dynamic_setitem_ellipsis_tensor(): fact.grad_impl() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @@ -548,7 +548,7 @@ def test_dynamic_setitem_slice_sequence(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -598,7 +598,7 @@ def test_dynamic_rank_setitem_with_single_basic_index(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -647,7 +647,7 @@ def test_dynamic_rank_setitem_tuple_with_basic_index(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -694,7 +694,7 @@ def test_dynamic_rank_setitem_tuple_with_mix_index(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -733,7 +733,7 @@ def test_dynamic_rank_setitem_tuple_with_multi_tensor_index(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -771,7 +771,7 @@ def test_dynamic_rank_setitem_tuple_with_empty_bool_tensor_index(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard @@ -807,7 +807,7 @@ def test_dynamic_rank_setitem_with_list_index(): @pytest.mark.skip(reason="Need to be fixed.") -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard diff --git a/tests/st/ops/test_eps_op.py b/tests/st/ops/test_eps_op.py index dc77e1137f8..158377a2d67 100644 --- a/tests/st/ops/test_eps_op.py +++ b/tests/st/ops/test_eps_op.py @@ -31,7 +31,7 @@ class NetEps(nn.Cell): return self.eps(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_func_binary_cross_entropy_with_logits.py b/tests/st/ops/test_func_binary_cross_entropy_with_logits.py index a667482077a..06da4ecd55a 100644 --- a/tests/st/ops/test_func_binary_cross_entropy_with_logits.py +++ b/tests/st/ops/test_func_binary_cross_entropy_with_logits.py @@ -25,7 +25,7 @@ class Net(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_func_dense.py b/tests/st/ops/test_func_dense.py index b2dd07d54a4..a32a8b25f7b 100644 --- a/tests/st/ops/test_func_dense.py +++ b/tests/st/ops/test_func_dense.py @@ -25,7 +25,7 @@ class Net(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_func_diagonal.py b/tests/st/ops/test_func_diagonal.py index 387d32499f4..c28f2315ac3 100644 --- a/tests/st/ops/test_func_diagonal.py +++ b/tests/st/ops/test_func_diagonal.py @@ -32,7 +32,7 @@ class Net(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -54,7 +54,7 @@ def test_ops_diagonal(mode): assert np.allclose(output.asnumpy(), expect_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_func_eq.py b/tests/st/ops/test_func_eq.py index 93ee882482b..379d2d71d4d 100644 --- a/tests/st/ops/test_func_eq.py +++ b/tests/st/ops/test_func_eq.py @@ -28,7 +28,7 @@ class Net(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_func_inner.py b/tests/st/ops/test_func_inner.py index b4382cd9b27..afa8c744aae 100644 --- a/tests/st/ops/test_func_inner.py +++ b/tests/st/ops/test_func_inner.py @@ -50,7 +50,7 @@ def test_inner_normal(mode): assert np.allclose(out.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_func_interpolate.py b/tests/st/ops/test_func_interpolate.py index 8a66c1e0ab5..ded762f3b95 100644 --- a/tests/st/ops/test_func_interpolate.py +++ b/tests/st/ops/test_func_interpolate.py @@ -223,7 +223,7 @@ def test_interpolate_bicubic(mode): assert np.allclose(output_4d_2.asnumpy(), except_4d_2, atol=1e-5, rtol=1e-5) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_func_nansum.py b/tests/st/ops/test_func_nansum.py index 55ec486f7df..d57452469a5 100644 --- a/tests/st/ops/test_func_nansum.py +++ b/tests/st/ops/test_func_nansum.py @@ -26,7 +26,7 @@ class Net(nn.Cell): return ops.nansum(x, axis=0, keepdims=True, dtype=ms.int64) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_func_one_hot.py b/tests/st/ops/test_func_one_hot.py index bc6fe476218..fdf8b62f5b3 100644 --- a/tests/st/ops/test_func_one_hot.py +++ b/tests/st/ops/test_func_one_hot.py @@ -25,7 +25,7 @@ class Net(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_func_select_scatter.py b/tests/st/ops/test_func_select_scatter.py index 2fff0240764..6330caf439d 100644 --- a/tests/st/ops/test_func_select_scatter.py +++ b/tests/st/ops/test_func_select_scatter.py @@ -74,7 +74,7 @@ def test_ops_select_scatter(mode): assert np.allclose(output.asnumpy(), expect_output1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_func_slice_scatter.py b/tests/st/ops/test_func_slice_scatter.py index b0360efa9ce..1b35ad43335 100644 --- a/tests/st/ops/test_func_slice_scatter.py +++ b/tests/st/ops/test_func_slice_scatter.py @@ -32,7 +32,7 @@ class Net(nn.Cell): return ops.slice_scatter(input_x, src, self.dim, self.start, self.end, self.step) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -66,7 +66,7 @@ def test_ops_slice_scatter(mode): assert np.allclose(output.asnumpy(), expect_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_func_squeeze.py b/tests/st/ops/test_func_squeeze.py index 879f00d09d4..2f33c3c24ca 100644 --- a/tests/st/ops/test_func_squeeze.py +++ b/tests/st/ops/test_func_squeeze.py @@ -31,7 +31,7 @@ class Net(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_func_sum.py b/tests/st/ops/test_func_sum.py index ec8da34bb0d..2de28d75505 100644 --- a/tests/st/ops/test_func_sum.py +++ b/tests/st/ops/test_func_sum.py @@ -78,7 +78,7 @@ def test_sum_keepdim(mode): assert np.allclose(out.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_func_vector_norm.py b/tests/st/ops/test_func_vector_norm.py index cd5f995b1f3..dfdd8b1d072 100644 --- a/tests/st/ops/test_func_vector_norm.py +++ b/tests/st/ops/test_func_vector_norm.py @@ -25,7 +25,7 @@ class Net(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_func_zeta.py b/tests/st/ops/test_func_zeta.py index 21f45234055..aa34844657b 100644 --- a/tests/st/ops/test_func_zeta.py +++ b/tests/st/ops/test_func_zeta.py @@ -24,7 +24,7 @@ class Net(nn.Cell): return ops.zeta(x, y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_gaussian_nll_loss.py b/tests/st/ops/test_gaussian_nll_loss.py index 42210902654..f0e42cd9b5c 100644 --- a/tests/st/ops/test_gaussian_nll_loss.py +++ b/tests/st/ops/test_gaussian_nll_loss.py @@ -30,7 +30,7 @@ class Net(nn.Cell): return loss -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -60,7 +60,7 @@ def test_gaussian_nll_loss_full(mode, full): assert np.allclose(output.asnumpy(), expected) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_lppool1d.py b/tests/st/ops/test_lppool1d.py index 41572e4daa5..25d90ad2e61 100644 --- a/tests/st/ops/test_lppool1d.py +++ b/tests/st/ops/test_lppool1d.py @@ -27,7 +27,7 @@ class Net(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_margin_ranking_loss.py b/tests/st/ops/test_margin_ranking_loss.py index 88f2777e3c8..c92fb93cdeb 100644 --- a/tests/st/ops/test_margin_ranking_loss.py +++ b/tests/st/ops/test_margin_ranking_loss.py @@ -31,7 +31,7 @@ class MarginRankingLoss(nn.Cell): return ops.margin_ranking_loss(x, y, label, margin=margin, reduction=self.reduction) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_ops_clip_func.py b/tests/st/ops/test_ops_clip_func.py index c98fd0e2ce0..f70e9628ec5 100644 --- a/tests/st/ops/test_ops_clip_func.py +++ b/tests/st/ops/test_ops_clip_func.py @@ -140,7 +140,7 @@ def test_ops_clip(mode): assert np.allclose(output_case_2.asnumpy(), expect_output_case_2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_ops_isposinf.py b/tests/st/ops/test_ops_isposinf.py index 7dde27c6cd1..5fc7dadb075 100644 --- a/tests/st/ops/test_ops_isposinf.py +++ b/tests/st/ops/test_ops_isposinf.py @@ -24,7 +24,7 @@ class Net(nn.Cell): return ops.isposinf(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/ops/test_split.py b/tests/st/ops/test_split.py index 4f990471f78..9edabb12897 100644 --- a/tests/st/ops/test_split.py +++ b/tests/st/ops/test_split.py @@ -77,7 +77,7 @@ def test_f_tensor_split_list(mode): assert np.allclose(res.asnumpy(), exp) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -183,7 +183,7 @@ class VSplitNet(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/optimizer/test_asgd.py b/tests/st/optimizer/test_asgd.py index 9ad635a4261..b443771a416 100644 --- a/tests/st/optimizer/test_asgd.py +++ b/tests/st/optimizer/test_asgd.py @@ -43,7 +43,7 @@ def test_default_asgd(mode): assert np.allclose(loss_default_asgd, loss, atol=1.e-5) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/optimizer/test_sgd.py b/tests/st/optimizer/test_sgd.py index e3839fd9bed..bbdaa9d1e19 100644 --- a/tests/st/optimizer/test_sgd.py +++ b/tests/st/optimizer/test_sgd.py @@ -19,7 +19,7 @@ import mindspore.context as context from .optimizer_utils import FakeNet, build_network -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/optimizer_ex/test_adam_cmp.py b/tests/st/optimizer_ex/test_adam_cmp.py index 8af87f2a3e3..35c6b5050d8 100644 --- a/tests/st/optimizer_ex/test_adam_cmp.py +++ b/tests/st/optimizer_ex/test_adam_cmp.py @@ -201,7 +201,7 @@ def test_adam_group(mode): fact.result_cmp() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -218,7 +218,7 @@ def test_adam_lr_dynamic(mode): fact.result_cmp() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -235,7 +235,7 @@ def test_adam_group_lr_dynamic(mode): fact.result_cmp() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/optimizer_ex/test_adamw_cmp.py b/tests/st/optimizer_ex/test_adamw_cmp.py index f9bbc90d2f9..cdaa7e7c968 100644 --- a/tests/st/optimizer_ex/test_adamw_cmp.py +++ b/tests/st/optimizer_ex/test_adamw_cmp.py @@ -186,7 +186,7 @@ def test_adamw_basic(mode): fact.result_cmp() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -220,7 +220,7 @@ def test_adamw_lr_dynamic(mode): fact.result_cmp() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -237,7 +237,7 @@ def test_adamw_group_lr_dynamic(mode): fact.result_cmp() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/optimizer_ex/test_lr_scheduler.py b/tests/st/optimizer_ex/test_lr_scheduler.py index 844b2fe4c20..f46c3874370 100644 --- a/tests/st/optimizer_ex/test_lr_scheduler.py +++ b/tests/st/optimizer_ex/test_lr_scheduler.py @@ -69,7 +69,7 @@ def test_sequential_lr_scheduler(mode): assert np.allclose(current_lr[0].asnumpy(), expect_list[i]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -102,7 +102,7 @@ def test_reduce_lr_on_plateau(mode): assert np.allclose(current_lr[0].asnumpy(), expect_list[i]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -134,7 +134,7 @@ def test_step_lr(mode): assert np.allclose(current_lr[0].asnumpy(), expect_list[i]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -166,7 +166,7 @@ def test_linear_lr(mode): assert np.allclose(current_lr[0].asnumpy(), expect_list[i]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -198,7 +198,7 @@ def test_exponential_lr(mode): assert np.allclose(current_lr[0].asnumpy(), expect_list[i]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -230,7 +230,7 @@ def test_polynomial_lr(mode): assert np.allclose(current_lr[0].asnumpy(), expect_list[i]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -272,7 +272,7 @@ def test_lambdalr_scheduler(mode): assert np.allclose([float(lr) for lr in current_lr], expect_list[i]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -335,7 +335,7 @@ def test_multistep_lr(mode): assert np.allclose(current_lr[0].asnumpy(), expect_list[i]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -367,7 +367,7 @@ def test_constant_lr(mode): assert np.allclose(current_lr[0].asnumpy(), expect_list[i]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -399,7 +399,7 @@ def test_cyclic_lr(mode): assert np.allclose(current_lr[0].asnumpy(), expect_list[i]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -434,7 +434,7 @@ def test_cosine_annealing_warm_restarts(mode): assert np.allclose(current_lr[0].asnumpy(), expect_list[epoch*iters+i]) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/pynative/data_parallel/test_pynative_hccl.py b/tests/st/pynative/data_parallel/test_pynative_hccl.py index a26426c33c8..929244bd668 100644 --- a/tests/st/pynative/data_parallel/test_pynative_hccl.py +++ b/tests/st/pynative/data_parallel/test_pynative_hccl.py @@ -58,7 +58,7 @@ def train_allreduce_8p(q, device_id, device_num): output = net(Tensor(input_x, mstype.float32)) q.put(np.allclose(output.asnumpy(), except_output)) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.env_single diff --git a/tests/st/pynative/data_parallel/test_pynative_nccl_all.py b/tests/st/pynative/data_parallel/test_pynative_nccl_all.py index 93039a16560..506e9879f19 100644 --- a/tests/st/pynative/data_parallel/test_pynative_nccl_all.py +++ b/tests/st/pynative/data_parallel/test_pynative_nccl_all.py @@ -16,7 +16,7 @@ import os import pytest -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.env_single def test_pynative_nccl_allreduce(): diff --git a/tests/st/pynative/hook/test_hook_function.py b/tests/st/pynative/hook/test_hook_function.py index 0fd21a64499..7a8f29828ad 100644 --- a/tests/st/pynative/hook/test_hook_function.py +++ b/tests/st/pynative/hook/test_hook_function.py @@ -73,7 +73,7 @@ def test_var_hook_forward(): assert np.allclose(out1.asnumpy(), out2.asnumpy(), 0.00001, 0.00001) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_cpu diff --git a/tests/st/pynative/hook/test_pynative_backward_hook.py b/tests/st/pynative/hook/test_pynative_backward_hook.py index d9dfe16b90e..9410a9578a9 100644 --- a/tests/st/pynative/hook/test_pynative_backward_hook.py +++ b/tests/st/pynative/hook/test_pynative_backward_hook.py @@ -122,7 +122,7 @@ class CmpNetFWHook(nn.Cell): return x -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/pynative/hook/test_pynative_forward_hook.py b/tests/st/pynative/hook/test_pynative_forward_hook.py index c283b4d46ce..9b5443c8e55 100644 --- a/tests/st/pynative/hook/test_pynative_forward_hook.py +++ b/tests/st/pynative/hook/test_pynative_forward_hook.py @@ -297,7 +297,7 @@ class CompareMultiNet2(nn.Cell): return x -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -520,7 +520,7 @@ class TestHookNet(nn.Cell): return x -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/pynative/jit/test_outmost_dyn_shape_jit.py b/tests/st/pynative/jit/test_outmost_dyn_shape_jit.py index 6402fa19bc2..0c3659d7565 100644 --- a/tests/st/pynative/jit/test_outmost_dyn_shape_jit.py +++ b/tests/st/pynative/jit/test_outmost_dyn_shape_jit.py @@ -75,7 +75,7 @@ def cmp_func(x, y): return x -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/pynative/jit/test_pynative_jit.py b/tests/st/pynative/jit/test_pynative_jit.py index 0788a2ac4bc..f247e6c3a08 100644 --- a/tests/st/pynative/jit/test_pynative_jit.py +++ b/tests/st/pynative/jit/test_pynative_jit.py @@ -358,7 +358,7 @@ def test_pynative_jit_empty_graph(): assert output.asnumpy() == 10 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/pynative/ops/test_ops.py b/tests/st/pynative/ops/test_ops.py index 6359e519437..153bbf1ba33 100644 --- a/tests/st/pynative/ops/test_ops.py +++ b/tests/st/pynative/ops/test_ops.py @@ -44,7 +44,7 @@ def expand_tensor(a, b): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard diff --git a/tests/st/pynative/test_tensor_getitem.py b/tests/st/pynative/test_tensor_getitem.py index 3c46593d2ff..63968590f26 100644 --- a/tests/st/pynative/test_tensor_getitem.py +++ b/tests/st/pynative/test_tensor_getitem.py @@ -242,7 +242,7 @@ class TensorItemByNone(Cell): return ret -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/pynative/test_tensor_setitem.py b/tests/st/pynative/test_tensor_setitem.py index 3736d070b08..f7fa9c09c0d 100644 --- a/tests/st/pynative/test_tensor_setitem.py +++ b/tests/st/pynative/test_tensor_setitem.py @@ -239,7 +239,7 @@ class TensorItemSetWithNumber(Cell): return ret -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_arm_ascend910b_training diff --git a/tests/st/rl/test_cell_list_getattr.py b/tests/st/rl/test_cell_list_getattr.py index c79f80b9b11..f534f23cac3 100644 --- a/tests/st/rl/test_cell_list_getattr.py +++ b/tests/st/rl/test_cell_list_getattr.py @@ -62,7 +62,7 @@ def verify_list_item_getattr(trainer, expect_res, expect_grad_res): assert np.array_equal(res2.asnumpy(), expect_grad_res.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -82,7 +82,7 @@ def test_list_item_getattr(): verify_list_item_getattr(trainer, expect_res, expect_grad_res) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -292,7 +292,7 @@ def test_cell_list_getattr3(): verify_list_item_getattr2(trainer, expect_res, expect_grad_res) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/scipy_st/test_grad.py b/tests/st/scipy_st/test_grad.py index bf326fc5368..f0473ec718e 100644 --- a/tests/st/scipy_st/test_grad.py +++ b/tests/st/scipy_st/test_grad.py @@ -24,7 +24,7 @@ from mindspore.scipy.ops import SolveTriangular from tests.st.scipy_st.utils import create_random_rank_matrix, create_sym_pos_matrix, gradient_check -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_x86_gpu_training @pytest.mark.env_onecard diff --git a/tests/st/sparse/test_control_flow.py b/tests/st/sparse/test_control_flow.py index 17ad0e7bd70..9245668f4d4 100644 --- a/tests/st/sparse/test_control_flow.py +++ b/tests/st/sparse/test_control_flow.py @@ -25,7 +25,7 @@ from .sparse_utils import compare_csr, get_csr_tensor, csr_add, get_csr_componen forward_grad_net -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @@ -103,7 +103,7 @@ def test_control_flow_while_if_continue_not_relevant_gt(): assert isinstance(grad_graph[0], CSRTensor) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @@ -141,7 +141,7 @@ def test_control_flow_for_while_return_in_while_no(): assert isinstance(grad_graph[1], CSRTensor) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard diff --git a/tests/st/sparse/test_coo.py b/tests/st/sparse/test_coo.py index eda174f4f04..b21fced7eb7 100644 --- a/tests/st/sparse/test_coo.py +++ b/tests/st/sparse/test_coo.py @@ -24,7 +24,7 @@ from mindspore.ops import functional as F from .sparse_utils import get_platform, compare_res, compare_coo -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -140,7 +140,7 @@ def test_coo_tensor_in_while(): assert out.shape == shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @@ -241,7 +241,7 @@ def test_dtype_coo_tensor(): assert out4 in [mstype.float32] -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/sparse/test_csr.py b/tests/st/sparse/test_csr.py index 7616098017f..8bde2e9ea78 100644 --- a/tests/st/sparse/test_csr.py +++ b/tests/st/sparse/test_csr.py @@ -27,7 +27,7 @@ from mindspore.ops.operations import _csr_ops from .sparse_utils import get_platform, compare_res, compare_csr -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -57,7 +57,7 @@ def test_make_csr(): compare_csr(csr3, csr2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @pytest.mark.platform_x86_gpu_training @@ -247,7 +247,7 @@ def test_csr_tensor_in_while_cpu(): assert shape == out.shape -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @@ -313,7 +313,7 @@ def test_batch_csr_ops(): assert np.allclose(graph_res_gather.asnumpy(), expect5) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard @@ -549,7 +549,7 @@ def test_dtype_csr_tensor(): assert out4 in [mstype.float32] -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard diff --git a/tests/st/sparse/test_sparse_dynshape.py b/tests/st/sparse/test_sparse_dynshape.py index a2881c4b14a..7d3ee1cfe89 100644 --- a/tests/st/sparse/test_sparse_dynshape.py +++ b/tests/st/sparse/test_sparse_dynshape.py @@ -116,7 +116,7 @@ def test_csr_to_dense_dshape(): test_dynamic.test_dynamic_grad_net(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard diff --git a/tests/st/syntax/simple_expression/test_assign.py b/tests/st/syntax/simple_expression/test_assign.py index 7647544b116..d7574e969f1 100644 --- a/tests/st/syntax/simple_expression/test_assign.py +++ b/tests/st/syntax/simple_expression/test_assign.py @@ -33,7 +33,7 @@ class Net(nn.Cell): return x -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/syntax/simple_expression/test_assignment_ops.py b/tests/st/syntax/simple_expression/test_assignment_ops.py index 04e3cb0d402..803a386066f 100644 --- a/tests/st/syntax/simple_expression/test_assignment_ops.py +++ b/tests/st/syntax/simple_expression/test_assignment_ops.py @@ -488,7 +488,7 @@ def test_number_assigndiv_tensor(): assert np.all(result.asnumpy() == expect.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/syntax/simple_expression/test_compare.py b/tests/st/syntax/simple_expression/test_compare.py index 8f16749ddb6..ac4be2ed91b 100644 --- a/tests/st/syntax/simple_expression/test_compare.py +++ b/tests/st/syntax/simple_expression/test_compare.py @@ -83,7 +83,7 @@ def test_compare_tensor_int_vs_tensor_float(): assert not ret -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -106,7 +106,7 @@ def test_multiple_compare(): assert ret -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -129,7 +129,7 @@ def test_multiple_compare_2(): assert not ret -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -152,7 +152,7 @@ def test_multiple_compare_3(): assert ret -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/syntax/simple_expression/test_sequence_getitem.py b/tests/st/syntax/simple_expression/test_sequence_getitem.py index 0d08a5c1ac7..5309d63d059 100644 --- a/tests/st/syntax/simple_expression/test_sequence_getitem.py +++ b/tests/st/syntax/simple_expression/test_sequence_getitem.py @@ -22,7 +22,7 @@ from mindspore.common.tensor import Tensor context.set_context(mode=context.GRAPH_MODE) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -43,7 +43,7 @@ def test_tuple_getitem_with_constant_bool_index(): assert ret2 == 1 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -64,7 +64,7 @@ def test_tuple_getitem_with_constant_bool_index_2(): assert np.all(ret2.asnumpy() == np.array([1, 2, 3, 4])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -89,7 +89,7 @@ def test_tuple_getitem_with_variable_bool_index(): assert np.all(ret2.asnumpy() == np.array([1])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -110,7 +110,7 @@ def test_tuple_getitem_with_variable_bool_index_2(): assert np.all(ret2.asnumpy() == np.array([1])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -131,7 +131,7 @@ def test_tuple_getitem_with_variable_bool_index_3(): assert ret2 == 1 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -152,7 +152,7 @@ def test_list_getitem_with_constant_bool_index(): assert ret2 == 1 -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -173,7 +173,7 @@ def test_list_getitem_with_constant_bool_index_2(): assert np.all(ret2.asnumpy() == np.array([1, 2, 3, 4])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training @@ -198,7 +198,7 @@ def test_list_getitem_with_variable_bool_index(): assert np.all(ret2.asnumpy() == np.array([1])) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/syntax/simple_expression/test_zip_operation.py b/tests/st/syntax/simple_expression/test_zip_operation.py index 0757dbb4741..3a880002049 100644 --- a/tests/st/syntax/simple_expression/test_zip_operation.py +++ b/tests/st/syntax/simple_expression/test_zip_operation.py @@ -60,7 +60,7 @@ def test_zip_operation_args_size(): assert np.all(out.asnumpy() == 1) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_arm_ascend_training @pytest.mark.platform_x86_ascend_training diff --git a/tests/st/tensor/test_abs.py b/tests/st/tensor/test_abs.py index 56f2690e59f..32f1cb9dbb7 100644 --- a/tests/st/tensor/test_abs.py +++ b/tests/st/tensor/test_abs.py @@ -23,7 +23,7 @@ class Net(nn.Cell): return abs(x) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -45,7 +45,7 @@ def test_tensor_abs(mode): assert np.allclose(output.asnumpy(), expect_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_cholesky_solve.py b/tests/st/tensor/test_cholesky_solve.py index 30b2f40652a..1c6fee22ba3 100644 --- a/tests/st/tensor/test_cholesky_solve.py +++ b/tests/st/tensor/test_cholesky_solve.py @@ -26,7 +26,7 @@ class CholeskySolveTensorNet(nn.Cell): return input1.cholesky_solve(input2, upper) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_diagonal.py b/tests/st/tensor/test_diagonal.py index 18b93f43577..d6303bbeead 100644 --- a/tests/st/tensor/test_diagonal.py +++ b/tests/st/tensor/test_diagonal.py @@ -31,7 +31,7 @@ class Net(nn.Cell): return out -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -53,7 +53,7 @@ def test_tensor_diagonal(mode): assert np.allclose(output.asnumpy(), expect_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_diagonal_scatter.py b/tests/st/tensor/test_diagonal_scatter.py index 05207aa5b82..09b4e1fedaa 100644 --- a/tests/st/tensor/test_diagonal_scatter.py +++ b/tests/st/tensor/test_diagonal_scatter.py @@ -25,7 +25,7 @@ class Net(nn.Cell): return x.diagonal_scatter(src, offset=1, dim1=1, dim2=2) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_eq.py b/tests/st/tensor/test_eq.py index 61ea57fa962..224df5592da 100644 --- a/tests/st/tensor/test_eq.py +++ b/tests/st/tensor/test_eq.py @@ -27,7 +27,7 @@ class Net(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training @@ -51,7 +51,7 @@ def test_eq(mode): assert np.allclose(expect_value, value.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_floor_divide.py b/tests/st/tensor/test_floor_divide.py index b01f66a78d1..b29a4713e27 100644 --- a/tests/st/tensor/test_floor_divide.py +++ b/tests/st/tensor/test_floor_divide.py @@ -24,7 +24,7 @@ class Net(nn.Cell): return x.floor_divide(y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_inner.py b/tests/st/tensor/test_inner.py index d496713c9c3..213def52122 100644 --- a/tests/st/tensor/test_inner.py +++ b/tests/st/tensor/test_inner.py @@ -72,7 +72,7 @@ def test_inner_with_scalar(mode): assert np.allclose(out.asnumpy(), expect_out) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_item.py b/tests/st/tensor/test_item.py index f0c01af669d..b00fd36a789 100644 --- a/tests/st/tensor/test_item.py +++ b/tests/st/tensor/test_item.py @@ -30,7 +30,7 @@ class Net(nn.Cell): return x.item(self.index) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_movedim.py b/tests/st/tensor/test_movedim.py index c47fc672dee..04aecfd992a 100644 --- a/tests/st/tensor/test_movedim.py +++ b/tests/st/tensor/test_movedim.py @@ -24,7 +24,7 @@ class Net(nn.Cell): return x.movedim(1, 0) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_permute.py b/tests/st/tensor/test_permute.py index a888f54f729..129e791bc76 100644 --- a/tests/st/tensor/test_permute.py +++ b/tests/st/tensor/test_permute.py @@ -55,7 +55,7 @@ def test_tensor_permute(mode): assert np.allclose(output.asnumpy(), expect_output) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_squeeze.py b/tests/st/tensor/test_squeeze.py index 68e695b636d..c88f7552f2f 100644 --- a/tests/st/tensor/test_squeeze.py +++ b/tests/st/tensor/test_squeeze.py @@ -30,7 +30,7 @@ class Net(nn.Cell): return output -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_sum.py b/tests/st/tensor/test_sum.py index 2c2f4fedff9..7d0675bb5a6 100644 --- a/tests/st/tensor/test_sum.py +++ b/tests/st/tensor/test_sum.py @@ -31,7 +31,7 @@ class Net(nn.Cell): return x.sum(self.axis, self.dtype, self.keepdims, self.initial) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_tolist.py b/tests/st/tensor/test_tolist.py index c3639a799fd..ce373dd3a82 100644 --- a/tests/st/tensor/test_tolist.py +++ b/tests/st/tensor/test_tolist.py @@ -24,7 +24,7 @@ class Net(nn.Cell): return x.tolist() -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_type.py b/tests/st/tensor/test_type.py index 7bd3ff755ee..c226a3c668d 100644 --- a/tests/st/tensor/test_type.py +++ b/tests/st/tensor/test_type.py @@ -30,7 +30,7 @@ class Net(nn.Cell): return x.type(self.dtype) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_type_as.py b/tests/st/tensor/test_type_as.py index 69b6d9be361..af030f4e3cc 100644 --- a/tests/st/tensor/test_type_as.py +++ b/tests/st/tensor/test_type_as.py @@ -24,7 +24,7 @@ class Net(nn.Cell): return x.type_as(y) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/tensor/test_var.py b/tests/st/tensor/test_var.py index 9e150bf4847..50ed007c3a7 100644 --- a/tests/st/tensor/test_var.py +++ b/tests/st/tensor/test_var.py @@ -25,7 +25,7 @@ class Net(nn.Cell): return x.var(axis=0, ddof=True, keepdims=True) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_cpu @pytest.mark.platform_arm_cpu @pytest.mark.platform_x86_gpu_training diff --git a/tests/st/vmap/test_vmap_graph.py b/tests/st/vmap/test_vmap_graph.py index 318b3c36ade..030c3f75da4 100644 --- a/tests/st/vmap/test_vmap_graph.py +++ b/tests/st/vmap/test_vmap_graph.py @@ -471,7 +471,7 @@ def test_vmap_as_vmap_input(): assert np.allclose(output.asnumpy(), expect_res.asnumpy()) -@pytest.mark.level0 +@pytest.mark.level1 @pytest.mark.platform_x86_gpu_training @pytest.mark.platform_x86_cpu @pytest.mark.env_onecard