diff --git a/docs/api/api_python/mindspore.context.rst b/docs/api/api_python/mindspore.context.rst index 8ebb754b381..ea8c123d1a6 100644 --- a/docs/api/api_python/mindspore.context.rst +++ b/docs/api/api_python/mindspore.context.rst @@ -99,7 +99,7 @@ MindSpore context,用于配置当前执行环境,包括执行模式、执行 内存重用: - **mem_Reuse**:表示内存复用功能是否打开。设置为True时,将打开内存复用功能。设置为False时,将关闭内存复用功能。 - 有关running data recoder和内存复用配置详细信息,请查看 `配置RDR和内存复用 `_。 + 有关running data recoder和内存复用配置详细信息,请查看 `配置RDR和内存复用 `_。 - **precompile_only** (bool) - 表示是否仅预编译网络。默认值:False。设置为True时,仅编译网络,而不执行网络。 diff --git a/docs/api/api_python/mindspore/mindspore.build_searched_strategy.rst b/docs/api/api_python/mindspore/mindspore.build_searched_strategy.rst index 8a04e3a27c6..9dbea494732 100644 --- a/docs/api/api_python/mindspore/mindspore.build_searched_strategy.rst +++ b/docs/api/api_python/mindspore/mindspore.build_searched_strategy.rst @@ -3,7 +3,7 @@ mindspore.build_searched_strategy .. py:class:: mindspore.build_searched_strategy(strategy_filename) - 构建网络中每个参数的策略,用于分布式推理。关于它的使用细节,请参考: `保存和加载模型(HyBrid Parallel模式) `_。 + 构建网络中每个参数的策略,用于分布式推理。关于它的使用细节,请参考: `保存和加载模型(HyBrid Parallel模式) `_。 **参数:** diff --git a/docs/api/api_python/mindspore/mindspore.merge_sliced_parameter.rst b/docs/api/api_python/mindspore/mindspore.merge_sliced_parameter.rst index 4c5b2f72259..459032f1858 100644 --- a/docs/api/api_python/mindspore/mindspore.merge_sliced_parameter.rst +++ b/docs/api/api_python/mindspore/mindspore.merge_sliced_parameter.rst @@ -3,7 +3,7 @@ mindspore.merge_sliced_parameter .. py:method:: mindspore.merge_sliced_parameter(sliced_parameters, strategy=None) - 将参数切片合并为一个完整的参数,用于分布式推理。关于它的细节,请参考:`保存和加载模型(HyBrid Parallel模式) `_。 + 将参数切片合并为一个完整的参数,用于分布式推理。关于它的细节,请参考:`保存和加载模型(HyBrid Parallel模式) `_。 **参数:** diff --git a/docs/api/api_python/train/mindspore.train.callback.Callback.rst b/docs/api/api_python/train/mindspore.train.callback.Callback.rst index dea2d583db1..2c58862e312 100644 --- a/docs/api/api_python/train/mindspore.train.callback.Callback.rst +++ b/docs/api/api_python/train/mindspore.train.callback.Callback.rst @@ -5,7 +5,7 @@ Callback函数可以在step或epoch开始前或结束后执行一些操作。 要创建自定义Callback,需要继承Callback基类并重载它相应的方法,有关自定义Callback的详细信息,请查看 - `Callback `_。 + `Callback `_。 .. py:method:: begin(run_context) diff --git a/mindspore/ccsrc/plugin/device/ascend/kernel/tbe/tbe_utils.cc b/mindspore/ccsrc/plugin/device/ascend/kernel/tbe/tbe_utils.cc index 3dbabbe68d0..a4a5336b1e2 100644 --- a/mindspore/ccsrc/plugin/device/ascend/kernel/tbe/tbe_utils.cc +++ b/mindspore/ccsrc/plugin/device/ascend/kernel/tbe/tbe_utils.cc @@ -122,7 +122,7 @@ std::string GetOpDebugLevel() { if (!TbeUtils::IsOneOf(value_ranges, std::stoul(env_level.c_str()))) { MS_LOG(WARNING) << "Invalid environment variable '" << kCOMPILER_OP_LEVEL << "': " << env_level << ", the value should be in [0, 1, 2, 3, 4], now using the default value 3." - "Get more detail info at https://www.mindspore.cn/docs/note/zh-CN/master/env_var_list.html"; + "Get more detail info at https://www.mindspore.cn/docs/zh-CN/master/note/env_var_list.html"; } else { op_debug_level = env_level; } diff --git a/mindspore/python/mindspore/ops/operations/comm_ops.py b/mindspore/python/mindspore/ops/operations/comm_ops.py index 369a4d3f83f..e1dc70afe97 100644 --- a/mindspore/python/mindspore/ops/operations/comm_ops.py +++ b/mindspore/python/mindspore/ops/operations/comm_ops.py @@ -664,7 +664,7 @@ class NeighborExchange(Primitive): This operator requires a full-mesh network topology, each device has the same vlan id, and the ip & mask are in the same subnet, please check the `details \ - `_. + `_. Args: send_rank_ids (list(int)): Ranks which the data is sent to. @@ -740,7 +740,7 @@ class AlltoAll(PrimitiveWithInfer): This operator requires a full-mesh network topology, each device has the same vlan id, and the ip & mask are in the same subnet, please check the `details \ - `_. + `_. Args: split_count (int): On each process, divide blocks into split_count number. @@ -831,7 +831,7 @@ class NeighborExchangeV2(Primitive): This operator requires a full-mesh network topology, each device has the same vlan id, and the ip & mask are in the same subnet, please check the `details \ - `_. + `_. Args: send_rank_ids (list(int)): Ranks which the data is sent to. 8 rank_ids represents 8 directions, if one diff --git a/mindspore/python/mindspore/train/callback/_callback.py b/mindspore/python/mindspore/train/callback/_callback.py index 99b211bd5eb..049e7b637b0 100644 --- a/mindspore/python/mindspore/train/callback/_callback.py +++ b/mindspore/python/mindspore/train/callback/_callback.py @@ -82,7 +82,7 @@ class Callback: Callback function can perform some operations before and after step or epoch. To create a custom callback, subclass Callback and override the method associated with the stage of interest. For details of Callback fusion, please check - `Callback `_. + `Callback `_. Examples: >>> import numpy as np @@ -249,7 +249,7 @@ class RunContext: Callback objects can stop the loop by calling request_stop() of run_context. This class needs to be used with :class:`mindspore.train.callback.Callback`. For details of Callback fusion, please check - `Callback `_. + `Callback `_. Args: original_args (dict): Holding the related information of model. diff --git a/mindspore/python/mindspore/train/serialization.py b/mindspore/python/mindspore/train/serialization.py index c885da2961c..b68f5404f40 100644 --- a/mindspore/python/mindspore/train/serialization.py +++ b/mindspore/python/mindspore/train/serialization.py @@ -1396,7 +1396,7 @@ def build_searched_strategy(strategy_filename): """ Build strategy of every parameter in network. Used in the case of distributed inference. For details of it, please check: - ``_. + ``_. Args: strategy_filename (str): Name of strategy file. @@ -1447,7 +1447,7 @@ def merge_sliced_parameter(sliced_parameters, strategy=None): """ Merge parameter slices into one parameter. Used in the case of distributed inference. For details of it, please check: - ``_. + ``_. Args: sliced_parameters (list[Parameter]): Parameter slices in order of rank id.