forked from mindspore-Ecosystem/mindspore
fix format of release
Signed-off-by: Ting Wang <kathy.wangting@huawei.com>
This commit is contained in:
parent
9c255de251
commit
2083014879
2
OWNERS
2
OWNERS
|
@ -27,7 +27,7 @@ files:
|
|||
- guoqi1024
|
||||
- baochong
|
||||
|
||||
".*\.md$":
|
||||
".*\\.md$":
|
||||
approvers:
|
||||
- gemini524
|
||||
- Hanshize
|
||||
|
|
14
RELEASE.md
14
RELEASE.md
|
@ -17,7 +17,7 @@
|
|||
|
||||
- [STABLE] Support dynamic weight decay for optimizers, that is weight decay value will change according to the increasing step during training.
|
||||
- [STABLE] Add four methods to create Tensor, which are `mindspore.numpy.rand()`, `mindspore.numpy.randn()`, `mindspore.numpy.randint()`, and `mindspore.ops.arange()`.
|
||||
- [STABLE] Add `mindspore.callback.History` in Callback.
|
||||
- [STABLE] Add `mindspore.train.callback.History` in Callback.
|
||||
- [BETA] Support custom operator implemented by Julia operator.
|
||||
- [STABLE] Support accessing attributes and methods of user-defined classes through `mindspore.ms_class` class decorator.
|
||||
- [STABLE] Support training when a network has side effect operations and control flow statements at the same time.
|
||||
|
@ -46,7 +46,7 @@
|
|||
|
||||
#### Executor
|
||||
|
||||
- [BETA] [Failure Recovery Under Data Parallel Training](https://www.mindspore.cn/tutorials/experts/en/master/parallel/train_gpu.html#%E5%AE%B9%E7%81%BE%E6%81%A2%E5%A4%8D) Support auto failure recovery under data parallel training mode.
|
||||
- [BETA] [Failure Recovery Under Data Parallel Training](https://www.mindspore.cn/tutorials/experts/zh-CN/master/parallel/train_gpu.html#%E5%AE%B9%E7%81%BE%E6%81%A2%E5%A4%8D) Support auto failure recovery under data parallel training mode.
|
||||
- [BETA] Support searching for the number of threads under the CPU to obtain the optimal number of threads for execution. The entire search process takes 50 steps, and the overall performance will reach a stable state after 50 steps. When testing performance, data after 50 steps need to be used as a standard.
|
||||
|
||||
#### DataSet
|
||||
|
@ -55,8 +55,8 @@
|
|||
- [STABLE] Python multiprocessing optimization and make processes exit normally.
|
||||
- [STABLE] Support [Dataset Autotune](https://www.mindspore.cn/tutorials/experts/en/master/debug/dataset_autotune.html) for tuning the speed of dataset pipeline automatically.
|
||||
- [BETA] [Dataset Offload](https://www.mindspore.cn/docs/en/master/design/dataset_offload.html) support new data augmentation operations: RandomColorAdjust, RandomSharpness, TypeCast.
|
||||
- Output a single data column when __getitem__/__next__ methods of GeneratorDataset return a single NumPy object.
|
||||
- Use ulimit -u 10240 to increase the number of threads/processes available to the current user when specify too many processes or threads for loading dataset may cause RuntimeError: can't start new thread.
|
||||
- Output a single data column when `__getitem__/__next__` methods of GeneratorDataset return a single NumPy object.
|
||||
- Use `ulimit -u 10240` to increase the number of threads/processes available to the current user when specify too many processes or threads for loading dataset may cause RuntimeError: can't start new thread.
|
||||
|
||||
### API Change
|
||||
|
||||
|
@ -65,9 +65,9 @@
|
|||
##### Python API
|
||||
|
||||
- Modify the gradient return value type of the hook corresponding to the register_backward_hook function, and change the gradient return value to the tuple type uniformly.([!31876](https://gitee.com/mindspore/mindspore/pulls/31876))
|
||||
- Deprecated usage: `import mindspore.dataset.engine.datasets as ds`. Use `import mindspore.dataset as ds` instead as recommended in [mindspore doc](https://www.mindspore.cn/docs/zh-CN/master/api_python/mindspore.dataset.html).
|
||||
- Deprecated usage: `import mindspore.dataset.engine.datasets as ds`. Use `import mindspore.dataset as ds` instead as recommended in [mindspore doc](https://www.mindspore.cn/docs/en/master/api_python/mindspore.dataset.html).
|
||||
- Add `mindspore.ms_class` interface, as class decorator for user-defined classes. It allows MindSpore to identify user-defined classes and access their attributes and methods([!30855](https://gitee.com/mindspore/mindspore/pulls/30855))
|
||||
- Deprecate `mindspore.SparseTensor` and use `mindspore.COOTensor` instead. ([!28505]())
|
||||
- Deprecate `mindspore.SparseTensor` and use `mindspore.COOTensor` instead. ([!28505](https://gitee.com/mindspore/mindspore/pulls/28505))
|
||||
|
||||
## MindSpore Lite
|
||||
|
||||
|
@ -175,7 +175,7 @@ Contributions of any kind are welcome!
|
|||
|
||||
###### The `set_auto_parallel_context()` interface of context is changed from
|
||||
|
||||
###### `set_auto_parallel_context(parallel_mode=AUTO_PARALLEL, auto_parallel_search_mode="dynamic_programming")` to ` set_auto_parallel_context(parallel_mode=AUTO_PARALLEL, search_mode="dynamic_programming")`
|
||||
###### `set_auto_parallel_context(parallel_mode=AUTO_PARALLEL, auto_parallel_search_mode="dynamic_programming")` to `set_auto_parallel_context(parallel_mode=AUTO_PARALLEL, search_mode="dynamic_programming")`
|
||||
|
||||
#### Collect Data and Create Landscape
|
||||
|
||||
|
|
|
@ -9,7 +9,7 @@
|
|||
#### OS
|
||||
|
||||
- [STABLE] 支持Python 3.8版本(Linux/Windows/Mac)。
|
||||
- [STABLE] 安装简化,提供详细安装指南以及自动化安装脚本。
|
||||
- [STABLE] 简化安装,提供详细安装指南和自动化安装脚本。
|
||||
- [STABLE] Windows版本支持算子多线程。
|
||||
- [STABLE] GCC兼容7.3到9.x版本。
|
||||
|
||||
|
@ -17,7 +17,7 @@
|
|||
|
||||
- [STABLE] 优化器支持动态权重衰减,即训练期间权重衰减值随着step的增加而变化。
|
||||
- [STABLE] 增加四种创建Tensor的方法,分别是`mindspore.numpy.rand()`、`mindspore.numpy.randn()`、`mindspore.numpy.randint()`和`mindspore.ops.arange ()`。
|
||||
- [STABLE] 增加一种callback方法 `mindspore.callback.History`。
|
||||
- [STABLE] 增加一种callback方法 `mindspore.train.callback.History`。
|
||||
- [BETA] 自定义算子支持Julia算子。
|
||||
- [STABLE] 通过 `mindspore.ms_class` 类装饰器,支持获取用户自定义类的属性和方法。
|
||||
- [STABLE] 支持同时存在副作用算子和控制流语句的网络的训练。
|
||||
|
@ -26,17 +26,17 @@
|
|||
|
||||
#### PyNative
|
||||
|
||||
- [STABLE] 在PyNative模式下支持hook函数功能,包括前向hook接口register_forward_pre_hook, register_forward_hook和反向hook接口register_backward_hook。
|
||||
- [STABLE] 优化PyNative模式执行性能,将前端Python与后端C++进行并行执行。
|
||||
- [STABLE] 在PyNative模式下支持hook函数功能,包括前向hook接口register_forward_pre_hook、register_forward_hook和反向hook接口register_backward_hook。
|
||||
- [STABLE] 优化PyNative模式执行性能,并行执行前端Python与后端C++。
|
||||
|
||||
#### Auto Parallel
|
||||
|
||||
- [STABLE] 在MoE场景中支持TopK的路由、数据并行和优化器切分。
|
||||
- [STABLE] 支持AllGather/ReduceScatter通信算子融合, 在DATA_PARALLEL模式支持AllReduce按数据量大小编译。
|
||||
- [STABLE] 支持AllGather/ReduceScatter通信算子融合,在DATA_PARALLEL模式支持AllReduce按数据量大小编译。
|
||||
- [STABLE] 在并行模式下支持ops.clip_by_global_norm。
|
||||
- [STABLE] 在并行模式下支持AdaSum优化器。
|
||||
- [STABLE] 支持自动优化器切分。
|
||||
- [STABLE] 支持AllltoAll可配置开启.支持自动插入VirtualDatasetCell。
|
||||
- [STABLE] 支持AlltoAll可配置开启,支持自动插入VirtualDatasetCell。
|
||||
- [STABLE] 在流水线并行训练中,支持自动推断可训练的参数。
|
||||
- [STABLE] 支持集群的设备数目不为2的幂次方。
|
||||
- [STABLE] 在自动并行模式中支持策略传播。
|
||||
|
@ -47,16 +47,16 @@
|
|||
#### Executor
|
||||
|
||||
- [BETA] [数据并行训练容灾](https://www.mindspore.cn/tutorials/experts/zh-CN/master/parallel/train_gpu.html#%E5%AE%B9%E7%81%BE%E6%81%A2%E5%A4%8D) 支持多卡数据并行训练容灾恢复。
|
||||
- [BETA] 支持在cpu下的线程数搜索,获取最优线程数进行执行。整个搜索过程需要耗时50个steps,整体的性能会在50个steps后达到稳定的状态。在测试性能的时候,需要以50个steps之后的数据作为标准。
|
||||
- [BETA] 支持在CPU下的线程数搜索,获取最优线程数来执行。整个搜索过程需要耗时50个steps,整体的性能会在50个steps后达到稳定的状态。在测试性能的时候,需要以50个steps之后的数据作为标准。
|
||||
|
||||
#### DataSet
|
||||
|
||||
- [STABLE] 增加了数据处理API的差异文档,比较TensorFlow.data与MindSpore.dataset部分算子的差异,详见 [对比文档](https://www.mindspore.cn/docs/zh-CN/master/note/api_mapping/tensorflow_api_mapping.html#tf-data).
|
||||
- [STABLE] 增加了数据处理API的差异文档,比较TensorFlow.data与MindSpore.dataset部分算子的差异,详见 [对比文档](https://www.mindspore.cn/docs/zh-CN/master/note/api_mapping/tensorflow_api_mapping.html#tf-data)。
|
||||
- [STABLE] Python多进程逻辑优化,保证不同异常场景的正常退出。
|
||||
- [STABLE] 支持[自动数据加速](https://www.mindspore.cn/tutorials/experts/zh-CN/master/debug/dataset_autotune.html),可以自适应调节数据处理管道的执行速度。
|
||||
- [BETA] [数据处理异构加速](https://www.mindspore.cn/docs/zh-CN/master/design/dataset_offload.html) 支持了新的数据增强操作: RandomColorAdjust, RandomSharpness, TypeCast。
|
||||
- GeneratorDataset加载自定义数据集时,当__getitem__/__next__方法返回单个NumPy对象,对应会输出单个数据列。
|
||||
- 用户在数据预处理中使用过多的进程数/线程数情况下,会出现错误RuntimeError: can't start new thread,可以通过 ulimit -u 10240 增加当前用户可用的线程/进程数解决。
|
||||
- [BETA] [数据处理异构加速](https://www.mindspore.cn/docs/zh-CN/master/design/dataset_offload.html) 支持了新的数据增强操作: RandomColorAdjust、RandomSharpness和TypeCast。
|
||||
- GeneratorDataset加载自定义数据集时,当`__getitem__/__next__`方法返回单个NumPy对象,对应会输出单个数据列。
|
||||
- 用户在数据预处理中使用过多的进程数/线程数情况下,会出现错误RuntimeError: can't start new thread,可以通过 `ulimit -u 10240` 增加当前用户可用的线程/进程数解决。
|
||||
|
||||
### API变更
|
||||
|
||||
|
@ -65,9 +65,9 @@
|
|||
##### Python API
|
||||
|
||||
- 修改register_backward_hook功能对应hook的梯度返回值类型,将梯度返回值统一改成tuple类型。([!31876](https://gitee.com/mindspore/mindspore/pulls/31876))
|
||||
- 弃用的import用法: `import mindspore.dataset.engine.datasets as ds` ,因其import目录过深且过度依赖Python目录结构。推荐使用官方推荐用法 `import mindspore.dataset as ds` ,更多参考详见 [API文档](https://www.mindspore.cn/docs/zh-CN/master/api_python/mindspore.dataset.html).
|
||||
- 弃用的import用法: `import mindspore.dataset.engine.datasets as ds` ,因其import目录过深且过度依赖Python目录结构。推荐使用 `import mindspore.dataset as ds` ,更多参考详见 [API文档](https://www.mindspore.cn/docs/zh-CN/master/api_python/mindspore.dataset.html)。
|
||||
- 新增`mindspore.ms_class` 接口,作为用户自定义类的类装饰器,使得MindSpore能够识别用户自定义类,并且获取这些类的属性和方法。([!30855](https://gitee.com/mindspore/mindspore/pulls/30855))
|
||||
- `mindspore.SparseTensor`接口废弃使用,对应新接口为`mindspore.COOTensor`. ([!28505]())
|
||||
- `mindspore.SparseTensor`接口废弃使用,对应新接口为`mindspore.COOTensor`。 ([!28505](https://gitee.com/mindspore/mindspore/pulls/28505))
|
||||
- Tensor新增一个入参`internal`,作为框架内部使用。
|
||||
|
||||
## MindSpore Lite
|
||||
|
@ -76,8 +76,8 @@
|
|||
|
||||
#### 后量化
|
||||
|
||||
- [STABLE] 后量化支持动态量化算法.
|
||||
- [BETA] 后量化模型支持在英伟达GPU推理.
|
||||
- [STABLE] 后量化支持动态量化算法。
|
||||
- [BETA] 后量化模型支持在英伟达GPU上执行推理。
|
||||
|
||||
### 贡献者
|
||||
|
||||
|
|
Loading…
Reference in New Issue