update release notes for r0.7.0

This commit is contained in:
jonyguo 2020-08-28 15:12:56 +08:00
parent 6c68e70cfe
commit b0f847072c
1 changed files with 115 additions and 5 deletions

View File

@ -1,3 +1,113 @@
# Release 0.7.0-beta
## Major Features and Improvements
### MindSpore Training and Inference Framework
#### Ascend 910
* New models
* TinyBert: a smaller and faster version of BERT using transformer distillation for natural language understanding on GLUE benchmark.
* SE-ResNet50: add Squeeze-and-Excitation blocks(SE-Blocks) to the resnet50 network to improve channel interdependencies for image classification on ImageNet 2012 dataset.
* Inception V3: the third version of Inception convolutional architectures for image classification on ImageNet 2012 dataset.
* Frontend and user interface
* Embedding operator high-level packaging to support segmented by field for Wide&Deep.
* Load multi-node checkpoint into single-process to support host-device hybrid inference.
* Support Concat/Tile/Strideslice distributed operators.
* Support cumulative gradient and batch training split.
* Support variable parameter input for Cell object.
* Parameter mixed calculation optimization for pynative mode.
* Deep Probabilistic Programming
* Support statistical distributions classes used to generate stochastic tensors.
* Support probabilistic inference algorithms.
* Support BNN layers used to construct BNN in Graph mode.
* Support interfaces for the transformation between BNN and DNN in Graph mode.
* Support uncertainty estimation to estimate epistemic uncertainty and aleatoric uncertainty.
* User interfaces change log
* change base class of parameter([!3473](https://gitee.com/mindspore/mindspore/pulls/3473))
* change binary to mindir([!4258](https://gitee.com/mindspore/mindspore/pulls/4258))
* change export from geir to air([!4269](https://gitee.com/mindspore/mindspore/pulls/4269))
* Init parameter data by default([!3967](https://gitee.com/mindspore/mindspore/pulls/3967))
* change IndexedSlices to RowTensor([!4031](https://gitee.com/mindspore/mindspore/pulls/4031))
* Must set or change parallel mode before any Initializer created([!4801](https://gitee.com/mindspore/mindspore/pulls/4801))
* Executor and performance optimization
* Minspore graph compilation process performance improved by 20%.
* Decoupling C++ and Python modules to achieve separate compilation of core modules.
* Serving module supports restful interface [POST].
* Data processing, augmentation, and save format
* Support automatic data augmentation
* Support GNN distributed cache in single node
* Support ConcatDataset using distributed sampler
#### Other Hardware Support
* GPU platform
* New model supported: VGG16, ResNet101, TinyBert, DeepFM.
* Support some distributed operators in ResNet50 and Wide&Deep.
* Support automatic parallel for Wide&Deep.
* Support function funcs[i](*inputs) (such as switch-case).
* Support distributed training with parameter server.
* Support GPU operator profiling.
* Performance optimization of the distributed training with allreduce.
* Performance optimization of the mixed precision training.
* Performance optimization of the pynative mode.
* Performance optimization of the convolution operator, batch normalization operator.
* CPU platform
* Support MobileNetV2 Re-Training: Re-train the network with different class number.
### MindSpore Lite
* Converter
* Support third party model, including TFLite/Caffe/ONNX.
* Add 93 TFLite op.
* Add 24 Caffe op.
* Add 62 ONNX op.
* Add support for windows.
* Add 11 optimized passes, include fusion/const fold.
* Support aware-training and Post-training quantization.
* CPU
* Add 100+opssupport fp32, int8/uint8, FP16 ops
* Support fast convolution algorithms: Sliding Window, Img2col + Gemm, Strassen, Winograd
* Support assembly/neon instruction.
* Support CPU fp16 and sdot on ARM v8.2+.
* GPU
* Add 20+ ops for OpenCL.
* Support image2D/buffer format.
* Optimize online initialization time.
* add optimized convolution1X1/3X3/depthwise/convolution_transposed for OpenCL.
* Tool & example
* Add benchmark and TimeProfile tools.
* Add image classification and object detection Android Demo.
## Bugfixes
* Models
* normalize the readme file([!5410](https://gitee.com/mindspore/mindspore/pulls/5410))
* fix a sink_size bug for transformer([!5393](https://gitee.com/mindspore/mindspore/pulls/5393))
* fix bool type optional for resnet50([!5363](https://gitee.com/mindspore/mindspore/pulls/5363))
* Python API
* improve interface '__bool__' for tensor([!4000](https://gitee.com/mindspore/mindspore/pulls/4000))
* fix GPU-ResizeNearestNeighbor([!3760](https://gitee.com/mindspore/mindspore/pulls/3760))
* fix topK multi dimention grad func([!3711](https://gitee.com/mindspore/mindspore/pulls/3711))
* fix scatterop error msg([!3699](https://gitee.com/mindspore/mindspore/pulls/3699))
* fix bug of cast dtype when using mix_presion in pynative mode([!3730](https://gitee.com/mindspore/mindspore/pulls/3730))
* Executor
* fix etsnet train error when UnsegmentSum's first input shape is (1,) ([!4573](https://gitee.com/mindspore/mindspore/pulls/4573))
* fix bug of result error in while control flow because of unsupporting for value reference ([!4103](https://gitee.com/mindspore/mindspore/pulls/4103))
* fix bug of the output tensor does not carry device data type ([!3774](https://gitee.com/mindspore/mindspore/pulls/3774))
* fix bug of avoiding multi attr value are eliminated in pynative mode ([!4225](https://gitee.com/mindspore/mindspore/pulls/4225))
* fix bug of AssignAdd unable to work normally in multi-cases ([!5171](https://gitee.com/mindspore/mindspore/pulls/5171))
* GPU platform
* improve the environment variable checking for nvcc compiler path ([!5140](https://gitee.com/mindspore/mindspore/pulls/5140))
* fix bug of error in cast operator conversion from fp16 to fp32 ([!4147](https://gitee.com/mindspore/mindspore/pulls/4147))
* fix bug of the array out of bound in case of make_tuple operator ([!5219](https://gitee.com/mindspore/mindspore/pulls/5219))
* Data processing and Pro
* fix GeneratorDataset time out([!3624](https://gitee.com/mindspore/mindspore/pulls/3624))
* fix concat operator get_dataset_size error([!4701](https://gitee.com/mindspore/mindspore/pulls/4701))
* fixing python validator for Repeat Op([!4366](https://gitee.com/mindspore/mindspore/pulls/4366))
## Contributors
Thanks goes to these wonderful people:
Adel, Alexey, andy, andy_wangrui, anthonyaje, anzhengqi, askmiao, avakh, baihuawei, bingyaweng, BowenK, buxue, caifubi, CaoJian, caozhou, Cathy, changzherui, chenfei, chengxianbin, chenhaozhe, chenjianping, chentingting, chenzomi, chenzupeng, chujinjin, cjh9368, Corleone, cristoval, danish, dengyutao, eric, Eric, ervinzhang, etone-chan, fangzehua, fary86, fuzhiye, gengdongjie, genglishuai, Giancarlo, gongdaguo, gukecai, guohongzilong, GuoMengHao, hangq, hanhaocheng, hanhuifeng2020, hanjun996, Harshvardhan, He, heleiwang, hesham, hexia, Hoai, hongxing, huangdongrun, huanghui, huangxinjing, islam_amin, Jesse, jianghui58, jiangzhiwen, jin-xiulang, jinyaohui, jjfeing, John, Jonathan, jonyguo, kai00, kingfo, kpy, kswang, laiyongqiang, leilei_snow, leopz, Li, liangzelang, lianliguang, lichen_101010, lichenever, lihongkang, lilei, limingqi107, ling, lingyunli63, linqingke, lirongzhen1, liubuyu, liuwenhao4, liuxiao78, liuxiao93, liuzhongkai, Lixia, lixian, liyong, lizhenyu, looop5, luoyang, lvchangquan, lvliang, lvwenyuan, lyvette, mahdi, Mahdi, mamba_ni, maning202007, Margaret_wangrui, mayang, meixiaowei, meng_chunyang, ms_yan, nhussain, panbingao, panfengfeng, panyifeng, Payne, Peilin, peixu_ren, pengyongrong, Pengyongrong, qianlong, qujianwei, root, shenwei41, shibeiji, simson, songhonglei413, Su, sunsuodong, suteng, tao_yunhao, TFbunny, tinazhang, tom__chen, tony_liu2, tronzhang, VectorSL, wandongdong, wangdongxu, wanghua, wangmin, wangshaocong, wangzhe, wanyiming, Wei, wenchunjiang, wilfChen, WilliamLian, wsc, wukesong, wuweikang, wuxuejian, wuyongkang, xiefangqi, xuanyue, Xun, xutianchun, xuyongfei, yanghaitao, yangjie159, YangLuo, yangruoqi713, yangyongjie, yangzhenzhang, yankai, yao_yf, yelihua, yeyunpeng, Yi, yoni, yoonlee666, yuchaojie, yujianfeng, yuximiao, zhangxuetong, zhaizhiqiang, Zhang, zhangxinfeng3, zhangxuetong, zhangyihui, zhangz0911gm, zhanke, zhanyuan, zhaodezan, zhaoting, zhaozhenlong, zhengjun10, zhongligeng, zhoufeng, zhousiyi, zhouyaqiang, zhouyuanshen, Zichun, Zirui, zjun, zongha, ZPaC, lijiaqi, liangchenghui, wangminggui
Contributions of any kind are welcome!
# Release 0.6.0-beta # Release 0.6.0-beta
## Major Features and Improvements ## Major Features and Improvements
@ -207,11 +317,11 @@ Contributions of any kind are welcome!
* Executor * Executor
* Fix dropouttopK and addn errors in PyNative mode ([!1285](https://gitee.com/mindspore/mindspore/pulls/1285), [!1138](https://gitee.com/mindspore/mindspore/pulls/1138), [!1033](https://gitee.com/mindspore/mindspore/pulls/1033)). * Fix dropouttopK and addn errors in PyNative mode ([!1285](https://gitee.com/mindspore/mindspore/pulls/1285), [!1138](https://gitee.com/mindspore/mindspore/pulls/1138), [!1033](https://gitee.com/mindspore/mindspore/pulls/1033)).
* Fix memory leaks after execution in PyNatvie mode ([!1201](https://gitee.com/mindspore/mindspore/pulls/1201)). * Fix memory leaks after execution in PyNatvie mode ([!1201](https://gitee.com/mindspore/mindspore/pulls/1201)).
* Fix HCCL failure in some special scenes ([!1204](https://gitee.com/mindspore/dashboard/projects/mindspore/mindspore/pulls/1204), [!1252](https://gitee.com/mindspore/dashboard/projects/mindspore/mindspore/pulls/1252)). * Fix HCCL failure in some special scenes ([!1204](https://gitee.com/mindspore/mindspore/pulls/1204), [!1252](https://gitee.com/mindspore/mindspore/pulls/1252)).
* Fix SSD network when Select failed, cann't find kernel info([!1449](https://gitee.com/mindspore/dashboard/projects/mindspore/mindspore/pulls/1449)). * Fix SSD network when Select failed, cann't find kernel info([!1449](https://gitee.com/mindspore/mindspore/pulls/1449)).
* Fix Topk operator selection strategy bug between aicore and aicpu([!1367](https://gitee.com/mindspore/dashboard/projects/mindspore/mindspore/pulls/1367)). * Fix Topk operator selection strategy bug between aicore and aicpu([!1367](https://gitee.com/mindspore/mindspore/pulls/1367)).
* Fix input memory size of 'assign' op unequal in control sink mode when assigning a data from one child graph to another child graph([!802](https://gitee.com/mindspore/dashboard/projects/mindspore/mindspore/pulls/802)). * Fix input memory size of 'assign' op unequal in control sink mode when assigning a data from one child graph to another child graph([!802](https://gitee.com/mindspore/mindspore/pulls/802)).
* Fix allreduce ir inconsistency([!989](https://gitee.com/mindspore/dashboard/projects/mindspore/mindspore/pulls/989)). * Fix allreduce ir inconsistency([!989](https://gitee.com/mindspore/mindspore/pulls/989)).
* GPU platform * GPU platform
* Fix summary for gradient collection ([!1364](https://gitee.com/mindspore/mindspore/pulls/1364)) * Fix summary for gradient collection ([!1364](https://gitee.com/mindspore/mindspore/pulls/1364))
* Fix the slice operator ([!1489](https://gitee.com/mindspore/mindspore/pulls/1489)) * Fix the slice operator ([!1489](https://gitee.com/mindspore/mindspore/pulls/1489))