Compare commits

...

397 Commits
master ... r0.7

Author SHA1 Message Date
mindspore-ci-bot 2139c7ddc6 !8172 【轻量级 PR】:update RELEASE.md.
5 years ago
mindspore-ci-bot c4e3fd70ee !8181 fix securec download links due to mistakes made by openeuler community
5 years ago
yanghaoran 28372fbb96 update graphengine, fix securec download links
5 years ago
shenwei41 81c2068635 update RELEASE.md.
5 years ago
mindspore-ci-bot 4a8f92e4be !6799 mobilenetv2 update and debug
6 years ago
Payne c53cd6bb22 align to docs and r1.0
6 years ago
mindspore-ci-bot 297e65e162 !6579 modify README_CN.md
6 years ago
wuweikang 1729a27bdb modify README_CN.md
6 years ago
mindspore-ci-bot 38da39b0be !6544 modify Dockerfile for devel
6 years ago
mindspore-ci-bot 56e22a5c82 !6535 【MD】r0.7 Branch: MD5 value update in the file - icu4c.cmake of branch r0.7
6 years ago
wuweikang b551623c86 modify Dockerfile for devel
6 years ago
mayang 5529d41769 MD5 value update in the file icu4c.cmake of branch r0.7
6 years ago
mindspore-ci-bot ab997f9e37 !6299 mobilenetv2 debug for load ckpt
6 years ago
mindspore-ci-bot 5027130939 !6360 modify Dockerfile for GPU version
6 years ago
wuweikang 30450e2177 modify Dockerfile for GPU version
6 years ago
Payne 93c4d2929c load_ckpt debug
6 years ago
mindspore-ci-bot e39775edfd !6218 [MS][LITE]delete thread header for fix 0.7r java api bugs
6 years ago
mindspore-ci-bot e0953c47c3 !6213 [MS][LITE]fix the compilation failure
6 years ago
mindspore-ci-bot 8e59d7449e !6187 make ref node's update edge using same address
6 years ago
yeyunpeng 41fd02baf3 delete thread header for fix 0.7r java api bugs
6 years ago
gongdaguo 02ba93492d fix the compilation failure and change the model download address
6 years ago
WilliamLian 683e4aba74 make ref edage using same address
6 years ago
mindspore-ci-bot d373efcf99 !6129 mobilenetv2 add incremental_learn func
6 years ago
Payne 4e2ccefe29 mobilenetv2 add incremental learning func
6 years ago
mindspore-ci-bot 6b4c5227ee !5942 adjust Dockerfile for nvcc support
6 years ago
mindspore-ci-bot 738ade501d !6014 update info with new platform
6 years ago
mindspore-ci-bot 527d36597c !6010 fix bug of aicpu device ptr release failed
6 years ago
jonyguo 7132be9ed2 update info with platform
6 years ago
caifubi 1c97d65f40 fix bug of aicpu device ptr release failed
6 years ago
mindspore-ci-bot 0ce4e3418c !5934 update supported platform info
6 years ago
wuweikang b83aa68071 adjust Dockerfile for nvcc support
6 years ago
jonyguo 86f5189323 update supported platform info
6 years ago
mindspore-ci-bot fc4bf19294 !5820 fix ub fusion
6 years ago
jjfeing bd1261e5e6 adapt tensorengin modify, fix ub fusion
6 years ago
mindspore-ci-bot 5de9578abb !5794 rewrite fp16 to fp32
6 years ago
cjh9368 16f41b1619 rewrite fp16 to fp32 and fp32 to fp16
6 years ago
mindspore-ci-bot 7abe5a231f !5764 fix release note word err
6 years ago
jonyguo 27807a21ea fix: RELEASE.md spelling word error
6 years ago
mindspore-ci-bot 40bf2493fa !5490 modify the format info of tensorAdd
6 years ago
mindspore-ci-bot a585177825 !5575 update run for br: r0.7
6 years ago
wuxuejian 06a6af7628 update aicpu proto and update module: graphengine
6 years ago
mindspore-ci-bot 9833a00e93 !5450 Fixbugfix for server shard range computation
6 years ago
mindspore-ci-bot 13fcfcdb78 !5592 [MS][Googlenet][README]googlenet readme have somewhere need to correct
6 years ago
CaoJian 96fad3b0a4 googlenet README update
6 years ago
mindspore-ci-bot a96646fc5a !5569 change readme file in yolov3_darknet53
6 years ago
mindspore-ci-bot 26aba3b74c !5499 Support manual convert to quantative network of resnet
6 years ago
mindspore-ci-bot 5120720bac !5562 bert scripts enhancement
6 years ago
mindspore-ci-bot 4382ce202c !5555 Modify read me of inceptionv3
6 years ago
mindspore-ci-bot 29fabd1324 !5561 Fix C++ coding standard problem
6 years ago
mindspore-ci-bot a43bd07c39 !5560 clean static check
6 years ago
mindspore-ci-bot d332e4d2b7 !5581 clean static check
6 years ago
mindspore-ci-bot 403cb700b2 !5565 update RELEASE
6 years ago
cristoval b2ffdf82c6 bugfix for server shard range computation
6 years ago
mindspore-ci-bot f5cba099d0 !5567 Delete some useless comments
6 years ago
mindspore-ci-bot 9d10d934c7 !5557 [MS][LITE][Develop] fix bug of arm cpu int8 op: pooling memory leak
6 years ago
lyvette 62e7746e77 clean
6 years ago
chenfei 298393b66b add manual quantative network of resnet50
6 years ago
meng_chunyang 26bf044dde update RELEASE
6 years ago
cjh9368 fb973778e5 fix bug
6 years ago
yoonlee666 d472ee3921 bert script enhancement
6 years ago
yangyongjie e0a128142b fix performance of yolov3_darknet53 in readme file.
6 years ago
yeyunpeng 794aeb8e2c fix static check problem
6 years ago
zhanyuan 6051e3cfec Delete some useless comments
6 years ago
mindspore-ci-bot 03ff5f334a !5487 add schema file for BERT and TinyBERT
6 years ago
yangruoqi713 dbe5cd9091 [MS][LITE][Develop] fix bug of arm cpu int8 op: pooling
6 years ago
zhouyaqiang 02fe2f9f1d modify read me of inception v3
6 years ago
mindspore-ci-bot 82dba19adf !5546 update README
6 years ago
meng_chunyang e256877be9 update README
6 years ago
mindspore-ci-bot 433eaab225 !5544 remove unused benchmark option
6 years ago
hangq 18d3c686dc remove unused benchmark option
6 years ago
mindspore-ci-bot 7ebecd8d09 !5541 update model zoon image_classification reame ,delete some error links
6 years ago
mindspore-ci-bot 6b062f2454 !5510 Modify read me of deeplabv3 and inceptionv3
6 years ago
mindspore-ci-bot 6dd79e2b78 !5543 warpctc and maskrcnn readme recitification
6 years ago
mindspore-ci-bot 9ac3e85045 !5515 vgg16 readme update
6 years ago
gengdongjie f366faa275 warpctc and maskrcnn readme recitification
6 years ago
Jolin Zhang46 4b7c6208a3 update model zoon image_classification readme
6 years ago
mindspore-ci-bot e70d664119 !5493 Modify name of the package generated on x86.
6 years ago
mindspore-ci-bot 60f9577103 !5535 add image_classification in model zoo
6 years ago
mindspore-ci-bot 92a1c55941 !5537 Fix some tiny bugs for converter in windows enviroment and conv ops
6 years ago
liuwenhao4 22e37013b9 Fix some tiny bugs for converter in windows enviroment and conv ops
6 years ago
mindspore-ci-bot fa97e24938 !5519 [MS][LITE][Develop]fix lite debug compile
6 years ago
mindspore-ci-bot 7714669469 !5498 modify yolov3-resnet18 README
6 years ago
mindspore-ci-bot 3f305b8afa !5511 modify README.md
6 years ago
mindspore-ci-bot 9fcf1e37b6 !5506 Add the REAMD.md in Wide&Deep model
6 years ago
zhangbiao31 7a30cd8784 !4 add image_classification
6 years ago
wangzhe c223bb3949 add image_classification
6 years ago
mindspore-ci-bot 1621e02b68 !5523 update release info for br: r0.7
6 years ago
wukesong 2fb88e8c46 modify readme
6 years ago
jonyguo 958ca51437 udpate release notes for r0.7
6 years ago
zhouyaqiang 61e8f10209 modify readme
6 years ago
sunsuodong 23e68ce790 fix_debug_compile
6 years ago
CaoJian 422a836cde vgg16 readme update
6 years ago
huangxinjing 330f18de5f Add README for multi-table of Wide&Deep
6 years ago
mindspore-ci-bot 0662256f23 !5470 bugfix on argparser for bool
6 years ago
mindspore-ci-bot 3a6749ab78 !5464 update resnet50 readme
6 years ago
wsc b7a22e1398 Modify name of the package generated on x86.
6 years ago
chengxianbin 6c055b96d1 modiy yolov3-resnet18 net README.md
6 years ago
mindspore-ci-bot a9943a382c !5466 remove bool parameter parser in wide_and_deep
6 years ago
limingqi107 0775db0940 modify the format info of tensorAdd
6 years ago
wanghua 6674a88de4 add schema file for BERT and TinyBERT
6 years ago
mindspore-ci-bot 13dd07c34e !5459 modify transformer & fasterrcnn fix bool arguement parse bug
6 years ago
mindspore-ci-bot 0178ecf785 !5440 lstm readme update
6 years ago
gengdongjie b15a48a53b bugfix on argpasr for bool
6 years ago
wilfChen 50324d3dda lstm readme update
6 years ago
yao_yf 0c175b2cc0 no bool parameter parser in wide_and_deep
6 years ago
mindspore-ci-bot 947c396f09 !5392 Modify read me of deeplabv3 and inceptionv3
6 years ago
VectorSL 48d669636a update resnet readme
6 years ago
mindspore-ci-bot 7d38a1fb7e !5456 sync lite to r0.7
6 years ago
mindspore-ci-bot 565b542886 !5437 Modify for resnet readme
6 years ago
yuchaojie 32787a3453 modify transformer & fasterrcnn fix bool arguement parse bug
6 years ago
xuanyue 0ce8708dee sync lite to r0.7
6 years ago
mindspore-ci-bot b5393e6628 !5429 fix yolov3-resnet18&ssd bool argument parse bug and modify README
6 years ago
chengxianbin 3da41e1091 modify yolov3-darknet53-quant README and fix bool arguement parse bug
6 years ago
qujianwei e0ac982589 modify for resnet readme
6 years ago
mindspore-ci-bot 52ed1ea725 !5404 update submodule akg to r0.7 branch
6 years ago
mindspore-ci-bot 2ff8de674c !5421 Fix sparse slicer leak
6 years ago
mindspore-ci-bot 13978a2c98 !5422 update release note for br:r0.7
6 years ago
jonyguo b0f847072c update release notes for r0.7.0
6 years ago
ZPaC 1ac075919a Fix sparse-slicer leak.
6 years ago
mindspore-ci-bot 6c68e70cfe !5400 yolov3_darknet53 script suit for gpu
6 years ago
mindspore-ci-bot 915d9a0d8f !5409 Fix bugs in c-api: rename, concat, take, sampler, duplicate column & Change docstring of OneHot
6 years ago
mindspore-ci-bot 03b9cf3c12 !5380 fix formula grammar error in comment of SmoothL1Loss
6 years ago
mindspore-ci-bot 321e8d3e8f !5410 ReadMe file normalize
6 years ago
luoyang 4f98ecb433 Fix bugs in c-api: rename, concat, take, sampler, duplicate column &
6 years ago
mindspore-ci-bot d24af4b181 !5396 fix googlenet performance
6 years ago
chenfei 1eab0cd71b fix README
6 years ago
zhouyaqiang c317c4643b fix readme of inceptionv3 and deeplabv3
6 years ago
mindspore-ci-bot 0c7c3c3e8d !5393 change enable_data_sink value to true for transformer
6 years ago
looop5 b0ce67fdd6 update submodule akg to r0.7 branch
6 years ago
mindspore-ci-bot 23fc178a5a !5363 Modify for resnet readme and fix bool type optional
6 years ago
hanhuifeng2020 1f0a760cdb yolov3_darknet53 suit for gpu
6 years ago
panfengfeng 3b8562aa3d fix googlenet performance
6 years ago
yuchaojie ed1340f1e5 set enable_data_sink value to true for transformer
6 years ago
mindspore-ci-bot 98528bbc16 !5370 fix laod input data error in pynative mode on gpu
6 years ago
mindspore-ci-bot 86cfa89c97 !5379 CutMixBatch and MixUpBatch fix for 2D one-hot labels for r0.7
6 years ago
mindspore-ci-bot beae07a9d1 !5364 gpu GoogleNet performance optimize
6 years ago
mindspore-ci-bot 4f9c6e77b7 !5273 Delete extra file.
6 years ago
ZPaC ac27f82607 Delete extra file
6 years ago
Xiao Tianci ed53f7f821 fix formula grammar error in comment of SmoothL1Loss
6 years ago
mahdi a5228c75c7 Fixed 2D one-hot label problems in CutMix and MixUp
6 years ago
mindspore-ci-bot 8533744d7c !5342 GPU FusedBatchNormEx add signature
6 years ago
mindspore-ci-bot f3444977f3 !5339 add type for get obj id & fix resource clear bug
6 years ago
mindspore-ci-bot 581788f040 !5344 Delete __del__func in Profiler.
6 years ago
mindspore-ci-bot 3d511f80d5 !5357 gpu GoogleNet performance optimize
6 years ago
mindspore-ci-bot cb88a43a24 !5288 yolov3-resnet18 and ssd net README file normalize
6 years ago
chujinjin 8948e55ae5 fix load input data error when input is a tuple
6 years ago
qujianwei 3adc06024c modify for resnet readme and fix bool type option
6 years ago
chengxianbin f3d87dd13e modiy ssd&yolov3-resnet18 net README.md
6 years ago
VectorSL c381430e71 slice support nhwc
6 years ago
mindspore-ci-bot e5780288e9 !5341 transfer tensor to tuple
6 years ago
mindspore-ci-bot d670bcb003 !5346 Fix bug of DepthwiseConv2d deleting arg 'group'
6 years ago
simson b00cdb2fe6 add type for get obj id & fix resource clear bug
6 years ago
limingqi107 3516447749 gpu GoogleNet performance optimize
6 years ago
李嘉琪 0e9815f63c modify 0.7
6 years ago
mindspore-ci-bot 2f189543b9 !5328 [bug]fix bugs when parameters updata r0.7
6 years ago
mindspore-ci-bot c0a184ae8f !5291 serving RESTful: opt performance, short timeout
6 years ago
mindspore-ci-bot ebfca60cdf !5176 fix data dump log
6 years ago
mindspore-ci-bot f01613508f !5257 Fix no attribute'_graph_data' error in graphdata.py
6 years ago
chenfei c545422384 can't delete arg 'group' of DepthwiseConv2d
6 years ago
VectorSL db3f387f14 FusedBatchNormEx add signature
6 years ago
yuximiao 6005091a09 fix __del__ in Profiler
6 years ago
Wei Luning 051b019c96 fix bug in parameter init
6 years ago
mindspore-ci-bot ee6ab2980d !5318 Fix wrong error
6 years ago
mindspore-ci-bot f9609e4ca7 !5252 add tinybert gpu readme
6 years ago
mindspore-ci-bot 3fdf9b72df !5306 [MS][LITE][Develop]fix arm32 compile
6 years ago
mindspore-ci-bot d142a8d944 !5232 modify readme for deepfm in r0.7
6 years ago
xuyongfei 78f88cde1b serving RESTful: opt for performance
6 years ago
sunsuodong 6385eafacf fix_arm32
6 years ago
leilei_snow 08c1d4bf44 fix index error
6 years ago
mindspore-ci-bot 5f3581aa69 !5303 Enlarge the threshold of resnet50 in pynative
6 years ago
mindspore-ci-bot e40fd0b4ed !5242 Fix remove internal output for unique device target
6 years ago
mindspore-ci-bot 620f5856ef !5293 Add test cases for uniform ops on GPU
6 years ago
mindspore-ci-bot e3b0ae75ae !5286 [MD] minddata gpu add circular_memory to device_queue
6 years ago
mindspore-ci-bot a47ce883e9 !5245 fix assign memory error in pynative mode
6 years ago
mindspore-ci-bot 3d2f761505 !5299 Add dtype check in uniform and normal distribution and fix docs
6 years ago
mindspore-ci-bot 647053ed4d !5296 Fix some doc errors in pp distributions and bijectors
6 years ago
mindspore-ci-bot 1c093f7db7 !5241 code refine for BN docs
6 years ago
mindspore-ci-bot ddf1b25f9b !5259 Fix some comments of api
6 years ago
lvliang 6f84bc57f4 enlarge-the-threshold-of-resnet50-in-pynative
6 years ago
Xun Deng 8ab9903125 add parameter type check in normal and uniform distribution
6 years ago
peixu_ren f8dde0c3c2 Fix some doc errors in pp distributions and bijectors
6 years ago
peixu_ren 85e67f193c Add test cases for uniform ops on GPU
6 years ago
mindspore-ci-bot be2a98e7c8 !5284 Fixes a bug in RandomSharpness cpp op for br:r0.7
6 years ago
mindspore-ci-bot d506630185 !5261 update api for optional - br:r0.7
6 years ago
xiefangqi 6d0ae1794f add circular_memory function to gpu
6 years ago
avakh 6300058cb6 fixing a bug in random sharpness
6 years ago
mindspore-ci-bot 71dd8a4a71 !5233 GPU add log in LoadInputData
6 years ago
mindspore-ci-bot 57e131a136 !5276 Fix the problem of resource clear v2 in r0.7
6 years ago
mindspore-ci-bot 81d67f6828 !5266 Revert "Avoid copy when create Tensor from numpy array"
6 years ago
simson 7c406fb3a0 fix risk of memory leak
6 years ago
mindspore-ci-bot c904bc2f00 !5251 Combine sparse embedding gradient
6 years ago
mindspore-ci-bot adeeda2fe1 !5236 transfer tensor to tuple
6 years ago
yangyongjie fcc5f77a77 modify readme for deepfm
6 years ago
mindspore-ci-bot bed0f5d3c3 !5249 Put elimminate AllReduce const pass back
6 years ago
He Wei eca64ab8e9 Revert "Avoid copy when create Tensor from numpy array"
6 years ago
mindspore-ci-bot 84989b0cbf !5218 SoftDvppDecodeRandomCropResizeJpeg do not support resize iimage to odd resolution
6 years ago
jonyguo f45e2c921f fix: update api comment for optional
6 years ago
zhangxinfeng3 758269d049 modify some comments of api
6 years ago
heleiwang 814fc1d03e fix graphdata error
6 years ago
mindspore-ci-bot b9dff22d6c !5104 fix multiple epoch data issue for gpu profiler
6 years ago
mindspore-ci-bot 54e615e904 !5207 fix bug for yolov3-resnet18 scripts
6 years ago
yujianfeng 499d81e573 Fix remove internal output for unique device target
6 years ago
cristoval 817bfed1ec combine sparse embedding gradient
6 years ago
hanhuifeng2020 a533147f52 add tinybert gpu readme
6 years ago
BowenK fcc0e263f3 Revert "Revert "Eliminate AllReduce when the input is a constant""
6 years ago
mindspore-ci-bot 46de719a12 !5209 Modify the name of parameters in uniform
6 years ago
chujinjin 7b6dd0b84d fix runopassignoutputmemory error
6 years ago
lizhenyu 8438221259 code refine for BN docs
6 years ago
李嘉琪 2f5a454ef4 transfer_tensor_to_tuple
6 years ago
qianlong 6b6409d910 DecodeAndCropAndResizeJepg do not support odd size
6 years ago
VectorSL ed9c63469b GPU add log in loadinputdata when tensor input != graph input
6 years ago
mindspore-ci-bot 049acf6d58 !5173 GPU fix getinputformat error
6 years ago
mindspore-ci-bot 55d997ec11 !5052 update readme for docker gpu environment deployment
6 years ago
mindspore-ci-bot ee7d9bf4ac !5213 Quick fix scalar add flaws in MindSpore
6 years ago
mindspore-ci-bot 6b706529c5 !5200 fix SE-Resnet50 infer to use 24 epoch and add SE-Resnet50 readme description
6 years ago
VectorSL 7884176df7 fix getinputformat error when input is not a realnode
6 years ago
mindspore-ci-bot bedfa8578b !5154 serving RESTful, disable http port reuse, update error msg output to user
6 years ago
peixu_ren d522d7ba46 Quick fix scalar add flaws in MindSpore
6 years ago
peixu_ren e701fbfa2e Modify the name of parameters in uniform
6 years ago
mindspore-ci-bot 7090e16df8 !5134 stream parallel support ctrl stream
6 years ago
mindspore-ci-bot 382f9a8ebb !5195 Fix CheckTensor in pynative mode
6 years ago
mindspore-ci-bot fd8ad73689 !5194 fix: padded dataset when no div and with repeat op for br:r0.7
6 years ago
mindspore-ci-bot 7bdb90a40b !5158 fix softdvpp coredump
6 years ago
mindspore-ci-bot 185b25c6de !5179 add create_dict_iterator parameter num_epochs api
6 years ago
chengxianbin 115a85114f clear the warning of execute run_distribute_train.sh
6 years ago
mindspore-ci-bot 0bf7f1f39c !5152 change group conv dtype in gpu resnext50
6 years ago
qujianwei d2588f1935 fix SE-ResNet50 infer to 24-epoch and add ReadMe description
6 years ago
mindspore-ci-bot 6017521219 !5133 modify readme.md for yolov3_darknet53 in r0.7
6 years ago
xuyongfei c0389eaea4 serving RESTful, disable http port reuse, update error msg output to user
6 years ago
Xun Deng 6bb61615fd fix checktensor in pynative mode
6 years ago
qianlong bc8aec007f fix softdvpp coredump
6 years ago
jonyguo d262c63214 fix: padded dataset with non div & repeat
6 years ago
mindspore-ci-bot 9e20e17590 !5191 Add note of limitation for prarmeters of uniform
6 years ago
mindspore-ci-bot 07103b98a9 !5101 Add erf and erfc as generic functions for all the backend
6 years ago
peixu_ren 13584ebc22 Add note of limitation for prarmeters of uniform
6 years ago
peixu_ren 5ce4bcf416 Add erf and erfc as generic functions for all the backend and fix notation in power_transform.
6 years ago
mindspore-ci-bot 92787df680 !5159 add cuda path checker
6 years ago
mindspore-ci-bot 8afbba7936 !5177 Revert "!5121 Fix the problem of resource clear in r0.7"
6 years ago
anzhengqi 3c4feaa4f4 add num_epochs api
6 years ago
simson 556f79d185 Revert "!5121 Fix the problem of resource clear in r0.7"
6 years ago
mindspore-ci-bot 1a42811748 !5144 Rectification of API comments of r0.7
6 years ago
mindspore-ci-bot cee889e426 !5126 Fix problem in RandomPosterize & CutMixBatch
6 years ago
jjfeing 3af2059342 fix data dump log info
6 years ago
mindspore-ci-bot a0a6463210 !5121 Fix the problem of resource clear in r0.7
6 years ago
lizhenyu 1d8e5a27b7 add cuda path checker
6 years ago
mindspore-ci-bot 0bcd75bd2b !5119 Avoid copy when create Tensor from numpy array
6 years ago
zhaoting 576c35d408 change group conv dtype in gpu resnext50
6 years ago
mindspore-ci-bot e05a5c8002 !5115 modify bert and tinybert scripts and README
6 years ago
bingyaweng 5b941df087 rectification of API comments
6 years ago
mindspore-ci-bot b77c7d2729 !5099 modify wrong characters in r07
6 years ago
mindspore-ci-bot ac0b72e67f !5089 Fix MASS and FasterRcnn CI Problem.
6 years ago
gukecai 4ace444f54 parallel ctrl
6 years ago
yangyongjie 32e7eb9148 modify readme for yolov3_darknet53
6 years ago
mindspore-ci-bot c481d45996 !5072 model_zoo README.md format change for googlenet
6 years ago
wanghua cb893dfca9 modify BERT and TinyBERT README
6 years ago
mindspore-ci-bot 6b55dc802d !5116 Modify readme of deeplabv3 and inceptionv3
6 years ago
zhouyaqiang 41de14bacd modify readme of deeplabv3 and inceptionv3
6 years ago
mindspore-ci-bot ee37dc52fa !5054 fix shape and data size
6 years ago
luoyang a75ac9c445 Add type check for RandomPosterize & Add Float tensor support for CutMixBatch
6 years ago
askmiao 1804ea246d fix multiple epoch data issue
6 years ago
mindspore-ci-bot 56350e71f0 !5080 modify sgd and momentum and withgradcell comments
6 years ago
mindspore-ci-bot 761cd8393e !5107 raise RuntimeError when set different mode after Initializer created
6 years ago
simson 90c004078e fix risk of memory leak
6 years ago
mindspore-ci-bot e9f2aae73d !5083 [bugfix]LSTM SyncDeviceToHost failed
6 years ago
He Wei a6690168a8 Avoid copy when create Tensor from numpy array
6 years ago
hexia 4eb1706307 shape_wrong_r0.7
6 years ago
Yi Huaijie 524cf0ed9a raise RuntimeError when set different mode after Initializer created
6 years ago
mindspore-ci-bot d8d7cebc8a !5090 Update fix custom exp/log ops cast logic to r0.7
6 years ago
linqingke 9f5a7939d4 fix fasterrcnn and mass bug.
6 years ago
caozhou 83557cd28d modify wrong characters in r07
6 years ago
Zichun Ye d29bd6862a modify custom_ops to pass pynative mode
6 years ago
mindspore-ci-bot 1b71d50953 !5068 resolve fronted layout
6 years ago
lizhenyu 68b37ee1b3 [bugfix]LSTM SyncDeviceToHost failed
6 years ago
李嘉琪 0238ba4d14 modify comments7
6 years ago
mindspore-ci-bot 891228bcbe !5062 GPU opt insert transpose pass
6 years ago
CaoJian 1f9c3bb044 model_zoo README.md format change for googlenet
6 years ago
caozhou 8287445f95 resolve frontend layput
6 years ago
mindspore-ci-bot 10c6fb5612 !5023 Update lastest fix of bernoulii probs to r0.7
6 years ago
mindspore-ci-bot fae225460c !5034 Add readme and fix some comments of api
6 years ago
VectorSL 54bb6ba58c gpu optimize transpose
6 years ago
wuweikang cec6206bc4 update readme for docker gpu environment deployment
6 years ago
mindspore-ci-bot e69a91b6b5 !5018 maskrcnn adopt smoothl1loss change
6 years ago
zhangxinfeng3 2869e5ace4 update some comments of api
6 years ago
mindspore-ci-bot 09cf1c1a54 !5014 add libevent copyright r0.7
6 years ago
Zichun Ye 04b5b8c737 fix bernoulli prob formula; fix some other minor bugs
6 years ago
mindspore-ci-bot 8d0b52fb13 !5008 fix bug of EraseAssign
6 years ago
hexia 2027f73eaf add libevent copyright
6 years ago
mindspore-ci-bot befc209480 !4805 [MS][LITE]add nnacl readme
6 years ago
mindspore-ci-bot c5279ecf35 !5001 [MS][LITE][Develop]fix fp16 kernel register
6 years ago
mindspore-ci-bot d2de60f1d5 !4991 [MS][LITE] arm cpu fp32 op: add common functions and slidewindow for conv depthwise
6 years ago
mindspore-ci-bot dc13718ce2 !4974 [MS][LITE][Develop]fp16 conv1x1 bug
6 years ago
mindspore-ci-bot 7dbe9f7067 !4986 optimize prule
6 years ago
fuzhiye 2c9daf0f14 optimize prule
6 years ago
sunsuodong bcd97d9751 fix fp16
6 years ago
mindspore-ci-bot 7b8229d644 !4913 transform bn to scale
6 years ago
mindspore-ci-bot e3c053c4ff !4961 Optimize the performance of BatchNorm and FusedBatchNorm, add Fp16 kernel
6 years ago
yangruoqi713 b4551670a9 [MS][LITE] arm cpu fp32 op: conv depthwise
6 years ago
sunsuodong 9734f2a88e batch_norm_fp16
6 years ago
zhengjun10 204ab11572 add bn convert scale pass
6 years ago
mindspore-ci-bot 80d570f003 !4999 change long to int64
6 years ago
yeyunpeng 6b46acb39e change long to int64
6 years ago
mindspore-ci-bot dde257592b !4870 add UnPack method in ops & remove anf_importer populater
6 years ago
hangq 28e3508718 add UnPack method in ops & remove anf_importer populater
6 years ago
mindspore-ci-bot 07a75658bf !4995 [MS][LITE][Develop]stack support int32
6 years ago
ling 0fac817a2d [MS][LITE][Develop]Fp16 conv1x1 bug
6 years ago
chenjianping babff262e3 stack support int32
6 years ago
wenchunjiang a221ee176b fix EraseAssign bug
6 years ago
mindspore-ci-bot 6d0bbb36a3 !4908 Modify the '-fvisibility' attribute during building of mindspore-lite.
6 years ago
mindspore-ci-bot 8219df7337 !4975 Fix bug and add tflite models
6 years ago
wsc 758130924f Modify the '-fvisibility' attribute of mindspore-lite project.
6 years ago
mindspore-ci-bot 33c7b49219 !4987 [bugfix]SyncDeviceToHost failed when device address size is zero
6 years ago
mindspore-ci-bot 0f362bb158 !4980 [MS][LITE][Develop]benchmark mem check fixed
6 years ago
mindspore-ci-bot faa0fed8bb !4989 add the hiai_cpu_face_attr caffe gate
6 years ago
mindspore-ci-bot 150b987898 !4989 add the hiai_cpu_face_attr caffe gate
6 years ago
mindspore-ci-bot 50877b586d !4971 [MS][LITE][Develop]move nnacl to lite/
6 years ago
mindspore-ci-bot 75fce54208 !4790 Fix a Pynative bug.
6 years ago
chenjianping d88a98658c move nnacl to lite/
6 years ago
mindspore-ci-bot add52da73e !4973 Fix errors in exp calculation logics
6 years ago
mindspore-ci-bot d2641bbf79 !4969 delete group parameter from nn.DepthwiseConv2d
6 years ago
zhaodezan 8339d5dae6 add the hiai_cpu_face_attr caffe gate
6 years ago
mindspore-ci-bot 94a109f476 !4898 Fix coredump caused by function call depth too large
6 years ago
mindspore-ci-bot 31a04ea1fe !4977 correct benchmark help info
6 years ago
jin-xiulang c246b177a6 Debug for Pynative mode.
6 years ago
kai00 a9771d63e1 benchmark mem check fixed
6 years ago
mindspore-ci-bot aedd6de6d5 !4927 fix bug for identity
6 years ago
lizhenyu 1becddf3a4 [bugfix]SyncDeviceToHost failed when device address size is zero
6 years ago
jianghui58 1d601b6924 fix bug and add two tflite models
6 years ago
fary86 04524b6bd3 Fix coredump caused by function call depth too large
6 years ago
mindspore-ci-bot 7098b5c5d5 !4979 fix the segment when the input data is null
6 years ago
mindspore-ci-bot d3733b3b04 !4942 Revert AllReduce const elimination
6 years ago
mindspore-ci-bot 90552c4933 !4861 [MS][LITE][Develop]add conv per channel support for int8
6 years ago
mindspore-ci-bot 8e360888d0 !4590 fix gpu matmul fp32 accuracy
6 years ago
mindspore-ci-bot aefca7b782 !4968 Fix some errors in API about ops and validator of input.
6 years ago
mindspore-ci-bot 1556450445 !4972 fix post training quant
6 years ago
kpy 4fa89408a1 pynative add identity primitive and add comment for set_grad
6 years ago
zhaozhenlong 488c991eba update benchmark help info
6 years ago
mindspore-ci-bot e2203bed01 !3957 Gpu StridedSlice dims exceeds
6 years ago
mindspore-ci-bot b52229379d !4931 add script
6 years ago
mindspore-ci-bot d184066b77 !4967 Resnext50 readme normalize
6 years ago
mindspore-ci-bot b23b957228 !4964 memory ascend while multi model.eval
6 years ago
mindspore-ci-bot 38c366306c !4937 vgg16: modify readme format and replace callback
6 years ago
zhaodezan fed8f406ac fix segment when the input data is null
6 years ago
zhaoting a5c16fc4ac resnext50 readme normalize
6 years ago
mindspore-ci-bot 0c60f7e6ac !4965 fix googlenet deepfm
6 years ago
mindspore-ci-bot 6a5c517d7b !4943 process the bn with phase in prototxt
6 years ago
mindspore-ci-bot 387f4d445f !4938 auto umap buffer for opencl and clean code
6 years ago
peixu_ren 03dac9b621 Fix errors in exp calculation logics
6 years ago
mindspore-ci-bot 2b78032605 !4910 Add Tflite Models
6 years ago
mindspore-ci-bot 42a092d687 !4915 [MS][LITE][Develop]mem check fixed
6 years ago
mindspore-ci-bot fd9be2ddc2 !4946 Get server rank id in python and fix multi server error.
6 years ago
mindspore-ci-bot 115c0cbf72 !4923 fix dts bug
6 years ago
ms_yan 0752c566b1 modify format in vgg16
6 years ago
xutianchun 1ae9f81c82 Fix Post Quant
6 years ago
mindspore-ci-bot fe11760834 !4944 Add tflite and caffe models
6 years ago
Corleone 4b60297832 auto unmap buffer for opencl and clean code
6 years ago
chenzomi 8337ae710e delete group parameter from nn.DepthwiseConv2d
6 years ago
zhaizhiqiang 9d98246dae readme for nnacl
6 years ago
jianghui58 defd9a784e add eight tflite models
6 years ago
kai00 15dff1c4a4 check mem fixing
6 years ago
mindspore-ci-bot 026bbc46ff !4941 [MS][LITE][Develop]rename caffeprelu to prelu
6 years ago
mindspore-ci-bot ac81886328 !4916 fix generator_dataset hangs and test_graphdata_distributed.py failing randomly
6 years ago
liuxiao93 cfe8859499 fix some bug in API and validator of input.
6 years ago
panfengfeng 30b69d3488 fix googlenet & deepfm
6 years ago
mindspore-ci-bot 77198f3182 !4914 Modify the problem list
6 years ago
ZPaC 830172201a Fix multi server precision error.
6 years ago
mindspore-ci-bot b366608a3f !4952 Fix errors in log calculation logics
6 years ago
lixian 1ffb095037 add conv per channel support for int8 kernel
6 years ago
anzhengqi dd942e1807 fix memory ascend while multi model.eval
6 years ago
mindspore-ci-bot 9b503e4f38 !4955 Fixes for Dynamic Augmentation Ops
6 years ago
mindspore-ci-bot 528fb81093 !4959 delete #include "src/ops/primitive_c.h" at model.h
6 years ago
yeyunpeng f4eee11dfa delete #include "src/ops/primitive_c.h" at model.h
6 years ago
mindspore-ci-bot e06dfaa80d !4854 fix bug
6 years ago
mindspore-ci-bot 9f19076788 !4956 Fix CheckTuple in pynative mode
6 years ago
mindspore-ci-bot ab45bec828 !4924 Modify API comments and fix error of st
6 years ago
peixu_ren 1c8eb9b15d Fix errors in log calculation logics
6 years ago
shenwei41 76518f7f13 Modify the problem list
6 years ago
mindspore-ci-bot 8ee136db18 !4953 C++ API: Minor fixes for dataset parameters
6 years ago
meng_chunyang 631aa8cf46 add tflite and caffe models
6 years ago
Xun Deng dc11fa9f53 Fixed CheckTuple issues and error message
6 years ago
Mahdi a5f9b8f92e Added fix for MixUpBatch and CutMixBatch and for RandomAffine
6 years ago
Cathy Wong 93810a0dc8 C++ API: Minor fixes for dataset parameters
6 years ago
mindspore-ci-bot 04decda0c5 !4951 Fix missing symbol, build failure issue with -B on
6 years ago
Harshvardhan Gupta ac457f3163 fix build failures and wp tensor search condition
6 years ago
mindspore-ci-bot 9d7250c483 !4776 Introduce 2 extra ctrl flags to DataBuffer in dataset, address remaining cmts to PR4632
6 years ago
Zirui Wu 74c1e6da60 introducing pause and quit flags to DataBuffer
6 years ago
mindspore-ci-bot 3eef4a4e06 !4948 Fix backend compilation error when -g is off
6 years ago
chenjianping 5a83d3a7df rename prelu to leakyrelu,rename caffeprelu to prelu
6 years ago
Jesse Lee ebd4cc5c0a Fix glog off compilation bug
6 years ago
wukesong 9784a58da2 add script shell
6 years ago
zhaodezan ad883e4384 process bn with phase
6 years ago
BowenK 7a7e499475 Revert "Eliminate AllReduce when the input is a constant"
6 years ago
xuanyue bbedc02700 fix bug
6 years ago
bingyaweng 3422f60d50 modify comments of API
6 years ago
cjh9368 70001a71ea fix dts bug
6 years ago
wilfChen 837aecf9af gpu stridedslice
6 years ago
heleiwang 4870abc848 1. fix generator_dataset hangs
6 years ago
qujianwei c21ffc0317 fix gpu matmul fp32 accuracy
6 years ago

@ -66,10 +66,12 @@ MindSpore offers build options across multiple backends:
| Hardware Platform | Operating System | Status |
| :---------------- | :--------------- | :----- |
| Ascend910 | Ubuntu-x86 | ✔️ |
| | Ubuntu-aarch64 | ✔️ |
| | EulerOS-x86 | ✔️ |
| | EulerOS-aarch64 | ✔️ |
| GPU CUDA 10.1 | Ubuntu-x86 | ✔️ |
| CPU | Ubuntu-x86 | ✔️ |
| | Ubuntu-aarch64 | ✔️ |
| | Windows-x86 | ✔️ |
For installation using `pip`, take `CPU` and `Ubuntu-x86` build version as an example:
@ -149,7 +151,23 @@ currently the containerized build options are supported as follows:
sudo apt-get update && sudo apt-get install -y nvidia-container-toolkit nvidia-docker2
sudo systemctl restart docker
```
Then edit the file daemon.json:
```
$ vim /etc/docker/daemon.json
{
"runtimes": {
"nvidia": {
"path": "nvidia-container-runtime",
"runtimeArgs": []
}
}
}
```
Restart docker again:
```
sudo systemctl daemon-reload
sudo systemctl restart docker
```
Then you can pull and run the latest stable image using the below command:
```
docker pull mindspore/mindspore-gpu:0.7.0-beta

@ -1,4 +1,4 @@
![MindSpore标志](docs/MindSpore-logo.png "MindSpore logo")
![MindSpore标志](docs/MindSpore-logo.png "MindSpore logo")
============================================================
[View English](./README.md)
@ -66,10 +66,12 @@ MindSpore提供跨多个后端的构建选项
| 硬件平台 | 操作系统 | 状态 |
| :------------ | :-------------- | :--- |
| Ascend 910 | Ubuntu-x86 | ✔️ |
| | Ubuntu-aarch64 | ✔️ |
| | EulerOS-x86 | ✔️ |
| | EulerOS-aarch64 | ✔️ |
| GPU CUDA 10.1 | Ubuntu-x86 | ✔️ |
| CPU | Ubuntu-x86 | ✔️ |
| | Ubuntu-aarch64 | ✔️ |
| | Windows-x86 | ✔️ |
使用`pip`命令安装,以`CPU`和`Ubuntu-x86`build版本为例
@ -120,10 +122,10 @@ MindSpore的Docker镜像托管在[Docker Hub](https://hub.docker.com/r/mindspore
| 硬件平台 | Docker镜像仓库 | 标签 | 说明 |
| :----- | :------------------------ | :----------------------- | :--------------------------------------- |
| CPU | `mindspore/mindspore-cpu` | `x.y.z` | 已经预安装MindSpore `x.y.z` CPU版本的生产环境。 |
| | | `devel` | 提供开发环境从源头构建MindSpore`CPU`后端。安装详情请参考https://www.mindspore.cn/install。 |
| | | `devel` | 提供开发环境从源头构建MindSpore`CPU`后端。安装详情请参考https://www.mindspore.cn/install 。 |
| | | `runtime` | 提供运行时环境安装MindSpore二进制包`CPU`后端)。 |
| GPU | `mindspore/mindspore-gpu` | `x.y.z` | 已经预安装MindSpore `x.y.z` GPU版本的生产环境。 |
| | | `devel` | 提供开发环境从源头构建MindSpore`GPU CUDA10.1`后端。安装详情请参考https://www.mindspore.cn/install。 |
| | | `devel` | 提供开发环境从源头构建MindSpore`GPU CUDA10.1`后端。安装详情请参考https://www.mindspore.cn/install 。 |
| | | `runtime` | 提供运行时环境安装MindSpore二进制包`GPU CUDA10.1`后端)。 |
| Ascend | <center>&mdash;</center> | <center>&mdash;</center> | 即将推出,敬请期待。 |
@ -148,7 +150,23 @@ MindSpore的Docker镜像托管在[Docker Hub](https://hub.docker.com/r/mindspore
sudo apt-get update && sudo apt-get install -y nvidia-container-toolkit nvidia-docker2
sudo systemctl restart docker
```
编辑文件 daemon.json:
```
$ vim /etc/docker/daemon.json
{
"runtimes": {
"nvidia": {
"path": "nvidia-container-runtime",
"runtimeArgs": []
}
}
}
```
再次重启docker:
```
sudo systemctl daemon-reload
sudo systemctl restart docker
```
使用以下命令获取并运行最新的稳定镜像:
```
docker pull mindspore/mindspore-gpu:0.7.0-beta

File diff suppressed because one or more lines are too long

@ -3693,6 +3693,72 @@ Copyright (c) 1999, Frank Warmerdam
Copyright (c) 1991-1996 Sam Leffler
Copyright (c) 1996 USAF Phillips Laboratory
Software: libevent 2.1.12
Copyright notice:
Copyright (C) 1998 - 2012, Daniel Stenberg, <daniel@haxx.se>, et al.
COPYRIGHT AND PERMISSION NOTICE
Copyright (c) 1996 - 2013, Daniel Stenberg, <daniel@haxx.se>.
Copyright (C) 2012, iSEC Partners.
Copyright (c) 1987, 1993, 1994, 1995
Copyright (c) 1987, 1993, 1994, 1996
Copyright 2002 Niels Provos <provos@citi.umich.edu>
Copyright (c) 2007-2012 Niels Provos and Nick Mathewson
Copyright (c) 2000-2007 Niels Provos <provos@citi.umich.edu>
Copyright (c) 2007-2012 Niels Provos, Nick Mathewson
Copyright (c) 2009-2012 Niels Provos and Nick Mathewson
Copyright (c) 2006-2007 Niels Provos <provos@citi.umich.edu>
Copyright (c) 2008-2012 Niels Provos and Nick Mathewson
Copyright (c) 1991, 1993
Copyright (c) 2009, Michihiro NAKAJIMA
Copyright 2000-2013 Kitware, Inc.
Copyright 2000-2011 Insight Software Consortium
notices of original copyright by their contributors; see each source
Copyright (C) 1996-2018 Free Software Foundation, Inc.
Copyright (c) 2010 Chris Davis, Niels Provos, and Nick Mathewson
Copyright (c) 2010-2012 Niels Provos and Nick Mathewson
Copyright (c) 1996, David Mazieres <dm@uun.org>
Copyright (c) 2008, Damien Miller <djm@openbsd.org>
Copyright (c) 2002-2007 Niels Provos <provos@citi.umich.edu>
Copyright (c) 2002-2006 Niels Provos <provos@citi.umich.edu>
Copyright (c) 2009-2012 Niels Provos, Nick Mathewson
Copyright 2000-2009 Niels Provos <provos@citi.umich.edu>
Copyright 2009-2012 Niels Provos and Nick Mathewson
Copyright 2000-2007 Niels Provos <provos@citi.umich.edu>
Copyright 2007-2012 Niels Provos, Nick Mathewson
Copyright 2003-2009 Niels Provos <provos@citi.umich.edu>
Copyright 2006-2007 Niels Provos
Copyright 2007-2012 Nick Mathewson and Niels Provos
Copyright (c) 2005-2007 Niels Provos <provos@citi.umich.edu>
Copyright (c) 2003-2009 Niels Provos <provos@citi.umich.edu>
Copyright 2007-2012 Niels Provos and Nick Mathewson
Copyright (c) 2007 Sun Microsystems. All rights reserved.
Copyright (c) 2008-2012 Niels Provos, Nick Mathewson
Copyright 2002 Christopher Clark
Copyright 2005-2012 Nick Mathewson
Copyright 2001-2007 Niels Provos <provos@citi.umich.edu>
Copyright (c) 2012 Niels Provos and Nick Mathewson
Copyright (c) 2000 Dug Song <dugsong@monkey.org>
Copyright (c) 1993 The Regents of the University of California.
Copyright (c) 1998 Todd C. Miller <Todd.Miller@courtesan.com>
Copyright (c) 2003 Michael A. Davis <mike@datanerds.net>
Copyright (c) 2007 Sun Microsystems
Copyright (c) 2002 Christopher Clark
Copyright (c) 2006 Maxim Yegorushkin <maxim.yegorushkin@gmail.com>
Copyright (c) 2010 BitTorrent, Inc.
Copyright (c) 2005-2012 Niels Provos and Nick Mathewson
Copyright (c) 1993
Copyright 2003 Michael A. Davis <mike@datanerds.net>
Copyright 2003-2007 Niels Provos <provos@citi.umich.edu>
Copyright 2008-2012 Niels Provos and Nick Mathewson
Copyright (c) 2003-2007 Niels Provos <provos@citi.umich.edu>
Copyright (c) 2013 Niels Provos and Nick Mathewson
Copyright (c) 2009-2012 Nick Mathewson and Niels Provos
Copyright (c) 2007-2013 Niels Provos and Nick Mathewson
Copyright (c) 2012 Ross Lagerwall <rosslagerwall@gmail.com>
tinytest.c -- Copyright 2009-2012 Nick Mathewson
tinytest.h -- Copyright 2009-2012 Nick Mathewson
tinytestmacros.h -- Copyright 2009-2012 Nick Mathewson
Software: opencv 4.2.0
Copyright notice:
Copyright (C) 2016, NVIDIA Corporation, all rights reserved.

2
akg

@ -1 +1 @@
Subproject commit 3bb6264188d0b1d6ff776a35a571bc7190df0800
Subproject commit 76a1ecf9da48fa463e25ad63c26281fb5867874d

@ -16,20 +16,20 @@
@title mindspore_build
SET BASEPATH=%CD%
IF NOT EXIST %BASEPATH%/build (
IF NOT EXIST "%BASEPATH%/build" (
md "build"
)
cd %BASEPATH%/build
set BUILD_PATH=%CD%
IF NOT EXIST %BUILD_PATH%/mindspore (
IF NOT EXIST "%BUILD_PATH%/mindspore" (
md "mindspore"
)
cd %CD%/mindspore
IF "%2%" == "lite" (
IF "%1%" == "lite" (
call :gene_gtest
call :run_cmake
IF errorlevel 1 (
@ -47,14 +47,17 @@ IF "%2%" == "lite" (
)
cd %BUILD_PATH%/mindspore
IF "%1%" == "" (
cmake --build . -- -j6
IF "%2%" == "" (
cmake --build . --target package -- -j6
) ELSE (
cmake --build . -- -j%1%
cmake --build . --target package -- -j%2%
)
IF errorlevel 1 (
echo "build fail."
goto run_fail
) ELSE (
cd %BASEPATH%/output
rd /s /q _CPack_Packages
)
) ELSE (
cmake -DCMAKE_BUILD_TYPE=Release -DENABLE_CPU=ON -DENABLE_MINDDATA=ON -DUSE_GLOG=ON ^
@ -83,7 +86,7 @@ goto run_eof
cd %BUILD_PATH%/mindspore
cmake -DBUILD_DEVICE=on -DBUILD_CONVERTER=on -DPLATFORM_ARM64=off -DSUPPORT_TRAIN=off ^
-DCMAKE_BUILD_TYPE=Release -DSUPPORT_GPU=off -DBUILD_MINDDATA=off -DOFFLINE_COMPILE=off ^
-G "CodeBlocks - MinGW Makefiles" %BASEPATH%/mindspore/lite
-G "CodeBlocks - MinGW Makefiles" "%BASEPATH%/mindspore/lite"
GOTO:EOF
:gene_gtest
@ -94,31 +97,31 @@ GOTO:EOF
GOTO:EOF
:gene_protobuf
SET PROTOC=%BASEPATH%/build/mindspore/_deps/protobuf-src/_build/protoc
SET PROTOC="%BASEPATH%/build/mindspore/_deps/protobuf-src/_build/protoc"
SET PROTO_SRC_DIR=%BASEPATH%/mindspore/lite/tools/converter/parser/caffe
SET PROTO_SRC_DIR="%BASEPATH%/mindspore/lite/tools/converter/parser/caffe"
cd %PROTO_SRC_DIR%
%PROTOC% *.proto --proto_path=%PROTO_SRC_DIR% --cpp_out=%PROTO_SRC_DIR%
SET PROTO_SRC_DIR=%BASEPATH%/mindspore/lite/tools/converter/parser/onnx
SET PROTO_SRC_DIR="%BASEPATH%/mindspore/lite/tools/converter/parser/onnx"
cd %PROTO_SRC_DIR%
%PROTOC% *.proto --proto_path=%PROTO_SRC_DIR% --cpp_out=%PROTO_SRC_DIR%
cd %BUILD_PATH%/mindspore
GOTO:EOF
:gene_flatbuffer
SET FLATC=%BASEPATH%/build/mindspore/_deps/flatbuffers-src/_build/flatc
SET FLATC="%BASEPATH%/build/mindspore/_deps/flatbuffers-src/_build/flatc"
SET FLAT_DIR=%BASEPATH%/mindspore/lite/schema
cd %FLAT_DIR%
IF EXIST inner rd /s /q inner
md inner
%FLATC% -c -b *.fbs
%FLATC% -c -b --reflect-types --gen-mutable --reflect-names --gen-object-api -o %FLAT_DIR%/inner *.fbs
%FLATC% -c -b --reflect-types --gen-mutable --reflect-names --gen-object-api -o "%FLAT_DIR%/inner" *.fbs
SET FLAT_DIR=%BASEPATH%/mindspore/lite/tools/converter/parser/tflite
cd %FLAT_DIR%
%FLATC% -c -b --reflect-types --gen-mutable --reflect-names --gen-object-api -o %FLAT_DIR% *.fbs
%FLATC% -c -b --reflect-types --gen-mutable --reflect-names --gen-object-api -o "%FLAT_DIR%" *.fbs
cd %BUILD_PATH%/mindspore
GOTO:EOF

@ -53,7 +53,7 @@ usage()
echo " -n Compile minddata lite"
echo " -M Enable MPI and NCCL for GPU training, gpu default on"
echo " -V Specify the minimum required cuda version, default CUDA 10.1"
echo " -I Compile lite"
echo " -I Enable compiling mindspore lite for arm64, arm32 or x86_64, default disable mindspore lite compiling"
echo " -K Compile with AKG, default on"
echo " -s Enable serving module, default off"
echo " -w Enable acl module, default off"
@ -393,7 +393,7 @@ build_mindspore()
CMAKE_VERBOSE="--verbose"
fi
cmake --build . --target package ${CMAKE_VERBOSE} -j$THREAD_NUM
echo "success to build mindspore project!"
echo "success building mindspore project!"
}
checkndk() {
@ -618,10 +618,12 @@ build_lite()
if [[ "${COMPILE_RET}" -ne 0 ]]; then
echo "---------------- mindspore lite: build failed ----------------"
exit 1
else
mv ${BASEPATH}/output/tmp/*.tar.gz* ${BASEPATH}/output/
rm -rf ${BASEPATH}/output/tmp/
echo "---------------- mindspore lite: build success ----------------"
exit 0
fi
}

@ -8,7 +8,7 @@ else()
VER 67.1
LIBS ${LIB_ICU_COMMON} ${LIB_ICU_DATA} ${LIB_ICU_I18N}
URL https://github.com/unicode-org/icu/archive/release-67-1.tar.gz
MD5 0c2662a2b0bc80b0eb56495205247c8f
MD5 fd525fb47d8827b0b7da78b51dd2d93f
CONFIGURE_COMMAND ${CMAKE_SOURCE_DIR}/scripts/build_icu4c.sh
)
include_directories(${icu4c_INC})

@ -1,13 +1,18 @@
include(CMakePackageConfigHelpers)
set(LIB_DIR ${MAIN_DIR}/lib)
set(INC_DIR ${MAIN_DIR}/include)
set(TURBO_DIR ${MAIN_DIR}/third_party/libjpeg-turbo)
set(OPENCV_DIR ${MAIN_DIR}/third_party/opencv)
set(PROTOBF_DIR ${MAIN_DIR}/third_party/protobuf)
set(FLATBF_DIR ${MAIN_DIR}/third_party/flatbuffers)
set(LIB_DIR ${MAIN_DIR}-${COMPONENT_NAME}/lib)
set(INC_DIR ${MAIN_DIR}-${COMPONENT_NAME}/include)
set(TURBO_DIR ${MAIN_DIR}-${COMPONENT_NAME}/third_party/libjpeg-turbo)
set(OPENCV_DIR ${MAIN_DIR}-${COMPONENT_NAME}/third_party/opencv)
set(PROTOBF_DIR ${MAIN_DIR}-${COMPONENT_NAME}/third_party/protobuf)
set(FLATBF_DIR ${MAIN_DIR}-${COMPONENT_NAME}/third_party/flatbuffers)
install(FILES ${TOP_DIR}/mindspore/lite/build/src/libmindspore-lite.so DESTINATION ${LIB_DIR} COMPONENT ${COMPONENT_NAME})
set(LIB_DIR_RUN_X86 ${MAIN_DIR}-${RUN_X86_COMPONENT_NAME}/lib)
set(INC_DIR_RUN_X86 ${MAIN_DIR}-${RUN_X86_COMPONENT_NAME}/include)
set(TURBO_DIR_RUN_X86 ${MAIN_DIR}-${RUN_X86_COMPONENT_NAME}/third_party/libjpeg-turbo)
set(OPENCV_DIR_RUN_X86 ${MAIN_DIR}-${RUN_X86_COMPONENT_NAME}/third_party/opencv)
set(PROTOBF_DIR_RUN_X86 ${MAIN_DIR}-${RUN_X86_COMPONENT_NAME}/third_party/protobuf)
set(FLATBF_DIR_RUN_X86 ${MAIN_DIR}-${RUN_X86_COMPONENT_NAME}/third_party/flatbuffers)
if (BUILD_MINDDATA)
install(DIRECTORY ${TOP_DIR}/mindspore/ccsrc/minddata/dataset/include/ DESTINATION ${INC_DIR} COMPONENT ${COMPONENT_NAME} FILES_MATCHING PATTERN "*.h")
install(FILES ${TOP_DIR}/mindspore/lite/build/minddata/libminddata-lite.so DESTINATION ${LIB_DIR} COMPONENT ${COMPONENT_NAME})
@ -30,25 +35,52 @@ if (BUILD_MINDDATA)
endif ()
if (PLATFORM_ARM64)
install(FILES ${TOP_DIR}/mindspore/lite/build/src/libmindspore-lite.so DESTINATION ${LIB_DIR} COMPONENT ${COMPONENT_NAME})
install(FILES ${TOP_DIR}/mindspore/core/ir/dtype/type_id.h DESTINATION ${INC_DIR}/ir/dtype COMPONENT ${COMPONENT_NAME})
install(DIRECTORY ${TOP_DIR}/mindspore/lite/include/ DESTINATION ${INC_DIR} COMPONENT ${COMPONENT_NAME} FILES_MATCHING PATTERN "*.h")
install(DIRECTORY ${TOP_DIR}/mindspore/lite/schema/ DESTINATION ${INC_DIR}/schema COMPONENT ${COMPONENT_NAME} FILES_MATCHING PATTERN "*.h" PATTERN "inner" EXCLUDE)
install(FILES ${TOP_DIR}/mindspore/lite/build/src/runtime/kernel/arm/nnacl/liboptimize.so DESTINATION ${LIB_DIR} COMPONENT ${COMPONENT_NAME})
install(FILES ${TOP_DIR}/mindspore/lite/build/nnacl/liboptimize.so DESTINATION ${LIB_DIR} COMPONENT ${COMPONENT_NAME})
install(DIRECTORY ${TOP_DIR}/third_party/flatbuffers/include DESTINATION ${FLATBF_DIR} COMPONENT ${COMPONENT_NAME})
elseif (PLATFORM_ARM32)
install(FILES ${TOP_DIR}/mindspore/lite/build/src/libmindspore-lite.so DESTINATION ${LIB_DIR} COMPONENT ${COMPONENT_NAME})
install(FILES ${TOP_DIR}/mindspore/core/ir/dtype/type_id.h DESTINATION ${INC_DIR}/ir/dtype COMPONENT ${COMPONENT_NAME})
install(DIRECTORY ${TOP_DIR}/mindspore/lite/include/ DESTINATION ${INC_DIR} COMPONENT ${COMPONENT_NAME} FILES_MATCHING PATTERN "*.h")
install(DIRECTORY ${TOP_DIR}/mindspore/lite/schema/ DESTINATION ${INC_DIR}/schema COMPONENT ${COMPONENT_NAME} FILES_MATCHING PATTERN "*.h" PATTERN "inner" EXCLUDE)
install(DIRECTORY ${TOP_DIR}/third_party/flatbuffers/include DESTINATION ${FLATBF_DIR} COMPONENT ${COMPONENT_NAME})
elseif (CMAKE_SYSTEM_NAME MATCHES "Windows")
get_filename_component(CXX_DIR ${CMAKE_CXX_COMPILER} PATH)
file(GLOB LIB_LIST ${CXX_DIR}/libstdc++-6.dll ${CXX_DIR}/libwinpthread-1.dll ${CXX_DIR}/libssp-0.dll ${CXX_DIR}/libgcc_s_seh-1.dll)
install(FILES ${TOP_DIR}/build/mindspore/tools/converter/converter_lite.exe DESTINATION ${TOP_DIR}/build/mindspore/package COMPONENT ${COMPONENT_NAME})
install(FILES ${LIB_LIST} DESTINATION ${TOP_DIR}/build/mindspore/package COMPONENT ${COMPONENT_NAME})
install(FILES ${TOP_DIR}/build/mindspore/tools/converter/libconverter_parser.a DESTINATION ${TOP_DIR}/build/mindspore/package COMPONENT ${PARSER_NAME})
else ()
install(FILES ${TOP_DIR}/third_party/protobuf/build/lib/libprotobuf.so.19.0.0 DESTINATION ${PROTOBF_DIR}/lib RENAME libprotobuf.so.19 COMPONENT ${COMPONENT_NAME})
install(DIRECTORY ${TOP_DIR}/mindspore/lite/include/ DESTINATION ${INC_DIR_RUN_X86} COMPONENT ${RUN_X86_COMPONENT_NAME} FILES_MATCHING PATTERN "*.h")
install(DIRECTORY ${TOP_DIR}/mindspore/lite/schema/ DESTINATION ${INC_DIR_RUN_X86}/schema COMPONENT ${RUN_X86_COMPONENT_NAME} FILES_MATCHING PATTERN "*.h" PATTERN "inner" EXCLUDE)
install(FILES ${TOP_DIR}/mindspore/core/ir/dtype/type_id.h DESTINATION ${INC_DIR_RUN_X86}/ir/dtype COMPONENT ${RUN_X86_COMPONENT_NAME})
install(DIRECTORY ${TOP_DIR}/third_party/flatbuffers/include DESTINATION ${FLATBF_DIR_RUN_X86} COMPONENT ${RUN_X86_COMPONENT_NAME})
install(FILES ${TOP_DIR}/mindspore/lite/build/src/libmindspore-lite.so DESTINATION ${LIB_DIR_RUN_X86} COMPONENT ${RUN_X86_COMPONENT_NAME})
install(FILES ${TOP_DIR}/third_party/protobuf/build/lib/libprotobuf.so.19.0.0 DESTINATION ${PROTOBF_DIR}/lib RENAME libprotobuf.so.19 COMPONENT ${COMPONENT_NAME})
endif ()
set(CPACK_GENERATOR TGZ)
if (CMAKE_SYSTEM_NAME MATCHES "Windows")
set(CPACK_GENERATOR ZIP)
else ()
set(CPACK_GENERATOR TGZ)
endif ()
set(CPACK_ARCHIVE_COMPONENT_INSTALL ON)
set(CPACK_COMPONENTS_ALL ${COMPONENT_NAME})
if (PLATFORM_ARM64 OR PLATFORM_ARM32)
set(CPACK_COMPONENTS_ALL ${COMPONENT_NAME})
elseif (WIN32)
set(CPACK_COMPONENTS_ALL ${COMPONENT_NAME})
else ()
set(CPACK_COMPONENTS_ALL ${COMPONENT_NAME} ${RUN_X86_COMPONENT_NAME})
endif ()
set(CPACK_PACKAGE_FILE_NAME ${MAIN_DIR})
set(CPACK_PACKAGE_DIRECTORY ${TOP_DIR}/output/tmp)
if (WIN32)
set(CPACK_PACKAGE_DIRECTORY ${TOP_DIR}/output)
else ()
set(CPACK_PACKAGE_DIRECTORY ${TOP_DIR}/output/tmp)
endif()
set(CPACK_PACKAGE_CHECKSUM SHA256)
include(CPack)

@ -6,6 +6,7 @@ MAINTAINER leonwanghui <leon.wanghui@huawei.com>
ENV PYTHON_ROOT_PATH /usr/local/python-3.7.5
ENV CMAKE_ROOT_PATH /usr/local/cmake-3.14.1
ENV PATH ${PYTHON_ROOT_PATH}/bin:${CMAKE_ROOT_PATH}/bin:/usr/local/bin:$PATH
ENV LD_LIBRARY_PATH ${PYTHON_ROOT_PATH}/lib
# Install base tools
RUN apt update \
@ -48,7 +49,7 @@ RUN apt install -y libffi-dev libssl-dev zlib1g-dev libbz2-dev libncurses5-dev \
&& tar -xvf v3.7.5.tar.gz \
&& cd /tmp/cpython-3.7.5 \
&& mkdir -p ${PYTHON_ROOT_PATH} \
&& ./configure --prefix=${PYTHON_ROOT_PATH} \
&& ./configure --prefix=${PYTHON_ROOT_PATH} --enable-shared \
&& make -j4 \
&& make install -j4 \
&& rm -f /usr/local/bin/python \

@ -1,4 +1,4 @@
FROM nvidia/cuda:10.1-cudnn7-runtime-ubuntu18.04
FROM nvidia/cuda:10.1-cudnn7-devel-ubuntu18.04
MAINTAINER leonwanghui <leon.wanghui@huawei.com>
@ -35,9 +35,7 @@ RUN DEBIAN_FRONTEND=noninteractive apt install -y \
autoconf \
libtool \
automake \
flex \
libnccl2=2.4.8-1+cuda10.1 \
libnccl-dev=2.4.8-1+cuda10.1
flex
# Set bash
RUN echo "dash dash/sh boolean false" | debconf-set-selections

@ -6,6 +6,7 @@ MAINTAINER leonwanghui <leon.wanghui@huawei.com>
ENV PYTHON_ROOT_PATH /usr/local/python-3.7.5
ENV CMAKE_ROOT_PATH /usr/local/cmake-3.14.1
ENV PATH ${CMAKE_ROOT_PATH}/bin:/usr/local/bin:$PATH
ENV LD_LIBRARY_PATH ${PYTHON_ROOT_PATH}/lib
# Install base tools
RUN apt update \
@ -51,7 +52,7 @@ RUN apt install -y libffi-dev libssl-dev zlib1g-dev libbz2-dev libncurses5-dev \
&& tar -xvf v3.7.5.tar.gz \
&& cd /tmp/cpython-3.7.5 \
&& mkdir -p ${PYTHON_ROOT_PATH} \
&& ./configure --prefix=${PYTHON_ROOT_PATH} \
&& ./configure --prefix=${PYTHON_ROOT_PATH} --enable-shared \
&& make -j4 \
&& make install -j4 \
&& rm -f /usr/local/bin/python \

@ -1,4 +1,4 @@
FROM nvidia/cuda:10.1-cudnn7-runtime-ubuntu18.04
FROM nvidia/cuda:10.1-cudnn7-devel-ubuntu18.04
MAINTAINER leonwanghui <leon.wanghui@huawei.com>
@ -35,9 +35,7 @@ RUN DEBIAN_FRONTEND=noninteractive apt install -y \
autoconf \
libtool \
automake \
flex \
libnccl2=2.4.8-1+cuda10.1 \
libnccl-dev=2.4.8-1+cuda10.1
flex
# Set bash
RUN echo "dash dash/sh boolean false" | debconf-set-selections

Binary file not shown.

After

Width:  |  Height:  |  Size: 54 KiB

@ -1 +1 @@
Subproject commit 622af6c1c50034bea5a08bd409c5a410782bfe53
Subproject commit 80f9c96ed3fe0f07bf40a91d1f41373681d0c0dc

@ -24,6 +24,7 @@
#include <memory>
#include <iostream>
#include <chrono>
#include <vector>
#ifndef ENABLE_ACL
#include "mindspore/core/utils/log_adapter.h"
@ -44,6 +45,19 @@ class LogStream {
return *this;
}
template <typename T>
LogStream &operator<<(const std::vector<T> &val) noexcept {
(*sstream_) << "[";
for (size_t i = 0; i < val.size(); i++) {
(*this) << val[i];
if (i + 1 < val.size()) {
(*sstream_) << ", ";
}
}
(*sstream_) << "]";
return *this;
}
LogStream &operator<<(std::ostream &func(std::ostream &os)) noexcept {
(*sstream_) << func;
return *this;

@ -17,8 +17,6 @@ import json
import os
import sys
from te.platform.cce_conf import te_set_version
from te.platform.fusion_manager import op_build_cfg_dis, op_build_cfg_en, set_current_op_name, \
init_op_pattern, set_op_params, set_op_build_type, get_op_pattern, set_current_op_func_name
from te.platform.fusion_util import fusion_op
from common import check_kernel_info, get_args, get_build_in_impl_path, get_ddk_version
@ -27,7 +25,6 @@ build_in_impl_path = get_build_in_impl_path()
# op function list
op_build = "compile"
op_pre_build = "pre_build"
fusion_pattern_start_flag = "fusion_pattern_start"
fusion_pattern_end_flag = "fusion_pattern_end"
@ -83,19 +80,7 @@ def build_op(build_type, json_str):
else:
op_module = __import__("impl."+op_name, globals(), locals(), [op_name], 0)
# get function
if build_type == op_pre_build:
# set op parameter
op_build_cfg_dis()
set_current_op_func_name(op_name)
set_current_op_name(kernel_name)
init_op_pattern()
set_op_params(*outputs_args, *attrs_args, kernel_name=kernel_name)
set_op_build_type('prebuild')
if custom_flag:
py_fn_name = kernel_info['op_info']['name']
else:
py_fn_name = op_name
elif build_type == op_build:
if build_type == op_build:
if custom_flag:
py_fn_name = kernel_info['op_info']['name']
else:
@ -106,13 +91,6 @@ def build_op(build_type, json_str):
if op_func is None:
raise ValueError("Op:{} function {} is not supported by Tbe.".format(op_name, build_type))
# pre build
if build_type == op_pre_build:
op_func(*inputs_args, *outputs_args, *attrs_args, kernel_name=kernel_name)
# disable only pattern configuration
op_build_cfg_en()
return get_op_pattern()
# call function
if kernel_name[0:19] == "bounding_box_encode":
return op_func(*inputs_args, *outputs_args, *attrs_args, kernel_name_val=kernel_name)
@ -120,8 +98,6 @@ def build_op(build_type, json_str):
return op_func(*inputs_args, *outputs_args, *attrs_args, kernel_name=kernel_name)
except Exception as e:
if build_type == op_pre_build:
op_build_cfg_en()
raise RuntimeError(e)
@ -136,14 +112,9 @@ def compile_fusion_op(json_str):
Exception: If specific keyword is not found.
"""
args = json.loads(json_str)
te_set_version(ddk_version)
if 'fusion_op' not in args or not args['fusion_op']:
raise ValueError("Json string Errors, key:fusion_op not found.")
if 'prebuild_ops' not in args or not args['prebuild_ops']:
raise ValueError("Json string Errors, key:prebuild_ops not found.")
pre_build_op_list = args['prebuild_ops']
for op in pre_build_op_list:
build_op(op_pre_build, json.dumps(op))
fusion_op_arg = args['fusion_op']
return fusion_op(json.dumps(fusion_op_arg))
@ -159,8 +130,6 @@ def compile_with_json(json_str):
json_info = json.loads(json_str)
if "fusion_op" in json_info:
ret = compile_fusion_op(json_str)
elif "compile_type" in json_info:
ret = build_op(op_pre_build, json_str)
else:
ret = build_op(op_build, json_str)
return ret

@ -20,6 +20,8 @@
#include <vector>
#include <memory>
#include <algorithm>
#include <map>
#include <climits>
#include "runtime/device/kernel_runtime.h"
#include "backend/kernel_compiler/aicpu/aicpu_kernel_mod.h"
#include "backend/kernel_compiler/akg/akg_kernel_build.h"
@ -218,7 +220,7 @@ void SetNodeInputs(const std::shared_ptr<AnfNode> &anf_node, mindspore::NodeDef
mindspore::TensorShape_Dim *dim = tensorShape->add_dim();
dim->set_size((::google::protobuf::int64)item);
}
node_inputs->set_tensor_type((mindspore::DataType)input_data_type);
node_inputs->set_tensor_type(input_data_type);
node_inputs->set_mem_device("HBM");
}
}
@ -245,7 +247,7 @@ void SetNodeOutputs(const std::shared_ptr<AnfNode> &anf_node, mindspore::NodeDef
}
TypeId output_type = AnfAlgo::GetOutputDeviceDataType(anf_node, output_index);
int32_t output_data_type = AicpuOpUtil::MsTypeToProtoType(output_type);
node_outputs->set_tensor_type((mindspore::DataType)output_data_type);
node_outputs->set_tensor_type(output_data_type);
node_outputs->set_mem_device("HBM");
}
}
@ -287,6 +289,109 @@ bool CreateNodeDefBytes(const std::shared_ptr<AnfNode> &anf_node,
return true;
}
bool CreateExtInfo(const std::shared_ptr<AnfNode> &anf_node, const std::shared_ptr<AicpuOpKernelMod> &kernel_mod_ptr) {
if (!anf_node->isa<CNode>()) {
return true;
}
if (!AnfAlgo::IsDynamicShape(anf_node)) {
return true;
}
MS_LOG(INFO) << "CreateExtInfo start, " << anf_node->fullname_with_scope();
int32_t unknown_shape_type = UnknowShapeOpType::DEPEND_COMPUTE;
uint64_t ext_info_head_len = kExtInfoHeadSize;
std::string ext_info;
size_t input_num = AnfAlgo::GetInputTensorNum(anf_node);
size_t output_num = AnfAlgo::GetOutputTensorNum(anf_node);
// 1.addr:unknown shape type
uint64_t ext_info_len = ext_info.size();
ext_info_len += ext_info_head_len + sizeof(int32_t);
// 2.addr:input ShapeAndType
ext_info_len += ext_info_head_len + input_num * sizeof(ShapeAndType);
// 3.addr:output ShapeAndType
ext_info_len += ext_info_head_len + output_num * sizeof(ShapeAndType);
uint64_t ext_info_offset = ext_info.size();
ext_info.resize(ext_info_len, 0);
char *ext_info_buf = ext_info.data();
// deal1: unknown shape type
ExtInfo *info = reinterpret_cast<ExtInfo *>(ext_info_buf + ext_info_offset);
info->infoType = FWK_ADPT_EXT_SHAPE_TYPE;
info->infoLen = sizeof(int32_t);
ext_info_offset += ext_info_head_len;
int32_t *shape_type = reinterpret_cast<int32_t *>(ext_info_buf + ext_info_offset);
*shape_type = unknown_shape_type;
ext_info_offset += info->infoLen;
// deal2:input ShapeAndType
info = reinterpret_cast<ExtInfo *>(ext_info_buf + ext_info_offset);
info->infoType = FWK_ADPT_EXT_INPUT_SHAPE;
info->infoLen = input_num * sizeof(ShapeAndType);
ext_info_offset += ext_info_head_len;
ShapeAndType *inputs = reinterpret_cast<ShapeAndType *>(ext_info_buf + ext_info_offset);
for (size_t input_index = 0; input_index < input_num; input_index++) {
TypeId input_type = AnfAlgo::GetInputDeviceDataType(anf_node, input_index);
std::vector<size_t> input_shape;
int32_t input_data_type;
if (input_type == kObjectTypeString) {
auto cnode = anf_node->cast<CNodePtr>();
MS_EXCEPTION_IF_NULL(cnode);
auto input_node = cnode->inputs()[input_index + 1];
auto value_ptr = GetValueNode(input_node);
auto value = GetValue<std::string>(value_ptr);
input_shape.push_back(1);
input_shape.push_back(value.size());
input_data_type = AicpuOpUtil::MsTypeToProtoType(kTypeUnknown);
} else {
input_shape = AnfAlgo::GetInputDeviceShape(anf_node, input_index);
input_data_type = AicpuOpUtil::MsTypeToProtoType(input_type);
}
inputs[input_index].type = input_data_type;
size_t input_shape_index = 0;
for (; input_shape_index < input_shape.size(); input_shape_index++) {
inputs[input_index].dims[input_shape_index] = SizeToLong(input_shape[input_shape_index]);
}
if (input_shape.size() < kMaxShapeDims) {
inputs[input_index].dims[input_shape_index] = LLONG_MIN;
}
}
ext_info_offset += info->infoLen;
// deal3:output ShapeAndType
info = reinterpret_cast<ExtInfo *>(ext_info_buf + ext_info_offset);
info->infoType = FWK_ADPT_EXT_OUTPUT_SHAPE;
info->infoLen = output_num * sizeof(ShapeAndType);
ext_info_offset += ext_info_head_len;
ShapeAndType *outputs = reinterpret_cast<ShapeAndType *>(ext_info_buf + ext_info_offset);
for (size_t output_index = 0; output_index < output_num; output_index++) {
std::vector<size_t> output_shape = AnfAlgo::GetOutputDeviceShape(anf_node, output_index);
TypeId output_type = AnfAlgo::GetOutputDeviceDataType(anf_node, output_index);
int32_t output_data_type = AicpuOpUtil::MsTypeToProtoType(output_type);
outputs[output_index].type = output_data_type;
size_t output_shape_index = 0;
for (; output_shape_index < output_shape.size(); output_shape_index++) {
outputs[output_index].dims[output_shape_index] = SizeToLong(output_shape[output_shape_index]);
}
if (output_shape_index < kMaxShapeDims) {
outputs[output_index].dims[output_shape_index] = LLONG_MIN;
}
}
// set ext info
kernel_mod_ptr->SetExtInfo(ext_info);
return true;
}
KernelModPtr AicpuOpBuild(const std::shared_ptr<AnfNode> &anf_node) {
MS_EXCEPTION_IF_NULL(anf_node);
std::string op_name = AnfAlgo::GetCNodeName(anf_node);
@ -300,6 +405,11 @@ KernelModPtr AicpuOpBuild(const std::shared_ptr<AnfNode> &anf_node) {
if (!CreateNodeDefBytes(anf_node, kernel_mod_ptr)) {
MS_LOG(EXCEPTION) << "Create nodeDefBytes faild!";
}
if (!CreateExtInfo(anf_node, kernel_mod_ptr)) {
MS_LOG(EXCEPTION) << "Create nodeDefBytes faild!";
}
if (!SetIOSize(anf_node, kernel_mod_ptr)) {
MS_LOG(EXCEPTION) << "Set input output size list failed.";
}

@ -43,6 +43,7 @@ AicpuOpKernelMod::~AicpuOpKernelMod() {
input_size_list_.clear();
output_size_list_.clear();
workspace_size_list_.clear();
ext_info_.clear();
}
void AicpuOpKernelMod::SetInputSizeList(const std::vector<size_t> &size_list) { input_size_list_ = size_list; }
@ -54,6 +55,7 @@ const std::vector<size_t> &AicpuOpKernelMod::GetWorkspaceSizeList() const { retu
void AicpuOpKernelMod::SetInputList(const std::vector<int64_t> &inputList) { inputList_ = inputList; }
void AicpuOpKernelMod::SetOutputList(const std::vector<int64_t> &outputList) { outputList_ = outputList; }
void AicpuOpKernelMod::SetNodeDef(const std::string &nodeDef) { (void)node_def_str_.assign(nodeDef); }
void AicpuOpKernelMod::SetExtInfo(const std::string &ext_info) { ext_info_ = ext_info; }
void AicpuOpKernelMod::SetNodeName(const std::string &node_name) { node_name_ = node_name; }
void AicpuOpKernelMod::SetAnfNode(const mindspore::AnfNodePtr &anf_node) {
MS_EXCEPTION_IF_NULL(anf_node);
@ -84,16 +86,30 @@ void AicpuOpKernelMod::CreateCpuKernelInfo(const std::vector<AddressPtr> &inputs
auto node_def_len = node_def_str_.length();
param_len += node_def_len;
param_len += sizeof(uint32_t);
AicpuParamHead aicpu_param_head;
aicpu_param_head.length = param_len;
aicpu_param_head.ioAddrNum = io_addrs_num;
if (ext_info_.empty()) {
MS_LOG(INFO) << "Static Shape Kernel";
aicpu_param_head.extInfoLength = 0;
aicpu_param_head.extInfoAddr = 0;
} else {
MS_LOG(INFO) << "Dynamic Kernel Ext Info size:" << ext_info_.size();
}
// Create taskArgs: AicpuParamHead + ioAddrs + notifyId + customizedAttr
AicpuParamHead paramHead = {static_cast<uint32_t>(param_len), static_cast<uint32_t>(io_addrs_num)};
args_.clear();
(void)args_.append(reinterpret_cast<const char *>(&paramHead), sizeof(AicpuParamHead));
(void)args_.append(reinterpret_cast<const char *>(&aicpu_param_head), sizeof(AicpuParamHead));
// TaskArgs append ioAddrs
if (io_addrs_size != 0) {
(void)args_.append(reinterpret_cast<const char *>(io_addrs.data()), io_addrs_size);
}
// size for node_def
args_.append(reinterpret_cast<const char *>(&node_def_len), sizeof(uint32_t));
// When it's aicpu customized ops, taskArgs should append customized attr
if (node_def_len != 0) {
(void)args_.append(reinterpret_cast<const char *>(node_def_str_.data()), node_def_len);
@ -145,8 +161,9 @@ std::vector<TaskInfoPtr> AicpuOpKernelMod::GenTask(const std::vector<AddressPtr>
node_name_ = kTopKV2;
}
AicpuTaskInfoPtr task_info_ptr = make_shared<ge::model_runner::AicpuTaskInfo>(
kernel_name_, stream_id, node_so_, node_name_, node_def_str_, input_data_addrs, output_data_addrs, NeedDump());
AicpuTaskInfoPtr task_info_ptr =
make_shared<ge::model_runner::AicpuTaskInfo>(kernel_name_, stream_id, node_so_, node_name_, node_def_str_,
ext_info_, input_data_addrs, output_data_addrs, NeedDump());
MS_LOG(INFO) << "AicpuOpKernelMod GenTask end";
return {task_info_ptr};

@ -36,6 +36,7 @@ class AicpuOpKernelMod : public AscendKernelMod {
void SetOutputList(const std::vector<int64_t> &outputList);
void SetAnfNode(const AnfNodePtr &anf_node);
void SetNodeDef(const std::string &nodeDef);
void SetExtInfo(const std::string &ext_info);
void SetNodeName(const std::string &node_name);
/**
@ -58,6 +59,7 @@ class AicpuOpKernelMod : public AscendKernelMod {
std::string node_def_str_;
std::string node_name_;
std::string node_so_;
std::string ext_info_;
std::vector<int64_t> inputList_;
std::vector<int64_t> outputList_;
AnfNodePtr anf_node_;

@ -21,7 +21,6 @@
#include <map>
#include <string>
#include "backend/kernel_compiler/kernel.h"
namespace mindspore {
namespace kernel {
constexpr auto kInitDataSetQueue = "InitDataSetQueue";
@ -50,6 +49,36 @@ struct AicpuParamHead {
uint64_t extInfoAddr; // extInfo address
} __attribute__((packed));
const uint32_t kExtInfoHeadSize = 8;
struct ExtInfo {
int32_t infoType; // extend type
uint32_t infoLen; // length for infoMsg
char infoMsg[0]; // extend value
} __attribute__((packed));
// Extent info ShapeAndType
const uint32_t kMaxShapeDims = 8;
struct ShapeAndType {
int32_t type;
int64_t dims[kMaxShapeDims];
} __attribute__((packed));
// Extend Info type for task
enum FWKTaskExtInfoType {
FWK_ADPT_EXT_SHAPE_TYPE = 0,
FWK_ADPT_EXT_INPUT_SHAPE,
FWK_ADPT_EXT_OUTPUT_SHAPE,
FWK_ADPT_EXT_INVALID
};
// for unknown shape op type
enum UnknowShapeOpType {
DEPEND_IN_SHAPE = 1, // op out shape get by input shape
DEPEND_CONST_VALUE = 2, // op out shape get by const op value
DEPEND_SHAPE_RANGE = 3, // op out shape get by range
DEPEND_COMPUTE = 4 // op out shape get by totally computing
};
class AicpuOpUtil {
public:
static int MsTypeToProtoType(TypeId ms_type);

@ -26,7 +26,7 @@ message AttrValue {
repeated int64 i = 3 [ packed = true ]; //"array(int)"
repeated float f = 4 [ packed = true ]; //"array(float)"
repeated bool b = 5 [ packed = true ]; //"array(bool)"
repeated DataType type = 6 [ packed = true ]; //"array(type)"
repeated int32 type = 6 [ packed = true ]; //"array(type)"
repeated TensorShape shape = 7; //"array(shape)"
repeated Tensor tensor = 8; //"array(tensor)"
}

@ -18,9 +18,16 @@ package mindspore;
import "attr.proto";
import "tensor.proto";
message DynamicIdx {
int32 idx = 1;
int32 num = 2;
}
message NodeDef {
string op = 2;
map<string, AttrValue> attrs = 3;
repeated Tensor inputs = 4;
repeated Tensor outputs = 5;
map<string, DynamicIdx> dym_inputs = 6;
map<string, DynamicIdx> dym_outputs = 7;
}

@ -26,9 +26,12 @@ message Tensor {
TensorShape tensor_shape = 1;
// tensor content data type
DataType tensor_type = 2;
int32 tensor_type = 2;
// tensor memory device
// data located memory device , "DDR" "HBM" OR "NONE"
string mem_device = 3;
string name = 4;
uint64 data_ptr = 5;
uint64 data_size = 6;
}

@ -31,5 +31,5 @@ message TensorShape {
bool unknown_rank = 3;
// data format "NHWC" "NCHW" "NC1HWC0" OR "NONE"
string data_format = 4;
int32 data_format = 4;
};

Some files were not shown because too many files have changed in this diff Show More

Loading…
Cancel
Save