5.5 KiB
INT8 MKL-DNN quantization
This document describes how to use Paddle inference Engine to convert the FP32 model to INT8 model on ResNet-50 and MobileNet-V1. We provide the instructions on enabling INT8 MKL-DNN quantization in Paddle inference and show the ResNet-50 and MobileNet-V1 results in accuracy and performance.
0. Install PaddlePaddle
Follow PaddlePaddle installation instruction to install PaddlePaddle. If you build PaddlePaddle yourself, please use the following cmake arguments.
cmake .. -DWITH_TESTING=ON -WITH_FLUID_ONLY=ON -DWITH_GPU=OFF -DWITH_MKL=ON -DWITH_MKLDNN=ON -DWITH_INFERENCE_API_TEST=ON -DON_INFER=ON
Note: MKL-DNN and MKL are required.
1. Enable INT8 MKL-DNN quantization
For reference, please examine the code of unit test enclosed in analyzer_int8_image_classification_tester.cc.
-
Create Analysis config
INT8 quantization is one of the optimizations in analysis config. More information about analysis config can be found here
-
Create quantize config by analysis config
We enable the MKL-DNN quantization procedure by calling an appropriate method from analysis config. Afterwards, all the required quantization parameters (quantization op names, quantization strategies etc.) can be set through quantizer config which is present in the analysis config. It is also necessary to specify a pre-processed warmup dataset and desired batch size.
//Enable MKL-DNN quantization
cfg.EnableMkldnnQuantizer();
//use analysis config to call the MKL-DNN quantization config
cfg.mkldnn_quantizer_config()->SetWarmupData(warmup_data);
cfg.mkldnn_quantizer_config()->SetWarmupBatchSize(100);
2. Accuracy and Performance benchmark
We provide the results of accuracy and performance measured on Intel(R) Xeon(R) Gold 6271 on single core.
Dataset: ILSVRC2012 Validation dataset
I. Top-1 Accuracy on Intel(R) Xeon(R) Gold 6271
Model | FP32 Accuracy | INT8 Accuracy | Accuracy Diff(FP32-INT8) |
---|---|---|---|
GoogleNet | 70.50% | 69.81% | 0.69% |
MobileNet-V1 | 70.78% | 70.42% | 0.36% |
MobileNet-V2 | 71.90% | 71.35% | 0.55% |
ResNet-101 | 77.50% | 77.42% | 0.08% |
ResNet-50 | 76.63% | 76.52% | 0.11% |
VGG16 | 72.08% | 72.03% | 0.05% |
VGG19 | 72.57% | 72.55% | 0.02% |
II. Throughput on Intel(R) Xeon(R) Gold 6271 (batch size 1 on single core)
Model | FP32 Throughput(images/s) | INT8 Throughput(images/s) | Ratio(INT8/FP32) |
---|---|---|---|
GoogleNet | 34.06 | 72.79 | 2.14 |
MobileNet-V1 | 80.02 | 230.65 | 2.88 |
MobileNet-V2 | 99.38 | 206.92 | 2.08 |
ResNet-101 | 7.38 | 27.31 | 3.70 |
ResNet-50 | 13.71 | 50.55 | 3.69 |
VGG16 | 3.64 | 10.56 | 2.90 |
VGG19 | 2.95 | 9.02 | 3.05 |
Notes:
-
Measurement of accuracy requires a model which accepts two inputs: data and labels.
-
Different sampling batch size data may cause slight difference on INT8 top accuracy.
-
CAPI performance data is better than python API performance data because of the python overhead. Especially for the small computational model, python overhead will be more obvious.
3. Commands to reproduce the above accuracy and performance benchmark
Two steps to reproduce the above-mentioned accuracy results, and we take GoogleNet benchmark as an example:
-
Prepare dataset
Running the following commands to download and preprocess the ILSVRC2012 Validation dataset.
cd /PATH/TO/PADDLE/build
python ../paddle/fluid/inference/tests/api/full_ILSVRC2012_val_preprocess.py
Then the ILSVRC2012 Validation dataset will be preprocessed and saved by default in ~/.cache/paddle/dataset/int8/download/int8_full_val.bin
-
Commands to reproduce benchmark
You can run test_analyzer_int8_imagenet_classification
with the following arguments to reproduce the accuracy result on GoogleNet.
./paddle/fluid/inference/tests/api/test_analyzer_int8_image_classification --infer_model=third_party/inference_demo/int8v2/resnet50/model --infer_data=/~/.cache/paddle/dataset/int8/download/int8_full_val.bin --batch_size=1 --paddle_num_threads=1
To verify all the 7 models, you need to set the parameter of --infer_model
to one of the following values in command line:
--infer_model /PATH/TO/PADDLE/build/third_party/inference_demo/int8v2/MODEL_NAME/model
MODEL_NAME=googlenet, mobilenetv1, mobilenetv2, resnet101, resnet50, vgg16, vgg19