You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: python/paddle/fluid/contrib/int8_inference/README.md
+8-7
Original file line number
Diff line number
Diff line change
@@ -6,7 +6,7 @@ PaddlePaddle supports offline INT8 calibration to accelerate the inference speed
6
6
You need to install at least PaddlePaddle-1.3 python package `pip install paddlepaddle==1.3`.
7
7
8
8
## 1. How to generate INT8 model
9
-
You can refer to the unit test in [test_calibration.py](../tests/test_calibration.py). Basically, there are three steps:
9
+
You can refer to the unit test in [test_calibration_resnet50.py](../tests/test_calibration_resnet50.py). Basically, there are three steps:
10
10
* Construct calibration object.
11
11
12
12
```python
@@ -68,18 +68,19 @@ Notes:
68
68
* The INT8 theoretical speedup is 4X on Intel® Xeon® Cascadelake Server (please refer to `The theoretical peak compute gains are 4x int8 OPS over fp32 OPS.` in [Reference](https://software.intel.com/en-us/articles/lower-numerical-precision-deep-learning-inference-and-training"Reference")). Therefore, op-level gain is 4X and topology-level is smaller.
0 commit comments