Research Article

A Novel Low-Bit Quantization Strategy for Compressing Deep Neural Networks

Table 1

Expect error of different values.

Scheme = 0.0625 = 0.125 = 0.25 = 0.5 = 1

0.40780.32980.21060.08250.0458
0.32980.21030.07950.02390.0443
0.21020.07910.02090.02230.0443
0.07900.02050.01930.02230.0443
0.02040.01890.01930.02230.0443
0.01890.01890.01930.02230.0443