TY - GEN
T1 - AutoReP
T2 - 2023 IEEE/CVF International Conference on Computer Vision, ICCV 2023
AU - Peng, Hongwu
AU - Huang, Shaoyi
AU - Zhou, Tong
AU - Luo, Yukui
AU - Wang, Chenghong
AU - Wang, Zigeng
AU - Zhao, Jiahui
AU - Xie, Xi
AU - Li, Ang
AU - Geng, Tony
AU - Mahmood, Kaleel
AU - Wen, Wujie
AU - Xu, Xiaolin
AU - Ding, Caiwen
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - The growth of the Machine-Learning-As-A-Service (MLaaS) market has highlighted clients' data privacy and security issues. Private inference (PI) techniques using cryptographic primitives offer a solution but often have high computation and communication costs, particularly with non-linear operators like ReLU. Many attempts to reduce ReLU operations exist, but they may need heuristic threshold selection or cause substantial accuracy loss. This work introduces AutoReP, a gradient-based approach to lessen non-linear operators and alleviate these issues. It automates the selection of ReLU and polynomial functions to speed up PI applications and introduces distribution-aware polynomial approximation (DaPa) to maintain model expressivity while accurately approximating ReLUs. Our experimental results demonstrate significant accuracy improvements of 6.12% (94.31%, 12.9K ReLU budget, CIFAR-10), 8.39% (74.92%, 12.9K ReLU budget, CIFAR-100), and 9.45% (63.69%, 55K ReLU budget, Tiny-ImageNet) over current state-of-the-art methods, e.g., SNL. Morever, AutoReP is applied to EfficientNet-B2 on ImageNet dataset, and achieved 75.55% accuracy with 176.1 × ReLU budget reduction. The codes are shared on Github https://github.com/HarveyP123/AutoReP.
AB - The growth of the Machine-Learning-As-A-Service (MLaaS) market has highlighted clients' data privacy and security issues. Private inference (PI) techniques using cryptographic primitives offer a solution but often have high computation and communication costs, particularly with non-linear operators like ReLU. Many attempts to reduce ReLU operations exist, but they may need heuristic threshold selection or cause substantial accuracy loss. This work introduces AutoReP, a gradient-based approach to lessen non-linear operators and alleviate these issues. It automates the selection of ReLU and polynomial functions to speed up PI applications and introduces distribution-aware polynomial approximation (DaPa) to maintain model expressivity while accurately approximating ReLUs. Our experimental results demonstrate significant accuracy improvements of 6.12% (94.31%, 12.9K ReLU budget, CIFAR-10), 8.39% (74.92%, 12.9K ReLU budget, CIFAR-100), and 9.45% (63.69%, 55K ReLU budget, Tiny-ImageNet) over current state-of-the-art methods, e.g., SNL. Morever, AutoReP is applied to EfficientNet-B2 on ImageNet dataset, and achieved 75.55% accuracy with 176.1 × ReLU budget reduction. The codes are shared on Github https://github.com/HarveyP123/AutoReP.
UR - http://www.scopus.com/inward/record.url?scp=85175135544&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85175135544&partnerID=8YFLogxK
U2 - 10.1109/ICCV51070.2023.00478
DO - 10.1109/ICCV51070.2023.00478
M3 - Conference contribution
AN - SCOPUS:85175135544
T3 - Proceedings of the IEEE International Conference on Computer Vision
SP - 5155
EP - 5165
BT - Proceedings - 2023 IEEE/CVF International Conference on Computer Vision, ICCV 2023
Y2 - 2 October 2023 through 6 October 2023
ER -