TY - JOUR
T1 - PriViT
T2 - Vision Transformers for Private Inference
AU - Dhyani, Naren
AU - Mo, Jianqiao Cambridge
AU - Yubeaton, Patrick
AU - Cho, Minsu
AU - Joshi, Ameya
AU - Garg, Siddharth
AU - Reagen, Brandon
AU - Hegde, Chinmay
N1 - Publisher Copyright:
© 2024, Transactions on Machine Learning Research. All rights reserved.
PY - 2024
Y1 - 2024
N2 - The Vision Transformer (ViT) architecture has emerged as the backbone of choice for state-of-the-art deep models for computer vision applications. However, ViTs are ill-suited for private inference using secure multi-party computation (MPC) protocols, due to the large number of non-polynomial operations (self-attention, feed-forward rectifiers, layer normalization). We develop PriViT, a gradient-based algorithm to selectively Taylorize nonlinearities in ViTs while maintaining their prediction accuracy. Our algorithm is conceptually very simple, easy to implement, and achieves improved performance over existing MPC-friendly transformer architectures in terms of the latency-accuracy Pareto frontier.
AB - The Vision Transformer (ViT) architecture has emerged as the backbone of choice for state-of-the-art deep models for computer vision applications. However, ViTs are ill-suited for private inference using secure multi-party computation (MPC) protocols, due to the large number of non-polynomial operations (self-attention, feed-forward rectifiers, layer normalization). We develop PriViT, a gradient-based algorithm to selectively Taylorize nonlinearities in ViTs while maintaining their prediction accuracy. Our algorithm is conceptually very simple, easy to implement, and achieves improved performance over existing MPC-friendly transformer architectures in terms of the latency-accuracy Pareto frontier.
UR - http://www.scopus.com/inward/record.url?scp=85214851057&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85214851057&partnerID=8YFLogxK
M3 - Article
AN - SCOPUS:85214851057
SN - 2835-8856
VL - 2024
JO - Transactions on Machine Learning Research
JF - Transactions on Machine Learning Research
ER -