Dynamic sparse attention for scalable transformer acceleration Jan 1, 2022· Liu Liu , Zheng Qu , Zhaodong Chen , Fengbin Tu , Yufei Ding Prof. Yuan Xie · 0 min read Cite Type Journal article Publication IEEE Transactions on Computers Last updated on Jan 1, 2022 Authors Prof. Yuan Xie Chair Professor Fang Professor of Engineering | Chair Professor | IEEE/ACM/AAAS Fellow ← Compact Multi-level Sparse Neural Networks with Input Independent Dynamic Rerouting Jan 1, 2022 EPQuant: A Graph Neural Network compression approach based on product quantization Jan 1, 2022 →