Feb. 22, 2023, 2:10 a.m. | Xinghua Xue, Cheng Liu, Haitong Huang, Ying Wang, Bing Yang, Tao Luo, Lei Zhang, Huawei Li, Xiaowei Li

cs.CR updates on arXiv.org arxiv.org

Vision Transformers (ViTs) with outstanding performance becomes a popular
backbone of deep learning models for the main-stream vision tasks including
classification, object detection, and segmentation. Other than the performance,
reliability is also a critical metric for the adoption of ViTs in
safety-critical applications such as autonomous driving and robotics. With the
observation that the major computing blocks in ViTs such as multi-head
attention and feed forward are usually performed with general matrix
multiplication (GEMM), we propose to adopt a classical …

adoption algorithm applications attention autonomous autonomous driving classification computing critical deep learning detection driving forward head main major object performance popular reliability robotics safety safety-critical segmentation stream tolerance transformers

CyberSOC Technical Lead

@ Integrity360 | Sandyford, Dublin, Ireland

Cyber Security Strategy Consultant

@ Capco | New York City

Cyber Security Senior Consultant

@ Capco | Chicago, IL

Sr. Product Manager

@ MixMode | Remote, US

Corporate Intern - Information Security (Year Round)

@ Associated Bank | US WI Remote

Senior Offensive Security Engineer

@ CoStar Group | US-DC Washington, DC