Feb. 22, 2023, 2:10 a.m. | Xinghua Xue, Cheng Liu, Haitong Huang, Ying Wang, Bing Yang, Tao Luo, Lei Zhang, Huawei Li, Xiaowei Li

cs.CR updates on arXiv.org arxiv.org

Vision Transformers (ViTs) with outstanding performance becomes a popular
backbone of deep learning models for the main-stream vision tasks including
classification, object detection, and segmentation. Other than the performance,
reliability is also a critical metric for the adoption of ViTs in
safety-critical applications such as autonomous driving and robotics. With the
observation that the major computing blocks in ViTs such as multi-head
attention and feed forward are usually performed with general matrix
multiplication (GEMM), we propose to adopt a classical …

adoption algorithm applications attention autonomous autonomous driving classification computing critical deep learning detection driving forward head main major object performance popular reliability robotics safety safety-critical segmentation stream tolerance transformers

Information Assurance Security Specialist (IASS)

@ OBXtek Inc. | United States

Cyber Security Technology Analyst

@ Airbus | Bengaluru (Airbus)

Vice President, Cyber Operations Engineer

@ BlackRock | LO9-London - Drapers Gardens

Cryptography Software Developer

@ Intel | USA - AZ - Chandler

Lead Consultant, Geology

@ WSP | Richmond, VA, United States

BISO Cybersecurity Director

@ ABM Industries | Alpharetta, GA, United States