Software/Hardware Co-design for Multi-modal Multi-task Learning in Autonomous Systems

被引:14
作者
Hao, Cong [1 ]
Chen, Deming [2 ]
机构
[1] Georgia Inst Technol, Atlanta, GA 30332 USA
[2] Univ Illinois, Urbana, IL USA
来源
2021 IEEE 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS) | 2021年
关键词
D O I
10.1109/AICAS51828.2021.9458577
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Optimizing the quality of result (QoR) and the quality of service (QoS) of AI-empowered autonomous systems simultaneously is very challenging. First, there are multiple input sources, e.g., multi-modal data from different sensors, requiring diverse data preprocessing, sensor fusion, and feature aggregation. Second, there are multiple tasks that require various AI models to run simultaneously, e.g., perception, localization, and control. Third, the computing and control system is heterogeneous, composed of hardware components with varied features, such as embedded CPUs, GPUs, FPGAs, and dedicated accelerators. Therefore, autonomous systems essentially require multi-modal multi-task (MMMT) learning which must be aware of hardware performance and implementation strategies. While MMMT learning has been attracting intensive research interests, its applications in autonomous systems are still underexplored. In this paper, we first discuss the opportunities of applying MMMT techniques in autonomous systems, and then discuss the unique challenges that must be solved. In addition, we discuss the necessity and opportunities of MMMT model and hardware co-design, which is critical for autonomous systems especially with power/resource-limited or heterogeneous platforms. We formulate the MMMT model and heterogeneous hardware implementation co-design as a differentiable optimization problem, with the objective of improving the solution quality and reducing the overall power consumption and critical path latency. We advocate for further explorations of MMMT in autonomous systems and software/hardware co-design solutions.
引用
收藏
页数:5
相关论文
共 62 条
[1]   A Deep Multi-task Contextual Attention Framework for Multi-modal Affect Analysis [J].
Akhtar, Md Shad ;
Chauhan, Dushyant Singh ;
Ekbal, Asif .
ACM TRANSACTIONS ON KNOWLEDGE DISCOVERY FROM DATA, 2020, 14 (03)
[2]  
[Anonymous], 2017, arXiv
[3]  
Aranjuelo N., 2018, AMDO
[4]  
Baker B., 2016, P INT C LEARN REPR I
[5]   Multimodal Machine Learning: A Survey and Taxonomy [J].
Baltrusaitis, Tadas ;
Ahuja, Chaitanya ;
Morency, Louis-Philippe .
IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2019, 41 (02) :423-443
[6]   Choosing the Best Sensor Fusion Method: A Machine-Learning Approach [J].
Brena, Ramon F. ;
Aguileta, Antonio A. ;
Trejo, Luis A. ;
Molino-Minero-Re, Erik ;
Mayora, Oscar .
SENSORS, 2020, 20 (08)
[7]  
Cao NY, 2019, ISSCC DIG TECH PAP I, V62, P222, DOI 10.1109/ISSCC.2019.8662311
[8]   Multitask learning [J].
Caruana, R .
MACHINE LEARNING, 1997, 28 (01) :41-75
[9]   MultiNet: Multi-Modal Multi-Task Learning for Autonomous Driving [J].
Chowdhuri, Sauhaarda ;
Pankaj, Tushar ;
Zipser, Karl .
2019 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2019, :1496-1504
[10]  
Crawshaw M, ARXIV 2020 ARXIV2009