Bioinspired In-Sensor Multimodal Fusion for Enhanced Spatial and Spatiotemporal Association

被引:10
作者
Ma, Sijie [1 ,2 ,3 ]
Zhou, Yue [1 ,2 ,3 ]
Wan, Tianqing [1 ,2 ,3 ]
Ren, Qinqi [1 ,2 ,3 ]
Yan, Jianmin [1 ,2 ,3 ]
Fan, Lingwei [1 ,2 ,3 ]
Yuan, Huanmei [4 ]
Chan, Mansun [4 ]
Chai, Yang [1 ,2 ,3 ]
机构
[1] Hong Kong Polytech Univ, Dept Appl Phys, Kowloon, Hong Kong 999077, Peoples R China
[2] Hong Kong Polytech Univ, Joint Res Ctr Microelect, Kowloon, Hong Kong 999077, Peoples R China
[3] Hong Kong Polytech Univ, Joint Res Ctr Microelect, Kowloon, Hong Kong 999077, Peoples R China
[4] Hong Kong Univ Sci & Technol, Dept Elect & Comp Engn, Kowloon, Hong Kong 999077, Peoples R China
关键词
in-sensor computing; neuromorphic computing; multimodal integration; two-dimensional (2D) semiconductor; edge computing; floating gate transistor; INTEGRATION; GENERATION;
D O I
10.1021/acs.nanolett.4c01727
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Multimodal perception can capture more precise and comprehensive information compared with unimodal approaches. However, current sensory systems typically merge multimodal signals at computing terminals following parallel processing and transmission, which results in the potential loss of spatial association information and requires time stamps to maintain temporal coherence for time-series data. Here we demonstrate bioinspired in-sensor multimodal fusion, which effectively enhances comprehensive perception and reduces the level of data transfer between sensory terminal and computation units. By adopting floating gate phototransistors with reconfigurable photoresponse plasticity, we realize the agile spatial and spatiotemporal fusion under nonvolatile and volatile photoresponse modes. To realize an optimal spatial estimation, we integrate spatial information from visual-tactile signals. For dynamic events, we capture and fuse in real time spatiotemporal information from visual-audio signals, realizing a dance-music synchronization recognition task without a time-stamping process. This in-sensor multimodal fusion approach provides the potential to simplify the multimodal integration system, extending the in-sensor computing paradigm.
引用
收藏
页码:7091 / 7099
页数:9
相关论文
empty
未找到相关数据