StairNet: visual recognition of stairs for human-robot locomotion

被引:3
|
作者
Kurbis, Andrew Garrett [1 ,3 ]
Kuzmenko, Dmytro [4 ]
Ivanyuk-Skulskiy, Bogdan [4 ]
Mihailidis, Alex [1 ,3 ]
Laschowski, Brokoslaw [2 ,3 ,5 ]
机构
[1] Univ Toronto, Inst Biomed Engn, Toronto, ON, Canada
[2] Univ Toronto, Robot Inst, Toronto, ON, Canada
[3] Toronto Rehabil Inst, KITE Res Inst, Toronto, ON, Canada
[4] Natl Univ Kyiv, Mohyla Acad, Dept Math, Kiev, Ukraine
[5] Univ Toronto, Dept Mech & Ind Engn, Toronto, ON, Canada
关键词
Computer vision; Deep learning; Wearable robotics; Prosthetics; Exoskeletons;
D O I
10.1186/s12938-024-01216-0
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Human-robot walking with prosthetic legs and exoskeletons, especially over complex terrains, such as stairs, remains a significant challenge. Egocentric vision has the unique potential to detect the walking environment prior to physical interactions, which can improve transitions to and from stairs. This motivated us to develop the StairNet initiative to support the development of new deep learning models for visual perception of real-world stair environments. In this study, we present a comprehensive overview of the StairNet initiative and key research to date. First, we summarize the development of our large-scale data set with over 515,000 manually labeled images. We then provide a summary and detailed comparison of the performances achieved with different algorithms (i.e., 2D and 3D CNN, hybrid CNN and LSTM, and ViT networks), training methods (i.e., supervised learning with and without temporal data, and semi-supervised learning with unlabeled images), and deployment methods (i.e., mobile and embedded computing), using the StairNet data set. Finally, we discuss the challenges and future directions. To date, our StairNet models have consistently achieved high classification accuracy (i.e., up to 98.8%) with different designs, offering trade-offs between model accuracy and size. When deployed on mobile devices with GPU and NPU accelerators, our deep learning models achieved inference speeds up to 2.8 ms. In comparison, when deployed on our custom-designed CPU-powered smart glasses, our models yielded slower inference speeds of 1.5 s, presenting a trade-off between human-centered design and performance. Overall, the results of numerous experiments presented herein provide consistent evidence that StairNet can be an effective platform to develop and study new deep learning models for visual perception of human-robot walking environments, with an emphasis on stair recognition. This research aims to support the development of next-generation vision-based control systems for robotic prosthetic legs, exoskeletons, and other mobility assistive technologies.
引用
收藏
页数:19
相关论文
共 50 条
  • [31] Recognition in Human-Robot Interaction: The Gateway to Engagement
    Brinck, Ingar
    Balkenius, Christian
    2019 JOINT IEEE 9TH INTERNATIONAL CONFERENCE ON DEVELOPMENT AND LEARNING AND EPIGENETIC ROBOTICS (ICDL-EPIROB), 2019, : 31 - 36
  • [32] GESTURE RECOGNITION FOR CONTROL IN HUMAN-ROBOT INTERACTIONS
    Reid, Chris
    Samanta, Biswanath
    ASME INTERNATIONAL MECHANICAL ENGINEERING CONGRESS AND EXPOSITION, 2014, VOL 4B, 2015,
  • [33] Facial Expression Recognition for Human-Robot Interaction
    Hsu, Shih-Chung
    Huang, Hsin-Hui
    Huang, Chung-Lin
    2017 FIRST IEEE INTERNATIONAL CONFERENCE ON ROBOTIC COMPUTING (IRC), 2017, : 1 - 7
  • [34] Human-Robot Teamwork using Activity Recognition and Human Instruction
    Cuntoor, Naresh P.
    Collins, Roderic
    Hoogs, Anthony J.
    2012 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2012, : 459 - 465
  • [35] Human Activity Recognition in the Context of Industrial Human-Robot Interaction
    Roitberg, Alina
    Perzylo, Alexander
    Somani, Nikhil
    Giuliani, Manuel
    Rickert, Markus
    Knoll, Alois
    2014 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA), 2014,
  • [36] MULTIMODAL HUMAN ACTION RECOGNITION IN ASSISTIVE HUMAN-ROBOT INTERACTION
    Rodomagoulakis, I.
    Kardaris, N.
    Pitsikalis, V.
    Mavroudi, E.
    Katsamanis, A.
    Tsiami, A.
    Maragos, P.
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 2702 - 2706
  • [37] Human-robot collaborative interaction with human perception and action recognition
    Yu, Xinyi
    Zhang, Xin
    Xu, Chengjun
    Ou, Linlin
    NEUROCOMPUTING, 2024, 563
  • [38] Gesture recognition based human-robot interactive control for robot soccer
    Shieh, Ming-Yuan
    Wang, Chen-Yang
    Wu, Wen-Lan
    Liang, Jing-Min
    MICROSYSTEM TECHNOLOGIES-MICRO-AND NANOSYSTEMS-INFORMATION STORAGE AND PROCESSING SYSTEMS, 2021, 27 (04): : 1175 - 1186
  • [39] Visual Exploration and Analysis of Human-Robot Interaction Rules
    Zhang, Hui
    Boyles, Michael J.
    VISUALIZATION AND DATA ANALYSIS 2013, 2013, 8654
  • [40] Vision based gesture recognition for human-robot symbiosis
    Bhuiyan, Md. Al-Amin
    Islam, Md. Ezharul
    Begum, Nasima
    Hasanuzzaman, Md.
    Liu, Chang Hong
    Ueno, Haruki
    PROCEEDINGS OF 10TH INTERNATIONAL CONFERENCE ON COMPUTER AND INFORMATION TECHNOLOGY (ICCIT 2007), 2007, : 418 - +