StairNet: visual recognition of stairs for human-robot locomotion

被引:3
|
作者
Kurbis, Andrew Garrett [1 ,3 ]
Kuzmenko, Dmytro [4 ]
Ivanyuk-Skulskiy, Bogdan [4 ]
Mihailidis, Alex [1 ,3 ]
Laschowski, Brokoslaw [2 ,3 ,5 ]
机构
[1] Univ Toronto, Inst Biomed Engn, Toronto, ON, Canada
[2] Univ Toronto, Robot Inst, Toronto, ON, Canada
[3] Toronto Rehabil Inst, KITE Res Inst, Toronto, ON, Canada
[4] Natl Univ Kyiv, Mohyla Acad, Dept Math, Kiev, Ukraine
[5] Univ Toronto, Dept Mech & Ind Engn, Toronto, ON, Canada
关键词
Computer vision; Deep learning; Wearable robotics; Prosthetics; Exoskeletons;
D O I
10.1186/s12938-024-01216-0
中图分类号
R318 [生物医学工程];
学科分类号
0831 ;
摘要
Human-robot walking with prosthetic legs and exoskeletons, especially over complex terrains, such as stairs, remains a significant challenge. Egocentric vision has the unique potential to detect the walking environment prior to physical interactions, which can improve transitions to and from stairs. This motivated us to develop the StairNet initiative to support the development of new deep learning models for visual perception of real-world stair environments. In this study, we present a comprehensive overview of the StairNet initiative and key research to date. First, we summarize the development of our large-scale data set with over 515,000 manually labeled images. We then provide a summary and detailed comparison of the performances achieved with different algorithms (i.e., 2D and 3D CNN, hybrid CNN and LSTM, and ViT networks), training methods (i.e., supervised learning with and without temporal data, and semi-supervised learning with unlabeled images), and deployment methods (i.e., mobile and embedded computing), using the StairNet data set. Finally, we discuss the challenges and future directions. To date, our StairNet models have consistently achieved high classification accuracy (i.e., up to 98.8%) with different designs, offering trade-offs between model accuracy and size. When deployed on mobile devices with GPU and NPU accelerators, our deep learning models achieved inference speeds up to 2.8 ms. In comparison, when deployed on our custom-designed CPU-powered smart glasses, our models yielded slower inference speeds of 1.5 s, presenting a trade-off between human-centered design and performance. Overall, the results of numerous experiments presented herein provide consistent evidence that StairNet can be an effective platform to develop and study new deep learning models for visual perception of human-robot walking environments, with an emphasis on stair recognition. This research aims to support the development of next-generation vision-based control systems for robotic prosthetic legs, exoskeletons, and other mobility assistive technologies.
引用
收藏
页数:19
相关论文
共 50 条
  • [21] Visual tracking of silhouettes for human-robot interaction
    Menezes, P
    Brèthes, L
    Lerasle, F
    Danès, P
    Dias, J
    PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS 2003, VOL 1-3, 2003, : 971 - 976
  • [22] Visual Coordination Task for Human-Robot Collaboration
    Khatib, Maram
    Al Khudir, Khaled
    De Luca, Alessandro
    2017 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2017, : 3762 - 3768
  • [23] Face recognition and tracking for human-robot interaction
    Song, KT
    Chen, WJ
    2004 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN & CYBERNETICS, VOLS 1-7, 2004, : 2877 - 2882
  • [24] Emotion in human-robot interaction: Recognition and display
    Wendt, Cornalia
    Kuehnlenz, Kolja
    Popp, Michael
    Karg, Michella
    INTERNATIONAL JOURNAL OF PSYCHOLOGY, 2008, 43 (3-4) : 578 - 578
  • [25] Gesture recognition for human-robot collaboration: A review
    Liu, Hongyi
    Wang, Lihui
    INTERNATIONAL JOURNAL OF INDUSTRIAL ERGONOMICS, 2018, 68 : 355 - 367
  • [26] Gesture spotting and recognition for human-robot interaction
    Yang, Hee-Deok
    Park, A-Yeon
    Lee, Seong-Whan
    IEEE TRANSACTIONS ON ROBOTICS, 2007, 23 (02) : 256 - 270
  • [27] Human-robot interaction - Facial gesture recognition
    Rudall, BH
    ROBOTICA, 1996, 14 : 596 - 597
  • [28] Facial Expression Recognition for Human-Robot Interface
    Khan, Mohammad Ibrahim
    Bhuiyan, Md. Al-Amin
    INTERNATIONAL JOURNAL OF COMPUTER SCIENCE AND NETWORK SECURITY, 2009, 9 (04): : 300 - 306
  • [29] Emotion Recognition in Human-Robot Interaction Using the NAO Robot
    Valagkouti, Iro Athina
    Troussas, Christos
    Krouska, Akrivi
    Feidakis, Michalis
    Sgouropoulou, Cleo
    COMPUTERS, 2022, 11 (05)
  • [30] Fuzzy visual detection for human-robot interaction
    Shieh, Ming-Yuan
    Hsieh, Chung-Yu
    Hsieh, Tsung-Min
    ENGINEERING COMPUTATIONS, 2014, 31 (08) : 1709 - 1719