Optimization of Neuroprosthetic Vision via End-to-End Deep Reinforcement Learning

被引:17
作者
Kucukoglu, Burcu [1 ]
Rueckauer, Bodo [1 ]
Ahmad, Nasir [1 ]
van Steveninck, Jaap de Ruyter [1 ]
Guclu, Umut [1 ]
van Gerven, Marcel [1 ]
机构
[1] Radboud Univ Nijmegen, Donders Inst Brain Cognit & Behav, Dept Artificial Intelligence, Heyendaalseweg 135, NL-6525 AJ Nijmegen, Gelderland, Netherlands
关键词
Visual neuroprosthesis; phosphene vision; deep reinforcement learning; end-to-end optimization; ENVIRONMENT;
D O I
10.1142/S0129065722500526
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual neuroprostheses are a promising approach to restore basic sight in visually impaired people. A major challenge is to condense the sensory information contained in a complex environment into meaningful stimulation patterns at low spatial and temporal resolution. Previous approaches considered task-agnostic feature extractors such as edge detectors or semantic segmentation, which are likely suboptimal for specific tasks in complex dynamic environments. As an alternative approach, we propose to optimize stimulation patterns by end-to-end training of a feature extractor using deep reinforcement learning agents in virtual environments. We present a task-oriented evaluation framework to compare different stimulus generation mechanisms, such as static edge-based and adaptive end-to-end approaches like the one introduced here. Our experiments in Atari games show that stimulation patterns obtained via task-dependent end-to-end optimized reinforcement learning result in equivalent or improved performance compared to fixed feature extractors on high difficulty levels. These findings signify the relevance of adaptive reinforcement learning for neuroprosthetic vision in complex environments.
引用
收藏
页数:16
相关论文
共 38 条
[1]  
Badia AP, 2020, PR MACH LEARN RES, V119
[2]   The role of computer vision in prosthetic vision [J].
Barnes, Nick .
IMAGE AND VISION COMPUTING, 2012, 30 (08) :478-479
[3]  
Bollen CJM, 2019, 2019 8TH INTERNATIONAL CONFERENCE ON AFFECTIVE COMPUTING AND INTELLIGENT INTERACTION WORKSHOPS AND DEMOS (ACIIW), P85, DOI [10.1109/ACIIW.2019.8925229, 10.1109/aciiw.2019.8925229]
[4]   Deep Reinforcement Learning and Its Neuroscientific Implications [J].
Botvinick, Matthew ;
Wang, Jane X. ;
Dabney, Will ;
Miller, Kevin J. ;
Kurth-Nelson, Zeb .
NEURON, 2020, 107 (04) :603-616
[5]  
Boyle J, 2001, ANZIIS 2001: PROCEEDINGS OF THE SEVENTH AUSTRALIAN AND NEW ZEALAND INTELLIGENT INFORMATION SYSTEMS CONFERENCE, P85
[6]  
Bradski G, 2000, DR DOBBS J, V25, P120
[7]   SENSATIONS PRODUCED BY ELECTRICAL STIMULATION OF VISUAL CORTEX [J].
BRINDLEY, GS ;
LEWIN, WS .
JOURNAL OF PHYSIOLOGY-LONDON, 1968, 196 (02) :479-&
[8]  
Castaño-Candamil S, 2019, IEEE ENG MED BIO, P2900, DOI [10.1109/embc.2019.8857533, 10.1109/EMBC.2019.8857533]
[9]   Shape perception via a high-channel-count neuroprosthesis in monkey visual cortex [J].
Chen, Xing ;
Wang, Feng ;
Fernandez, Eduardo ;
Roelfsema, Pieter R. .
SCIENCE, 2020, 370 (6521) :1191-1196
[10]   Real and virtual mobility performance in simulated prosthetic vision [J].
Dagnelie, Gislin ;
Keane, Pearse ;
Narla, Venkata ;
Yang, Liancheng ;
Weiland, James ;
Humayun, Mark .
JOURNAL OF NEURAL ENGINEERING, 2007, 4 (01) :S92-S101