Controllable Visual-Tactile Synthesis

被引:4
作者
Gao, Ruihan [1 ]
Yuan, Wenzhen [1 ]
Zhu, Jun-Yan [1 ]
机构
[1] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
来源
2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV | 2023年
关键词
GRASP;
D O I
10.1109/ICCV51070.2023.00648
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep generative models have various content creation applications such as graphic design, e-commerce, and virtual try-on. However, current works mainly focus on synthesizing realistic visual outputs, often ignoring other sensory modalities, such as touch, which limits physical interaction with users. In this work, we leverage deep generative models to create a multi-sensory experience where users can touch and see the synthesized object when sliding their fingers on a haptic surface. The main challenges lie in the significant scale discrepancy between vision and touch sensing and the lack of explicit mapping from touch sensing data to a haptic rendering device. To bridge this gap, we collect high-resolution tactile data with a GelSight sensor and create a new visuotactile clothing dataset. We then develop a conditional generative model that synthesizes both visual and tactile outputs from a single sketch. We evaluate our method regarding image quality and tactile rendering accuracy. Finally, we introduce a pipeline to render high-quality visual and tactile outputs on an electroadhesion-based haptic device for an immersive experience, allowing for challenging materials and editable sketch inputs.
引用
收藏
页码:7017 / 7029
页数:13
相关论文
共 93 条
[1]   Pose with Style: Detail-Preserving Pose-Guided Image Synthesis with Conditional StyleGAN [J].
Albahar, Badour ;
Lu, Jingwan ;
Yang, Jimei ;
Shu, Zhixin ;
Shechtman, Eli ;
Huang, Jia-Bin .
ACM TRANSACTIONS ON GRAPHICS, 2021, 40 (06)
[2]   Constant-roll in the Palatini-R2 models [J].
Antoniadis, Ignation ;
Lykkas, Angelos ;
Tamvakis, Kyriakos .
JOURNAL OF COSMOLOGY AND ASTROPARTICLE PHYSICS, 2020, (04)
[3]  
Azadi S, 2019, ARXIV191111357
[4]  
Beheshti Elham, 2019, ACM INT C INT DES CH
[5]   NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis [J].
Ben Mildenhall ;
Srinivasan, Pratul P. ;
Tancik, Matthew ;
Barron, Jonathan T. ;
Ramamoorthi, Ravi ;
Ng, Ren .
COMPUTER VISION - ECCV 2020, PT I, 2020, 12346 :405-421
[6]   Data-Driven Haptic Modeling of Normal Interactions on Viscoelastic Deformable Objects Using a Random Forest [J].
Bhardwaj, Amit ;
Cha, Hojun ;
Choi, Seungmoon .
IEEE ROBOTICS AND AUTOMATION LETTERS, 2019, 4 (02) :1379-1386
[7]  
Birnholtz Jeremy, 2015, INT C HUM COMP INT M
[8]   Perceptual Constancy in the Reproduction of Virtual Tactile Textures With Surface Displays [J].
Bochereau, Serena ;
Sinclair, Stephen ;
Hayward, Vincent .
ACM TRANSACTIONS ON APPLIED PERCEPTION, 2018, 15 (02)
[9]  
Burns David Arthur, 2021, IEEE WORLD HAPT C WH
[10]   GAN-based image-to-friction generation for tactile simulation of fabric material [J].
Cai, Shaoyu ;
Zhao, Lu ;
Ban, Yuki ;
Narumi, Takuji ;
Liu, Yue ;
Zhu, Kening .
COMPUTERS & GRAPHICS-UK, 2022, 102 :460-473