no code implementations • 9 Jan 2025 • Mohammad Amin Mirzaee, Hung-Jui Huang, Wenzhen Yuan
Scanning large-scale surfaces is widely demanded in surface reconstruction applications and detecting defects in industries' quality control and maintenance stages.
1 code implementation • 9 Dec 2024 • Ruihan Gao, Kangle Deng, Gengshan Yang, Wenzhen Yuan, Jun-Yan Zhu
We design a lightweight 3D texture field to synthesize visual and tactile textures, guided by 2D diffusion model priors on both visual and tactile domains.
no code implementations • 10 Oct 2024 • Irving Fang, Kairui Shi, Xujin He, Siqi Tan, Yifan Wang, Hanwen Zhao, Hung-Jui Huang, Wenzhen Yuan, Chen Feng, Jing Zhang
Humans effortlessly integrate common-sense knowledge with sensory input from vision and touch to understand their surroundings.
1 code implementation • 24 Sep 2024 • Jiacheng Ruan, Wenzhen Yuan, Zehao Lin, Ning Liao, Zhiyu Li, Feiyu Xiong, Ting Liu, Yuzhuo Fu
CamObj-Instruct is collected for fine-tuning the LVLMs with improved instruction-following capabilities, and it includes 11, 363 images and 68, 849 conversations with diverse instructions.
1 code implementation • ICCV 2023 • Ruihan Gao, Wenzhen Yuan, Jun-Yan Zhu
Deep generative models have various content creation applications such as graphic design, e-commerce, and virtual Try-on.
no code implementations • 22 Nov 2022 • Fengyu Yang, Chenyang Ma, Jiacheng Zhang, Jing Zhu, Wenzhen Yuan, Andrew Owens
The ability to associate touch with sight is essential for tasks that require physically interacting with objects in the world.
1 code implementation • 12 Sep 2022 • Shubham Kanitkar, Helen Jiang, Wenzhen Yuan
To facilitate the study of how holding poses affect grasp stability, we present PoseIt, a novel multi-modal dataset that contains visual and tactile data collected from a full cycle of grasping an object, re-positioning the arm to one of the sampled poses, and shaking the object.
1 code implementation • CVPR 2022 • Ruohan Gao, Zilin Si, Yen-Yu Chang, Samuel Clarke, Jeannette Bohg, Li Fei-Fei, Wenzhen Yuan, Jiajun Wu
We present ObjectFolder 2. 0, a large-scale, multisensory dataset of common household objects in the form of implicit neural representations that significantly enhances ObjectFolder 1. 0 in three aspects.
1 code implementation • 24 Dec 2020 • Arpit Agarwal, Tim Man, Wenzhen Yuan
Tactile sensing has seen a rapid adoption with the advent of vision-based tactile sensors.
Robotics Graphics
1 code implementation • 24 Oct 2019 • Tingguang Li, Krishnan Srinivasan, Max Qing-Hu Meng, Wenzhen Yuan, Jeannette Bohg
Finally, we show how our approach generalizes to objects of other shapes.
no code implementations • 28 Sep 2019 • Negin Heravi, Wenzhen Yuan, Allison M. Okamura, Jeannette Bohg
Therefore, it is challenging to model the mapping from material and user interactions to haptic feedback in a way that generalizes over many variations of the user's input.
1 code implementation • 8 Apr 2019 • Ruoyu Wang, Shiheng Wang, Songyu Du, Erdong Xiao, Wenzhen Yuan, Chen Feng
Soft bodies made from flexible and deformable materials are popular in many robotics applications, but their proprioceptive sensing has been a long-standing challenge.
Robotics
no code implementations • 9 Aug 2018 • Shaoxiong Wang, Jiajun Wu, Xingyuan Sun, Wenzhen Yuan, William T. Freeman, Joshua B. Tenenbaum, Edward H. Adelson
Perceiving accurate 3D object shape is important for robots to interact with the physical world.
no code implementations • 28 May 2018 • Roberto Calandra, Andrew Owens, Dinesh Jayaraman, Justin Lin, Wenzhen Yuan, Jitendra Malik, Edward H. Adelson, Sergey Levine
This model -- a deep, multimodal convolutional network -- predicts the outcome of a candidate grasp adjustment, and then executes a grasp by iteratively selecting the most promising actions.
1 code implementation • 21 Feb 2018 • Shan Luo, Wenzhen Yuan, Edward Adelson, Anthony G. Cohn, Raul Fuentes
In this paper, addressing for the first time (to the best of our knowledge) texture recognition from tactile images and vision, we propose a new fusion method named Deep Maximum Covariance Analysis (DMCA) to learn a joint latent space for sharing features through vision and tactile sensing.
1 code implementation • 16 Oct 2017 • Roberto Calandra, Andrew Owens, Manu Upadhyaya, Wenzhen Yuan, Justin Lin, Edward H. Adelson, Sergey Levine
In this work, we investigate the question of whether touch sensing aids in predicting grasp outcomes within a multimodal sensing framework that combines vision and touch.
1 code implementation • CVPR 2017 • Wenzhen Yuan, Shaoxiong Wang, Siyuan Dong, Edward Adelson
A thin flexible fabric, when draped, tends to look different from a heavy stiff fabric.