Abstract:
In this work, we tackle the simulated to real (Sim2Real) gap in vision-based tactile sensors for surface classification. Specifically, we target four surface types: flat,...Show MoreMetadata
Abstract:
In this work, we tackle the simulated to real (Sim2Real) gap in vision-based tactile sensors for surface classification. Specifically, we target four surface types: flat, curved, edge, and corner. We first train a diffusion model (DM) with a small set of real-world, unlabeled images gathered from everyday objects using the DIGIT sensor. Next, we employ the TACTO simulator to generate images by uniformly sampling object surfaces from the YCB model set. The images are transformed into the real domain using the DM and automatically labeled, allowing us to train a surface classifier without the need of manual annotations. To further align features across the real and simulated domains, we use an adversarial approach during training. The evaluation on tactile images from 15 3D-printed YCB objects shows an accuracy of 83.2%, significantly higher than the 35.9% achieved by training only on simulated images, confirming the effectiveness of our method. Testing on data acquired with different DIGIT sensors yields 81.6% accuracy on average, outperforming training with labeled real data. Moreover, our method demonstrates robustness also when applied to a different vision-based tactile sensor, GelSight Mini, obtaining a remarkable 83.3% accuracy on a balanced dataset. Lastly, we validate our approach with a 6-D object pose estimation task using tactile data.
Published in: IEEE Sensors Journal ( Volume: 25, Issue: 5, 01 March 2025)