Multimodal zero-shot learning for tactile texture recognition



Cao, Guanqun, Jiang, Jiaqi, Bollegala, Danushka, Li, Min and Luo, Shan ORCID: 0000-0003-4760-0372
(2024) Multimodal zero-shot learning for tactile texture recognition. Robotics and Autonomous Systems, 176. p. 104688.

Access the full-text of this item by clicking on the Open Access link.

Abstract

Tactile sensing plays an irreplaceable role in robotic material recognition. It enables robots to distinguish material properties such as their local geometry and textures, especially for materials like textiles. However, most tactile recognition methods can only classify known materials that have been touched and trained with tactile data, yet cannot classify unknown materials that are not trained with tactile data. To solve this problem, we propose a tactile Zero-Shot Learning framework to recognise materials when they are touched for the first time, using their visual and semantic information, without requiring tactile training samples. The biggest challenge in tactile Zero-Shot Learning is to recognise disjoint classes between training and test materials, i.e., the test materials that are not among the training ones. To bridge this gap, the visual modality, providing tactile cues from sight, and semantic attributes, giving high-level characteristics, are combined together and act as a link to expose the model to these disjoint classes. Specifically, a generative model is learnt to synthesise tactile features according to corresponding visual images and semantic embeddings, and then a classifier can be trained using the synthesised tactile features for zero-shot recognition. Extensive experiments demonstrate that our proposed multimodal generative model can achieve a high recognition accuracy of 83.06% in classifying materials that were not touched before. The robotic experiment demo and the FabricVST dataset are available at https://sites.google.com/view/multimodalzsl.

Item Type: Article
Divisions: Faculty of Science and Engineering > School of Electrical Engineering, Electronics and Computer Science
Depositing User: Symplectic Admin
Date Deposited: 17 Apr 2024 09:56
Last Modified: 17 Apr 2024 09:57
DOI: 10.1016/j.robot.2024.104688
Open Access URL: https://doi.org/10.1016/j.robot.2024.104688
Related URLs:
URI: https://livrepository.liverpool.ac.uk/id/eprint/3180403