Abstract
Fine-grained tactile perception of objects is significant for robots to explore the unstructured environment. Recent years have seen the success of Convolutional Neural Networks (CNNs)-based methods for tactile perception using high-resolution optical tactile sensors. However, CNNs-based approaches may not be efficient for processing tactile image data and have limited interpretability. To this end, we propose a Graph Neural Network (GNN)-based approach for tactile recognition using a soft biomimetic optical tactile sensor. The obtained tactile images can be transformed into graphs, while GNN can be used to analyse the implicit tactile information among the tactile graphs. The experimental results indicate that with the proposed GNN-based method, the maximum tactile recognition accuracy can reach 99.53%. In addition, Gradient-weighted Class Activation Mapping (Grad-CAM) and Unsigned Grad-CAM (UGrad-CAM) methods are used for visual explanations of the models. Compared to traditional CNNs, we demonstrated that the generated features of the GNN-based model are more
intuitive and interpretable.
intuitive and interpretable.
Original language | English |
---|---|
Title of host publication | 2022 27th International Conference on Automation and Computing |
Subtitle of host publication | Smart Systems and Manufacturing, ICAC 2022 |
Editors | Chenguang Yang, Yuchun Xu |
Publisher | Institute of Electrical and Electronics Engineers (IEEE) |
Number of pages | 6 |
ISBN (Electronic) | 978-1-6654-9807-4 |
ISBN (Print) | 978-1-6654-9808-1 |
DOIs | |
Publication status | Published - 10 Oct 2022 |
Event | Proceedings of the 27th IEEE International Conference on Automation and Computing (ICAC2022) - Duration: 1 Sept 2022 → 3 Sept 2022 |
Conference
Conference | Proceedings of the 27th IEEE International Conference on Automation and Computing (ICAC2022) |
---|---|
Period | 1/09/22 → 3/09/22 |
Bibliographical note
Publisher Copyright:© 2022 IEEE.