Abstract

The paper focuses on surface material classification with unbalanced visual and haptic data, which is important in teleoperation and robotic recognition. For this problem, existent classification methods inevitably suffer from performance degradation, as they tend to emphasize the major classes and ignore the minor ones. To overcome such an issue, we address this classification problem by the double deep Q-learning network (DDQN) method which not only offers strong representation ability, but also avoids over-estimation. Specifically, we first transform haptic accelerations to their spectrograms by Short-Time Fourier Transform (STFT), and respectively feed visual images and haptic spectrograms to pre-trained ResNet50 for extracting low-dimensional feature vectors. Then, the hybrid visual-haptic feature vectors are fed into DDQN as a sequence of states. With respect to each state, DDQN assigns it with an estimated class label. By comparing with the true label, DDQN earns a reward that is dependent on the imbalance ratio of the class. Through maximization of the cumulative rewards, DDQN enhances its classification performance. For the purposes of validation, numerical evaluations are carried out on TUM69, LMT108 and HaTT datasets. The results show that DDQN outperforms existent methods in both classification performance and computational complexity.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call