The article titled, “Adversarial Unsupervised Domain Adaptation for Hand Gesture Recognition using Thermal Images,” has been accepted for publication in the IEEE Sensors Journal (2023).

Aveen Dayal, Aishwarya M., Abhilash S., C. Krishna Mohan, Abhinav Kumar, Linga Reddy Cenkeramaddi, “Adversarial Unsupervised Domain Adaptation for Hand Gesture Recognition using Thermal Images,” has been accepted for publication in the IEEE Sensors Journal (2023).

Adversarial Unsupervised Domain Adaptation for Hand Gesture Recognition using Thermal Images

Abstract: Hand gesture recognition has a wide range of applications, including in the automotive and industrial sectors, health assistive systems, authentication, and so on. Thermal images are more resistant to environmental changes than red–green–blue (RGB) images for hand gesture recognition. However, one disadvantage of using thermal images for the aforementioned task is the scarcity of labeled thermal datasets. To tackle this problem, we propose a method that combines unsupervised domain adaptation (UDA) techniques with deep-learning (DL) technology to remove the need for labeled data in the learning process. There are several types and methods for implementing UDA, with adversarial UDA being one of the most common. In this article, the first time in this field, we propose a novel adversarial UDA model that uses channel attention and bottleneck layers to learn domain-invariant features across RGB and thermal domains. Thus, the proposed model leverages the information from the labeled RGB data to solve the hand gesture recognition task using thermal images. We evaluate the proposed model on two hand gesture datasets, namely, Sign Digit Classification and Alphabet Gesture Classification, and compare it to other benchmark models in terms of accuracy, model size, and model parameters. Our model outperforms the other state-of-the-art methods on the Sign Digit Classification and Alphabet Gesture Classification datasets and achieves 91.32% and 80.91% target test accuracy, respectively.

DOI: 10.1109/JSEN.2023.3235379

Keywords: Task analysis, Gesture recognition, Adaptation models, Feature extraction, Thermal imaging sensors, Thermal imaging camera, Data models, Training

The article, “Deep Learning based Sign Language Digits Recognition from Thermal Images with Edge Computing System,” has been accepted for publication in the IEEE Sensors Journal.

Daniel S. Breland, Simen B. Skriubakken, Aveen Dayal, Ajit Jha, Phaneendra K. Yalavarthy, and Linga R. Cenkeramaddi, “Deep Learning based Sign Language Digits Recognition from Thermal Images with Edge Computing System,” IEEE Sensors Journal 2021 (in press).

Keywords: Gesture recognition, Cameras, Pins, Integrated circuit modeling, Assistive technology, Three-dimensional displays, Lighting

Abstract: The sign language digits based on hand gestures have been utilized in various applications such as human-computer interaction, robotics, health and medical systems, health assistive technologies, automotive user interfaces, crisis management and disaster relief, entertainment, and contactless communication in smart devices. The color and depth cameras are commonly deployed for hand gesture recognition, but the robust classification of hand gestures under varying illumination is still a challenging task. This work presents the design and deployment of a complete end-to-end edge computing system that can accurately provide the classification of hand gestures captured from thermal images. A thermal dataset of 3200 images was created with each sign language digit having 320 thermal images. The solution presented here utilizes live images taken from a low-resolution thermal camera of 32×32 pixels, feeding into a novel light weight deep learning model based on bottleneck motivated from deep residual learning for classification of hand gestures. The edge computing system presented here utilizes Raspberry pi with a thermal camera making it highly portable. The designed system achieves an accuracy of 99.52% on the test data set with an added advantage of accuracy being invariable to background lighting conditions as it is based on thermal imaging.

More details:DOI: 10.1109/JSEN.2021.3061608