1. Introduction
Hand-based gesture recognition is one of the hottest research fields, since it is of great significance in designing artificially intelligent human computer interfaces. Driving a modern car is an extremely difficult task [
1]. A driver has to perform multi-tasking, such as observing the road, monitoring the vehicle’s status, Global Positioning System (GPS) monitoring, operating numerous electronic and mechanical devices and using audio entertainment. The gesture interface inside a car can assist the driver to perform various tasks. Different sensors have been used for gesture recognition, such as camera, radio-frequency identification (RFID), data-gloves, etc. [
2,
3,
4,
5,
6,
7,
8,
9]. Cameras, however, have a number of line of sight-related challenges that may prevent gesture recognition from being effective. For example, poorly-lit environments may have a negative impact on the image quality and in turn degrade the performance of gesture detection through the camera. The other main issue with camera-based gesture recognition is privacy [
10]. An alternate method for gesture recognition is glove-based sensors. The data-glove-based methods use sensor devices for digitizing hand and finger motions into multi-parametric data [
5]. The extra sensors make it easy to collect hand movement and configuration. However, the devices are quite expensive and bring much cumbersome experience to the users [
6]. The environment inside a vehicle is usually dark at night, and it is inconvenient to wear something during driving; therefore, the above-mentioned techniques are not suitable for vehicular applications.
To overcome the above problems, radar-based gesture recognition can be used as a user interface inside a vehicle. Radar-based gesture recognition techniques have the advantage of better performance in dark environments, do not have privacy issues and do not require wearing sensors. In [
11,
12,
13,
14,
15], the researchers have used Doppler radar sensors for gesture recognition. Molchanov et al. [
11] used multiple sensors, including a depth camera and Doppler, for gesture recognition inside a vehicle. Portable radar sensors for gesture recognition in smart home applications are discussed in [
12]. Kim Youngwook et al. [
13] have performed hand-based gesture recognition using Doppler radar using machine learning techniques; however, the results are too dependent on the orientation and distance between hand and radar.
In addition to Doppler radars, UWB radars have been in the spotlight in recent years. There are many advantages of using IR-UWB radar, such as high range resolution and robustness to multipath due to the extremely wide bandwidth [
16]. The major application areas of UWB radar technology are sensors and communications, localization, tracking and biomedical research [
17]. IR-UWB sensor has been used in various radar applications, such as non-invasive vital sign monitoring [
18,
19,
20], multiple object counting [
21] and direction recognition of moving targets [
22], and it has the ability to detect stationary or slowly moving targets [
23]. However, there is very little reference work available in the literature about gesture recognition based on IR-UWB radar sensors. Ren Nan et al. [
24] have presented an algorithm for big gesture recognition through IR-UWB radar, but the gestures detected in that work were simply based on the position difference of the hand and may not be useful in practical applications. Junbum Park et al. [
25] used an IR-UWB radar sensor for detecting hand-based gestures through machine learning techniques. Although the results show high accuracy, there was an overfitting problem, and the gestures testing in a real environment showed much lower accuracy. Furthermore, there is no method included for the distance compensation or robustness of the algorithm to a change in distance or the orientation of the hand.
The main problem noted in the past radar-based gesture recognition algorithms was that they were vulnerable to distance and orientation; and the feature extraction through machine learning caused the overfitting problem in some cases, which made them error prone. To overcome these problems, we have presented a robust algorithm for hand-based gesture recognition using an IR-UWB radar sensor in this paper. We do not use the completely raw data as an input to the classifier in order to avoid the overfitting problem. We extracted three robust features, i.e., the variance of the pdf of the magnitude histogram, frequency and the variance of time of arrival (TOA) from the pre-processed signal reflected from the human hand. The features extracted were robust and showed better performance even if we changed the orientation of the hand. After the feature extraction, we used the K-means clustering algorithm for classification of the gestures. In order to make the algorithm robust against the distance and orientation variation, we have integrated the TOA-based distance information into the clustering algorithm.
In order to differentiate the gesture motion from some random hand or body motion, we included a data-fitting algorithm. Since the gesture motion defined in our work is almost periodic, therefore we fit the received gesture signal into a sinusoid and check the R-square value. If the R-square value is above a certain threshold, then it is supposed to be periodic and, hence, classified as a gesture signal; otherwise, it is classified as a non-gesture motion. The process block diagram of our algorithm is shown in
Figure 1.
The main contribution of our work is that it is the first real-time IR-UWB-based gesture recognition technique, which avoids the overfitting problem and shows robustness when a change in distance or orientation of the hand occurs, because of the selection of robust parameters and the integration of the TOA information into the clustering algorithm. Additionally, we proposed an algorithm for the detection of only intended gestures while ignoring any random movement in front of the radar sensor. Considering these advantages, this method would be an important technology of the car user interface as one of the core technologies of the future autonomous vehicles.
The hand-based gestures for our work are shown in
Figure 2. The first gesture (Gesture 0) is the empty gesture when there is no hand movement in front of the radar.
Table 1 shows the detailed explanation of the defined gestures. Gestures 1, 2 and 3 are broadly classified as small gestures, while Gestures 4 and 5 are classified as big gestures with larger displacements. The rest of the paper is organized as follows. In
Section 2 of the paper, the feature extraction and classification are discussed. In
Section 3, the results of gesture training and classification are presented, and conclusions are given in
Section 4 of the paper. References are given at the end of the paper.