Next Article in Journal
Electric Susceptibility at Partial Coverage of a Circular One-Side Access Capacitive Sensor with Rigid Polyurethane Foams
Previous Article in Journal
Research on Bearing Surface Scratch Detection Based on Improved YOLOV5
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Human Activity Recognition Algorithm with Physiological and Inertial Signals Fusion: Photoplethysmography, Electrodermal Activity, and Accelerometry

1
Department of Electrical and Computer Engineering, University of Central Florida, Orlando, FL 32816, USA
2
School of Engineering, University of North Florida, Jacksonville, FL 32224, USA
*
Authors to whom correspondence should be addressed.
Submission received: 11 March 2024 / Revised: 23 April 2024 / Accepted: 7 May 2024 / Published: 9 May 2024
(This article belongs to the Section Wearables)

Abstract

:
Inertial signals are the most widely used signals in human activity recognition (HAR) applications, and extensive research has been performed on developing HAR classifiers using accelerometer and gyroscope data. This study aimed to investigate the potential enhancement of HAR models through the fusion of biological signals with inertial signals. The classification of eight common low-, medium-, and high-intensity activities was assessed using machine learning (ML) algorithms, trained on accelerometer (ACC), blood volume pulse (BVP), and electrodermal activity (EDA) data obtained from a wrist-worn sensor. Two types of ML algorithms were employed: a random forest (RF) trained on features; and a pre-trained deep learning (DL) network (ResNet-18) trained on spectrogram images. Evaluation was conducted on both individual activities and more generalized activity groups, based on similar intensity. Results indicated that RF classifiers outperformed corresponding DL classifiers at both individual and grouped levels. However, the fusion of EDA and BVP signals with ACC data improved DL classifier performance compared to a baseline DL model with ACC-only data. The best performance was achieved by a classifier trained on a combination of ACC, EDA, and BVP images, yielding F1-scores of 69 and 87 for individual and grouped activity classifications, respectively. For DL models trained with additional biological signals, almost all individual activity classifications showed improvement (p-value < 0.05). In grouped activity classifications, DL model performance was enhanced for low- and medium-intensity activities. Exploring the classification of two specific activities, ascending/descending stairs and cycling, revealed significantly improved results using a DL model trained on combined ACC, BVP, and EDA spectrogram images (p-value < 0.05).

1. Introduction

With recent technological advancements in developing multi-sensor wearable devices, research related to physiological signals has grown rapidly. Classifying human physical activities has recently received a lot of attention due to its connection to physical and mental health. Physical activity (PA) is not only essential for preventing obesity; it might also confer neuroprotection in Alzheimer’s (AD), Parkinson’s (PD), and Huntington’s (HD) diseases via the upregulation of synaptic signaling pathways [1]. It also helps in rehabilitation processes such as cardiorespiratory fitness (CRF) [2]. Wearable physical activity (PA) monitoring systems have also been developed for monitoring the activity levels of a specific population, such as athletes or elderly individuals. E-textile sensors combining accelerometer and goniometer data assist in estimating joint angles; smart insoles are utilized for gait analysis; and smart garments are used in measuring respiratory rate during PA [3,4,5]. In general, knowledge of the frequency and intensity of an individual’s activity plays a significant role in improving his or her lifestyle, which is achievable using a reliable classifier.
Several factors must be considered in improving HAR classification performance, including the placement of wearable [6,7,8,9]. Using data acquired from the wrist is challenging due to the often non-activity-related natural hand/wrist movements. However, the wrist is the recommended collection site for increased wear time [7,8].
Among physiological signals, accelerometer (ACC) data are the main signal in proposed algorithms. In a recent survey of 163 selected HAR studies, it was shown that 149 deployed at least one accelerometer or an accelerometer in conjunction with another sensing modality (gyroscope, magnetometer, body temperature sensor, electrocardiograph, electromyography, etc.) [10]. The prevalent use of accelerometers in HAR may be explained by the low cost and small size of the device as standalone sensors, and additionally, their low energy consumption and high feature performance, as demonstrated in modern smartphones, further contribute to their widespread adoption [11,12].
Recently, the concept of multi-modal HAR has been explored, and it was shown that multi-modal-based methods can improve classification results by leveraging the complementary characteristics between the single modes [13]. For example, accelerometers were used in conjunction with other sensors including gyroscopes [14], magnetometers [15], heart rate (HR) and respiration rate sensors [16], electrocardiographs [17], surface electromyography (sEMG) [18], and skin conductance measurement (EDA) [19].
Jia and Liu [20] proposed a classification method to harness data from a 7-lead ECG worn on the chest and an accelerometer magnitude vector recorded from waist. Time- and frequency-domain features were calculated for classifying lying, sitting, standing, walking, walking upstairs, walking downstairs, and running activities. Average classification accuracy from accelerometer data alone was 93.83%, and it was increased to 99% when fused with ECG data.
Photoplethysmography (PPG) is a less-obtrusive alternative to ECG for obtaining a continuous cardiac signal. It was also shown that ECG signals can be reconstructed from PPG signals [21]. Furthermore, advancements in optical sensors have made PPG an alternative for measuring pulse rate variability features, as surrogates for heart rate variability (HRV) indexes [22]. These techniques offer greater convenience and less intrusion compared to traditional methods.
The derived signal from the PPG sensor is called the blood volume pulse (BVP), and it is the measurement of changes in blood volume in arteries and capillaries. Interestingly, changes in the BVP signal can reflect changes in blood pressure [23], which may correspond to levels of exertion characterizing some types of activities [24]. These attributes make it a promising modality for a wide array of applications, including but not limited to human activity monitoring [17,25,26,27].
However, classifier performance improvement related to non-ACC signal data may be activity-dependent. The use of accelerometer data alone can be insufficient for recognition of activities that require little movement but considerable energy consumption [28]. An example is an activity such as cycling, when a wrist-worn accelerometer is measuring mostly the slight movements of the wrist. Interestingly, a study conducted by Weippert et al. demonstrated that heart rate variability measures could be used to distinguish static exercise (supine leg press) from dynamic exercise (cycling) when the heart rate for each was similar [29]. Another study shows that combining selected HRV features from a single-lead ECG resulted in improved performance in separating sitting from standing and walking from ascending-walking classes over that obtained from accelerometer alone [28]. ECG-derived signals such as PPG data, although not proven to be informative enough on its own as an effectively discriminating input, have provided the capability to improve the separation of activities (when combined with a triaxial accelerometer) involving similar movement yet requiring different levels of exertion, such as cycling and cycling with added resistance [30].
A recent study employed an ensemble method of pre-trained deep learning models (Resnet50V2, MobileNetV2, and Xception), fused at the feature level, to classify four activities (running, walking, and high/low resistance cycling), for the purpose of comparing the efficacy of PPG input data to ECG [25]. The resulting accuracies of the ensemble model tested on the ECG and PPG data were around 94% and 89%, respectively, demonstrating a potential significant improvement in classification accuracy when PPG is used in deep learning models (CNNs) over machine learning models.
Recording a high-quality PPG/BVP signal is complex, since it is influenced by changes in ambient light, electrical noise from the device, and sensor movement [26]. Methods to alleviate the motion artifact (MA) contained in PPG signals due to physical activity have been proposed at the cost of increasing complexity to the signal processing phase of activity classification [31,32]. However, the PPG signal’s susceptibility to corruption by noise introduced by MA has been exploited to predict activities [27].
Another additional signal that has the potential to improve HAR performance is electrodermal activity (EDA). EDA represents changes in the electrical conductivity of the skin. The most salient characteristic of an EDA signal is the skin conductance responses (SCRs) resulting from an underlying sympathetic reaction to a stimulus [33]. The SCRs are the rapid-onset and smooth, exponential-like, transient events noticeable in the EDA signal. EDA data have been used in the prediction of seizures [34], human emotion recognition [35], menstrual cycle analysis [36], and in affective computing [37]. Somewhat recently, EDA has been used in the classification of activity intensity as perceived by the subject (relative PA intensity) [19,38].
Poli et al. [19] trained SVM and bagged tree models on features taken from triaxial accelerometry, HR, inter-beat interval (IBI), skin temperature, and EDA. Using the bagged tree classification model trained on EDA data alone resulted in an average F1-score of 73.8%—which increased to 93.9% when all modalities were included—for classification of sedentary, moderate, and vigorous activities.
Despite the popularity of classifying physical activities using time-series data, some studies focused on using image representation of activities. Such a classification system typically employs a CNN for class prediction. Short-time Fourier transform (STFT), reduced interference distribution with Hanning kernel (RIDHK), and smoothed pseudo Wigner–Ville distribution (SPWVD) have been used in the classification of radar echo signals produced by various physical activities with averaged accuracy (over six physical activities) of 96.61%, 94.72%, and 91.06%, respectively [39]. Accuracy was increased to 98.11% when all three images were vertically concatenated and used as input to a VGG16 network.
In another study, accelerometer and gyroscope data from the Sussex-Huawei Locomotion (SHL) dataset [40] were used to generate FFT spectrograms as inputs to a NN classifier [41]. The network was trained separately on spectrograms corresponding to each axis of both sensors resulting in F1-scores of 90.5%, 91.1%, 90.1%, and 92.8% for accelerometer spectrograms (x, y, z, and magnitude, respectively) and 84.7%, 87.8%, and 83.7% for gyroscope spectrograms (x, y, and z axes, respectively). It should be noted that the SHL dataset contains certain activity classes (riding in a car, bus, train, and on the subway) which are not human physical activities per se.
This study aims to explore the significance of non-inertial signals in enhancing the recognition of human activities within both feature-based and deep learning classifiers. The subsequent sections delve into the processes of collecting physiological data, preprocessing signals, extracting features, generating spectrogram images, and employing classification techniques. These methods are then evaluated against inertial-based classifiers to assess their performance. Furthermore, the impact of blood volume pulse (BVP) and electrodermal activity (EDA) signals on enhancing the recognition of cycling and stairs-climbing activities is investigated and reported.

2. Materials and Methods

The interest of this study is in determining whether a classification system can be improved by the fusion of multiple physiological inputs with triaxial ACC data. The performances of a random forest (RF) classifier and a pre-trained deep learning (DL) model (ResNet18) were analyzed using combinations of physiological signals. RF models were trained on features, and DL models were trained on short-time Fourier transform (STFT) images, engineered from the following modalities: 0-ACC, 1-ACC, and BVP; 2-ACC and EDA; and 3-ACC, BVP, and EDA signals. These datasets will be referred to as 0, 1, 2, and 3, respectively, and the corresponding classification models will be referred to as C0, C1, C2, and C3 for both RF and DL classifications. Workflow of the HAR algorithm development—demonstrating its general steps, including data collection, data processing, dataset generation—and the two approaches that were taken in this paper to develop RF and DL algorithms are shown in Figure 1. Each step is described in detail in the following sections.

2.1. Physiological Data Collection

In this study, the Empatica E4 (Empatica Inc., Boston MA) wristband was used to collect physiological data from subjects while performing a set of activities. E4 sensors include (1) a MEMS-type triaxial (x, y, and z) accelerometer that measures continuous gravitational force on the scale ±2 g with sampling frequency of 32 Hz; (2) a PPG sensor, from which BVP and HR data are derived (sampling frequencies of 64 Hz and 1 Hz, respectively); (3) an EDA sensor recording data at a sampling frequency of 4 Hz; and (4) a skin temperature sensor recording data at a sampling frequency of 4 Hz [42].
An Institutional Review Board protocol (#1796115) was approved for data collection at the University of North Florida and University of Central Florida. Twenty-three subjects (20.7 ± 1.6 years old), including ten females and thirteen males, were recruited to wear the wristband during the experiment. Before starting the recording session, subjects were provided with all information regarding this study, as well as the consent form to review and sign. The subjects were helped to set up the wristband and to ensure the proper placement. The device was worn on the non-dominant wrist. A fifteen-minute time duration was considered for the device to warm-up before starting the session. Subjects were asked to perform a series of activities, each of five-minute duration or—if less—for a duration that the subject was comfortable performing a given activity. Subjects were instructed on how to perform each activity; however, to simulate free-living conditions, there were no strict requirements defining each activity. A short break between activities was given for rest, hydration, and to allow the subject’s heart rate to return to as close to his/her baseline as possible. The baseline heart rate was obtained during the sitting activity. The activities were performed in the following order: sitting, standing, lying, walking, brisk walking, jogging, running (on the indoor track), cycling (stationary bicycle at a rate of around 80 revolutions per minute without resistance), and using stairs (ascending and descending), with the non-dominant hand free to swing. The total data recorded (756.25 min) consisted of 112.95, 110.27, 110.2, 105.81, 91.25, 122.63, 103.14, and 121.42 minutes in brisk walking, cycling, jogging, lying, running, stairs, standing, and walking, respectively.
All activities were performed in the University of North Florida Student Wellness Complex or the University of Central Florida Recreation and Wellness Center. All the activity start times were labeled by tapping the E4’s button (<1 s), and the type of activity was recorded by a proctor. The timestamps later were extracted from a downloaded .csv file that contains a record of the times of the events marked during a session.

2.2. Data Processing

Upon completion of the activity sessions, the subject’s data were inspected via the E4 Connect session viewer. Event markers denoting the start and stop time of each activity performed were verified against the times marked by hand in the session. The data corresponding to the first and last five seconds of an activity were discarded along with occasional segments of the recording during which it was noted that the subject either momentarily stopped performing the activity or the subject’s wrist movement (sensor-worn wrist) became excessive (for example, during standing or lying).
For each session, a timestamp series was generated, determining the start and end time of each activity. For data segmentation, initially, window sizes ranging from 5 to 20-second with a step size of 5-second were generated and tested. Ultimately, we selected a 10-second window size to ensure sufficient frequency resolution for electrodermal activity (EDA) features and to improve overall performance. Additionally, a 50% overlap was used to achieve both an informative window size and acceptable dataset size [43].
All pre-processing steps (filtering and segmentation) and feature computations were completed using Python v. 3.8.11. A second-order Butterworth bandpass filter (0.3–10 Hz) was used to remove the DC component and frequencies above 10 Hz in ACC data (x, y, and z components), noting that 98% of the spectral amplitudes corresponding to human activities lie below 10 Hz [44]. PPG and EDA signals were filtered using a fourth-order Chebyshev-II band pass filter (0.4–5 Hz [45]), and a fourth-order Butterworth high-pass filter (0.05–2 Hz), respectively. Datasets 0, 1, 2, and 3 were generated from the processed and segmented data for all subjects. Time- and frequency-domain features computed for RF classification are listed by modality in Table S8. STFT images were generated from 10-second data segments of BVP, EDA, and ACC magnitude, with a 50% overlap.

2.3. HAR Classifier Development

A HAR system which included two independent algorithms was assembled. First, a random forest algorithm—which is among the popular classification algorithms for HAR models [17,46]—was trained on features extracted from 10-second data segments. Then, as a DL classifier, a convolutional neural network obtained via transfer learning was trained on spectrogram images of 10-second pre-processed ACC, BVP, and EDA data segments.

2.3.1. Random Forest Algorithm

The RF classifier from the sklearn.ensemble module with 500 estimators (trees) was used in this study. Grid search was performed over a range of 100 to 1000 with a step size of 100 to optimize this hyperparameter. The remaining default parameters were used, including min_samples_split of 2, and max_feautres of sqrt. Four RF algorithms (C0, C1, C2, and C3) were trained to classify eight individual activities (lying, standing, walking, brisk walking, jogging, running, stairs, and cycling) using features from datasets 0 through 3, respectively. Next, activities were grouped according to the body movement involved in the activity and activity–intensity overlap, which may have likely been induced by the relaxed experimental environment. Therefore, standing and lying, walking and brisk walking, and jogging and running engendered low-, medium-, and high-intensity groups. Cycling and stairs remained separate activity classes. RF algorithms C0 through C3 were also trained on grouped-activity datasets.
Features, both low-level and handcrafted, extracted from ACC, BVP, and EDA data, are reported in Table S8. The feature set is represented by statistical estimators, such as those used in previous HAR studies [17,19,47], and some experimentally and intuitively derived [48]; for example, a range of frequency bands associated with respective levels of cardiac output.
Training and testing were conducted in leave-one-subject-out mode. Training data were further divided into 90%/10% randomized train/validation datasets. Data were then normalized (z-score, sklearn.preprocessing.StandardScaler), and a single test was conducted for each test subject.

2.3.2. Convolutional Neural Network (CNN)

Pre-trained DL image classification neural networks eliminate the laborious task of designing a DL network and provide a re-trainable model which has learned rich feature representations from more than a million images. MATLAB’s (R2021b) Deep Learning Toolbox Model for ResNet-18 [49] was used in this study to perform the DL experiments (individual activity classification and group-based classification), analogous to the machine learning experiments described in Section 2.3.1.
STFT images were generated from datasets 0 through 3. The magnitude vector of ACC segments (i.e., ⅓ sqrt(ACCx2 + ACCy2 + ACCz2) was used to construct STFT images for dataset 0. Datasets 1 through 3 were created through vertical concatenation of the STFT images. The original ACC and EDA image size was 220 × 220 pixels; however, the BVP spectrogram image was cropped to a 220-width-by-110-height size due to the excessive near-zero-valued amplitudes present in the STFT higher frequencies. After vertical concatenation, the combined image was resized to 224 × 224, as was required by the ResNet-18 input shape. New classification and fully connected layers were created, and they replaced the two last ResNet-18 network layers to reduce the number of outputs from one thousand to eight classes. No other changes were made to the network. Time series signals are shown in Figure 2, along with their spectrograms, for eight activities.
The initial learning rate and batch size were set to 0.00125 and 75, after implementing a hyperparameter grid search over the ranges of 0.0005 to 0.002 with step size of 0.00025 and 25 to 125, with step size of 0.00025, respectively. Training data were divided into 90%/10% randomized train/validation datasets. All data were normalized, applying zero-center normalization, which normalizes each pixel value between [–1, 1]. Classifiers were trained for 50 epochs using the Adam optimizer.

3. Results

Data from twenty-three subjects performing eight activities (lying, standing, walking, brisk walking, jogging, running, ascending/descending stairs, and cycling) were included in the analysis of HAR classifier performances. First, individual activity classification was conducted. Then, for a less-granular classification task, six of the eight activities were grouped by their similar intensity (standing/lying (low), brisk walking/walking (medium), and jogging/running (high)). RF and CNN algorithms were trained to classify activities at both individual and group levels. Classifiers were trained on ACC (C0); ACC and BVP (C1); ACC and EDA (C2); and ACC, BVP, and EDA (C3) datasets. CNN was trained on STFT spectrogram images, while RF was trained on featurized data. To assess and compare classifiers performance, the area under the ROC curve and F1-score were calculated. Table 1 presents the performance of each classifier by input, both for individual and grouped activities. Due to variation in CNN model prediction from trial to trial, the average of three CNN trials is reported.
The results show that the RF classifier outperforms CNN; however, EDA and BVP signals improve the performance of the CNN classifier. AUC-ROC and F1-scores are reported in Table 2, Table 3, Table 4 and Table 5 for individual and grouped activity classification. Additional signals do not appear to improve RF classification results in individual or grouped activities (Table 2 and Table 4); however, for the CNN model, nearly all individual activity classifications were improved with additional signals (Table 3). In grouped activity classifications, CNN performance was improved for low- and medium-intensity activities, as well as for cycling and stairs (Table 5).
To determine if the improvement in performance was significant, a comparison of the three signal-augmented models to a baseline ACC-only CNN model was conducted using a mid-p-value McNemar test at the 5% significance level. The McNemar test showed statistically significant improvement in the classification results (p < 0.05) of C3 for all classes (as is shown in Table 3 and Table S2), except running, when compared to C0. Statistically significant improvement (p < 0.005) was also observed in grouped low- and medium-intensity stairs and cycling activities for C1, and in cycling for C2 when compared to C0 (Table 5 and Table S3). Adding both EDA and BVP inputs (C3) improved the performance of classification of all groups significantly (p < 0.005), except high-intensity activities.
Among common activities [10], stationary cycling and ascending/descending stairs appear to be two of the more difficult classes to discriminate from other modes of ambulation—or, in the case of cycling, from sedentary activities such as sitting—when using accelerometer data acquired from the wrist [7,17,50,51]. At a granular individual activity classification level, a Wilcoxon rank sum test was conducted to compare the classification accuracy of both cycling and stairs by comparing signal-augmented models to a baseline ACC-only CNN model. In C3, misclassification of stairs as standing and walking activities (p < 0.05) and cycling as lying and standing activities (p < 0.005) was significantly improved (Tables S4 and S5). Misclassification of cycling as standing was also decreased using C2 (p < 0.05). Misclassification of stairs and cycling with respect to the rest of the activities was negligible.
For grouped activity classification, a similar rank sum test was conducted. Using C3, misclassification of stairs as a low-intensity activity group (p < 0.05) and cycling as low- and medium-intensity activities (p < 0.005 and p < 0.05 respectively) was reduced. The combined ACC and BVP inputs improve classification error rates of both cycling and stairs, which were misclassified as low-intensity activities by C0 (Tables S6 and S7). Misclassification of stairs and cycling with respect to other activities not mentioned above was negligible.

4. Discussion and Conclusions

The application of wearable sensors in HAR system development continues to enjoy multi-disciplinary interest due to its importance in the monitoring of human well-being. Although most of the available algorithms rely on accelerometer sensor data, the purpose of this paper is to investigate the importance of BVP and EDA signals in HAR algorithms. Therefore, a multi-modal sensing wristband was employed for data collection. It should be noted that other signals, such as EMG and ECG, have the potential to improve PA classification results; however, the purpose of this study was limited to the data derived from signals obtained from a single Empatica E4 wristband.
Data from PPG, EDA, and triaxial ACC sensors were preprocessed and used to train machine learning and deep learning classifiers, with input modalities including only ACC data (C0); ACC and BVP (C1); ACC and EDA (C2); and ACC, EDA, and BVP (C3) to examine the importance of these non-ACC signals.
No significant improvement was observed in the RF classifier by adding additional signals. However, significant improvements were observed in overall F1-scores for CNN models trained on additional bio-signals and were confirmed by applying a McNemar test (p < 0.05). In individual activity classification, the average CNN F1-score was improved from 64.22% with C0 to 69.01% with C1, 67.7% with C2, and 69.89% with C3. The corresponding F1-scores for group classification also were improved from 80.65% with C0 to 86.34% with C1, 83.77% with C2, and 87.51% with C3. Regarding individual activity classification, a mid-p-value McNemar Test showed that performance of C3 was improved over C0 for all activities except running (p < 0.005); performance of C1 was improved over C0 for standing, lying, cycling, and stairs (p < 0.05); and performance of C2 was improved over C0 for cycling, jogging, lying, and stairs classes (p < 0.05). For the grouped, five-activity classification task, the McNemar Test showed that performance of both C1 and C3 was improved over C0 for all grouped activity classes except high-intensity (p < 0.005); and the performance of C2 was improved over C0 for cycling class (p < 0.05). This may suggest that there is a link between improved classification of cycling activity and EDA data facilitated by minimum wrist movement and, consequently, lower motion artifact. It is shown in Table 1 (also Table S1) that training the CNN model on BVP and/or EDA inputs consistently improves the average F1-scores, Matthews Correlation Coefficient (MCC), and Cohen’s kappa score.
Although there is no significant improvement in classification by a feature-based classifier such as the RF trained on bio-signal features, statistically significant improvement in classification by a multi-modal CNN model was observed. This improvement might be due to the ability of DL models to directly learn robust features from raw data for specific applications, whereas ML models such as RF require expertly extracted or engineered features for satisfactory performance [52]. Specifically for EDA and PPG, which are susceptible to noise, statistical features are less suitable. Other ML classification challenges may be (1) the issue of “overlapping” statistical characteristics of time- and frequency-domain signals, corresponding to certain activities (e.g., running and jogging) in free-living conditions (reduction of inter-class variability); (2) that for time spent in activity, even one as moderate as walking, cardiac output increases initially over a period of time and then tends to plateau [53]. This would imply that the beginning and end of a single activity may vary significantly in bio-signal amplitudes.
Motion artifacts also affect the performance of multimodal classifiers. In the BVP signal, the main artifacts composed of wrist acceleration in the x, y, and z directions and the changes in space between the skin and the PPG sensor, if considerably regular or periodic, are likely correlated with motion-intensity, which may characterize the subject’s activity. A spectral representation of a PPG signal has a dominant quasi-periodic signal component resulting from rhythmic wrist movement/foot–ground contact, and another component corresponding to heartbeat. If the two components have close periods, due to leakage, the spectral peak associated with the heartbeat may not be distinguishable from the peak associated with the wrist swing rhythm [26]. This condition may not be helpful to ML models but could be exploited by DL models.
The findings of our study, regarding the improved classification performance of a CNN despite being trained on MA-corrupted BVP inputs, may be supported by conclusions from research which used a combined convolutional and recurrent layer network to classify standing, walking, jogging, jumping, and sitting classes [32]. In this study, each subject’s PPG signal was decomposed into a cardiac component (0.5–4 Hz), a respiratory component (0.2–0.35 Hz), and a MA noise component (≥0.1 Hz) and converted to the frequency-domain using the FFT. When comparing models trained on data composed of the cardiac, respiratory, and MA components (individually and in combination), results suggested that the MA signal is a better predictor than the cardiac and respiration signal. Further, the predictability of activities from PPG does not only result from the elevation of heart rate and respiration rates but also from the MA generated when performing the activities [32]. Results from the C1 classifier in our study might suggest that the added BVP data could make classification by DL models robust against MA, given that the model is trained on noisy BVP data yet model accuracy significantly improves.
Among common activities in ML HAR datasets [10], stationary cycling and ascending/descending stairs appear to be more difficult to discriminate from other activities when using accelerometer data acquired from the wrist [7,17,50,51]. A somewhat heavily MA-affected activity, ascending/descending stairs is mostly misclassified as walking using ACC-based classifiers. However, the CNN model trained on BVP, EDA, and ACC data (C3) demonstrated significant improvement in separating stair use from walking (p < 0.05). On the other hand, cycling, which may be less affected by MA than ascending/descending stairs, is commonly misclassified as a sedentary activity such as sitting or standing. Results from our study show that the cycling class becomes more separable from lying and standing (p < 0.005) for CNN model C3. These results may point to a potential improvement in the classification of these challenging classes with the addition of bio-signal inputs when a suitable DL model is used.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/s24103005/s1, Table S1: Classifier evaluation metrics: Matthews Correlation Coefficient, Log-loss, and Cohen’s Kappa Score. Table S2: McNemar Test results, comparing fused CNN models to accelerometer-only CNN models in classifying individual activities (p-value). Table S3: McNemar Test results, comparing fused CNN models to accelerometer-only CNN models in classifying grouped activities (p-value). Table S4: Wilcoxon Rank Sum Test results at the 5% significance level, indicating individual activity classification error improvement of fused CNN models with respect to stairs class. Only activities which are improved significantly by at least one classifier are shown. Table S5: Wilcoxon Rank Sum Test results at the 5% significance level, indicating individual activity classification error improvement of fused CNN models with respect to cycling class. Only activities which are improved significantly by at least one classifier are shown. Table S6: Wilcoxon Rank Sum Test results at the 5% significance level, indicating grouped activity classification error improvement of fused CNN models with respect to stairs class. Only activities which are improved significantly by at least one classifier are shown. Table S7: Wilcoxon Rank Sum Test results at the 5% significance level, indicating grouped activity classification error improvement of fused CNN models with respect to cycling class. Only activities which are improved significantly by at least one classifier are shown. Table S8: Random forest data set features.

Author Contributions

Conceptualization, J.G. and M.N.; methodology, J.G. and M.N.; formal analysis, J.G.; data curation, J.G.; writing—original draft preparation, J.G.; writing—review and editing, M.N. All authors have read and agreed to the published version of the manuscript.

Funding

The APC was funded by University of North Florida.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of University of North Florida (#1796115; date of initial approval 9 July 2021).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Santiago, J.A.; Quinn, J.P.; Potashkin, J.A. Physical activity rewires the human brain against neurodegeneration. Int. J. Mol. Sci. 2022, 23, 6223. [Google Scholar] [CrossRef]
  2. Hannan, A.L.; Harders, M.P.; Hing, W.; Climstein, M.; Coombes, J.S.; Furness, J. Impact of wearable physical activity monitoring devices with exercise prescription or advice in the maintenance phase of cardiac rehabilitation: Systematic review and meta-analysis. BMC Sports Sci. Med. Rehabil. 2019, 11, 14. [Google Scholar] [CrossRef] [PubMed]
  3. Tognetti, A.; Lorussi, F.; Carbonaro, N.; De Rossi, D. Wearable goniometer and accelerometer sensory fusion for knee joint angle measurement in daily life. Sensors 2015, 15, 28435–28455. [Google Scholar] [CrossRef]
  4. Tian, Y.; Zhang, L.; Zhang, C.; Bao, B.; Li, Q.; Wang, L.; Song, Z.; Li, D. Deep-learning enabled smart insole system aiming for multifunctional foot-healthcare applications. Exploration 2024, 4, 20230109. [Google Scholar] [CrossRef]
  5. Massaroni, C.; Di Tocco, J.; Bravi, M.; Carnevale, A.; Presti, D.L.; Sabbadini, R.; Schena, E. Respiratory monitoring during physical activities with a multi-sensor smart garment and related algorithms. IEEE Sens. J. 2019, 20, 2173–2180. [Google Scholar] [CrossRef]
  6. Zhang, S.; Rowlands, A.V.; Murray, P.; Hurst, T.L. Physical activity classification using the GENEA wrist-worn accelerometer. Med. Sci. Sports Exerc. 2012, 44, 742–748. [Google Scholar] [CrossRef]
  7. Mannini, A.; Intille, S.S.; Rosenberger, M.; Sabatini, A.M.; Haskell, W. Activity recognition using a single accelerometer placed at the wrist or ankle. Med. Sci. Sports Exerc. 2013, 45, 2193–2203. [Google Scholar] [CrossRef] [PubMed]
  8. Doherty, A.; Jackson, D.; Hammerla, N.; Plötz, T.; Olivier, P.; Granat, M.H.; Wareham, N.J. Large scale population assessment of physical activity using wrist worn accelerometers: The UK biobank study. PLoS ONE 2017, 12, e0169649. [Google Scholar] [CrossRef]
  9. Troiano, R.; Mc Clain, J. Objective measures of physical activity, sleep, and strength in US National Health and Nutrition Examination Survey (NHANES) 2011–2014. In Proceedings of the 8th International Conference on Diet and Activity Methods, Rome, Italy, 14–17 May 2012. [Google Scholar]
  10. Demrozi, F.; Pravadelli, G.; Bihorac, A.; Rashidi, P. Human activity recognition using inertial, physiological and environmental sensors: A comprehensive survey. IEEE Access 2020, 8, 210816–210836. [Google Scholar] [CrossRef]
  11. Grouios, G.; Ziagkas, E.; Loukovitis, A.; Chatzinikolaou, K.; Koidou, E. Accelerometers in our pocket: Does smartphone accelerometer technology provide accurate data? Sensors 2022, 23, 192. [Google Scholar] [CrossRef]
  12. Nweke, H.F.; Teh, Y.W.; Mujtaba, G.; Al-Garadi, M.A. Data fusion and multiple classifier systems for human activity detection and health monitoring: Review and open research directions. Inf. Fusion 2019, 46, 147–170. [Google Scholar] [CrossRef]
  13. Shuchang, Z. A Survey on Human Action Recognition. arXiv 2022, arXiv:2301.06082. [Google Scholar]
  14. Wang, A.; Chen, G.; Yang, J.; Zhao, S.; Chang, C.Y. A comparative study on human activity recognition using inertial sensors in a smartphone. IEEE Sens. J. 2016, 16, 4566–4578. [Google Scholar] [CrossRef]
  15. Altun, K.; Barshan, B. Human activity recognition using inertial/magnetic sensor units. In Proceedings of the Human Behavior Understanding: First International Workshop, HBU 2010, Istanbul, Turkey, 22 August 2010; Proceedings 1. Springer: Berlin/Heidelberg, Germany, 2010; pp. 38–51. [Google Scholar]
  16. Martín, H.; Bernardos, A.M.; Tarrío, P.; Casar, J.R. Enhancing activity recognition by fusing inertial and biometric information. In Proceedings of the 14th International Conference on Information Fusion, Chicago, IL, USA, 5–8 July 2011; pp. 1–8. [Google Scholar]
  17. Afzali Arani, M.S.; Costa, D.E.; Shihab, E. Human activity recognition: A comparative study to assess the contribution level of accelerometer, ECG, and PPG signals. Sensors 2021, 21, 6997. [Google Scholar] [CrossRef]
  18. Meng, L.; Zhang, A.; Chen, C.; Wang, X.; Jiang, X.; Tao, L.; Chen, W. Exploration of human activity recognition using a single sensor for stroke survivors and able-bodied people. Sensors 2021, 21, 799. [Google Scholar] [CrossRef] [PubMed]
  19. Poli, A.; Gabrielli, V.; Ciabattoni, L.; Spinsante, S. Cross-Domain Classification of Physical Activity Intensity: An EDA-Based Approach Validated by Wrist-Measured Acceleration and Physiological Data. Electronics 2021, 10, 2159. [Google Scholar] [CrossRef]
  20. Jia, R.; Liu, B. Human daily activity recognition by fusing accelerometer and multi-lead ECG data. In Proceedings of the 2013 IEEE International Conference on Signal Processing, Communication and Computing (ICSPCC 2013), Yunnan, China, 5–8 August 2013; pp. 1–4. [Google Scholar]
  21. Zhu, Q.; Tian, X.; Wong, C.W.; Wu, M. ECG reconstruction via PPG: A pilot study. In Proceedings of the 2019 IEEE EMBS International Conference on Biomedical & Health Informatics (BHI), Chicago, IL, USA, 19–22 May 2019; pp. 1–4. [Google Scholar]
  22. Pinheiro, N.; Couceiro, R.; Henriques, J.; Muehlsteff, J.; Quintal, I.; Goncalves, L.; Carvalho, P. Can PPG be used for HRV analysis? In Proceedings of the 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Orlando, FL, USA, 16–20 August 2016; pp. 2945–2949. [Google Scholar]
  23. Shi, W.; Zhou, C.; Zhang, Y.; Li, K.; Ren, X.; Liu, H.; Ye, X. Hybrid modeling on reconstitution of continuous arterial blood pressure using finger photoplethysmography. Biomed. Signal Process. Control 2023, 85, 104972. [Google Scholar] [CrossRef]
  24. Peper, E.; Harvey, R.; Lin, I.M.; Tylova, H.; Moss, D. Is there more to blood volume pulse than heart rate variability, respiratory sinus arrhythmia, and cardiorespiratory synchrony? Biofeedback 2007, 35, 54. [Google Scholar]
  25. Almanifi, O.R.A.; Khairuddin, I.M.; Razman, M.A.M.; Musa, R.M.; Majeed, A.P.A. Human activity recognition based on wrist PPG via the ensemble method. ICT Express 2022, 8, 513–517. [Google Scholar] [CrossRef]
  26. Lee, J. Motion artifacts reduction from PPG using cyclic moving average filter. Technol. Health Care 2014, 22, 409–417. [Google Scholar] [CrossRef]
  27. Boukhechba, M.; Cai, L.; Wu, C.; Barnes, L.E. ActiPPG: Using deep neural networks for activity recognition from wrist-worn photoplethysmography (PPG) sensors. Smart Health 2019, 14, 100082. [Google Scholar] [CrossRef]
  28. Park, H.; Dong, S.Y.; Lee, M.; Youn, I. The role of heart-rate variability parameters in activity recognition and energy-expenditure estimation using wearable sensors. Sensors 2017, 17, 1698. [Google Scholar] [CrossRef] [PubMed]
  29. Weippert, M.; Behrens, K.; Rieger, A.; Stoll, R.; Kreuzfeld, S. Heart rate variability and blood pressure during dynamic and static exercise at similar heart rate levels. PloS ONE 2013, 8, e83690. [Google Scholar] [CrossRef] [PubMed]
  30. Biagetti, G.; Crippa, P.; Falaschetti, L.; Orcioni, S.; Turchetti, C. Human activity recognition using accelerometer and photoplethysmographic signals. In Intelligent Decision Technologies 2017: Proceedings of the 9th KES International Conference on Intelligent Decision Technologies (KES-IDT 2017)–Part II 9; Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 53–62. [Google Scholar]
  31. Zhang, Z.; Pi, Z.; Liu, B. TROIKA: A general framework for heart rate monitoring using wrist-type photoplethysmographic signals during intensive physical exercise. IEEE Trans. Biomed. Eng. 2014, 62, 522–531. [Google Scholar] [CrossRef] [PubMed]
  32. Fukushima, H.; Kawanaka, H.; Bhuiyan, M.S.; Oguri, K. Estimating heart rate using wrist-type photoplethysmography and acceleration sensor while running. In Proceedings of the 2012 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, San Diego, CA, USA, 28 August–1 September 2012; pp. 2901–2904. [Google Scholar]
  33. Posada-Quintero, H.F.; Chon, K.H. Innovations in electrodermal activity data collection and signal processing: A systematic review. Sensors 2020, 20, 479. [Google Scholar] [CrossRef] [PubMed]
  34. Nasseri, M.; Pal Attia, T.; Joseph, B.; Gregg, N.M.; Nurse, E.S.; Viana, P.F.; Worrell, G.; Dümpelmann, M.; Richardson, M.P.; Freestone, D.R.; et al. Ambulatory seizure forecasting with a wrist-worn device using long-short term memory deep learning. Sci. Rep. 2021, 11, 21935. [Google Scholar] [CrossRef]
  35. Ganapathy, N.; Veeranki, Y.R.; Kumar, H.; Swaminathan, R. Emotion recognition using electrodermal activity signals and multiscale deep convolutional neural network. J. Med. Syst. 2021, 45, 49. [Google Scholar] [CrossRef]
  36. Sides, K.; Kilungeja, G.; Tapia, M.; Kreidl, P.; Brinkmann, B.H.; Nasseri, M. Analyzing physiological signals recorded with a wearable sensor across the menstrual cycle using circular statistics. Front. Netw. Physiol. 2023, 3, 1227228. [Google Scholar] [CrossRef] [PubMed]
  37. Motogna, V.; Lupu-Florian, G.; Lupu, E. Strategy for affective computing based on HRV and EDA. In Proceedings of the 2021 International Conference on e-Health and Bioengineering (EHB), Iasi, Romania, 18–19 November 2021; pp. 1–4. [Google Scholar]
  38. Chowdhury, A.K.; Tjondronegoro, D.; Chandran, V.; Zhang, J.; Trost, S.G. Prediction of relative physical activity intensity using multimodal sensing of physiological data. Sensors 2019, 19, 4509. [Google Scholar] [CrossRef]
  39. Liu, Z.; Xu, L.; Jia, Y.; Guo, S. Human activity recognition based on deep learning with multi-spectrogram. In Proceedings of the 2020 IEEE 5th International Conference on Signal and Image Processing (ICSIP), Nanjing, China, 23–25 October 2020; pp. 11–15. [Google Scholar]
  40. Gjoreski, H.; Ciliberto, M.; Wang, L.; Morales FJ, O.; Mekki, S.; Valentin, S.; Roggen, D. The university of Sussex-huawei locomotion and transportation dataset for multimodal analytics with mobile devices. IEEE Access 2018, 6, 42592–42604. [Google Scholar] [CrossRef]
  41. Ito, C.; Cao, X.; Shuzo, M.; Maeda, E. Application of CNN for human activity recognition with FFT spectrogram of acceleration and gyro sensors. In Proceedings of the 2018 ACM International Joint Conference and 2018 International Symposium on Pervasive and Ubiquitous Computing and Wearable Computers, Singapore, 8–12 October 2018; pp. 1503–1510. [Google Scholar]
  42. Empatica Inc, Empatica E4, Revision 001, 19 November 2014. Available online: https://box.empatica.com/documentation/20141119_E4_TechSpecs.pdf (accessed on 6 May 2024).
  43. Dehghani, A.; Sarbishei, O.; Glatard, T.; Shihab, E. A quantitative comparison of overlapping and non-overlapping sliding windows for human activity recognition using inertial sensors. Sensors 2019, 19, 5026. [Google Scholar] [CrossRef] [PubMed]
  44. Khusainov, R.; Azzi, D.; Achumba, I.E.; Bersch, S.D. Real-time human ambulation, activity, and physiological monitoring: Taxonomy of issues, techniques, applications, challenges and limitations. Sensors 2013, 13, 12852–12902. [Google Scholar] [CrossRef] [PubMed]
  45. Liang, Y.; Elgendi, M.; Chen, Z.; Ward, R. An optimal filter for short photoplethysmogram signals. Sci. Data 2018, 5, 180076. [Google Scholar] [CrossRef]
  46. Zhu, J.; San-Segundo, R.; Pardo, J.M. Feature extraction for robust physical activity recognition. Hum. Centric Comput. Inf. Sci. 2017, 7, 16. [Google Scholar] [CrossRef]
  47. Rosati, S.; Balestra, G.; Knaflitz, M. Comparison of different sets of features for human activity recognition by wearable sensors. Sensors 2018, 18, 4189. [Google Scholar] [CrossRef] [PubMed]
  48. Posada-Quintero, H.F.; Reljin, N.; Mills, C.; Mills, I.; Florian, J.P.; VanHeest, J.L.; Chon, K.H. Time-varying analysis of electrodermal activity during exercise. PLoS ONE 2018, 13, e0198328. [Google Scholar] [CrossRef] [PubMed]
  49. He, Z.; Ren, S.; He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision And Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  50. Logacjov, A.; Bach, K.; Kongsvold, A.; Bårdstu, H.B.; Mork, P.J. HARTH: A human activity recognition dataset for machine learning. Sensors 2021, 21, 7853. [Google Scholar] [CrossRef]
  51. Mehrang, S.; Pietilä, J.; Korhonen, I. An activity recognition framework deploying the random forest classifier and a single optical heart rate monitoring and triaxial accelerometer wrist-band. Sensors 2018, 18, 613. [Google Scholar] [CrossRef]
  52. Zhang, S.; Li, Y.; Zhang, S.; Shahabi, F.; Xia, S.; Deng, Y.; Alshurafa, N. Deep learning in human activity recognition with wearable sensors: A review on advances. Sensors 2022, 22, 1476. [Google Scholar] [CrossRef]
  53. Laughlin, M.H. Cardiovascular response to exercise. Adv. Physiol. Educ. 1999, 277, S244. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions, and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions, or products referred to in the content.
Figure 1. Flowchart for HAR algorithm development; ACC, BVP, and EDA data were collected from a wristband device. RF algorithms were trained on features extracted from signals, and DL algorithms were trained on spectrogram images.
Figure 1. Flowchart for HAR algorithm development; ACC, BVP, and EDA data were collected from a wristband device. RF algorithms were trained on features extracted from signals, and DL algorithms were trained on spectrogram images.
Sensors 24 03005 g001
Figure 2. Paired time and time–frequency plots (STFT) of each activity corresponding to their ACC, BVP, and EDA segments (top to bottom) from a 20-year-old female subject. Activities represented are lying, standing, walking, brisk walking, jogging, running, stairs, and cycling (left to right). The colors in spectrogram represents the amplitude of a specific frequency at a given time, ranging from dark blues for low amplitudes and dark red, indicating stronger amplitudes.
Figure 2. Paired time and time–frequency plots (STFT) of each activity corresponding to their ACC, BVP, and EDA segments (top to bottom) from a 20-year-old female subject. Activities represented are lying, standing, walking, brisk walking, jogging, running, stairs, and cycling (left to right). The colors in spectrogram represents the amplitude of a specific frequency at a given time, ranging from dark blues for low amplitudes and dark red, indicating stronger amplitudes.
Sensors 24 03005 g002
Table 1. Random forest and CNN classifiers performance by inputs for both individual and grouped classes. Results are the mean ± standard deviation of the F1-score and area under the receiver operating characteristic (ROC) curve for 23 subjects (leave-one-subject-out mode).
Table 1. Random forest and CNN classifiers performance by inputs for both individual and grouped classes. Results are the mean ± standard deviation of the F1-score and area under the receiver operating characteristic (ROC) curve for 23 subjects (leave-one-subject-out mode).
ClassifierClassifier InputsIndividual
Activity
Classification
F1-Score
Grouped
Activity
Classification
F1-Score
Individual
Activity
Classification
AUC
Grouped
Activity
Classification
AUC
RFACC (C0)77.42 ± 23.0192.52 ± 9.6198.63 ± 2.3999.56 ± 0.77
ACC + BVP (C1)76.10 ± 22.9892.31 ± 9.5698.57 ± 2.4899.63 ± 0.64
ACC + EDA (C2)77.95 ± 22.9093.29 ± 10.0398.81 ± 2.5399.82 ± 0.43
ACC + BVP + EDA(C3)77.21 ± 23.3293.22 ± 10.5398.72 ± 2.6199.82 ± 0.47
CNNACC (C0)64.22 ± 19.2080.65 ± 11.0994.29 ± 4.1896.53 ± 4.08
ACC + BVP (C1)69.01 ± 20.2786.34 ± 11.1696.54 ± 3.5798.03 ± 3.49
ACC + EDA (C2)67.70 ± 22.9783.77 ± 11.8495.93 ± 4.3497.54 ± 2.54
ACC + BVP + EDA(C3)69.89 ± 18.5987.51 ± 9.5596.54 ± 4.2198.48 ± 2.24
Table 2. Random forest classification performance for eight individual activities.
Table 2. Random forest classification performance for eight individual activities.
Classifier InputACC (C0)ACC + BVP (C1)ACC + EDA (C2)ACC + BVP + EDA (C3)
ActivityF1-ScoreAUC ROCF1-ScoreAUC ROCF1-ScoreAUC ROCF1-ScoreAUC ROC
Brisk walking83.92 ± 21.7198.25 ± 4.4381.34 ± 23.1798.17 ± 4.7682.43 ± 22.4998.11 ± 7.1679.46 ± 27.6797.98 ± 6.92
Cycling88.24 ± 19.2199.37 ± 1.1287.60 ± 16.4299.57 ± 1.0089.19 ± 19.2799.66 ± 0.7990.44 ± 15.8599.70 ± 0.86
Jogging65.60 ± 31.8897.66 ± 3.5663.98 ± 32.1597.74 ± 3.3865.15 ± 32.3397.84 ± 3.4065.11 ± 31.1997.68 ± 3.54
Lying83.22 ± 16.5699.10 ± 1.7580.27 ± 18.5098.96 ± 2.0680.71 ± 20.0899.11 ± 1.7579.38 ± 21.2999.01 ± 1.97
Running59.69 ± 35.0998.44 ± 3.0168.54 ± 34.2798.41 ± 2.7760.49 ± 33.6998.57 ± 2.6669.45 ± 33.9198.54 ± 2.60
Stairs85.97 ± 11.8199.08 ± 1.5085.66 ± 11.6399.11 ± 1.4089.65 ± 11.0099.66 ± 0.8789.01 ± 12.0199.64 ± 0.92
Standing74.00 ± 22.1097.89 ± 2.3472.46 ± 20.1497.57 ± 3.0974.32 ± 21.1798.09 ± 2.4874.24 ± 18.2897.86 ± 2.91
Walking79.36 ± 25.7799.14 ± 1.3678.93 ± 27.5399.07 ± 1.4281.69 ± 23.2099.44 ± 1.1780.58 ± 26.3999.38 ± 1.17
Table 3. CNN classification performance for eight individual activities.
Table 3. CNN classification performance for eight individual activities.
Classifier InputACC (C0)ACC + BVP (C1)ACC + EDA (C2)ACC + BVP + EDA (C3)
ActivityF1-ScoreAUC ROCF1-ScoreAUC ROCF1-ScoreAUC ROCF1-ScoreAUC ROC
Brisk walking 365.78 ± 19.7195.18 ± 4.2365.06 ± 18.6495.29 ± 3.6164.89 ± 20.1894.93 ± 5.7068.72 ± 18.5595.83 ± 3.98
Cycling 12367.53 ± 16.6095.39 ± 3.4878.94 ± 14.6798.11 ± 2.2475.49 ± 16.5497.41 ± 2.7580.54 ± 12.96 98.32 ± 2.26
Jogging 2355.53 ± 27.2092.38 ± 4.0856.44 ± 25.0295.65 ± 3.5760.72 ± 23.2196.53 ± 2.3658.77 ± 23.0396.09 ± 3.55
Lying 12373.75 ± 17.1997.28 ± 3.4981.05 ± 20.3098.66 ± 2.1376.99 ± 18.8997.75 ± 3.7180.11 ± 18.3498.48 ± 2.43
Running51.10 ± 21.8495.37 ± 3.0255.61 ± 19.2895.77 ± 3.4756.35 ± 18.33 95.42 ± 3.86 54.73 ± 20.11 94.76 ± 7.29
Stairs 12375.19 ± 9.2296.23 ± 4.0180.52 ± 12.8997.37 ± 2.9878.81 ± 12.6796.48 ± 5.1381.68 ± 13.3998.07 ± 3.12
Standing 1366.94 ± 16.6296.24 ± 2.9174.36 ± 17.4497.54 ± 3.3669.63 ± 18.8296.49 ± 3.3973.08 ± 18.1397.67 ± 2.73
Walking 357.95 ± 25.2491.32 ± 8.2460.08 ± 24.9992.92 ± 7.2158.72 ± 25.1192.45 ± 7.8061.52 ± 24.2393.87 ± 8.35
1 Indicates significant improvement with added BVP. 2 Indicates significant improvement with added EDA. 3 Indicates significant improvement with added BVP and EDA.
Table 4. RF classification performance for grouped, stairs, and cycling activities.
Table 4. RF classification performance for grouped, stairs, and cycling activities.
Classifier InputACC (C0)ACC + BVP (C1)ACC + EDA (C2)ACC + BVP + EDA (C3)
ActivityF1-ScoreAUC ROCF1-ScoreAUC ROCF1-ScoreAUC ROCF1-ScoreAUC ROC
Low-intensity *96.36 ± 4.5899.85 ± 0.2795.46 ± 5.5899.89 ± 0.2496.66 ± 4.2899.93 ± 0.1496.67 ± 4.1999.94 ± 0.12
Medium-intensity92.96 ± 8.11 99.50 ± 0.8093.05 ± 10.4599.57 ± 0.54 92.96 ± 11.75 99.86 ± 0.2291.86 ± 16.57 99.86 ± 0.25
High-intensity99.33 ± 2.31100.0 ± 0.0099.39 ± 2.22100.0 ± 0.0099.30 ± 2.30100.0 ± 2.28 × 10−599.38 ± 2.22100.0 ± 1.14 × 10−3
Cycling87.87 ± 20.9199.41 ± 1.1087.18 ± 17.4899.60 ± 0.9689.06 ± 19.8699.64 ± 0.9189.93 ± 17.2499.68 ± 0.98
Stairs86.09 ± 12.1199.06 ± 1.6786.46 ± 12.0899.12 ± 1.4588.48 ± 11.9599.69 ± 0.86 88.24 ± 12.4699.62 ± 1.01
* Low-intensity: lying/standing; medium-intensity: brisk walking/walking; high-intensity: jogging/running.
Table 5. CNN classification performance for grouped, stairs, and cycling activities.
Table 5. CNN classification performance for grouped, stairs, and cycling activities.
Classifier InputACC (C0)ACC + BVP (C1)ACC + EDA (C2)ACC + BVP + EDA (C3)
ActivityF1-ScoreAUC ROCF1-ScoreAUC ROCF1-ScoreAUC ROCF1-ScoreAUC ROC
Low-intensity *1385.76 ± 8.74 96.66 ± 3.3091.89 ± 8.0697.55 ± 1.6888.27 ± 7.8396.98 ± 2.5391.70 ± 7.9598.15 ± 1.53
Medium-intensity 1382.61 ± 9.6892.89 ± 2.8886.37 ± 9.3597.53 ± 3.0683.59 ± 12.2296.07 ± 3.9587.81 ± 8.3697.60 ± 2.31
High-intensity95.16 ± 8.7999.66 ± 0.9894.65 ± 11.50 99.55 ± 1.4494.54 ± 12.1299.53 ± 1.4495.89 ± 7.5799.67 ± 1.03
Cycling 12365.07 ± 18.1497.52 ± 9.1178.76 ± 13.9898.29 ± 2.7774.82 ± 15.8298.29 ± 5.2980.63 ± 13.2499.13 ± 3.27
Stairs 1374.64 ± 10.1195.90 ± 4.1380.05 ± 12.9397.21 ± 3.7477.65 ± 11.2096.83 ± 4.2381.51 ± 10.6297.87 ± 3.05
1 Indicates significant improvement with added BVP. 2 Indicates significant improvement with added EDA. 3 Indicates significant improvement with added BVP and EDA. * Low intensity: lying/standing; medium-intensity: brisk walking/walking; high-intensity: jogging/running.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gilmore, J.; Nasseri, M. Human Activity Recognition Algorithm with Physiological and Inertial Signals Fusion: Photoplethysmography, Electrodermal Activity, and Accelerometry. Sensors 2024, 24, 3005. https://doi.org/10.3390/s24103005

AMA Style

Gilmore J, Nasseri M. Human Activity Recognition Algorithm with Physiological and Inertial Signals Fusion: Photoplethysmography, Electrodermal Activity, and Accelerometry. Sensors. 2024; 24(10):3005. https://doi.org/10.3390/s24103005

Chicago/Turabian Style

Gilmore, Justin, and Mona Nasseri. 2024. "Human Activity Recognition Algorithm with Physiological and Inertial Signals Fusion: Photoplethysmography, Electrodermal Activity, and Accelerometry" Sensors 24, no. 10: 3005. https://doi.org/10.3390/s24103005

APA Style

Gilmore, J., & Nasseri, M. (2024). Human Activity Recognition Algorithm with Physiological and Inertial Signals Fusion: Photoplethysmography, Electrodermal Activity, and Accelerometry. Sensors, 24(10), 3005. https://doi.org/10.3390/s24103005

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop