Next Article in Journal
Surface-Plasmon-Resonance Amplification of FMD Detection through Dendrimer Conjugation
Next Article in Special Issue
Development of Wearable Devices for Collecting Digital Rehabilitation/Fitness Data from Lower Limbs
Previous Article in Journal
Wearable Sensors Assess the Effects of Human–Robot Collaboration in Simulated Pollination
Previous Article in Special Issue
Multi-Sensors for Human Activity Recognition
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Evaluation of Diverse Vision Systems towards Detecting Human Pose in Collaborative Robot Applications

Laboratory for Advanced Manufacturing Simulation and Robotics, School of Mechanical and Materials Engineering, University College Dublin, Belfield, D04 V1W8 Dublin, Ireland
*
Author to whom correspondence should be addressed.
Submission received: 19 November 2023 / Revised: 26 December 2023 / Accepted: 5 January 2024 / Published: 17 January 2024
(This article belongs to the Special Issue Multi-sensor for Human Activity Recognition: 2nd Edition)

Abstract

:
Tracking human operators working in the vicinity of collaborative robots can improve the design of safety architecture, ergonomics, and the execution of assembly tasks in a human–robot collaboration scenario. Three commercial spatial computation kits were used along with their Software Development Kits that provide various real-time functionalities to track human poses. The paper explored the possibility of combining the capabilities of different hardware systems and software frameworks that may lead to better performance and accuracy in detecting the human pose in collaborative robotic applications. This study assessed their performance in two different human poses at six depth levels, comparing the raw data and noise-reducing filtered data. In addition, a laser measurement device was employed as a ground truth indicator, together with the average Root Mean Square Error as an error metric. The obtained results were analysed and compared in terms of positional accuracy and repeatability, indicating the dependence of the sensors’ performance on the tracking distance. A Kalman-based filter was applied to fuse the human skeleton data and then to reconstruct the operator’s poses considering their performance in different distance zones. The results indicated that at a distance less than 3 m, Microsoft Azure Kinect demonstrated better tracking performance, followed by Intel RealSense D455 and Stereolabs ZED2, while at ranges higher than 3 m, ZED2 had superior tracking performance.

1. Introduction

Industry 4.0 principles have been evolving in parallel with working environments that encapsulate human skills (e.g., cognition, decision making) and capabilities of robotic systems (e.g., dexterity, robustness, accuracy). Industry 4.0 technologies are on the advert of becoming an integral part of the current manufacturing ecosystem [1]. This causes multiple concerns about safety, ergonomics, and task optimisation [2], rendering the modelling of humans and their activities a critical aspect to be considered. Towards this direction, the utilisation of multiple sensors skeleton and Internet of Things (IoT) technologies, has gained popularity within manufacturing environments for various applications. The collection, transfer, and exchange of data from IoT devices via communication networks enable real-time interaction and cooperation among physical objects [3]. A series of virtual simulation-based solutions have been proposed, such as Digital Twin (DT), Cyber-Physical Systems (CPSs), and Digital Human Modelling (DHM), paving the way for fully digitising industrial shop floors [4,5,6].
The safety of operators within collaborative workspaces where they may share tasks with robots [7] is of paramount importance [8]. The design of these human–robot applications can be rendered more efficient with the utilisation of vision systems [9] as their technology enables the constant tracking and monitoring of human operators’ joints and the adaptive human–robot cooperation and interaction [10]. Typically, vision-based tracking solutions are widely categorised into marker-based and markerless. While marker-based tracking attains high accuracy, its increased cost, rigorous preparation requirements, and complexity have restricted its applicability [11]. On the other hand, markerless depth sensors with skeleton tracking capabilities have become increasingly popular due to their portability, generic applicability, and affordable cost [12]. These sensors exhibit Human Activity Recognition capabilities, by tracking the human pose and changes occurring in the environment and are applied to diverse research applications. However, Human Activity Recognition still remains a challenging area of research in computer vision [13].
Nevertheless, before considering vision systems as a viable option to operate in conjunction with or in lieu of ISO-certified sensor devices, such as safety camera systems, laser scanners, and proximity sensors [14], rigorous testing scenarios and methods are required to investigate their efficiency. The main goal would be to allow vision-based human tracking technologies to complement and work synergistically with built-in safety sensors that commercially available collaborative robots carry in order to overcome unforeseen circumstances to carry out complex tasks for instance shown in Ref. [15]. The identification or monitoring of specific safety features is required in industrial applications as per ISO/TS 15066 [16]. The specific standard provides a series of safety guidelines depending on the level of interaction and can be used complementarily to other ISO guidelines associated with robotic processes, such as ISO 10218-1:2011 [17], ISO 10218-2:2011 [18], and ISO 13855 [19,20]. Yet, it should be noted that the maturity and readiness of vision systems in industrial environments are still under review, since in some specific scenarios the detection of an operator may be prevented due to occlusions [21,22]. In these cases, using various types of sensors in conjunction with sensor fusion algorithms has been reported as a method to improve the overall perception of the process of human pose estimation for collaborative robotic applications [23].
In the present study, the main focus is placed on applications where understanding the complete body pose is crucial for effective human-robot cooperation. Three widely used vision sensors with high-depth accuracy were applied to detect human skeleton joints in two different poses, i.e., Azure Kinect (AK), Stereolabs ZED2, and Inter RealSense D455. Instead of detecting the pose of a perfectly planar object [24], the benchmarking in the present study involved the tracking of a human skeleton, aiming to investigate the performance of the sensors in a collaborative workspace in terms of accuracy and repeatability. The study estimates the coordinates of an operator’s joints at various depth levels from the cameras and compares them with the ground truth, calculating the average RMSE of the depth data. Furthermore, the position of the pelvis joint is tracked (being the parent joint of the skeleton data) to find its accuracy and RMSE with respect to the global frame. In addition, the RMSE of the position of the operator’s wrist is tracked to provide the error estimation with respect to the same global frame of reference. Finally, Kalman-based filtering is applied to fuse the data from the vision sensors at distinct collaborative zones assigned based on the analysed RMSE result. The authors also proposed a feasible control strategy of human motion tracking for Human Robot Collaboration (HRC) applications in a collaborative workspace.

2. State of the Art Review

This section presents recent publications that use human tracking systems based on vision sensors in collaborative environments. In most cases, the robot was fixed, and the human operator worked in proximity to complete independent tasks or interact with it. However, applications involving collaborative mobile robot and dual-arm mobile robots [25] exhibited a significantly increased complexity of the tracking strategy using vision systems due to occlusion, and various other hindrances.
A widespread application of such systems is ensuring that there is no collision between the end effector and a human or object and, to a lesser extent, between a human or object and the other joints of the robot. Bonci et al. [26] presented a proof of concept, dealing with the human-obstacle collision avoidance involving a collaborative fixed-base manipulator, utilising an Acusense RGB-D (Red Green Blue-Depth) camera. The collision avoidance strategy depended on the distance between the fixed robot and the operator. For short distances, it relied on the data collected from the depth sensor, while for longer distances (out of the range of the depth sensor) on the processing of the RGB frames using a You Only Look Once (YOLO)-based Convolutional Neural Network (CNN). Their proposed methodology claimed to reduce the amount of processed data while enhancing the operator’s safety. Scimmi et al. [27] approached the same problem using two Kinect v2 RGB-D cameras to acquire the position of the operator and avoid problems related to the occlusions of the sensors. Each camera extracted 25 joints of a human skeleton. The data collected from the two sets of coordinates were fused using a fusion algorithm developed to obtain the optimal skeleton poses. It was found that the proposed strategy could effectively alter the planned trajectory and prevent human–robot collisions in two case studies. Chen and Song [28] also used two Kinect V2 RGB-D cameras to develop a collision-free motion planning algorithm applied to a robotic arm. Initially, the acquired depth images were used to generate point cloud segmented objects, which were subsequently merged into a single cloud using a K-Nearest Neighbour (KNN) algorithm, aiming to identify the closest point from an obstacle to the robot. Moreover, a Kalman filter was applied in the process of estimating the obstacle motion parameters (velocity, position). It was found that the robotic manipulator managed to avoid collision with an obstacle and preserve the desired trajectory of the effector while following the proposed control design during a Cartesian hexagon task. Furthermore, Pupa et al. [29] applied an effective two-layered strategy for trajectory planning and velocity scaling in a six-DoF manipulator, aiming to enhance a safe HRC. The first layer planned dynamically the initial nominal trajectory, examined its feasibility at maximum velocity, and amended it based on human tracking information captured by six OptiTrack Prime cameras. The second layer adjusted the robot velocity to ensure that its limits adhered to ISO safety constraints. The system architecture was validated experimentally in two scenarios: when the operator hinders the motion or path of the robot and when the two agents are in proximity.
Several researchers have also investigated the use of cameras in conjunction with other sensors to implement dynamic obstacle avoidance strategies. For instance, Gatesichapakorn et al. [30] combined a laser localisation sensor with an RGB-D camera to navigate an autonomous mobile robot. The generation of the static map was implemented in the Robot Operating System (ROS) using a 2D laser-based Simultaneous Localisation and Mapping (SLAM) package. The experimentation in an indoor public space demonstrated the ability of the robot to adapt its motion to the appearance of a human obstacle and subsequently recover its trajectory. Another system that enabled the operation of an anthropomorphic robot through multiple sensors was proposed by Cherubini et al. [31] aiming to implement smart logistic tasks transporting automotive parts. It involved one RGB-D and four RGB cameras, two laser scanners, two force sensors, ten tactile sensors, and two stereo vision sensors where the individual tasks, including the target detection and obstacle mapping, were performed by different sensors. The robotic system was significantly accurate in recognising hand gestures, and therefore the authors proposed a real-time programming strategy based on sign language for intuitive robot control. It should be considered though that the use of such a high number of sensors increased the cost of the infrastructure significantly. Gradolewski et al. [32] presented a real-time safety system that proposes actions to a collaborative robot based on human detection and localisation. An HD vision camera was used for motion detection, together with an ultrasound sensor for proximity estimation. These devices, along with the controller, constituted the detection unit. Moreover, the authors estimated and compared three machine learning algorithms in terms of detection efficiency and maximum latency, concluding that YOLO outperformed Histogram of Oriented Gradients (HOGs) and Viola-Jones.
The improvement of the computational capabilities of Graphical Processing Unit (GPU) technology has significantly facilitated the integration of parallel computation into motion planning algorithms over the last years. Cefalo et al. [33] proposed an algorithm for collision detection to solve a Task-Constrained Motion planning problem [34], and applied it to a robotic arm. The proposed algorithm utilised two real-time images that presented the obstacle mapping (real depth image) and the future robot configuration (virtual depth image) obtained from a Kinect camera and the robot CAD model, respectively. The possibility of the collision scenario was processed in parallel by comparing the two images. Tölgyessy et al. [35] evaluated the Azure Kinect with its predecessors, namely, Kinect V1 and Kinect V2, focusing on precision and noise generation. Their study reported that the performance indicators of Azure Kinect lie within the range indicated in the official documentation. The study concluded that the Azure Kinect may not be suitable for outdoor applications due to limitations of the time-of-flight technology and requires a warm-up time of at least 40–50 min to give stable results.
Human pose detection with vision sensors is another key feature towards the enhancement of HRC activities. Johnson et al. utilised a vision-inertial-based fusion algorithm to initialise and calibrate a forward kinematic model of an arm, which tracks the position and orientation of the arm: the combination of using vision- and IMU-based sensors overcomes the drifts thereby improving the accuracy of tracking the pose of the human arm [36]. Similarly, a visual-inertial sensor-based approach with three sensor modules with each module comprising IMU and ArUco marker attached to three parts of the body mainly, to the trunk, upper arm, and forearm provides a simpler solution for the assessment of movement during robot-assisted training; the ArUco marker, which can be captured by the camera and the driftless orientation of the modules is computed via the visual-inertial sensor fusion algorithm [37]. An HRI framework using a vision-based system together with a three-axis accelerometer, trained on activity classification with a library of 22 gestures and six behaviours, demonstrated a 95% success in the recognition of gesture and 97% in the recognition of behaviour. The intelligent system integrates static and dynamic gestures using ANN and hidden Markov models [38]. Furthermore, a similar approach applied to a case study involving online robot teleoperation to assemble pins in car doors has been demonstrated [39]. An activity recognition strategy using Gaussian mixed HMM, using Microsoft Kinect, was able to detect the human activity with a recall accuracy of 84% with previously seen models and 78% with unseen models [40]. Also, Hernández et al. [41] compared the estimation of shoulder and elbow angles as captured by a webcam in rehabilitation exercises using markerless pose estimators from two CNN frameworks, OpenPose and Detectron2. The data collected from two Kinect V2 RGB-D cameras were fused to generate the ground truth for the upper body joint. OpenPose was found to identify the angles of the limbs more accurately than Detectron2 in all different scenarios. The tracking of the human body orientation with depth cameras, namely, Kinect V2, Azure Kinect, and ZED2i, for the detection of socially occupied space while interacting with people was investigated by Sosa-Leόn et al. [42]. Related approaches that identify the orientation of human body poses may be used in cases of Human–Robot Collaboration for real-time decision making and path planning to carry out tasks. Similarly, De Feudis et al. [43] assessed four different vision systems for hand tool pose estimation: ArUco, OpenPose, Azure Kinect Body Tracking, and YOLO network were used with HTC Vive as a benchmarking system. Further, in a study presented in [44], Azure Kinect and Intel RealSense D435i were compared where the Intel RealSense was reported to show poorer performance in the estimation after 2 m, while the Azure Kinect performed better. Furthermore, the study reported that the depth accuracy of Azure Kinect largely depends on the emissivity of the object, while the RealSense remained unaffected.
The experimentation involved three different motion scenarios of a human operator handling a cordless drill with its mandrel considered as the point of interest to be tracked [43]. The mean square point-to-point distance (D.RMS) and the multivariate R2 were used as the accuracy evaluation criteria. The authors found that the Azure Kinect Body Tracking attained the overall lowest performance, being particularly inaccurate to track the right- and left-hand joints. On the other hand, ArUco generated the most accurate results with the lowest standard deviation of D.RMS for all three scenarios. Similarly, another study [45] uses RGB data for task predictions within a collaborative workspace to manage an assembly process, which is validated by a demonstrator used to assemble a mechanical component. On evaluating four different frameworks, namely, Faster R-CNN, ResNet-50, and ResNet-101, YOLOv2 and YOLOv3, the YOLOv3 framework performed the best with an average mean performance of 72.26% when completing the assembly task.

3. Models and Methods

This paper proposes a new approach for comparing the performance of different vision systems, while taking advantage of the diverse capabilities of the associated hardware and software components, thus leading to the better human pose detection.

3.1. Experimental Setup

The skeleton pose detection was carried out using three depth-based vision sensors: Azure Kinect, Stereolabs ZED2, and Intel RealSense D455. Their key features are extensively presented in Table 1. The sensors were connected to a desktop computer with Intel i7-11th Gen 8 Core processor, 32 GB RAM, and 8 GB NVIDIA RTX 3070 graphic card. Each sensor uses a different depth-sensing technology. More specifically, AK utilises time of flight, i.e., emits and detects backscattered modulated light, translating the phase difference into depth distance for each pixel [46]. ZED2 uses a Convolutional Neural Network (CNN) algorithm for stereo matching [47], while Intel RealSense 455 [48] interprets the scene by comparing images acquired from two known and slightly different positions.
The markerless approach for skeleton tracking is primarily based on CNN approaches. Firstly, in the case of Azure Kinect, the Infrared Sensor (IR) data are fed into a Neural Network, which extracts a silhouette of users and 2D joint coordinates. Combining 2D joint pixel values with the depth data provides the 3D joint information of the skeleton joints [49]. Secondly, the ZED2 body tracking SDK uses neural networks to detect keypoints or the skeleton joints, which are combined with the depth and positional tracking provided by the SDK of ZED2 to obtain a 3D pose estimate of the persons in the scene. Finally, OpenPose, a popular pose estimation model [50] coupled with the Intel Realsense D455, is used to detect keypoints or parts to identify the human joints. Therefore, three sensors that are capable of skeleton-based tracking as well as of providing human key points [51,52,53] in 3D are used in this study.
A 2D pose estimation uses multi-stage CNN to predict Part Affinity Fields (PAFs) and confidence maps. The 2D joint pose estimation is converted into 3D information using depth data, if available [54]. The body tracking SDKs of Azure and ZED2 provide information about the individual joint positions and orientations, while in the case of the OpenPose framework [54] used in conjunction with the Intel D455, the skeleton information comprises exclusively 3D joint positions. Depending on the number of keypoints (joints) required, BODY_25 or COCO format could be chosen as the output of the OpenPose framework [55]. BODY_25 was preferred in this study as it attained faster detection by approximately 30% and higher accuracy by 3% compared to COCO [56]. In the case of the other two vision sensors, the default outputs of the SDK’s skeleton joint data were retained for the study. The authors individually compared the performances of the three body-tracking SDKs for the evaluation of the pose accuracy at different depths from the camera in order to find a suitable device that has the potential to be used in collaborative mobile robotic applications.
The options that the three depth cameras offer in terms of colour and depth resolution are presented in Table 2, along with the modes used in the current study. The experiments were performed within the ROS framework using the respective drivers of each sensor [57,58,59]. The joints information was acquired in the ROS network at a frequency of 18.5, 12, and 18 Hz for AK, ZED2, and Intel D455, respectively. In the case of AK, the NFOV (Narrow Field of View) mode with a range of 0.5–3.86 m was chosen for comparison with other vision sensors as NFOV covers more depth compared to WFOV (Wide Field of View) and attains superior pixel overlap as indicated by the manufacturer [60]. Furthermore, Tölgyessy et al. [9] tested various modes of AK body tracking SDK and reported that the data acquired using NFOV data were more stable than the WFOV mode. The resolution parameters selected for ZED2 and Intel D455 were based on the available computation power of the desktop computer and the requirement for the simultaneous operation of the three vision systems [61].
The experiments were carried out in a confined laboratory environment (7.8 × 3.4 × 4.5 m3) under physical lighting conditions involving natural sunlight and artificial roof light (Figure 1a). The various distance levels from the cameras (i.e., 1.5, 2.0, 3.0, 4.0, 5.0, 6.0 m) were marked on the reference line using a Bosch Laser Measure device (BLM) with ±1.5 mm (0.0015 m) accuracy to guide the operator. Moreover, two poles with a height of 1.274 m (Figure 1b) were placed on both sides of the reference line, serving as a guide for pose estimation involving the wrist joint.
The three cameras and the BLM were clamped on a desk camera mount, as seen in (Figure 2a), ensuring that they were aligned to the XY plane. The data were acquired with respect to the global frame (Reference Frame), as shown in Figure 2b. According to the ROS conventions, the coordinate frames X, Y, and Z were represented in red, green, and blue, respectively. The global frame from RViz (visualisation tool in ROS) with the individual coordinates of Azure, Intel D455, and the ZED2 camera is shown in Figure 2b. The position of the coordinate frames of the cameras was measured using the BLM and was configured in the ROS launch files of each vision sensor to ensure the setup is similar in the real and the virtual world, i.e., by measuring the offset from the Reference Frame to AK, AK to Intel, and AK to ZED2.
After the initialisation of the cameras, the operator moved on each marked point, standing with the hands down (Figure 3, Pose A) and subsequently repeated the same with the wrist on top of the pole (Figure 3, Pose B). Next, the BLM device (Figure 2a) is connected to a smartphone via Bluetooth to estimate the distance between the camera and the operator (ground truth) and calculate the RMSE values.
Then, the cameras started to provide the skeleton joint coordinates published as ROS messages. Overall, 50 samples were collected for each camera, pose, and distance level.
The sequence of data collection was carried out as follows:
  • The vision sensor initialised, and the operator moved to the floor marker.
  • The operator recorded the ground truth depth using the BLM device.
  • The operator moved to Pose A, and the camera started to record the data. First, 50 samples of joint coordinates were collected (XYZ) from each device with respect to the global frame of reference.
  • The process was repeated for Pose B.

3.2. Skeleton Tracking Information

The skeleton joints available for tracking are shown in Figure 4, along with the corresponding names reported in Table 3 based on the documentation of the respective SDKs. Overall, AK, ZED2, and OpenPose provide skeleton data for 32, 34, and 25 joints, respectively. The joints that pertain to the eyes, ears, nose, the tip of the thumbs, and toes were not considered in the evaluation process as they do not affect or contribute to the operator’s pose (see Table 3).
Initially, the datatype acquired from the SDKs via the ROS drivers of AK, ZED2, and Intel D455 was analysed. It was noted that the data of joints belonged to two different types, i.e., MarkerArray in the case of AK and List in the case of ZED2 and Intel D455. Therefore, it was processed and published as TF frames, as shown in Figure 5, for the calculation of translation (X, Y, Z) and rotation (quaternion or roll, pitch, and yaw) of various joints with respect to the reference frame (Figure 2b). Each of the joints used for evaluation in this study is shown in Figure 6.
At a distance of 1.5 m, the Intel D455 camera could capture only the upper body joints (pelvis included) (Table 1) due to the restricted field of view of the captured image data. However, in the case of AK and ZED2, the body tracking algorithm could predict the position of the lower joints of the operator and provide information with low accuracy. Furthermore, as the operator moved further away from the cameras (>1.5 m), the joints below the pelvis were also visible.
Apart from tracking the overall skeleton, particular focus was given to the tracking accuracy of the pelvis and wrist (right and left) joints. The reason is that the pelvis is the first parent joint of the skeleton pose; therefore, its accuracy and stability are critical. In addition, the tracking stability of the wrist joints is important, especially in the case of extension of the limbs (e.g., Pose B), and should be primarily considered when the HRC’s effectiveness is assessed.

3.3. Preliminary Test—Evaluation of Raw Data

After the camera setup, a preliminary procedure was devised to test the raw data. It was observed that during the tracking of joints in RViz, as the operator moved away from the camera, the skeleton gradually levitated from the ground in the case of AK (see Figure 7).
To investigate this further, additional tests were performed with the pelvis joint being tracked while the operator was moving along the reference line, starting from a distance of 1.5 m. In this way, the height (Z)–depth (Y) plot was defined (Figure 7a) with a noticeable slope to be observed exclusively in the case of AK. The skeleton poses are shown indicatively in Figure 7b for a distance level of 3 m. It can be observed that the skeleton coordinates acquired by AK are higher than the respective ZED2 and Intel D455.
The corresponding slope was analytically estimated at −0.110417. As a result, the final height (Z′) obtained by the AK coordinated data was calculated based on the real-time values of Z and Y as updated within the published TF data using Equation (1):
Z = Z s l o p e × Y
In addition, a moving average filter with a window size of 30 data was applied to the real-time data to minimise the noise. The obtained results are shown in Figure 8a, where the slope of Azure Kinect is significantly reduced, while the respective skeleton poses are shown in Figure 8b, with the pelvis joint of AK closely aligned with the pelvis joints of the other vision sensors.

4. Results and Discussion

This section assesses the accuracy of the depth (Y) estimation for the three cameras resulting from the tracked skeleton joint and evaluates their performance while capturing the two poses at various depths (distance levels from the camera). Two data sets are presented: (i) the raw skeleton data from the cameras in Section 4.1 and (ii) the filtered data (after applying the moving average filter) in Section 4.2. In both cases, the AK slope was compensated to minimise the levitation from the ground, as previously explained (Section 3.3). For the further evaluation of Pose B, the left and right wrists were selected as the common joints of all three cameras (see Table 3, joint numbers: Azure Kinect: 7, 14, ZED2: 7, 14, and Intel D455: 4, 7).

4.1. Accuracy Estimation of the Raw Data

4.1.1. Evaluation of the Depth Accuracy

The average RMSE values of the operator’s depth (Y) in Pose A for 50 iterations are shown in Figure 9a, while Figure 10a presents the results for the operator in Pose B.
The box plots showed that the AK body tracking attained the lowest RMSE, followed by Intel D455 and ZED2 (Figure 9a and Figure 10a). This increase in the average RMSE may be due to the inverse relationship between the disparity-depth pixel information [62]. Furthermore, the perspective foreshortening effect may have affected the accuracy of the skeleton poses in the case of stereo cameras [63,64,65]. Moreover, as the operator moved further away from the camera, the AK and Intel D455 joint data became unstable, and the deviation of the acquired skeletons from the original poses became significant at distances higher than 4 m (Figure 9b and Figure 10b). On the other hand, in the case of ZED2, the acquired skeleton was relatively consistent for both poses and all distance levels.

4.1.2. Overall Performance of the Skeleton Pose Estimation—Pose A and Pose B

The tracking of the overall skeleton joints obtained from the three sensors is depicted in Figure 11.
In Figure 11, as obtained from all vision sensors, it was noted that the overall skeleton poses of the operator presented gradual deviation along the X-axis with respect to the global frame, as indicated by the red rectangle on the pelvis joint. This trend was obtained for Pose A (Figure 11a–f) and Pose B (Figure 11g–l).
In the case of ZED2 and for both poses, the X-RMSE reduced as the operator moved away from the camera, as depicted in Figure 12 indicatively for Pose A. More specifically, the RMSE of the unfiltered pelvis joint data from AK and Intel D455 was lower than ZED2 by approximately 43% and 74%, respectively, at depth ranges of less than 2.5 m. However, at higher depths (>3 m), ZED2 demonstrated superior tracking performance in this distance range especially considering the tendency of AK and Intel D455 to deform the tracked skeleton significantly (Figure 9 and Figure 10).

4.1.3. Pose Accuracy Estimation by Tracking Wrist Joint—Pose B

Following the evaluation of the overall skeleton of the operator in Poses A and B, an additional evaluation was performed to estimate the position of the wrist joint using the poles as fixed objects, i.e., with known positions with respect to the reference line. As a result, the RMSE of the Y and Z for the discrete distance levels is presented in Figure 13 and Figure 14 for the left and right wrist, respectively.
Overall, an increase in the average RMSE of the wrist joint was observed as the operator–sensor distance increased. The deterioration of the tracking accuracy of the limbs with the tracking distance has also been confirmed by Romeo et al. [66] who reported that the acquired data of AK that pertained to the limbs (wrist, hands) were less accurate compared to the data of the upper body joints such as the pelvis, chest, and neck.
As the vision sensors utilise similar AI-based body tracking approaches to train their data, the results of ZED2 and Intel D455 resemble AK data. Training AI-based pose estimation neural networks with synthetic data in realistic conditions accounting for various extrinsic factors, image disparity, occlusion, and foreshortening may improve the overall accuracy of pose estimation.

4.2. Accuracy Estimation of the Filtered Data

This section shows the results of the second data set, filtered in real time using a moving average filter to minimise noise, jitter, and outliers.

4.2.1. Evaluation of the Depth Accuracy

The average RMSE values of the operator’s depth (Y) in Poses A and B after data filtering are presented in Figure 15a and Figure 16a. Figure 15b and Figure 16b present the overall posture of the skeleton in the two poses. In general, RMSE follows the same trend with the unfiltered data, i.e., increases as the camera–operator distance increases. Moreover, the filter had an overall positive effect on the capturing of Pose B in the case of Intel D455 and a negative in the case of AK, especially at longer distances. ZED2 had consistent skeleton tracking in most cases.

4.2.2. Overall Performance of the Skeleton Pose Estimation—Pose A and Pose B

The 3D plots of the overall skeleton poses are presented in Figure 17.
It can be deduced that the operator’s pose shifts gradually toward the positive X as it moves further away from the camera, similar to the results of unfiltered pelvis joint data as presented in Figure 11. Nevertheless, in the case of filtered data, the skeleton shift appears to take place more gradually. This effect may occur due to the minor standard deviation in the filtered data explained in the following section, which compares the raw and filtered data results.
The X-RMSE curve of the pelvis joints (Pose A) for Intel D455 was higher than AK and ZED2 (see Figure 18). Therefore, it can be stated that applying a real-time filter to Intel D455 data did not contribute to the reduction in its X-RSME values, while it lowered the error in the case of AK (see also Figure 12). Also, beyond 4 m, the tracking of the pelvis joint became unstable in the case of AK and Intel D455. In the case of AK, this may happen due to the limitations of the hardware’s tracking capabilities. The significant increase in the filtered X-RMSE of Intel D455 may have been caused due to an external disturbance that pertains to the extrinsic conditions of the laboratory, leading to poor accuracy. For instance, certain settings of the Intel D455 camera were not modified, e.g., the exposure was set to auto mode. However, this does not impact the tracked depth but affects the quality of the output image [67]. Furthermore, since OpenPose is primarily a 2D pose estimation algorithm, which uses colour images, this may have impacted the X-RMSE value.

4.2.3. Pose Accuracy Estimation by Tracking Wrist Joint—Pose B

The RMSE of the operator’s left and right wrist joints after the data filtering is shown in Figure 19 and Figure 20, respectively. The application of a low pass filter, such as a moving average filter, reduced the error of the wrist joints with respect to the Y- and Z-axis by lowering the random noises that affect the acquired data in the case of AK and ZED2. However, in the case of Intel D455, the overall RMSE is much higher than AK and ZED2, which indicates a minor effect of the applied filter, which extrinsic factors may cause during the tests. In addition, the postprocessing filter, namely, the temporal filter, was applied to the RealSense data configured in the camera’s ROS initialisation file. Therefore, the extrinsic factors and the postprocessing filter may have had no effect on reducing the overall RMSE value in the case of Intel D455. However, fine tuning the postprocessing filters under controlled light settings may reduce the RMSE error of the Intel D455 camera.

4.3. Unfiltered vs. Filtered Data

This section presents a comparison of the raw and filtered data. For example, the X and Y values of the operator’s pelvis joint at a depth of 3 m are presented in Figure 21, before and after applying a filter. The significance of its application is indicated by the conversion of the raw (noisy) data curve to a smooth (filtered) curve in the case of all vision sensors.
The authors also estimated the percent error (δ) of the depth (Y) at all distance levels in Pose A. The obtained results are reported in Figure 22. For Intel D455 and ZED2, δ was estimated at less than 2% at 4 m (Figure 22a), indicating its compliance with the respective values reported by the product specification [68,69]. In the case of ZED2, at short distances, the estimated δ was slightly higher than the one reported by the manufacturer.
Although data filtering has slightly increased the RMSE in the estimation of overall poses, its application in HRC scenarios may be preferable due to the resulting reduction in the Standard Deviation (σ) (Figure 22b).
The average RMSE values of the overall joint depth data of the operator in Poses A and B are depicted in Figure 23.
Similarly, the average Z-RMSE values of the wrist joints before and after filtering are shown in Figure 24. In this case, as the operator–camera distance increased, there was an increase in the overall average RSME data. The applied filter significantly improved the AK data compared to the rest of the output of the sensor, followed by ZED2.

4.4. Data Fusion in the Collaborative Zones

Based on the results obtained from the assessment of the performance of the cameras, the authors defined three collaborative zones and proposed a Kalman-based sensor fusion approach to combine the joint data and reconstruct the skeleton pose of the operator. The proposed approach was tested with the operator in Pose A.

4.4.1. Classification of Collaborative Zones and Sensor Fusion

The design of collaborative zones aimed to minimise the error of the joints and facilitate a safer HRC. Therefore, they were classified as Zone I (1.5 m to 2.0 m), Zone 2 (2.0 to 3.5), and Zone 3 (3.5 m and beyond), depending on the distance from the vision sensors, as presented in Figure 25. These limitations were defined considering the capabilities of the vision sensors as reported by the manufacturers (see also Table 1).
In Zones 1 and 2 (Figure 25a), 23 joints were available to reconstruct the skeleton pose using the data obtained from both AK and Intel D455, as they demonstrated a better performance in depths of this range (see Figure 9 and Figure 15). The common joints (indicated in brown) were fused using a Kalman filter, while the rest (shown in blue in Figure 25) were used as obtained from the AK. In Zone 3, ZED2 was explicitly used to track the skeleton pose (shown in red in Figure 25) due to its capability to track accurately at far distances, as explained in Section 4.1 and Section 4.2.

4.4.2. Pose Accuracy Estimation of Fused Data in the Collaborative Zone

This section provides the box plot of the results after the Kalman-based fusion of joints in the collaborative workspace. Figure 26a illustrates the average RMSE of the joint data derived from AK and Intel D455 and the RMSE of nine fused joints in Zone 1, which appears to be the lowest. The fused average RMSE of the joint depth values was estimated at 0.0389 m, with AK and Intel D455 values at 0.0472 m and 0.0649 m, respectively. Figure 26b shows the skeleton pose of the operator at approximately 1.826 m from the camera with AK, Intel D455, and fused joints to be depicted in blue, black, and brown, respectively.
Similarly, Figure 27 illustrates the RMSE of the skeleton joints in Zone 2 with 15 joint data fused in the case of Pose A. The common joints were analytically listed in Table 3.
At a distance of 2.699 m (Zone 2), the average RMSE values from AK and Intel D455 were 0.0784 m and 0.1078 m, respectively, while the RMSE corresponding to the fused joints was 0.08721 m (see Figure 27). This increase in the error of the fused joints may be caused due to extrinsic conditions. However, with further tuning of the sensors’ parameters, such as exposure, resolution, and noise filtering, as well as with the application of available postprocessing techniques, this error may be further reduced.

5. Conclusions

This study aimed to determine the accuracy of skeleton pose estimation at various depths using three different commercial vision systems and frameworks of skeleton pose tracking. One of the goals of the study was to compare various spatial computation kits, which differ in terms of hardware devices and associated software frameworks for tracking human operators. The comparison focused on evaluating the devices and frameworks leading in terms of human operator pose accuracy.
Based on the obtained results, the performance of the sensors from highest to lowest (in the order depth tracking: closer to far distance range) was assessed as follows: AK, Intel D455, and ZED2. The initial evaluation of the raw pelvis data demonstrated that AK data showed a linear levitation trend in height (Z) of the skeleton pose as the operator–camera distance increased. An analytical approach was used to minimise this slope. The obtained results showed that as the operator–camera distance increases, the skeleton pose gradually transitions with respect to the global frame. This phenomenon may affect safety and may be crucial in HRC applications. The deployment of multi-vision-based tracking systems can contribute to the minimisation of such an error.
Comparing the depth accuracy of raw and filtered data, it can be inferred that at a range shorter than 3 m, the AK and Intel D455 demonstrated better performance than ZED2, with the latter providing better tracking results beyond this range. However, at a distance approximately higher than 3.5 m, the tracking of AK becomes unstable due to the constraints of the NFOV mode, which has an operating range of 3.86 m. Therefore, when it comes to detecting entire skeleton poses beyond the range of 3.5 m, it is safer to utilise ZED2 to track entire human body poses and use bounding boxes. Further, with the additional functional tracking features, such as the velocity of the human operators, provided by the SDK of ZED2, this information can be easily used for collaborative mobile robotic applications for long-range tracking in shopfloor environments.
The tracking accuracy relies on various extrinsic parameters, such as the lighting conditions, the colour of the dress or jackets worn by the operator, the background colour, the resolution of the cameras, and the available computational power. Also, more variable parameters of the vision sensor are involved when multiple sensors are present in a scene. In addition, installing different SDKs and dependencies packages can be tedious and may lead to longer building time and runtime errors. Hence, extra attention should be given while performing these tasks involving different configurations of CMake flags, CUDA, and cuDNN versions. Furthermore, as more operators may be present in the scene, the computer requires more processing power to detect the operators’ skeleton joints without compromising the FPS rate.
Finally, developing such sophisticated algorithms utilises different software libraries (open source or commercially licensed), software packages, tools, etc., that contain thousands of lines of codes that have been independently tested. Hence, in the cases of deployment of various spatial computation frameworks, a constant tracking of updates is required in order to keep up to date with the latest features and functionalities provided by the SDKs. For instance, Sterolabs (ZED2) provided more frequent software updates with features and bug fixes, which in turn enhances the performance of the vision sensors.
When deployed, the capabilities of AI-based tracking of the human operator, on the whole, may vary in each scenario; this can be a risk and one of the significant challenges to consider when deploying similar tracking solutions, especially when compared against more conventional, safety-certified solutions.
As AI markerless tracking demonstrates moderate results regarding accuracy, its use on the workspace of a shop floor and its adoption by manufacturing companies is still limited. Their deployment in HRC scenarios in conjunction with additional ISO-certified safety sensors is still preferred in industry. Along with the skeleton tracking, additional features of SDKs, such as the object detection module of ZED2 SDK, may be used to determine the bounding box, the absolute velocity, and the operator’s position. The data collected could be used in conjunction with a sensor data processing or fusion algorithm, such as the Kalman filter or the particle filter algorithm to localise the operator’s position within a collaborative workplace. In addition, the position and velocity information could be used to sync the movements of a collaborative mobile robot with the movement of a human operator. As the pelvis is the parent joint that connects the rest of the skeleton joints, additional work in this area could involve its marker-based tracking in order to improve the overall skeleton accuracy. Future work includes the setup of a controlled lighting condition with LED and testing the performance of the vision sensor under various settings such as resolution, FPS, and brightness. Other classical or machine learning-based methods of determining the position of the human body, including, for instance the use of the pictorial structure framework approach or deep learning methods, could also be tested and benchmarked in the future using diverse hardware or software configurations.

Author Contributions

Conceptualization, A.K.R. and N.P.; methodology, A.K.R. and N.P.; software, A.K.R.; validation, A.K.R.; formal analysis, A.K.R. and M.K.; investigation, A.K.R. and N.P.; resources, N.P.; data curation, A.K.R.; writing—original draft preparation, A.K.R., M.K., B.F. and N.P.; writing—review and editing, M.K., A.K.R. and N.P.; visualization, A.K.R. and M.K.; supervision, N.P.; project administration, N.P.; funding acquisition, N.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the European Union Horizon 2020 Framework Programme—project SHERLOCK under Grant Agreement: 820689.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bilberg, A.; Malik, A.A. Digital Twin Driven Human–Robot Collaborative Assembly. CIRP Annals. 2019, 68, 499–502. [Google Scholar] [CrossRef]
  2. Aaltonen, I.; Salmi, T. Experiences and Expectations of Collaborative Robots in Industry and Academia: Barriers and Development Needs. Procedia Manuf. 2019, 38, 1151–1158. [Google Scholar] [CrossRef]
  3. Frank, A.G.; Dalenogare, L.S.; Ayala, N.F. Industry 4.0 Technologies: Implementation Patterns in Manufacturing Companies. Int. J. Prod. Econ. 2019, 210, 15–26. [Google Scholar] [CrossRef]
  4. Paul, G.; Abele, N.D.; Kluth, K. A Review and Qualitative Meta-Analysis of Digital Human Modeling and Cyber-Physical-Systems in Ergonomics 4.0. IISE Trans. Occup. Ergon. Hum. Factors. 2021, 9, 111–123. [Google Scholar] [CrossRef]
  5. Ramasubramanian, A.K.; Mathew, R.; Kelly, M.; Hargaden, V.; Papakostas, N. Digital Twin for Human–Robot Collaboration in Manufacturing: Review and Outlook. Appl. Sci. 2022, 12, 4811. [Google Scholar] [CrossRef]
  6. Yin, M.-Y.; Li, J.-G. A Systematic Review on Digital Human Models in Assembly Process Planning. Int. J. Adv. Manuf. Technol. 2023, 125, 1037–1059. [Google Scholar] [CrossRef]
  7. Wang, L.; Gao, R.; Váncza, J.; Krüger, J.; Wang, X.V.; Makris, S.; Chryssolouris, G. Symbiotic Human-Robot Collaborative Assembly. CIRP Annals. 2019, 68, 701–726. [Google Scholar] [CrossRef]
  8. Chemweno, P.; Pintelon, L.; Decre, W. Orienting Safety Assurance with Outcomes of Hazard Analysis and Risk Assessment: A Review of the ISO 15066 Standard for Collaborative Robot Systems. Saf. Sci. 2020, 129, 104832. [Google Scholar] [CrossRef]
  9. Tölgyessy, M.; Dekan, M.; Chovanec, L. Skeleton Tracking Accuracy and Precision Evaluation of Kinect V1, Kinect V2, and the Azure Kinect. Appl. Sci. 2021, 11, 5756. [Google Scholar] [CrossRef]
  10. Ramasubramanian, A.K.; Aiman, S.M.; Papakostas, N. On Using Human Activity Recognition Sensors to Improve the Performance of Collaborative Mobile Manipulators: Review and Outlook. Procedia CIRP 2021, 97, 211–216. [Google Scholar] [CrossRef]
  11. Nguyen, M.H.; Hsiao, C.C.; Cheng, W.H.; Huang, C.C. Practical 3D Human Skeleton Tracking Based on Multi-View and Multi-Kinect Fusion. Multimed. Syst. 2022, 28, 529–552. [Google Scholar] [CrossRef]
  12. Yeung, L.F.; Yang, Z.; Cheng, K.C.C.; Du, D.; Tong, R.K.Y. Effects of Camera Viewing Angles on Tracking Kinematic Gait Patterns Using Azure Kinect, Kinect v2 and Orbbec Astra Pro V2. Gait Posture 2021, 87, 19–26. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, H.-B.; Zhang, Y.-X.; Zhong, B.; Lei, Q.; Yang, L.; Du, J.-X.; Chen, D.-S. A Comprehensive Survey of Vision-Based Human Action Recognition Methods. Sensors 2019, 19, 1005. [Google Scholar] [CrossRef] [PubMed]
  14. Arkouli, Z.; Kokotinis, G.; Michalos, G.; Dimitropoulos, N.; Makris, S. AI-Enhanced Cooperating Robots for Reconfigurable Manufacturing of Large Parts. In Proceedings of the IFAC-PapersOnLine, Magaliesburg, South Africa, 7–8 December 2021; Volume 54, pp. 617–622. [Google Scholar]
  15. Ramasubramanian, A.K.; Papakostas, N. Operator—Mobile Robot Collaboration for Synchronized Part Movement. Procedia CIRP 2021, 97, 217–223. [Google Scholar] [CrossRef]
  16. ISO/TS 15066:2016; Robots and Robotic Devices—Collaborative Robots. ISO: Geneva, Switzerland, 2016; Volume 2016.
  17. ISO 10218-1:2011; Robots and Robotic Devices Requirements for Industrial Robots 1: Robots. ISO: Geneva, Switzerland, 2011; Volume 2016.
  18. ISO 10218-2:2011; Robots and Robotic Devices Requirements for Industrial Robots 2: Robot Systems and Integration. ISO: Geneva, Switzerland, 2011.
  19. Bdiwi, M.; Pfeifer, M.; Sterzing, A. A New Strategy for Ensuring Human Safety during Various Levels of Interaction with Industrial Robots. CIRP Ann. Manuf. Technol. 2017, 66, 453–456. [Google Scholar] [CrossRef]
  20. ISO 13855:2010; Safety of Machinery Positioning of Safeguards with Respect to the Approach Speeds of Parts of the Human Body. ISO: Geneva, Switzerland, 2010.
  21. Halme, R.J.; Lanz, M.; Kämäräinen, J.; Pieters, R.; Latokartano, J.; Hietanen, A. Review of Vision-Based Safety Systems for Human-Robot Collaboration. Procedia CIRP 2018, 72, 111–116. [Google Scholar] [CrossRef]
  22. Rodrigues, I.R.; Barbosa, G.; Oliveira Filho, A.; Cani, C.; Sadok, D.H.; Kelner, J.; Souza, R.; Marquezini, M.V.; Lins, S. A New Mechanism for Collision Detection in Human–Robot Collaboration Using Deep Learning Techniques. J. Control. Autom. Electr. Syst. 2022, 33, 406–418. [Google Scholar] [CrossRef]
  23. Amorim, A.; Guimares, D.; Mendona, T.; Neto, P.; Costa, P.; Moreira, A.P. Robust Human Position Estimation in Cooperative Robotic Cells. Robot. Comput.-Integr. Manuf. 2021, 67, 102035. [Google Scholar] [CrossRef]
  24. Kurillo, G.; Hemingway, E.; Cheng, M.L.; Cheng, L. Evaluating the Accuracy of the Azure Kinect and Kinect V2. Sensors 2022, 22, 2469. [Google Scholar] [CrossRef]
  25. Ibarguren, A.; Daelman, P. Path Driven Dual Arm Mobile Co-Manipulation Architecture for Large Part Manipulation in Industrial Environments. Sensors 2021, 21, 6620. [Google Scholar] [CrossRef]
  26. Bonci, A.; Cheng, P.D.C.; Indri, M.; Nabissi, G.; Sibona, F. Human-Robot Perception in Industrial Environments: A Survey. Sensors 2021, 21, 1571. [Google Scholar] [CrossRef] [PubMed]
  27. Scimmi, L.S.; Melchiorre, M.; Mauro, S.; Pastorelli, S. Multiple Collision Avoidance between Human Limbs and Robot Links Algorithm in Collaborative Tasks. In Proceedings of the ICINCO 2018—Proceedings of the 15th International Conference on Informatics in Control, Automation and Robotics, Porto, Portugal, 29–31 July 2018; Volume 2, pp. 291–298. [Google Scholar] [CrossRef]
  28. Chen, J.; Song, K. Collision-Free Motion Planning for Human-Robot Collaborative Safety under Cartesian Constraint. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), Brisbane, Australia, 21–25 May 2018. [Google Scholar]
  29. Pupa, A.; Arrfou, M.; Andreoni, G.; Secchi, C. A Safety-Aware Kinodynamic Architecture for Human-Robot Collaboration. IEEE Robot Autom. Lett. 2021, 6, 4465–4471. [Google Scholar] [CrossRef]
  30. Gatesichapakorn, S.; Takamatsu, J.; Ruchanurucks, M. ROS Based Autonomous Mobile Robot Navigation Using 2D LiDAR and RGB-D Camera. In Proceedings of the 2019 1st International Symposium on Instrumentation, Control, Artificial Intelligence, and Robotics, ICA-SYMP 2019, Bangkok, Thailand, 16–18 January 2019; pp. 151–154. [Google Scholar] [CrossRef]
  31. Cherubini, A.; Passama, R.; Navarro, B.; Sorour, M.; Khelloufi, A.; Mazhar, O.; Tarbouriech, S.; Zhu, J.; Tempier, O.; Crosnier, A.; et al. A Collaborative Robot for the Factory of the Future: BAZAR. Int. J. Adv. Manuf. Technol. 2019, 105, 3643–3659. [Google Scholar] [CrossRef]
  32. Gradolewski, D.; Maslowski, D.; Dziak, D.; Jachimczyk, B.; Mundlamuri, S.T.; Prakash, C.G.; Kulesza, W.J. A Distributed Computing Real-Time Safety System of Collaborative Robot. Elektron. Elektrotechnika 2020, 26, 4–14. [Google Scholar] [CrossRef]
  33. Cefalo, M.; Magrini, E.; Oriolo, G. Parallel Collision Check for Sensor Based Real-Time Motion Planning. In Proceedings of the IEEE International Conference on Robotics and Automation, Singapore, 29 May–3 June 2017; pp. 1936–1943. [Google Scholar] [CrossRef]
  34. Oriolo, G.; Vendittelli, M. A Control-Based Approach to Task-Constrained Motion Planning. In Proceedings of the 2009 IEEE/RSJ International Conference on Intelligent Robots and Systems, IROS 2009, St. Louis, MO, USA, 10–15 October 2009; pp. 297–302. [Google Scholar] [CrossRef]
  35. Tölgyessy, M.; Dekan, M.; Chovanec, Ľ.; Hubinský, P. Evaluation of the Azure Kinect and Its Comparison to Kinect v1 and Kinect V2. Sensors 2021, 21, 413. [Google Scholar] [CrossRef] [PubMed]
  36. Johnson, A.; Ramasubramanian, A.K.; Mathew, R.; Mulkeen, B.; Papakostas, N. Forward Kinematic Based Approach Using Sensor Fusion for Tracking the Human Limb. In Proceedings of the 2022 IEEE 28th International Conference on Engineering, Technology and Innovation (ICE/ITMC) & 31st International Association for Management of Technology (IAMOT) Joint Conference, Nancy, France, 19–23 June 2022; pp. 1–6. [Google Scholar]
  37. Li, T.; Yu, H. Upper Body Pose Estimation Using a Visual-Inertial Sensor System with Automatic Sensor-to-Segment Calibration. IEEE Sens. J. 2023, 23, 6292–6302. [Google Scholar] [CrossRef]
  38. Mendes, N.; Ferrer, J.; Vitorino, J.; Safeea, M.; Neto, P. Human Behavior and Hand Gesture Classification for Smart Human-Robot Interaction. Procedia Manuf. 2017, 11, 91–98. [Google Scholar] [CrossRef]
  39. Mendes, N.; Neto, P.; Safeea, M.; Moreira, A. Online Robot Teleoperation Using Human Hand Gestures: A Case Study for Assembly Operation. In Robot 2015: Second Iberian Robotics Conference: Advances in Robotics; Springer: Berlin/Heidelberg, Germany, 2016; Volume 2, ISBN 978-3-319-27148-4. [Google Scholar]
  40. Piyathilaka, L.; Kodagoda, S. Gaussian Mixture Based HMM for Human Daily Activity Recognition Using 3D Skeleton Features. In Proceedings of the 2013 IEEE 8th Conference on Industrial Electronics and Applications (ICIEA), Melbourne, Australia, 19–21 June 2013; pp. 567–572. [Google Scholar]
  41. Hernández, Ó.G.; Morell, V.; Ramon, J.L.; Jara, C.A. Human Pose Detection for Robotic-Assisted and Rehabilitation Environments. Appl. Sci. 2021, 11, 4183. [Google Scholar] [CrossRef]
  42. Sosa-León, V.A.L.; Schwering, A. Evaluating Automatic Body Orientation Detection for Indoor Location from Skeleton Tracking Data to Detect Socially Occupied Spaces Using the Kinect v2, Azure Kinect and Zed 2i. Sensors 2022, 22, 3798. [Google Scholar] [CrossRef]
  43. De Feudis, I.; Buongiorno, D.; Grossi, S.; Losito, G.; Brunetti, A.; Longo, N.; Di Stefano, G.; Bevilacqua, V. Evaluation of Vision-Based Hand Tool Tracking Methods for Quality Assessment and Training in Human-Centered Industry 4.0. Appl. Sci. 2022, 12, 1796. [Google Scholar] [CrossRef]
  44. Rijal, S.; Pokhrel, S.; Om, M.; Ojha, V.P. Comparing Depth Estimation of Azure Kinect and Realsense D435i Cameras. Ann. Ig. 2023. [Google Scholar] [CrossRef]
  45. Garcia, P.P.; Santos, T.G.; Machado, M.A.; Mendes, N. Deep Learning Framework for Controlling Work Sequence in Collaborative Human-Robot Assembly Processes. Sensors 2023, 23, 553. [Google Scholar] [CrossRef] [PubMed]
  46. Bamji, C.S.; Mehta, S.; Thompson, B.; Elkhatib, T.; Wurster, S.; Akkaya, O.; Payne, A.; Godbaz, J.; Fenton, M.; Rajasekaran, V.; et al. IMpixel 65nm BSI 320MHz Demodulated TOF Image Sensor with 3 μm Global Shutter Pixels and Analog Binning. In Proceedings of the Digest of Technical Papers—IEEE International Solid-State Circuits Conference, San Francisco, CA, USA, 11–15 February 2018; Volume 61, pp. 94–96. [Google Scholar]
  47. Wang, J.; Gao, Z.; Zhang, Y.; Zhou, J.; Wu, J.; Li, P. Real-Time Detection and Location of Potted Flowers Based on a ZED Camera and a YOLO V4-Tiny Deep Learning Algorithm. Horticulturae 2022, 8, 21. [Google Scholar] [CrossRef]
  48. Servi, M.; Mussi, E.; Profili, A.; Furferi, R.; Volpe, Y.; Governi, L.; Buonamici, F. Metrological Characterization and Comparison of D415, D455, L515 Realsense Devices in the Close Range. Sensors 2021, 21, 7770. [Google Scholar] [CrossRef]
  49. Liu, Z. 3D Skeletal Tracking on Azure Kinect—Azure Kinect Body Tracking SDK. Available online: https://www.microsoft.com/en-us/research/uploads/prod/2020/01/AKBTSDK.pdf (accessed on 18 June 2022).
  50. Human Pose Estimation with Deep Learning—Ultimate Overview in 2024. Available online: https://Viso.Ai/Deep-Learning/Pose-Estimation-Ultimate-Overview/ (accessed on 25 December 2023).
  51. Lee, S.; Lee, D.-W.; Jun, K.; Lee, W.; Kim, M.S. Markerless 3D Skeleton Tracking Algorithm by Merging Multiple Inaccurate Skeleton Data from Multiple RGB-D Sensors. Sensors 2022, 22, 3155. [Google Scholar] [CrossRef]
  52. Chung, J.-L.; Ong, L.-Y.; Leow, M.-C. Comparative Analysis of Skeleton-Based Human Pose Estimation. Future Internet 2022, 14, 380. [Google Scholar] [CrossRef]
  53. Lee, K.M.; Krishna, A.; Zaidi, Z.; Paleja, R.; Chen, L.; Hedlund-Botti, E.; Schrum, M.; Gombolay, M. The Effect of Robot Skill Level and Communication in Rapid, Proximate Human-Robot Collaboration. In Proceedings of the 2023 ACM/IEEE International Conference on Human-Robot Interaction, Stockholm, Sweden, 13–16 March 2023; Association for Computing Machinery: New York, NY, USA, 2023; pp. 261–270. [Google Scholar]
  54. Cao, Z.; Hidalgo, G.; Simon, T.; Wei, S.-E.; Sheikh, Y. OpenPose: Realtime Multi-Person 2D Pose Estimation Using Part Affinity Fields. IEEE Trans. Pattern. Anal. Mach. Intell. 2021, 43, 172–186. [Google Scholar] [CrossRef]
  55. OpenPose 1.7.0 The First Real-Time Multi-Person System to Jointly Detect Human Body, Hand, Facial, and Foot Keypoints. Available online: https://cmu-perceptual-computing-lab.github.io/openpose/web/html/doc/md_doc_02_output.html#pose-output-format-coco (accessed on 18 June 2022).
  56. OpenPose Doc—Release Notes. Available online: https://github.com/CMU-Perceptual-Computing-Lab/openpose/blob/master/doc/08_release_notes.md (accessed on 18 June 2022).
  57. Azure Kinect ROS Driver. Available online: https://github.com/microsoft/Azure_Kinect_ROS_Driver (accessed on 18 June 2022).
  58. Stereolabs ZED Camera. Available online: https://github.com/stereolabs/zed-ros-wrapper (accessed on 18 June 2022).
  59. Brian ROS OpenPose. Available online: https://github.com/ravijo/ros_openpose (accessed on 7 January 2024).
  60. Azure Kinect DK Hardware Specifications. Available online: https://docs.microsoft.com/en-us/azure/kinect-dk/hardware-specification (accessed on 12 June 2022).
  61. Ramasubramanian, A.K.; Mathew, R.; Preet, I.; Papakostas, N. Review and Application of Edge AI Solutions for Mobile Collaborative Robotic Platforms. Procedia CIRP 2022, 107, 1083–1088. [Google Scholar] [CrossRef]
  62. Jang, M.; Yoon, H.; Lee, S.; Kang, J.; Lee, S. A Comparison and Evaluation of Stereo Matching on Active Stereo Images. Sensors 2022, 22, 3332. [Google Scholar] [CrossRef]
  63. Karakaya, U.B. Algorithms for 3D Data Estimation from Single-Pixel ToF Sensors and Stereo Vision Systems. Master’s Thesis, Università degli studi di Padova, Padua, Italy, 2021. [Google Scholar]
  64. Wnuczko, M.; Singh, K.; Kennedy, J.M. Foreshortening Produces Errors in the Perception of Angles Pictured as on the Ground. Atten. Percept Psychophys 2016, 78, 309–316. [Google Scholar] [CrossRef]
  65. Questionable ZED Accuracy? Available online: https://github.com/stereolabs/zed-examples/issues/44 (accessed on 31 August 2022).
  66. Romeo, L.; Marani, R.; Malosio, M.; Perri, A.G.; D’Orazio, T. Performance Analysis of Body Tracking with the Microsoft Azure Kinect. In Proceedings of the 2021 29th Mediterranean Conference on Control and Automation, MED 2021, Puglia, Italy, 22–25 June 2021; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA; National Research Council of Italy (CNR), Institute of Intelligent Industrial Technologies and Systems for Advanced Manufacturing (STIIMA): Bari, Italy; pp. 572–577. [Google Scholar]
  67. Grunnet-Jepsen, A.; Sweetser, J.N.; Woodfill, J. Tuning Depth Cameras for Best Performance. Available online: https://dev.intelrealsense.com/docs/tuning-depth-cameras-for-best-performance (accessed on 4 July 2022).
  68. Stereolabs Inc. ZED2 Depth Settings. Available online: https://www.stereolabs.com/docs/depth-sensing/depth-settings/#depth-stabilization (accessed on 4 July 2022).
  69. Intel RealSense Depth Camera D455. Available online: https://www.intelrealsense.com/depth-camera-d455/ (accessed on 4 July 2022).
Figure 1. (a) Panoramic view of the laboratory with markings of discreet interval for estimating the pose of the operator at various depths, (b) the poles used for the pose estimation.
Figure 1. (a) Panoramic view of the laboratory with markings of discreet interval for estimating the pose of the operator at various depths, (b) the poles used for the pose estimation.
Sensors 24 00578 g001
Figure 2. (a) Setup of the vision sensors on the desk camera mount, (b) the global reference frame and the frames of the vision sensors as depicted in RViz.
Figure 2. (a) Setup of the vision sensors on the desk camera mount, (b) the global reference frame and the frames of the vision sensors as depicted in RViz.
Sensors 24 00578 g002
Figure 3. Experimental setup and procedure implemented to capture the poses of the operator at different depths.
Figure 3. Experimental setup and procedure implemented to capture the poses of the operator at different depths.
Sensors 24 00578 g003
Figure 4. The skeleton joints with joint numbers shown in Table 3 below that can be tracked by (a) Azure Kinect, (b) ZED2, (c) Intel D455.
Figure 4. The skeleton joints with joint numbers shown in Table 3 below that can be tracked by (a) Azure Kinect, (b) ZED2, (c) Intel D455.
Sensors 24 00578 g004
Figure 5. Transformation frames (TFs) of individual joints information in Rviz from vision sensor. (a) Azure Kinect, (b) Intel D455, (c) ZED2.
Figure 5. Transformation frames (TFs) of individual joints information in Rviz from vision sensor. (a) Azure Kinect, (b) Intel D455, (c) ZED2.
Sensors 24 00578 g005
Figure 6. Illustration of the joints used for the evaluation of the three cameras.
Figure 6. Illustration of the joints used for the evaluation of the three cameras.
Sensors 24 00578 g006
Figure 7. (a) Relation between the height (Z) and depth (Y) of pelvis joint, (b) corresponding skeleton joint data from each camera at a distance level of 3 m (plot view−upper elevated angle).
Figure 7. (a) Relation between the height (Z) and depth (Y) of pelvis joint, (b) corresponding skeleton joint data from each camera at a distance level of 3 m (plot view−upper elevated angle).
Sensors 24 00578 g007
Figure 8. (a) Relation between the height (Z) and depth (Y) of the pelvis joint after the slope compensation, (b) corresponding skeleton joint data from each camera at a distance level of 3 m.
Figure 8. (a) Relation between the height (Z) and depth (Y) of the pelvis joint after the slope compensation, (b) corresponding skeleton joint data from each camera at a distance level of 3 m.
Sensors 24 00578 g008
Figure 9. The unfiltered joint data of the three vision sensors capturing the human skeleton in Pose A: (a) average RMSE values of unfiltered joint data with a 3D skeleton at different depths, (b) the tracked skeleton joints of the operator at the various depth values.
Figure 9. The unfiltered joint data of the three vision sensors capturing the human skeleton in Pose A: (a) average RMSE values of unfiltered joint data with a 3D skeleton at different depths, (b) the tracked skeleton joints of the operator at the various depth values.
Sensors 24 00578 g009
Figure 10. The unfiltered joint data of the three vision sensors capturing the human skeleton in Pose B: (a) average RMSE values of unfiltered joint data with a 3D skeleton at different depths, (b) the tracked skeleton joints of the operator at the various depth values.
Figure 10. The unfiltered joint data of the three vision sensors capturing the human skeleton in Pose B: (a) average RMSE values of unfiltered joint data with a 3D skeleton at different depths, (b) the tracked skeleton joints of the operator at the various depth values.
Sensors 24 00578 g010
Figure 11. Evaluation of RMSE of pelvis joint data in Poses A and B with deviation of pelvis joint along the depth axis from the vision sensors in the range from 1.5 m to 6.0 m.
Figure 11. Evaluation of RMSE of pelvis joint data in Poses A and B with deviation of pelvis joint along the depth axis from the vision sensors in the range from 1.5 m to 6.0 m.
Sensors 24 00578 g011
Figure 12. RMSE of the unfiltered pelvis joint position along the X-axis for the three vision sensors in Pose A.
Figure 12. RMSE of the unfiltered pelvis joint position along the X-axis for the three vision sensors in Pose A.
Sensors 24 00578 g012
Figure 13. RMSE of left wrist joint of the operator in Pose B. (a) Y-axis (depth) data, (b) Z-axis (height) data.
Figure 13. RMSE of left wrist joint of the operator in Pose B. (a) Y-axis (depth) data, (b) Z-axis (height) data.
Sensors 24 00578 g013
Figure 14. RMSE of right wrist joint of the operator in Pose B. (a) Y-axis data, (b) Z-axis data.
Figure 14. RMSE of right wrist joint of the operator in Pose B. (a) Y-axis data, (b) Z-axis data.
Sensors 24 00578 g014
Figure 15. The filtered joint data of the three vision sensors capturing the human skeleton in Pose A: (a) average RMSE values of unfiltered joint data with 3D skeleton at different depths, (b) the tracked skeleton joints of the operator at the various depth values.
Figure 15. The filtered joint data of the three vision sensors capturing the human skeleton in Pose A: (a) average RMSE values of unfiltered joint data with 3D skeleton at different depths, (b) the tracked skeleton joints of the operator at the various depth values.
Sensors 24 00578 g015
Figure 16. The filtered joint data of the three vision sensors capturing the human skeleton in Pose B: (a) average RMSE values of unfiltered joint data with 3D skeleton at different depths, (b) the tracked skeleton joints of the operator at the various depth values.
Figure 16. The filtered joint data of the three vision sensors capturing the human skeleton in Pose B: (a) average RMSE values of unfiltered joint data with 3D skeleton at different depths, (b) the tracked skeleton joints of the operator at the various depth values.
Sensors 24 00578 g016
Figure 17. Evaluation of RMSE of pelvis joint in Poses A and B with deviation of pelvis data along the depth axis from vision sensors in the range from 1.5 m to 6.0 m.
Figure 17. Evaluation of RMSE of pelvis joint in Poses A and B with deviation of pelvis data along the depth axis from vision sensors in the range from 1.5 m to 6.0 m.
Sensors 24 00578 g017
Figure 18. RMSE of the filtered pelvis joint position along the X-axis for the three vision sensors in Pose A.
Figure 18. RMSE of the filtered pelvis joint position along the X-axis for the three vision sensors in Pose A.
Sensors 24 00578 g018
Figure 19. RMSE of left wrist joint of the operator in Pose B (filtered). (a) Y-axis (depth) data, (b) Z-axis (height) data.
Figure 19. RMSE of left wrist joint of the operator in Pose B (filtered). (a) Y-axis (depth) data, (b) Z-axis (height) data.
Sensors 24 00578 g019
Figure 20. RMSE of right wrist joint of the operator in Pose B (filtered). (a) Y-axis (depth) data, (b) Z-axis (height) data.
Figure 20. RMSE of right wrist joint of the operator in Pose B (filtered). (a) Y-axis (depth) data, (b) Z-axis (height) data.
Sensors 24 00578 g020
Figure 21. Differences in filtered and unfiltered data of pelvis joint at 3 m: (a) X-axis data, (b) Y-axis data.
Figure 21. Differences in filtered and unfiltered data of pelvis joint at 3 m: (a) X-axis data, (b) Y-axis data.
Sensors 24 00578 g021
Figure 22. (a) Absolute percent error (δ) of average depth measurement of joints in Pose A, (b) standard deviation (σ) of the average depth data of skeleton joints in Pose A.
Figure 22. (a) Absolute percent error (δ) of average depth measurement of joints in Pose A, (b) standard deviation (σ) of the average depth data of skeleton joints in Pose A.
Sensors 24 00578 g022
Figure 23. Average RMSE of joint depth values of two poses before and after applying moving average filter. (a) Pose A, (b) Pose B.
Figure 23. Average RMSE of joint depth values of two poses before and after applying moving average filter. (a) Pose A, (b) Pose B.
Sensors 24 00578 g023
Figure 24. Average RMSE of joint height values of the wrist data before and after applying moving average filter. (a) Left wrist, (b) right wrist.
Figure 24. Average RMSE of joint height values of the wrist data before and after applying moving average filter. (a) Left wrist, (b) right wrist.
Sensors 24 00578 g024
Figure 25. (a) Classification of zones for HRC tasks using multiple vision-based tracking systems, (b) example of fused output—Zone 1 at 1.8 m.
Figure 25. (a) Classification of zones for HRC tasks using multiple vision-based tracking systems, (b) example of fused output—Zone 1 at 1.8 m.
Sensors 24 00578 g025
Figure 26. (a) RMSE of fused joint depth values in Zone 1 with the corresponding skeleton pose from AK (blue), Intel (black), and combined skeleton (brown), (b) fused and reconstructed skeleton joints.
Figure 26. (a) RMSE of fused joint depth values in Zone 1 with the corresponding skeleton pose from AK (blue), Intel (black), and combined skeleton (brown), (b) fused and reconstructed skeleton joints.
Sensors 24 00578 g026
Figure 27. (a) RMSE of fused joint depth values in Zone 2 with the corresponding skeleton pose from AK (blue), Intel (black), and combined skeleton (brown), (b) fused and reconstructed skeleton joints.
Figure 27. (a) RMSE of fused joint depth values in Zone 2 with the corresponding skeleton pose from AK (blue), Intel (black), and combined skeleton (brown), (b) fused and reconstructed skeleton joints.
Sensors 24 00578 g027
Table 1. Comparison of the depth sensor specifications.
Table 1. Comparison of the depth sensor specifications.
Azure KinectZED2RealSense D455
Released dateJune 2019October 2020October 2020
Price EUR 370EUR 463EUR 432
Depth sensing technologyTime of flightNeural Stereo Depth SensingStereoscopic
Body tracking SDKAzure Kinect Body Tracking SDKZED Body tracking SDKOpenPose v1.7.0 Framework
Field of view (depth image)NFOV unbinned
75° × 65°
110° × 70°87° × 58°
Specified measuring distanceNFOV unbinned
0.5–3.86 m
0.3–20 m0.6–6 m
Table 2. The colour and depth resolution of the cameras used in the experiments.
Table 2. The colour and depth resolution of the cameras used in the experiments.
Azure KinectZED2Intel RealSense D455
SDK Version1.1.03.7.1v2.50.0
Colour resolution640 × 576 @ 30 fps720p @ 30 fps640 × 480 @ 30 fps
Depth resolution/modeNFOV unbinned
640 × 576 @ 30 fps
Ultra640 × 480 @ 30 fps
Table 3. The skeleton joints are tracked from various cameras.
Table 3. The skeleton joints are tracked from various cameras.
Joint No.Azure KinectZED2Intel RealSense D455
0Pelvis ′,″PelvisNose *
1Spine NavalNaval SpineNeck
2Spine ChestChest SpineRight Shoulder ′,″
3NeckNeckRight Elbow ′,″
4Clavicle LeftLeft ClavicleRight Wrist ′,″
5Shoulder Left ′,″Left ShoulderLeft Shoulder ′,″
6Elbow Left ′,″Left ElbowLeft Elbow ′,″
7Wrist Left ′,″Left wristLeft Wrist ′,″
8Hand LeftLeft HandMid Hip (Pelvis) ′,″
9Handtip LeftLeft HandtipRight Hip ′,″
10Thumb Left *Left Thumb *Right Knee ″
11Clavicle RightRight ClavicleRight Ankle ″
12Shoulder Right ′,″Right ShoulderLeft Hip ′,″
13Elbow Right ′,″Right ElbowLeft Knee ″
14Wrist Right ′,″Right WristLeft Ankle ″
15Hand RightRight HandRight Eye
16Handtip RightRight HandtipLeft Eye
17Thumb Right *Right Thumb *Right Ear
18Hip Left ′,″Left HipLeft Ear
19Knee Left ″Left KneeLeft Big Toe
20Ankle Left ″Left AnkleLeft Small Toe *
21Foot Left ″Left FootLeft Heel ″
22Hip Right ′,″Right HipRight Big Toe
23Knee Right ″Right KneeRight Small Toe *
24Ankle Right ″Right AnkleRight Heel ″
25Foot Right ″Right FootBackground *
26HeadHead-
27Nose *Nose *-
28Eye Left *Left Eye *-
29Ear Left *Left Ear *-
30Eye Right *Right Eye *-
31Ear Right *Right Ear *-
32-Left Heel *-
33-Right Heel *-
′ Common joints in Zone 1 fused using Kalman filter; ″ common joints in Zone 2 fused using Kalman filter; * joints excluded from the overall experiment as they do not affect the skeleton pose.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ramasubramanian, A.K.; Kazasidis, M.; Fay, B.; Papakostas, N. On the Evaluation of Diverse Vision Systems towards Detecting Human Pose in Collaborative Robot Applications. Sensors 2024, 24, 578. https://doi.org/10.3390/s24020578

AMA Style

Ramasubramanian AK, Kazasidis M, Fay B, Papakostas N. On the Evaluation of Diverse Vision Systems towards Detecting Human Pose in Collaborative Robot Applications. Sensors. 2024; 24(2):578. https://doi.org/10.3390/s24020578

Chicago/Turabian Style

Ramasubramanian, Aswin K., Marios Kazasidis, Barry Fay, and Nikolaos Papakostas. 2024. "On the Evaluation of Diverse Vision Systems towards Detecting Human Pose in Collaborative Robot Applications" Sensors 24, no. 2: 578. https://doi.org/10.3390/s24020578

APA Style

Ramasubramanian, A. K., Kazasidis, M., Fay, B., & Papakostas, N. (2024). On the Evaluation of Diverse Vision Systems towards Detecting Human Pose in Collaborative Robot Applications. Sensors, 24(2), 578. https://doi.org/10.3390/s24020578

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop