Next Article in Journal
Spectral Efficiency Maximization for Mixed-Structure Cognitive Radio Hybrid Wideband Millimeter-Wave Transceivers in Relay-Assisted Multi-User Multiple-Input Multiple-Output Systems
Previous Article in Journal
Preliminary Experience with Three Alternative Motion Sensors for 0.55 Tesla MR Imaging
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Video Abnormal Behavior Recognition and Trajectory Prediction Based on Lightweight Skeleton Feature Extraction

by
Ling Wang
1,
Cong Ding
1,
Yifan Zhang
1,
Tie Hua Zhou
1,*,
Wei Ding
2,3,
Keun Ho Ryu
4,5,6 and
Kwang Woo Nam
7
1
Department of Computer Science and Technology, School of Computer Science, Northeast Electric Power University, Jilin 132013, China
2
Key Laboratory of Computing Power Network and Information Security, Ministry of Education, Shandong Computer Science Center (National Supercomputer Center in Jinan), Qilu University of Technology (Shandong Academy of Sciences), Jinan 250000, China
3
Shandong Provincial Key Laboratory of Computer Networks, Shandong Fundamental Research Center for Computer Science, Jinan 250000, China
4
Data Science Laboratory, Faculty of Information Technology, Ton Duc Thang University, Ho Chi Minh City 700000, Vietnam
5
Research Institute, Bigsun System Co., Ltd., Seoul 06266, Republic of Korea
6
Database and Bioinformatics Laboratory, College of Electrical and Computer Engineering, Chungbuk National University, Cheongju 28644, Republic of Korea
7
Department of Computer and Information Engineering, Kunsan National University, Gunsan 54150, Republic of Korea
*
Author to whom correspondence should be addressed.
Submission received: 9 May 2024 / Revised: 4 June 2024 / Accepted: 6 June 2024 / Published: 7 June 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Video action recognition based on skeleton nodes is a highlighted issue in the computer vision field. In real application scenarios, the large number of skeleton nodes and behavior occlusion problems between individuals seriously affect recognition speed and accuracy. Therefore, we proposed a lightweight multi-stream feature cross-fusion (L-MSFCF) model to recognize abnormal behaviors such as fighting, vicious kicking, climbing over the wall, et al., which could obviously improve recognition speed based on lightweight skeleton node calculation, and improve recognition accuracy based on occluded skeleton node prediction analysis in order to effectively solve the behavior occlusion problem. The experiments show that our proposed All-MSFCF model has a video action recognition average accuracy rate of 92.7% for eight kinds of abnormal behavior recognition. Although our proposed lightweight L-MSFCF model has an 87.3% average accuracy rate, its average recognition speed is 62.7% higher than the full-skeleton recognition model, which is more suitable for solving real-time tracing problems. Moreover, our proposed Trajectory Prediction Tracking (TPT) model could real-time predict the moving positions based on the dynamically selected core skeleton node calculation, especially for the short-term prediction within 15 frames and 30 frames that have lower average loss errors.

1. Introduction

Video surveillance technology is widely used in daily life for public safety management. Abnormal behavior recognition and tracking technology, as an important video surveillance application technology, has a great deal of research significance and practical value. Abnormal behavior has different definitions in different scenarios, and in our research the abnormal behavior studied is based on the field of public security. According to the definition of “behavior” and “abnormal”, the abnormal behavior classification is clarified. “Behavior” [1] refers to the most basic and meaningful interactions with their surrounding environment. “Abnormal” [2] refers to phenomena that are different from the normal state. Therefore, the following definition of “abnormal behavior” is given: all actions, gestures, or events made in the current scene that are not suitable.
Abnormal behavior recognition and tracking technology can improve the efficiency and accuracy of video surveillance, reduce operators’ workloads, and detect and handle abnormal behavior events early. Abnormal recognition and tracking based on skeleton nodes is one of the important methods in current research. The detection effect of skeleton nodes is continuously improved, and it is increasingly becoming one of the core technologies of intelligent video surveillance. However, in real application scenarios, the large number of skeleton nodes or the occlusion between individuals seriously affects the abnormal behavior recognition speed and accuracy, and limits the application of abnormal behavior recognition and tracking algorithms. Therefore, we propose the L S F E model. Before feature extraction, for the problem of occluded skeleton nodes information, the L- M S F C F model is proposed to utilize the skeleton nodes information of past frames to predict the occluded skeleton node information, thus improving accuracy. When tracking abnormal targets, in order to address the issue of targets becoming occluded or disappearing from view, we suggest the T P T model.
The structure of this paper is as follows: The history of the current quest for abnormal behavior recognition is provided in Section 1, and the current status of the field is introduced in Section 2. The study’s purpose, materials, and procedures are all presented in Section 3. The L S F E approach is described in full in Section 4. A comprehensive explanation of the L- M S F C F model is given in Section 5. The T P T model is explained in depth in Section 6. The findings and analysis of the experiment are presented in Section 7. A thorough discussion is given in Section 8, and the study is concluded in Section 9.

2. Related Work

Computer vision techniques are gradually becoming the mainstream of abnormal behavior recognition, and the main challenge is to accurately extract and analyze representative appearance features and dynamic motions. In the early stages of the study, researchers typically thought of objects as particles. By simulating the tension in every pixel, Mehran R et al. [3] created a particle flow network to extract the interaction of force as features from video data. In order to capture the space-time properties of a crowd, new global features were proposed by Xie S et al. [4] to describe the position, speed, and direction of particles. Furthermore, Yu B et al. [5] enhanced the representation capacity of the particles by utilizing several comparable particles to describe objects. However, these feature-extraction techniques were unable to recover the subtle aspects of motions. In order to better capture motion information, numerous researchers have turned to feature extraction from space-time cubes, whereas Sabokrou M et al. [6] considered the sub-region of continuous frames as a space-time cube and extracted 3D gradient characteristics for cubes, and Fayyaz M et al. [7] collected global features from space-time cubes using an automatic encoder. Martinel N et al. [8] extracted deep features by rebuilding the interesting cubes using stacked sparse automatic encoders. Since features extracted from space-temporal cubes do not maintain the correlation of motion features between the cubes, Coşar S et al. [9] learned velocity and trajectory from real tracking data pixels and clustered the trajectories using a clustering tree to predict the most probable paths of the tracked objects. Xu M et al. [10] captured both groups and personal trajectories at the same time, and performed separate abnormal behavior detection.
Target motion trajectory prediction algorithms are vital to research in computer vision and robotics areas. The purpose is to forecast a target’s future motion trajectory utilizing existing motion data and environmental information so that a robot or other intelligent system can react appropriately. Kerdvibulvech C et al. [11,12] proposed a method for 3D human motion analysis for reconstruction and recognition. They used 3D gait signatures computed from 3D data that are obtained from a triangulation-based projector–camera system. Results demonstrated that the proposed 3D gait signatures-based biometrics provide valid results on real-world 3D data. Combining trajectory prediction due to maneuver recognition with trajectory prediction owing to constant slew rate and acceleration motion models was conducted by Houenou A et al. [13] and Czyz J et al. [14] proposed a mixed-value sequence state estimation algorithm. Shao X et al. [15] presented a unique filtering technique, which is to follow a target’s mobility utilizing GPS sensors. Vashishtha D et al. [16] and Kapania S et al. [17] improved particle filtering, combined color sequences, and constrained Bayesian state estimation to achieve motion trajectory prediction of the target. Choi D et al. [18] proposed a method using maximum likelihood multi-filter to obtain an overall estimate to predict the target trajectory by combining independent multiple kinematic model correlation estimates through a great likelihood rule. Predicting target trajectories by building kinematic and kinetic models will not affect the accuracy even if losing a large part of the data, but part of the target motion trajectory is nonlinear and prone to many curvilinear trajectories, which means that the model-based trajectory prediction algorithms will have the problem of low accuracy. Another important model is based on data-driven trajectory prediction studies [19]. The model uses both classification and regression algorithms to treat the trajectory prediction issues. Semwal et al. [20] suggested a target trajectory prediction technique, long short-term memory networks ( L S T M s ), and convolutional neural networks ( C N N s ). Deep neural network by Shirazi M S et al. [21], Faster R-CNN by Zhou H et al. [22], and YOLO network by Yoon Y C et al. [23] also have good performance in target trajectory prediction.
To solve the occlusion problem, Sabokrou M et al. [24] first applied fully convolutional neural networks to abnormal behavior detection. They completed it by utilizing AlexNet’s fully convolutional layer to extract deep features and by cascading Gaussian classifiers to identify abnormal behaviors. Chu W et al. [25] extracted temporal characteristics using 3D convolutional neural networks. Liqian Yan [26] suggested a 3D convolutional residual network structure in light of this. In order to lessen the effect of the network, Fang Z et al. [27] defined the motion characteristics of the footage using a visual system to define spatial features and the multi-scale histogram of the optical flow. Ye O et al. [28] extracted initial features through the C N N - L S T M network and used feature expectation subgraph to filter unexpected feature values. The values of the remaining predicted features were sent into S V M to detect abnormal behavior, and Tay N C et al. [29] created a shallow convolutional neural network to extract appearance characteristics, added spatial attention, and integrated it with a L S T M network.
In previous conclusions, although the target trajectory prediction and occlusion problems can be solved effectively, the accuracy and time complexity need to be improved. Therefore, we need further research on abnormal behavior recognition.

3. Materials and Methods

3.1. Motivation

In our research, lightweighting the skeleton is an effective way to cope with the effects of an excessive number of skeleton nodes. In addition, we found that the multi-stream feature cross-fusion method has significant advantages in feature extraction. Therefore, the flowchart of abnormal behavior recognition and tracking is demonstrated in Figure 1.

3.2. Datasets

The experiment utilized the human3.6m dataset, comprising 3.6 million 3D human posture examples and their related photos. These data were collected from six males and five females across 17 diverse scenes such as discussions, smoking, taking photos, and more. The video captures were from four calibrated cameras capable of capturing precise 3D joint positions and joint angles. For more information about the human 3.6 m dataset, see [30]. The UCF-Crime dataset [31], a vast collection of actual surveillance footage containing 1900 long, unedited recordings with 13 distinct kinds of abnormal events, was also used in the studies. Furthermore, the ShanghaiTech Campus dataset [32] was employed. It included over 270,000 training frames and 130 occurrences of abnormal events.

3.3. Methods

This paper is a study of abnormal behavior recognition and tracking in surveillance videos. First, the skeleton nodes of various behaviors are lightened by the L S F E method and then construct the optimal skeleton node architecture graphs for various behaviors. Second, they construct the L- M S F C F model for abnormal behavior recognition; after predicting the information of the occluded skeleton nodes, it takes the lightweight feature skeleton to coordinate information and the skeleton vectors as the dual-stream inputs and uses the cross-feature fusion to carry out the feature extraction. Finally, the T P T model is proposed for trajectory prediction. It provides a reference for the tracking of abnormal behavior targets.

4. Skeleton Feature Extraction

We proposed a lightweight skeleton feature extraction ( L S F E ) method to solve the problem of a large number of skeleton nodes. Firstly, we design the adaptive computation of the video frame window and find out the optimal video frame window length for optimizing the skeleton nodes; secondly, we design the formula for calculating the skeleton nodes of the proposed trigonometry and triangulated all the skeleton nodes, and find out the motion law of the individual behavioral movement process by using the association rule mining algorithm under the optimal video frame window length; finally, we find out the skeleton nodes that can represent the action by data mining and filter out the redundant skeleton nodes, so as to achieve the purpose of skeleton nodes optimization.

4.1. Data Preprocessing

The definition of behavior in this paper divides behavior into two categories: normal behavior and abnormal behavior. We constructs a normal behavior video database and an abnormal behavior video database. The definitions are as follows (Table 1).
Before action recognition based on skeleton joint point, it is necessary to convert the original video data into skeleton joint data, the space representation of the action is detected and recognized in the original video data. Through the existing posture evaluation algorithm, the video data can be transformed into corresponding skeleton joint data and the skeleton corresponding to each number in Figure 2 is represented as follows: 0—nose, 1—neck, 2-right shoulder, 3—right elbow, 4—right hand, 5—left shoulder, 6—left elbow, 7—left hand, 8—right hip, 9—right knee, 10—right foot, 11—left hip, 12—left knee, 13—left foot, 14—stomach, and 15—head.
Due to the different angular positions of the sportsman relative to the camera, resulting in possible differences in the coordinate origin, to facilitate the study, we do a harmonized coordinate transformation for the skeleton data. We reconstruct the coordinate with a triangle formed by the three points v 1 , v 2 , and v 3 in Figure 3.
v 1 , v 2 , v 3 in space constitute a triangle and its three sides are l 1 , l 2 , l 3 , v t , set as the projection point on the line l 3 . Through Equation (1) we can obtain three basis vectors by transforming the coordinate.
U t = v 3 v t v 1 v t v 3 v t v 1 v t , v 3 v t v 3 v t , v 1 v t v 1 v t
v 1 , v 2 , v 3 —the three skeleton joints in Figure 3;
v t —the projection point on the line l 3 ;
U t —the three basis vectors of the transformed coordinate.
The conversion process of the coordinates also needs the three basis vectors of the original coordinate. It is represented in Equation (2).
U 0 = 1 0 0 0 1 0 0 0 1
Using Equation (3), the three basis vectors of the original coordinate and the three basis vectors of the transformed coordinate are operated to obtain the corresponding transformation matrix.
R = U t 1 U 0
U t 1 —the inverse matrix of U t .
Using the transformation matrix, the original coordinates are transformed by Equation (4).
v 1 = R v 1 0 1 v 1
v—a skeleton node in the original coordinate;
v —the corresponding transformation node;
R—transformation matrix;
v 1 —new coordinate origin.
During the transformation, we designate the origin of the new coordinate as v 1 in the existing coordinate; moreover, all 3D skeleton coordinates are transformed into the new skeleton data with v 1 as the origin by the above equation.

4.2. Adaptive Sliding Window Selection Calculation

Activities are characterized by continuity and periodicity. However, the length of this cycle cannot be determined; therefore, this paper utilizes adaptive sliding window selection calculation to determine the cycle length that meets the requirements. First, we apply the method to the segmentation of action sequences. For the action sequence a = a 1 , a 2 , a n , we set the window width to T and step size to K, and each window contains R i = r i 1 , r i 2 r i T . In this way, the initial action sequence is divided into F T action segments, which can be represented as R = A 1 , A 2 A F T . Each segment contains K poses describing the local information of the body. Figure 4 shows the complete process of segmenting an action sequence by adaptive sliding window selection calculation and the original video frames are cited by the Human3.6m dataset [30].
The window width parameter T determines the size and the number of segments that can be segmented in an action sequence. A larger K means that each segment contains more poses and a coarser description of the movement; on the contrary, a smaller K means that each segment contains fewer poses and a more accurate description of the movement. Although smaller K describes movements more accurately, this means that smaller segments are more susceptible to noise in the 3D skeleton position tracking results, which in turn affects the recognition of movements. Defining the set of stored window sizes as L, we calculate each action accuracy in the window size from 3 to 23. The top three action accuracies are stored in the set L; we can obtain the calculation of all the actions and then take the size of the window with the highest number of occurrences as the window length. Finally, the most frequent occurrence is 15. Therefore, the window length of T = 15 is taken as a basic action sequence in this paper.

4.3. Lightweight Skeleton Feature Extraction Method (LSFE Method)

We put forward the lightweight skeleton feature extraction ( L S F E ) method. The method is based on the association rule mining of similar vectors. It converts the 3D skeleton data into a series of vectors with a length of 15 frames by using adaptive sliding window selection calculation and then utilizes vector similarity to mine the similarity association rule set of each node. If there exists a similarity association rule set, the skeleton node is considered to be a strongly associated skeleton node of the action. The computation is as follows:
Step 1: Obtain any one of its skeleton nodes v i based on the original skeleton node data V. Define the node data of the last two frames as v i + 1 , v i + 2 . Define the 3D coordinates of v i , v i + 1 , v i + 2 as x i y i z i , x i + 1 y i + 1 z i + 1 , x i + 2 y i + 2 z i + 2 .
Step 2: Calculate the angle change in the skeleton nodes v i , v i + 1 , v i + 2 and the three points in the time dimension according to Equation (5).
α = tan 1 x i x i + 2 z i z i + 2 β = tan 1 x i + 1 x i + 2 z i + 1 z i + 2 θ = tan 1 x i x i + 1 y i y i + 1
α , β , θ —the value of the angle change;
The plane ρ can be obtained through the skeleton nodes v i , v i + 1 , v i + 2 , which are represented as shown in Equation (6).
ρ : a x + b y + c z + d = 0
a, b, c, d—the plane equation parameters;
It is clear that the normal vector n = a , b , c of the plane ρ can be obtained, then the distance R from the origin of the space coordinates to this plane is obtained through Equation (7).
R = d a 2 + b 2 + c 2
After obtaining the height R of the proposed triangular pyramid, the volume X i is then calculated according to Equation (8).
X i = 1 3 × 2 × R × sin α 2 × sin β 2 × sin θ 2
α , β , θ —the three included angles of the proposed triangle pyramid;
R—height of the proposed triangular pyramid;
X i —volume of the proposed triangular pyramid.
According to Equation (8), we can find the set of vectorized data X j , M i = X i for a certain skeleton node v i of a certain action single video M j , then the vectorized dataset for all the data of a certain skeleton node of the action is denoted as Equation (9).
M i = { { X 1 , M 1 } , { X 2 , M 2 } { X j 1 , M i 1 } , { X j , M i } }
X j , M i —the vectorized dataset of the ith skeleton node of the jth video data under the action classification;
M i —a vectorized set of all data for a certain skeleton node.
Step 3: Construct the frequent item set. Scan all the X i data in the M i set in a single pass to determine the support of each X i . Since X i is a vector datum, it utilizes a calculation rule of similar vectors: if two vectors are similar vectors then their frequency adds one. The similar vectors are shows in Equation (10).
cos δ = Σ t = 1 n { X a } { X b } Σ t = 1 n X a 2 t = 1 n X b 2
X a —data a of a certain skeleton node in the frequent item set;
X b —data b of a skeleton node in the frequent item set;
cos δ —similarity of vectors X a and vectors X b .
X a and X b are two vectors of the same length, cos δ is between 0 and 1. When cos δ > 0.9, the vector X a is considered to be similar to the vector X b .
Step 4: Mining association rule sets. Define the association rule set of an individual’s behavior as a set in the form of key-value pairs J = { ( v i : Y i ) | v i V , Y i } . Mine the 16 skeleton nodes of an individual to obtain the set of association rules L for a single node v i , and L is added to Y i . When Y i = , ( v i : Y i ) , ( v i ; Y i ) will be stored in the association rule set J as key-value pairs; Y i = means that the current skeleton nodes do not have an obvious regularity; they cannot represent the behavior action and should be discarded. Define the final association rule set J = { ( v i : Y i ) | v i V , Y i } , the set of all skeleton nodes in the set J is v i , v i is a non-empty subset of V.
Step 5: Determine lightweight skeleton nodes. The maximum frequent item set in J is n. When n > μ , then v i is considered as the skeleton node of the current action. Finally, calculate all feature skeleton nodes.
The highest accuracy of skeleton node recognition is when μ = 3 . The extracted lightweight skeleton nodes for each action are shown in Table 2.
Step 6: Based on the above lightweight skeleton nodes, construct the L S F E model to recognize actions and verify the feasibility of lightweight skeleton nodes.

5. Lightweight Multi-Stream Features Cross-Fusion Model (L-MSFCF Model)

Lightweight feature skeleton node extraction is a core processing step for supporting the L- M S F C F model, which could help to greatly reduce the model parameter numbers and computation time than full skeleton processing. In fact, as shown in the experiment Section 7.2 (Section 4 L S F E model testing), only considering the optimized skeleton nodes to recognize the video behaviors, the accuracy is not ideal. In order to improve the recognition accuracy and further reduce the computation time, our proposed L- M S F C F model has enhanced lightweight features based on a multi-stream feature cross-fusion process in order to obtain more behavior feature information.

5.1. L-MSFCF Model Abnormal Behavior Recognition Process

The L- M S F C F model is different from the traditional multi-stream feature fusion action recognition method. The L- M S F C F model processes the occluded skeleton nodes and also utilizes the feature cross-fusion extraction method. Firstly, lightweight the skeleton nodes. Secondly, predict the occluded skeleton node information by utilizing the skeleton node data information of past frames. Finally, obtain action features through the skeleton stream, nodes stream, and feature cross-fusion stream. The L- M S F C F model strengthens the recognition ability of abnormal behaviors.
The L- M S F C F model abnormal behavior recognition process mainly has two steps: the first step is occluded skeleton nodes prediction and lightweight processing; the second part is lightweight skeleton data feature extraction through dual-stream, and then feature fusion is performed on all the features to finally obtain the classification results. Figure 5 shows the flowchart of the L- M S F C F model. The following are the steps:
Step 1: Preprocessing the skeleton data. Create a skeleton joint dataset and a skeleton vector dataset. Because skeleton vectors are composed of two skeleton nodes and the whole skeleton point graph is not a ring structure, this results in the number of skeleton vectors always being less than the number of skeleton nodes in the generation process by 1. We add an empty skeleton with the value of 0 to skeleton vectors so that there are as many skeleton nodes.
Step 2: Lightweighting the skeletons. Lightweight skeleton data are based on lightweight characteristic skeleton nodes for each action. Skeleton node data and skeleton vector data are processed similarly, taking skeleton node data processing as an example. The process is as follows:
According to Table 2, retain the corresponding characteristic skeleton node information and set other skeleton information to 0. Take fighting as an example; its original skeleton data of a certain frame is expressed as Equation (11).
v t i = [ g 1 , g 2 , g 3 , g 4 , g 5 , g 6 , g 7 , g 8 , g 9 , g 10 , g 11 , g 12 , g 13 ]
V t i —the original skeleton dataset of a frame;
g—a skeleton node in the current frame.
The lightweight skeleton nodes for the fighting in Table 2 are [3, 4, 6, 7, 9, 10, 12, 13], and the result of lightweight processing is shown in Equation (12).
v t i = [ 0 , 0 , g 3 , g 4 , 0 , g 6 , g 7 , 0 , g 9 , g 10 , 0 , g 12 , g 13 ]
v t i —skeleton dataset after lightweight processing;
g—a skeleton node in the current frame.
Step 3: Determine whether the lightweight skeleton node data are occluded or not; if the space coordinates of this skeleton node data are all 0, it is determined that this skeleton node data are occluded, then the skeleton node data are predicted.
Step 4: Process the skeleton node data and skeleton vector data separately by convolution to obtain features that can represent each action.
Step 5: Combine the skeleton node features and skeleton vector features to form the overall action features utilizing feature fusion.

5.2. Occluded Skeleton Node Prediction

Occluded skeleton nodes can cause noise to the abnormal behavior recognition, affecting the accuracy. To solve the problem, we suggest a generative network-based method for occluded skeleton node prediction, which utilizes the skeleton node data from past frames to predict the skeleton node information of the next frame.
The advantages over existing methods are: the G R U at the lowest level can learn the motion information of the smallest unit frame without interference from higher levels, and the higher levels can capture different features of the motion of specific length frames; moreover, the latest G R U outputs from different levels are used as inputs during the prediction period of each time step, which makes the motion information more adequate and the features of the next frame more comprehensive.
In Figure 6, the skeleton data information of the previous, the current, and the future frame is represented by the vectors e t + 1 , e t , e t + 1 . The expected skeleton data information at moments t and t + 1 is e t and e t + 1 . The skeleton data information for every time step is used as a series of input G R U units at the first layer. Define K distinct G R U unit sequences at the second level, each of which will only accept similar inputs from the first level’s G R U units that have been time-step-modeled. If K = 2 , for instance, the second layer would contain two G R U sequences: the first would be derived from time frame data t = { 1 , 3 , 5 , } , whereas the second would come from time frame data t = { 2 , 4 , 6 , } . G R U s at the same hierarchical level share weights, improving the characteristics of the skeleton data to improve long-term dependent learning. There are a total of K 2 G R U sequences on the third layer because for every K G R U sequence on the second layer, there are K different G R U sequences corresponding to it in the third layer. Each G R U sequence uses the same complex modulo K inputs from it. Up to level ( M 1 ) , where a G R U sequence of K M 1 will exist in level M, the hierarchy’s process of creating new, higher-level G R U sequences continues. In order to produce skeleton vector predictions for the associated hidden units in all hierarchies, a two-layer connected network is finally introduced. The inputs for these projected skeleton vectors will then contribute to the skeleton vector prediction process for upcoming frames.

5.3. Lightweight Multi-Stream Feature Cross-Fusion Process

Behavior recognition method networks with multi-stream feature fusion, such as dual-stream networks, 2 s - A G C N [33], typically utilize single-stream networks to extract characteristics independently before fusing them. The feature fusion method performs weight fusion at the end, and the average pooling layer will overrun the fusion step, making the network unable to fully perform each dependent feature.To solve this question, this subsection proposes a L- M S F C F model, which performs feature cross-fusion during pooling to fully utilize each tributary feature. There are two parts to introduce the model: the network architecture and the basic convolution module.
The whole network of L- M S F C F consists of three sub-stream networks: skeleton vector stream network, skeleton joint stream network, and features cross-fusion stream network. Each sub-stream network utilizes the 2 s - A G C N graph convolution network as the backbone. Either joints or skeletons can be used as input data. Formally, the skeleton sequence data are V = R C × T × S , D ϵ C × T × S , and C, T, S separately denote channel dimension, time dimension, and space dimension. Space characteristics may be extracted from the input data via the spatial stream network. Shallow sub-networks have a lot of inaccurate and localized data in their features. Conversely, features located in the network’s deeper levels have less false information and more global information. Many conventional networks are bottom-up and end-to-end systems that only employ a subset of top-layer characteristics. These methods lack local information that facilitates action recognition classification. For this reason, the network proposed in this paper selects features from multiple layers. The features extracted from different levels have different feelings and contain various local and global information.
The whole process of feature fusion is as follows:
Step 1: Mark the skeleton vector features collected from the skeleton vector stream network, denoted as f b v 1 , f b v 2 , f b v 3 , , f b v L . The skeleton joint stream network is almost identical to the skeleton stream network, where the extracted features are, respectively, denoted as f b n 1 , f b n 2 , f b n 3 , , f b n L . L is the maximum layer of elements. In the experiment part, we set L to 3.
Step 2: Calculate the weights of the skeleton vector stream network and the skeleton joint stream network. The skeleton vector stream network N b v ( D ) and the skeleton joint stream network N b n ( D ) are represented as shown in Equations (13) and (14).
N b v ( D ) = ( f b v 1 , f b v 2 , , f b v L ) · p b v
N b n ( D ) = ( f b n 1 , f b n 2 , , f b n L ) · p b n
N b v ( D ) —skeleton vector stream network;
N b n ( D ) —skeleton joint stream network;
P b v —skeleton vector stream network weight;
P b n —skeleton joint stream network weight.
Step 3: The fusion stream network inputs the features collected from the basic dual-stream network, and the weights of the fusion stream network are calculated. As an example, for the case where L is 3, the fusion stream network is represented as shown in Equation (15).
N f u s ( f b v 1 , f b v 2 , f b v 3 , f b n 1 , f b n 2 , f b n 3 ) = p f u s
N f u s —fusion stream network;
p f u s —fusion stream network weight.
Step 4: Use weighted average fusion function w ( · ) to compute the prediction weight of the whole network.
w ( p b v , p b n , p f u s ) = α p b v + β p b n + γ p f u s α + β + γ
α , β , γ —weighted average fusion function fixed weight parameters.
Step 5: The feature data of the three tributaries are fused in the fusion layer by weighted average fusion, and finally in the fully connected layer by S o f t m a x function. Fuse all the information to finally output a feature that can represent the whole action.
The convolution module’s goal is to extract deep features. This paper utilizes an adaptive graph convolutional network, and the advantage is that the whole process is a bottleneck structure, aiding in first reducing noise and then obtaining extremely effective information. Its specific structure is shown in Figure 7.
The entire convolutional block can be represented as:
f o u t = k K v W k f i n ( A k + δ C k )
f i n —input features;
f o u t —output features;
K v —kernel size in space dimensions;
W k —1 × 1 convolution operation;
A k —N × N adjacency matrix, its elements indicate whether a vertex is in a subset of another vertex;
δ —weighting parameter;
B k —data-driven matrix.
Throughout the computation, we set the K v -space dimension’s kernel size to 3. A k = ρ A ^ k ρ , and A k is N × N adjacency matrix whose elements indicate whether the weak feature skeleton nodes are in the subset of lightweight feature skeleton nodes or not. ρ is the normalized diagonal matrix, ρ k = Σ j A ^ k + σ . δ is set to 0.001 to avoid blank lines. W k denotes a 1 × 1 convolution operation with weights in the shape of B k × B k × 1 × 1 . B k is a data-driven matrix in the shape of N × N. B k × B k is a non-local block that goes through the computation of Figure 8 once before participating in a second computation. The value of δ directly determines the impact of B k on the quadratic convolution. In the experiment, we set δ = 0.3 to obtain high-level valid information, and if the parameters and elements in the matrix were not initialized, its value was set to 0.01.

6. Trajectory Prediction Tracking Model (TPT Model)

6.1. Five-Bit Skeleton Screening Method

To lower the entire model’s time complexity and reduce the impact of skeleton node occlusion on trajectory prediction, this paper proposes a five-bit screening method. First, the skeleton nodes are divided into five parts, A, B, C, D, and E, and their partitions are in Figure 9. Then, the skeleton nodes in each partition are sorted, utilizing lightweight feature skeleton node extraction results in Table 2. After that, select the feature skeleton nodes that can represent each partition. Finally, find out the mass point that is based on the five featured skeleton data, regard it as the starting point for trajectory prediction tracking.
Calculate the probability of their occurrence based on feature skeleton node extraction results in Table 2, and sort the skeleton nodes in each partition by top; the results are as follows:
We select the characteristic points in each partition and then denote them as e A , e B , e C , e D , e E . Form a pentagon e x by the five points. The calculation is in Equation (18).
v x = 1 N v Σ 1 N v v i
N v —vertices number;
v i —vertices space coordinates;
v x —mass point space coordinates.
In most cases, we can essentially detect complete skeleton nodes. However, in some cases, some are not detected. For example, the skeleton node 10 is occluded in the partition D of Figure 9. To solve this problem, we can select the top-ordered skeleton nodes of the partition in turn, and according to Table 3, node 9 should be selected as the representative node of the partition. When a partition is occluded, set it to the same position detected in the previous frame.
Considering that trajectory is a vector with velocity and direction, this paper calculates the change of direction and velocity of the mass point e x for each frame. The velocity of the mass point e x at T frames is represented in Equation (19).
V v x = ( v x v x 1 ) 2
v x —mass point e x space coordinates at frame T;
v x 1 —mass point e x space coordinates at frame T 1 .
Define the space coordinates at T 2 , T 1 , and T frames as v x 2 , v x 1 , v x , the angle of the mass point at T frames is the angle between the vector v x 2 v x 1 and v x 1 v x , as shown in Figure 10. The cosine function of this angle is expressed as Equation (20). In the experiment part, the angle of frame 1 and frame 2 is generally set to 0.
cos θ = v x 2 v x 1 . v x 1 v x | v x 2 v x 1 | | v x 1 v x | .
In addition, we need to calculate the absolute velocities of e A , e B , e C , e D , and e E ; their computation is comparable to the e x velocity calculation. Once all the above data are calculated, we track the trajectory of e x and extract its position and motion characteristics to make a prediction of the next trajectory. Our inputs include the position information and absolute velocity of e A , e B , e C , e D , and e E and the angle change of e x .

6.2. TPT Modeling Architecture

The T P T model is autoregressive, the model predicts the future frames’ trajectory by taking as inputs the previous cyclic state as well as features describing the earlier trajectory at each time step. The entire model forecasts the trajectory’s state in the upcoming K frames using the current frame data as input. The T P T network model consists of two G R U layers, each containing 1000 hidden units and a linear activation function Linear. Mul denotes the product of the two matrices. The purpose of regularization and normalization is to avoid overfitting the model and reduce the algorithm generalization error. Finally, after the Sigmoid activation, the mass points prediction coordinates are generated (Figure 11).

7. Experiment and Results

The experiments were conducted on a Windows 10 system with an Intel(R) xeon(R) E5-2640 v4 @ 2.40 GHz processor with 32G RAM and the graphics card was an NVIDIA GeForce RTX 2080Ti. The codes were written in Python 3.7, and the entire training and testing were conducted on the PyCharm.

7.1. Datasets

The experiment had three parts. Firstly, we compared the L S F E model and R N N model time complexity and accuracy. To confirm the reasonableness of the L- M S F C F model, the A l l - M S F C F model and 2 s - A G C N were compared. Lastly, we compared the T P T model with other models regarding the number of parameters and the final average loss error prediction.
The datasets we chose in this experiment were the human3.6m dataset [30], the UCF-Crime dataset [31], and the ShanghaiTech Campus dataset [32], as detailed in Section 3.2. Based on these datasets, we organized eight categories: walking, running, stooping, fighting, vicious kicking, climbing over walls, throwing suspicious objects, and slashing devices. Each video is between 0 and 10 s and the video format is avi. The entire dataset contains 3146 videos and we selected 314 videos as the test set and the rest as the training set.

7.2. LSFE Model Test

The purpose of lightweight skeleton nodes was to increase the action recognition speed, so we compared time complexity, as shown in Figure 12. Taking 15 frames as a recognition unit, the outcomes supported that the L S F E model’s time complexity was less than the R N N model’s [34]. The results show that the time complexity of L S F E is significantly better than that of the R N N model, and the average recognition speed is about 86.5% higher.
In order to further validate the effectiveness of the feature skeleton nodes, this paper compares the recognition accuracy rate between the L S F E model and the R N N model (Figure 13).
The results show that the accuracy of the R N N model is higher than that of the L S F E model for all eight actions. In Table 4, we compared the average accuracy and time in detail. Although the average accuracy of the L S F E model is 4.5% lower than that of the R N N model, its average recognition speed is 86.5% higher. Thus, there was merit in abnormal behavior recognition based on lightweight skeleton nodes.

7.3. L-MSFCF Model Test

The L- M S F C F model had a good performance in accuracy and loss value after training, the loss rate of the L- M S F C F model is displayed in Figure 14, and the accuracy rate is shown in Figure 15.
From Figure 14, the initial loss of the L- M S F C F model was as high as around 2.2 during the neural network’s training phase. However, it decreased rapidly after 500 iterations, and the convergence gradually slowed down after about 500 iterations, and the final loss value reached a smaller value. It indicated that the L- M S F C F model had a good learning effect.
From Figure 15, in the first 1000 iterations, the accuracy curve of L- M S F C F converged rapidly, and then after 1500 iterations the accuracy basically stayed stable at a value of about 0.84 (Figure 16).
In addition, we assessed the algorithm’s performance utilizing a confusion matrix. The vertical coordinate represented the real value, whereas the horizontal coordinate indicated the projected value. The diagonal element was the percentage of the projected value to the true value.
The confusion matrix diagram demonstrated that running and walking had higher similarity among normal behaviors, probably because of the higher similarity of the lightweight skeleton nodes in their recognition process. Meanwhile, fighting and running had high similarities in abnormal behavior. Overall, normal behavior was more accurately recognized than abnormal behavior, and the recognition accuracy is 87.3%, so the overall recognition effect basically meets the expectation.
The L- M S F C F model took lightweight skeleton nodes as inputs; we regarded the model that had the same network architecture but took full skeleton nodes as the A l l - M S F C F model.
We compared the L- M S F C F model, the A l l - M S F C F model, the 2 s - A G C N model [33], and the L S F E model. Table 5 presented the findings. The results demonstrated that when we input lightweight skeleton nodes, the L- M S F C F model’s recognition accuracy outperformed both the 2 s - A G C N model and the L S F E model by a large margin. However, the L- M S F C F model’s accuracy was lower than that of the A l l - M S F C F models.
We also compared the time complexity of the L- M S F C F model, A l l - M S F C F model, and 2 s - A G C N model to further assess the feasibility of the L- M S F C F model (Table 6).
Taking 15 frames as a recognition unit, the L- M S F C F model recognition speed was clearly higher than that of the A l l - M S F C F model and the 2 s - A G C N network model. Compared with the A l l - M S F C F model, the average recognition speed was more than two times higher. Compared with the 2 s - A G C N model, the average recognition speed is about 62.7% higher.
The L- M S F C F model was much more efficient than the 2 s - A G C N model, superior to the 2 s - A G C N network model in both recognition speed and accuracy. Even though the L- M S F C F model’s accuracy was 5.4% less than the A l l - M S F C F model’s, the recognition speed was improved by nearly one time. This showed that the L- M S F C F model had merit.
Finally, the method of this paper demonstrates the effect of recognizing some abnormal behaviors. The line in Figure 17 indicates the skeleton outline of the body in each frame. From Figure 17, it can be seen that the method can identify abnormal behavior more accurately.

7.4. TPT Model Test

We contrasted the T P T model with the P I F [35] and the S- G A N -P [36] model in order to emphasize its superiority. Each of them set up the training parameters, trained on the same datasets, predicted the trajectories in the future 15 frames, 30 frames, and 45 frames, and took the final average loss errors as the evaluation indexes. The experiment results are displayed in Table 7.
The findings demonstrated that the T P T model’s predictions for the next 15 and 30 frames outperform those of the other two network models, but as time goes on, the gap between the errors of the T P T model and S- G A N -P model gradually decreases, and after 45 frames the error of S- G A N -P model is smaller than the T P T model. It shows that T P T has an advantage in short-term prediction, but its advantage gradually decreases with the increase of time, so the later research can focus on the following long-time prediction.
The S- G A N -P model has 46.3k parameters and is the smallest model. However, the T P T model only has 17.6k parameters, which is about one-third of the parameters in S- G A N -P. Regarding inference speed, the quickest approach was S- G A N -P, which takes 0.0968 s for each inference step. The T P T model has an inference time of 0.0235 s for each inference step, four times faster than the S- G A N -P model. Table 8 shows that the T P T model has a considerable advantage both in parameter numbers and prediction time. We chose lightweight skeleton nodes and a redesigned convolutional architecture nicely circumvented the problem of substantial data and the usage of a cyclic architecture.
Finally, this paper shows part of the visualization results based on trajectory prediction tracking. In order to make the visualization results clear, we process the mass point e x in the visualization by keeping the X-axis and Y-axis coordinates unchanged, but the Z-axis coordinates are subtracted by half of the height of the body.
Firstly, Figure 18 and Figure 19 provide the visualization graphs of trajectory prediction results for some of the intact skeleton nodes. In the complete skeleton nodes, we utilize 4, 10, 7, 13, and 15, which represent the right hand, right foot, left hand, left foot, and head as the basis points to calculate the mass points. Figure 18 shows the walking posture prediction visualization results. Figure 19 shows the fighting posture prediction visualization results.
Secondly, Figure 20 and Figure 21 provide the visualization graphs of trajectory prediction results for some of the obscured skeleton nodes. In Figure 20, due to occlusion or other problems, it lacks skeleton nodes 6, 7, 12, and 13, which represent left elbow, left hand, left knee, and left foot. According to the Top order, the left shoulder and left hip are selected as one of the five bases. The other occluded skeleton information in Figure 20 and Figure 21 are sequentially calculated according to the five-bit skeleton screening method.

8. Discussion

This experiment verifies that the lightweight skeleton nodes process efficiently increases the timeliness of video action recognition. The L S F E method has a huge advantage in time complexity. The L- M S F C F model improves abnormal behavior recognition accuracy by predicting occluded skeletons and using feature fusion. We propose a method based on the previous frame for predicting skeleton data, which reduces the noise generated by the current skeleton data prediction from the distant skeleton data.
In this paper, 15 frames were selected as the most appropriate length of the action sequence. The advantages include a smaller amount of data and less noise, which enables a more accurate capture of video abnormal behaviors. Thus, the T P T model has the highest prediction accuracy under 15 frames, indicating that it is most effective in short-term prediction. The model can predict the trajectory of abnormal behaviors efficiently and quickly, showing its significant advantages in real-time applications in the computer vision field.
In terms of video abnormal behavior recognition and tracking, the research in this paper has achieved some milestones, but the applicability and reliability of the method in complex scenarios have not yet been fully discussed, and the stability of the algorithm’s performance in multiple datasets or scenarios, as well as the applicability of the algorithm to different numbers of skeleton nodes, have not been fully discussed. This paper only studies the video without combining audio, sensors, and other multiple data sources for comprehensive analysis. Future research should focus on the algorithm’s real-time, multimodal fusion and interpretability.

9. Conclusions

In this paper, we addressed the problem of a large number of skeleton nodes as well as behavioral occlusion between individuals degrading the abnormal behavior recognition speed and accuracy. We proposed a lightweight multi-stream feature cross-fusion (L- M S F C F ) model. The model adopted lightweight skeleton node computation, which significantly improved the recognition speed; at the same time, it improved the recognition accuracy by predicting the occluded skeleton nodes and effectively coped with the behavioral occlusion problem. Experiments show that our model achieves an average accuracy of 87.3% for abnormal behavior recognition. In addition, we also proposed the Trajectory Prediction Tracking ( T P T ) model, which can predict the movement position in real time based on core skeleton nodes, and its short-term prediction average loss error is small. In conclusion, our research effectively solves the behavioral occlusion problem while improving the recognition speed and accuracy, providing new ideas and methods for the development of the video action recognition field. The proposed model is expected to be applied to various practical applications in the fields of security and surveillance.

Author Contributions

Conceptualization, L.W., T.H.Z. and K.W.N.; formal analysis, L.W., C.D. and Y.Z.; funding acquisition, L.W., T.H.Z. and K.W.N.; investigation, C.D. and Y.Z.; methodology, L.W., C.D. and Y.Z.; software, W.D. and Y.Z.; validation, C.D. and Y.Z.; resources, T.H.Z., K.H.R. and K.W.N.; data curation, C.D. and Y.Z.; writing—original draft, L.W., C.D. and Y.Z.; writing—review and editing, L.W. and T.H.Z.; visualization, Y.Z. and W.D.; supervision, T.H.Z., K.H.R. and K.W.N. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (No. 62102076), by the Science and Technology Development Plan of Jilin Province, China (No. 20220402033GH), and by the Korea Agency for Infrastructure Technology Advancement (KAIA) grant funded by the Ministry of Land, Infrastructure and Transport (Grant RS-2022-00143336).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

A publicly available dataset was analyzed in this study. The Human3.6m dataset can be found here: http://vision.imar.ro/human3.6m/description.php (accessed on 18 August 2018). Dataset UCF-Crime can be found here: https://www.dropbox.com/sh/75v5ehq4cdg5g5g/AABvnJSwZI7zXb8_myBA0CLHa?dl=0 (accessed on 18 September 2018). Dataset ShanghaiTech Campus can be found here: https://svip-lab.github.io/dataset/campus_dataset.html (accessed on 20 April 2022).

Acknowledgments

The authors would like to thank the reviewers for their essential suggestions to help improve the manuscript.

Conflicts of Interest

The authors declare no conflicts of interests. The company was not involved in the study design, collection, analysis, interpretation of data, the writing of this article or the decision to submit it for publication.

References

  1. Herath, S.; Harandi, M.; Porikli, F. Going deeper into action recognition: A survey. Image Vis. Comput. 2017, 60, 4–21. [Google Scholar] [CrossRef]
  2. Wu, X.; Ou, Y.; Qian, H.; Xu, Y. A detection system for human abnormal behavior. In Proceedings of the 2005 IEEE/RSJ International Conference on Intelligent Robots and Systems, Edmonton, AB, Canada, 2–6 August 2005; pp. 1204–1208. [Google Scholar]
  3. Mehran, R.; Oyama, A.; Shah, M. Abnormal crowd behavior detection using social force model. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 935–942. [Google Scholar]
  4. Xie, S.; Zhang, X.; Cai, J. Video crowd detection and abnormal behavior model detection based on machine learning method. Neural Comput. Appl. 2019, 31, 175–184. [Google Scholar] [CrossRef]
  5. Yu, B.; Liu, Y.; Sun, Q. A content-adaptively sparse reconstruction method for abnormal events detection with low-rank property. IEEE Trans. Syst. Man Cybern. Syst. 2016, 47, 704–716. [Google Scholar] [CrossRef]
  6. Sabokrou, M.; Fathy, M.; Hoseini, M.; Klette, R. Real-time anomaly detection and localization in crowded scenes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Boston, MA, USA, 7–12 June 2015; pp. 56–62. [Google Scholar]
  7. Sabokrou, M.; Fayyaz, M.; Fathy, M.; Klette, R. Deep-cascade: Cascading 3d deep neural networks for fast anomaly detection and localization in crowded scenes. IEEE Trans. Image Process. 2017, 26, 1992–2004. [Google Scholar] [CrossRef]
  8. Martinel, N.; Micheloni, C.; Piciarelli, C.; Foresti, G.L. Camera selection for adaptive human–computer interface. IEEE Trans. Syst. Man Cybern. Syst. 2013, 44, 653–664. [Google Scholar] [CrossRef]
  9. Coşar, S.; Donatiello, G.; Bogorny, V.; Garate, C.; Alvares, L.O.; Brémond, F. Toward abnormal trajectory and event detection in video surveillance. IEEE Trans. Circuits Syst. Video Technol. 2016, 27, 683–695. [Google Scholar] [CrossRef]
  10. Xu, M.; Li, C.; Lv, P.; Lin, N.; Hou, R.; Zhou, B. An efficient method of crowd aggregation computation in public areas. IEEE Trans. Circuits Syst. Video Technol. 2017, 28, 2814–2825. [Google Scholar] [CrossRef]
  11. Kerdvibulvech, C.; Yamauchi, K. 3D human motion analysis for reconstruction and recognition. In Articulated Motion and Deformable Objects, Proceedings of the 8th International Conference, AMDO 2014, Palma de Mallorca, Spain, 16–18 July 2014; Proceedings 8; Springer: Cham, Switzerland, 2014; pp. 118–127. [Google Scholar]
  12. Kerdvibulvech, C.; Yamauchi, K. Structural human shape analysis for modeling and recognition. In Structural, Syntactic, and Statistical Pattern Recognition, Proceedings of the Joint IAPR International Workshop, S+ SSPR 2014, Joensuu, Finland, 20–22 August 2014; Proceedings; Springer: Cham, Switzerland, 2014; pp. 282–290. [Google Scholar]
  13. Houenou, A.; Bonnifait, P.; Cherfaoui, V.; Yao, W. Vehicle trajectory prediction based on motion model and maneuver recognition. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 4363–4369. [Google Scholar]
  14. Czyz, J.; Ristic, B.; Macq, B. A particle filter for joint detection and tracking of color objects. Image Vis. Comput. 2007, 25, 1271–1281. [Google Scholar] [CrossRef]
  15. Shao, X.; Huang, B.; Lee, J.M. Constrained Bayesian state estimation—A comparative study and a new particle filter based approach. J. Process Control 2010, 20, 143–157. [Google Scholar] [CrossRef]
  16. Vashishtha, D.; Panda, M. Maximum likelihood multiple model filtering for path prediction in intelligent transportation systems. Procedia Comput. Sci. 2018, 143, 635–644. [Google Scholar] [CrossRef]
  17. Kapania, S.; Saini, D.; Goyal, S.; Thakur, N.; Jain, R.; Nagrath, P. Multi object tracking with UAVs using deep SORT and YOLOv3 RetinaNet detection framework. In Proceedings of the 1st ACM Workshop on Autonomous and Intelligent Mobile Systems, Bangalore, India, 11 January 2020; pp. 1–6. [Google Scholar]
  18. Choi, D.; Yim, J.; Baek, M.; Lee, S. Machine learning-based vehicle trajectory prediction using v2v communications and on-board sensors. Electronics 2021, 10, 420. [Google Scholar] [CrossRef]
  19. Liu, W.; Sun, J.; Wang, G.; Bullo, F.; ChenFellow, J. Data-driven self-triggered control via trajectory prediction. IEEE Trans. Autom. Control. 2023, 68, 6951–6958. [Google Scholar] [CrossRef]
  20. Semwal, V.B.; Jain, R.; Maheshwari, P.; Khatwani, S. Gait reference trajectory generation at different walking speeds using LSTM and CNN. Multimed. Tools Appl. 2023, 82, 33401–33419. [Google Scholar] [CrossRef]
  21. Shokrolah Shirazi, M.; Morris, B.T. Trajectory prediction of vehicles turning at intersections using deep neural networks. Mach. Vis. Appl. 2019, 30, 1097–1109. [Google Scholar] [CrossRef]
  22. Zhou, H.; Ouyang, W.; Cheng, J.; Wang, X.; Li, H. Deep continuous conditional random fields with asymmetric inter-object constraints for online multi-object tracking. IEEE Trans. Circuits Syst. Video Technol. 2018, 29, 1011–1022. [Google Scholar] [CrossRef]
  23. Yoon, Y.C.; Kim, D.Y.; Song, Y.M.; Yoon, K.; Jeon, M. Online multiple pedestrians tracking using deep temporal appearance matching association. Inf. Sci. 2021, 561, 326–351. [Google Scholar] [CrossRef]
  24. Sabokrou, M.; Fayyaz, M.; Fathy, M.; Moayed, Z.; Klette, R. Deep-anomaly: Fully convolutional neural network for fast anomaly detection in crowded scenes. Comput. Vis. Image Underst. 2018, 172, 88–97. [Google Scholar] [CrossRef]
  25. Chu, W.; Xue, H.; Yao, C.; Cai, D. Sparse coding guided spatiotemporal feature learning for abnormal event detection in large videos. IEEE Trans. Multimed. 2018, 21, 246–255. [Google Scholar] [CrossRef]
  26. Yan, L. Researching on human abnormal behavior recognition algorithm. In Proceedings of the 2023 IEEE 6th Information Technology, Networking, Electronic and Automation Control Conference (ITNEC), Chongqing, China, 24–26 February 2023; Volume 6, pp. 1765–1769. [Google Scholar]
  27. Fang, Z.; Fei, F.; Fang, Y.; Lee, C.; Xiong, N.; Shu, L.; Chen, S. Abnormal event detection in crowded scenes based on deep learning. Multimed. Tools Appl. 2016, 75, 14617–14639. [Google Scholar] [CrossRef]
  28. Ye, O.; Deng, J.; Yu, Z.; Liu, T.; Dong, L. Abnormal event detection via feature expectation subgraph calibrating classification in video surveillance scenes. IEEE Access 2020, 8, 97564–97575. [Google Scholar] [CrossRef]
  29. Tay, N.C.; Tee, C.; Ong, T.S.; Teh, P.S. Abnormal behavior recognition using CNN-LSTM with attention mechanism. In Proceedings of the 2019 1st International Conference on Electrical, Control and Instrumentation Engineering (ICECIE), Kuala Lumpur, Malaysia, 25 November 2019; pp. 1–5. [Google Scholar]
  30. Ionescu, C.; Papava, D.; Olaru, V.; Sminchisescu, C. Human 3.6M. Available online: http://vision.imar.ro/human3.6m/description.php (accessed on 18 August 2018).
  31. Sultani, W.; Chen, C.; Shah, M. Real-world anomaly detection in surveillance videos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6479–6488. [Google Scholar]
  32. Liu, W.; Luo, W.; Lian, D.; Gao, S. Future frame prediction for anomaly detection—A new baseline. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6536–6545. [Google Scholar]
  33. Shi, L.; Zhang, Y.; Cheng, J.; Lu, H. Two-stream adaptive graph convolutional networks for skeleton-based action recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12026–12035. [Google Scholar]
  34. Martinez, J.; Black, M.J.; Romero, J. On human motion prediction using recurrent neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2891–2900. [Google Scholar]
  35. Liang, J.; Jiang, L.; Niebles, J.C.; Hauptmann, A.G.; Fei-Fei, L. Peeking into the future: Predicting future person activities and locations in videos. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5725–5734. [Google Scholar]
  36. Gupta, A.; Johnson, J.; Fei-Fei, L.; Savarese, S.; Alahi, A. Social gan: Socially acceptable trajectories with generative adversarial networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2255–2264. [Google Scholar]
Figure 1. Abnormal behavior recognition and tracking flowchart.
Figure 1. Abnormal behavior recognition and tracking flowchart.
Sensors 24 03711 g001
Figure 2. Skeleton node transformation schematic diagram. (a) Video data diagram; (b) skeleton data diagram.
Figure 2. Skeleton node transformation schematic diagram. (a) Video data diagram; (b) skeleton data diagram.
Sensors 24 03711 g002
Figure 3. Core skeleton node triangle calculation in 3D coordinates.
Figure 3. Core skeleton node triangle calculation in 3D coordinates.
Sensors 24 03711 g003
Figure 4. Adaptive sliding window intervals selection calculation example.
Figure 4. Adaptive sliding window intervals selection calculation example.
Sensors 24 03711 g004
Figure 5. L- M S F C F model flowchart.
Figure 5. L- M S F C F model flowchart.
Sensors 24 03711 g005
Figure 6. Occluded skeleton node prediction architecture.
Figure 6. Occluded skeleton node prediction architecture.
Sensors 24 03711 g006
Figure 7. L- M S F C F model architecture.
Figure 7. L- M S F C F model architecture.
Sensors 24 03711 g007
Figure 8. L- M S F C F convolutional block.
Figure 8. L- M S F C F convolutional block.
Sensors 24 03711 g008
Figure 9. Skeleton node partitions.
Figure 9. Skeleton node partitions.
Sensors 24 03711 g009
Figure 10. Mass e x angle schematic at frame T.
Figure 10. Mass e x angle schematic at frame T.
Sensors 24 03711 g010
Figure 11. T P T model architecture diagram.
Figure 11. T P T model architecture diagram.
Sensors 24 03711 g011
Figure 12. Time complexity comparison of R N N model and L S F E model.
Figure 12. Time complexity comparison of R N N model and L S F E model.
Sensors 24 03711 g012
Figure 13. Accuracy rate comparison of R N N model and L S F E model.
Figure 13. Accuracy rate comparison of R N N model and L S F E model.
Sensors 24 03711 g013
Figure 14. L- M S F C F model training iterations versus loss function.
Figure 14. L- M S F C F model training iterations versus loss function.
Sensors 24 03711 g014
Figure 15. L- M S F C F model training iterations versus accuracy.
Figure 15. L- M S F C F model training iterations versus accuracy.
Sensors 24 03711 g015
Figure 16. Confusion matrix diagram.
Figure 16. Confusion matrix diagram.
Sensors 24 03711 g016
Figure 17. Abnormal behavior recognition effect.
Figure 17. Abnormal behavior recognition effect.
Sensors 24 03711 g017
Figure 18. Walking posture prediction visualization results.
Figure 18. Walking posture prediction visualization results.
Sensors 24 03711 g018
Figure 19. Fighting posture prediction visualization results.
Figure 19. Fighting posture prediction visualization results.
Sensors 24 03711 g019
Figure 20. Walking posture trajectory prediction visualization results with occluded skeleton nodes.
Figure 20. Walking posture trajectory prediction visualization results with occluded skeleton nodes.
Sensors 24 03711 g020
Figure 21. Fighting posture trajectory prediction visualization results with occluded skeleton nodes.
Figure 21. Fighting posture trajectory prediction visualization results with occluded skeleton nodes.
Sensors 24 03711 g021
Table 1. Behavior classification.
Table 1. Behavior classification.
IdentifiersAction TypeCategory DefinitionBehaviors
NNormal behaviorWhen behavior is consistent with usual behavior, it is quite normal for an individual.Walking, running, stooping.
AAbnormal behaviorCan be divided into two kinds: one is the disturbance of order in public places, and the other refers to criminal acts.Fighting, vicious kicking, climbing over walls, throwing suspicious objects, and slashing devices.
Table 2. Lightweight feature skeleton node extraction results.
Table 2. Lightweight feature skeleton node extraction results.
Action TypeLightweight Feature Skeleton Nodes
Walking3, 4, 6, 7, 9, 10, 12, 13
Running3, 4, 6, 7, 9, 10, 12, 13
Stooping0, 1, 2, 3, 4, 5, 6, 7, 14
Fighting3, 4, 6, 7, 9, 10, 12, 13
Vicious kicking2, 3, 4, 5, 6, 7
Climbing over walls3, 4, 6, 7, 8, 9, 10, 11, 12, 13
Throwing suspicious objects3, 4, 6, 7, 10, 13
Slashing devices2, 3, 4, 5, 6, 7, 10, 13, 0
Table 3. Skeleton node sequences for each partition.
Table 3. Skeleton node sequences for each partition.
AreaSkeleton Node Sequences
A0, 1, 14, 15
B4, 3, 2
C7, 6, 5
D10, 9, 8
E13, 12, 11
Table 4. Model prediction accuracy and time comparison.
Table 4. Model prediction accuracy and time comparison.
ModelAverage AccuracyAverage Time
R N N Model80.7%362 ms
L S F E Model76.2%194 ms
Table 5. Behavior recognition rate accuracy comparison.
Table 5. Behavior recognition rate accuracy comparison.
ModelsL-MSFCFAll-MSFCF2s-AGCN
Behaviors
Walking0.940.960.87
Running0.920.950.88
Stooping0.910.930.84
Fighting0.810.870.73
Vicious kicking0.820.890.78
Climbing over walls0.860.950.83
Throwing suspicious objects0.850.940.83
Slashing devices0.870.930.84
Average accuracy0.8730.9270.825
Table 6. Time- consuming comparison.
Table 6. Time- consuming comparison.
ModelsParameter NumbersAverage Recognition Time
A l l - M S F C F 377.3 k347 ms
2 s - A G C N 117.9 k258 ms
L- M S F C F 62.6 k162 ms
Table 7. The model’s average loss errors in each frame prediction.
Table 7. The model’s average loss errors in each frame prediction.
Model15 Frame Errors30 Frame Errors45 Frame Errors
P I F 0.330.430.65
S- G A N -P0.220.350.51
T P T 0.190.320.57
Table 8. Model prediction times comparison.
Table 8. Model prediction times comparison.
ModelParameter NumberPrediction Time
P I F 360.3k132 ms
S- G A N -P46.3k97 ms
T P T 17.6 k23 ms
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, L.; Ding, C.; Zhang, Y.; Zhou, T.H.; Ding, W.; Ryu, K.H.; Nam, K.W. Video Abnormal Behavior Recognition and Trajectory Prediction Based on Lightweight Skeleton Feature Extraction. Sensors 2024, 24, 3711. https://doi.org/10.3390/s24123711

AMA Style

Wang L, Ding C, Zhang Y, Zhou TH, Ding W, Ryu KH, Nam KW. Video Abnormal Behavior Recognition and Trajectory Prediction Based on Lightweight Skeleton Feature Extraction. Sensors. 2024; 24(12):3711. https://doi.org/10.3390/s24123711

Chicago/Turabian Style

Wang, Ling, Cong Ding, Yifan Zhang, Tie Hua Zhou, Wei Ding, Keun Ho Ryu, and Kwang Woo Nam. 2024. "Video Abnormal Behavior Recognition and Trajectory Prediction Based on Lightweight Skeleton Feature Extraction" Sensors 24, no. 12: 3711. https://doi.org/10.3390/s24123711

APA Style

Wang, L., Ding, C., Zhang, Y., Zhou, T. H., Ding, W., Ryu, K. H., & Nam, K. W. (2024). Video Abnormal Behavior Recognition and Trajectory Prediction Based on Lightweight Skeleton Feature Extraction. Sensors, 24(12), 3711. https://doi.org/10.3390/s24123711

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop