Next Article in Journal
Effect of the Long-Term Mean and the Temporal Stability of Water-Energy Dynamics on China’s Terrestrial Species Richness
Next Article in Special Issue
Reducing Building Conflicts in Map Generalization with an Improved PSO Algorithm
Previous Article in Journal
An Original Application of Image Recognition Based Location in Complex Indoor Environments
Previous Article in Special Issue
Mapping Comparison and Meteorological Correlation Analysis of the Air Quality Index in Mid-Eastern China
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Identifying Different Transportation Modes from Trajectory Data Using Tree-Based Ensemble Classifiers

1
Key Laboratory of Spatial Information Processing and Application System Technology, Institute of Electronics, Chinese Academy of Sciences, Beijing 100190, China
2
School of Electronic, Electrical and Communication Engineering, University of Chinese Academy of Science, Beijing 100190, China
*
Author to whom correspondence should be addressed.
ISPRS Int. J. Geo-Inf. 2017, 6(2), 57; https://doi.org/10.3390/ijgi6020057
Submission received: 3 February 2017 / Revised: 3 February 2017 / Accepted: 18 February 2017 / Published: 22 February 2017
(This article belongs to the Special Issue Intelligent Spatial Decision Support)

Abstract

:
Recognition of transportation modes can be used in different applications including human behavior research, transport management and traffic control. Previous work on transportation mode recognition has often relied on using multiple sensors or matching Geographic Information System (GIS) information, which is not possible in many cases. In this paper, an approach based on ensemble learning is proposed to infer hybrid transportation modes using only Global Position System (GPS) data. First, in order to distinguish between different transportation modes, we used a statistical method to generate global features and extract several local features from sub-trajectories after trajectory segmentation, before these features were combined in the classification stage. Second, to obtain a better performance, we used tree-based ensemble models (Random Forest, Gradient Boosting Decision Tree, and XGBoost) instead of traditional methods (K-Nearest Neighbor, Decision Tree, and Support Vector Machines) to classify the different transportation modes. The experiment results on the later have shown the efficacy of our proposed approach. Among them, the XGBoost model produced the best performance with a classification accuracy of 90.77% obtained on the GEOLIFE dataset, and we used a tree-based ensemble method to ensure accurate feature selection to reduce the model complexity.

1. Introduction

Understanding user mobility from trajectory data has received considerable attention among users and developers. As a central component of understanding user mobility [1], the identification of transportation modes used by people can provide rich information for some tasks and utilized in many applications. First, the information of transportation modes used by users is applicable to the travel behavior research field [2,3]. Travel behavior is concerned with the transportation people take, the routes they choose, and so on. According to the user’s daily transportation mode, we can discover the user’s behavior, and thus carry out personalized recommendations and provide intelligent services. For example, the system may recommend residential or transportation services if somebody is traveling on a train. Second, it is a significant part of transportation planning [4] and traffic management, where transportation mode selection greatly influences transportation planning. In the past, this type of information was usually collected traditionally through questionnaires or telephone surveys, which were always expensive, inconvenient, time-consuming and inaccurate. However, if we can automatically identify the transportation modes accurately, it will also be a great convenience to people who trace their outdoor activity. Third, transportation mode identification can also serve other applications, such as GPS trajectories sharing applications [5], sports products and intelligent surveillance systems. For example, in trajectory sharing applications, the information on transportation modes could ensure greater meaningful sharing and assist it in calculating the number of calories burnt.
Many studies in the previous decade have focused on inferring transportation modes from trajectory data by using machine learning methods to identify transportation modes from GPS or accelerometer data, which have achieved great results. However, there are still several limitations in these studies. First, many studies promote performance mainly through two ways: (1) The use of more sensors (gyroscope, and rotation vector) or more accurate sensors (professional device) to collect trajectory data [6,7,8,9], which improves performance, as more accurate data means more difference between transportation modes. However, it is difficult for people to carry many sensors around in their daily lives, therefore, it is not a realistic option for most people; (2) The use Geographic Information System information (GIS) to create more discriminative features [10,11]. However, it is not always easy to access GIS information and is not suitable for all cities. Second, almost all studies use traditional machine learning algorithms such as Decision Tree, K-Nearest Neighbor and Support Vector Machines [6,7,10,11,12,13,14,15,16,17]. However, these methods were found to be worse than ensemble methods, not only in terms of accuracy but also model robustness [18,19,20,21].
Given the above limitations, we propose a method to infer hybrid transportation modes only from GPS data to achieve a good performance in a real life trajectory dataset collected by Microsoft Geolife Project [14,15]. The unique contributions of this research effort are as follows:
(1)
We used the statistics method [22] to generate some discriminative global features and then extracted local features from sub-trajectories after trajectory segmentation [23]. These features were then combined into a feature vector fed into the classification stage, which proved more robust than other used features and helped improve classification accuracy.
(2)
We used tree-based ensemble models to classify the transportation modes, which proved to outperform other traditional methods (K-Nearest Neighbor, Support Vector Machines, and Decision Tree) in many evaluations.
(3)
We extracted a lot of features, then used tree-based ensemble models to obtain feature importance, and then reduced the model complexity by removing the unimportant features, which our performance was better than the Principal Component Analysis (PCA) method.
(4)
Using our approach, we could identify six different transportation modes including walking, cycling, driving a car, taking a train and taking subway. The highest accuracy was 90.77% and was obtained only through GPS data.
The rest of our paper is organized as follows. The relevant literature is reviewed in the following section, before a brief introduction on how the classification models were developed. Related experiments of the study and discussion are then presented before our conclusions on the study.

2. Related Work

This section provides a summary of related work in the area of transportation mode identification, with almost all studies taking advantage of machine learning methods such as Decision tree [7,13,14,15,24,25], K-Nearest Neighbor [13], Support vector machines [12,13,14,17,23], Artificial neural networks [26,27], Random forest [13,24], Fuzzy systems [28,29] and hidden Markov models [25,30,31]. Among them, the Decision Tree and SVM methods were identified as the best when compared to other methods [13,14,16]. Other than these machine learning methods, statistical methods were also applied, such as Random Subspace Method [32]. In addition, Endo et al. [33] proposed a method using a deep neural network (DNN) by converting the raw trajectory data structure into image data structure.
Many studies wanting to improve model performance primarily depend on the following methods, such as extracting more discriminative features from GPS trajectories. For example, Zheng et al. [14] used common statistical features such as mean velocity, expectation of velocity, expectation of velocity, top three velocities and top three accelerations to identify four different transportation modes (Bike, Bus, Car, Walk). Zheng et al. [15] further introduced more advanced features including the heading change rate (HCR), the stop rate (SR), and the velocity change rate (VCR), which achieved a more accurate performance. Additionally, using extra sensors such as accelerometers, gyroscopes, rotation vectors to collect more accurate data provides greater detailed information between the different transportation modes. For example, Nick [16], Jahangiri [12,13], and Widhalm [31] used mobile phone sensors to collect accelerometer data of different transportation modes, which achieved a better performance than only using GPS data. Other examples in References [8,9] used body temperature, heart rate, light intensity obtained from other sensors as features, and were able to build a model for predicting user activities including walking, running, rowing and cycling. Furthermore, GIS data can be used to create more distinguished features, as detailed in References [24,28]. For example, Stenneth [24] derived the relevant features relating to transportation network information and improved the classification effectiveness of detection by 17% in comparison with only GPS data, which included real time bus locations, spatial rail and bus stop information. Table 1 summarizes the related work in the field of transportation modes identification.
It appears that these methods have at least one of the following constraints. Firstly, these methods are overly dependent on using multiple sensors to collect more accurate data or matching GIS information to generate more discriminative features; however, often only the GPS data are available as it is difficult to store the data from the different sensors at specific times, and it was not possible to access GIS data in many cities. Therefore, this study only used GPS data, which can also achieve high accuracy. Although Gonzalez [26,27] achieved 91.23% accuracy, only three transportation modes were identified. Secondly, most of these methods need to consider feature scaling as the values of raw data vary widely. In some machine learning algorithms, objective functions will not work properly without normalization; however, the tree-based ensemble method does not require feature scaling as these methods are invariant to monotonic transformations of individual features so feature scaling will not change anything in these methods. Thirdly, most of these methods do not use feature selection methods as some redundant and unimportant features are useless, which also adds the model complexity. Some studies use the PCA method for feature reduction; however, we used the tree-based ensemble methods to conduct feature selection, which performed better than the PCA method. Fourthly, most studies use single stronger classifiers (Decision, Tree, K-Nearest Neighbor, and Support Vector Machines) to identify transportation modes, but ensemble methods have been found to do better than these classifiers as ensemble methods are a combination set of weak learners, which are superior to a single stronger learner, and have been widely proven in many machine learning challenges [18,19,20]. Although some of them also use the random forest method, our method is different from these studies as we use other ensemble methods including GBDT [34] and XGBoost [19] besides the RF method. Furthermore, we used the ensemble method to conduct feature selection, which significantly differs from other recent work.
In this paper, we propose an approach using a tree-based ensemble method to infer hybrid transportation modes from only GPS data, which does not require any feature scaling because of the nature of tree-based methods. As tree-based methods can give importance to input features, we can have a minimum number of features with maximum predictive power by deleting the unimportant features.

3. Methodology

In this paper, we propose a method to identify six different transportation modes including walking, cycling, driving a car, taking a train and taking subway using GPS data. The main modules of our methodology for identifying transportation modes is shown in Figure 1 and includes data preprocessing, feature extraction, model classification and model evaluation. Raw GPS data were first preprocessed into trajectories, and then global and local features were extracted before using the ensemble method to classify the different transportation modes. From the ensemble method, we obtained the important extracted features and removed the unimportant features, before results were finally obtained. Each step is detailed in the following sections.

3.1. Data Preprocessing

For a better performance, two data preprocessing techniques were employed in this study. First, we removed the duplicate data in the dataset as some GPS points were recorded more than once due to recording errors on the GPS device. Second, according to common sense, we removed some outlier trajectories, which were deemed abnormal. For instance, if the average speed of a trajectory marked as “walking” exceeded 10 m/s, or if the average speed of a trajectory marked as “biking” exceeded 25 m/s, we identified them as abnormal trajectories and removed them from the dataset.

3.2. Feature Extraction

Feature engineering plays an important role in recognizing the transportation modes. Feature engineering is the process of using domain knowledge to create features to enable machine learning methods to work well; to achieve differentiated results in the recognition of transportation modes and to extracted a large number of features from the processed GPS trajectories. These features can then be categorized into global features and local features. Global features refer to descriptive statistics for the entire trajectory, which makes trajectories more comparable, and the local features extracted by profile decomposition reveal more detail in movement behavior. The following subsections explain each category feature in detail.

3.2.1. Global Features

First, we calculated four movement parameters from the original trajectories: speed, acceleration, turn angle (the direction of the two consecutive points) and sinuosity (its winding path divided by the distance). Following that, we used the statistics method [22] to extract global features based on the above-mentioned movement parameters and are listed as follows:
(1)
Mean: This is a measure of the central tendency of the data sets.
(2)
Stand deviation: This is a measure that is used to describe the dispersion of a set of data values.
(3)
Mode: The mode represents the value that occurs most frequently in the sets; the parameters are first rounded into integers as the parameters are continuous values.
(4)
Top three value, Minimum three value: These parameters are to reduce the error brought by the abnormal point with positional errors.
(5)
Value ranges: The maximum value minus the minimum value.
(6)
Percentile: A measure that represents the value below a given percentage of observations which is used in statistics. In this paper, we select the 25th percentile (lower quartiles) and 75th percentile (upper quartiles).
(7)
Interquartile range: This parameter is equal to the difference between the lower and upper quartiles.
(8)
Skewness: This is a measure of asymmetry of the probability distribution of a real-valued random variable about its mean, which can be used to describe the distribution of movement parameters, and is defined by:
S k e w n e s s = E [ ( X μ σ ) 3 ]
where μ represents the average of the data, and σ represents the standard deviation of the data.
(9)
Kurtosis: A measure of the “tailedness” of the probability distribution of a real-valued random variable and can also be used to describe the distribution of movement parameters compared to the normal distribution, and is defined by:
K u r t o s i s = E [ ( X m s ) 4 ] 3
where μ represents the average of the data, and σ represents the standard deviation of the data.
(10)
Coefficient of variation: This is a standardized measure of dispersion of probability distribution or frequency distribution, when the measurement scale difference is too large, it is inappropriate to use standard deviation to measure the parameters as the standard deviation is always understood in the context of the mean of the data but the C V is independent of the unit. Therefore, it can handle the problem of different unit or large measurement scale, and is defined by:
C V = σ μ
where μ represents the average of the data, and σ represents the standard deviation of the data.
(11)
Autocorrelation coefficient: This is the cross-correlation of a signal with itself at different points in time (what the cross stand for). First the coefficient of auto-covariance is calculated by:
c k = 1 N i = 1 N k ( x i μ ) ( x i + k μ )
where μ represents the average of the data.
Then, we can calculate the autocorrelation coefficient, which is defined by:
A u t o = c 1 c 0
It is a mathematical tool for finding repeating patterns, and represents the correlation between the values of the process at different times.
Secondly, in addition to these statistical features, we used other features proposed by Zheng et al. [15], which proved robust to traffic conditions and are listed as follows:
(1)
Heading change rate: This measure is a discrimination feature to distinguish different transportation modes as proposed by Zheng et al. [15]. It can be considered as the frequency in which people change their direction to certain threshold within unit distance, which can be used to distinguish motorized and non-motorized transportation modes, and is defined by:
H C R = | P c | / D i s t a n c e
where P c represents the set of GPS points at which user changes direction extending to a given threshold, and | P c | stands for the number of elements in P c .
(2)
Stop rate: The stop rate represents for the number of points at which the velocity of the user under a certain threshold in unit distance mentioned by Zheng et al. [15], is defined by:
S R = | P S | / D i s t a n c e
where P S represents the set of points at which a user’s velocity below a giving threshold.
(3)
Velocity change rate: This feature mentioned by Zheng et al. [15] is defined by:
V C R = | P V | / D i s t a n c e
where P V represents the collection of points at which the user’s velocity changes above a giving threshold, and the velocity change is defined by:
P V = | v 2 v 1 | v 1
(4)
Trajectory length: The total distance of the trajectory.

3.2.2. Local Features

We adopted the profile decomposition algorithm mentioned by Dodge [23] to generate several local features. As the movement parameters (velocity, acceleration, and turning angle) change over time when a person is moving in the space, this becomes a profile or a function. First, the movement parameter is expressed as a time series, where the amplitude and change frequency will provide proof to describe the movement behavior and physics of the person. Second, we used two measures to decompose the movement parameters, where a deviation from the central line represents the amplitude variations over time. Sinuosity reflects the frequency variations with time change. These points were then divided into four categories including “low sinuosity and low deviation”, “high sinuosity and low deviation”, “low sinuosity and high deviation” and “high sinuosity and high deviation”, which are labeled 0, 1, 2, and 3, respectively. More details can be found in Reference [23].
Next, a series of features were extracted from the profile decomposition algorithm. This included the mean and standard statistics of segment length per decomposition class and per parameter; the count of changes of decomposition classes; and the proportion per decomposition class account for the total number of points.
A summary of the global features and local features is shown in Table 2, where every trajectory extracts the global features and local features is mentioned. In Figure 2, we plotted the distribution of six features mentioned above included four global features (a–d) and two local features (e and f); however, we found it difficult to differentiate between the different transportation modes using only a single feature. Nevertheless, these features can provide useful information in transportation identification when combined together. Therefore, 111 features were combined into the classification procedure, which is shown in Table 2.

3.3. Model Classification and Model Evluation

Ensemble methods train multiple base learners to solve the same problems, which differs from base learning algorithms. The basic structure of ensemble learning is shown in Figure 3. Ensemble methods were highlighted by two pioneering works and have since become major machine learning algorithms. One empirical work, by Hansen and Salamon [35], found that the combination of a set of classifiers often perform better than the best single classifier. The other pioneering work was theoretical, where Schapire [36] proved that a combination set of weak learners was superior to a single stronger learner. Ensemble methods are widely used in many machine learning challenges; for example, Random Forest was used in 2010 Knowledge Discovery and Data Mining (KDD) Cup to win the first prize [20], the Gradient Boosting Decision Tree (GBDT) was used in the Netflix prize [18], and XGBoost was successfully used in Kaggle competition and 2015 KDD Cup [19].
K-Nearest Neighbor [13,25], Decision Tree (DT) [7,13,14,15,24,25] and Support Vector Machine (SVM) [12,13,14,17,23,25] have been widely used to classify transportation modes and have resulted in good performance. However, in this paper, we choose to use the tree ensemble models to classify transportation modes instead of using the above-mentioned methods. The reasons why we selected the tree ensemble methods are as follows. Firstly, tree ensemble methods perform better than traditional methods (KNN, DT, SVM) not only in higher prediction accuracy, but also in superior efficiency, which has been proved in many machine learning challenges. Secondly, the parameter tuning in tree ensemble methods was easy to conduct. Thirdly, deep learning or neural network was not used as the data volume is insufficient and the parameter tuning in deep learning was too complex. The ensemble methods used in this paper were Random Forest (RF) [37], Gradient Boosting Decision Tree (GBDT) [34] and XGBoost [19].
In the next section, we introduce the basic principles of these three models briefly where more detail can be found in the corresponding references, and then the evaluations of our method are explained.

3.3.1. Random Forest

Random Forest was proposed in 2001 [37] and used the bagging method to ensemble many decision trees with the final classification result obtained based on the majority votes. Random Forest was widely applied at each classification task, and achieved great performance. The main steps are as follow:
(1)
The column sampling technique was used to select the train set for growing the tree. Not all samples fit into the tree growing, but the boostrap method was used to generate the training set.
(2)
The row sampling technique was used to select m (m << M) variables from M input variables, where the best spilt was found on these variables, to help contribute in overcome overfitting.
(3)
Each tree in the forest grew completely without any pruning.

3.3.2. Gradient Boosting Decision Tree

The GBDT is a widely used method for classification and regression problems proposed by Friedman [34], which produces a learning model in the form of a combination of a set of weak prediction models, typically decision trees. The difference between GBDT and the traditional Boost is that every iteration in GBDT is to reduce the residual of the last iteration. In order to eliminate the residual error, the GBDT method establishes a new model on the direction of gradient. The GBDT is also a similar method to RF in that it combines a set of weak learners. The main difference between the GBDT and RF is that the tree in the GBDT is fit on the residual of the former tree, so that the GBDT model can try to reduce the bias, whereas RF tries to reduce the variance. Thus, RF can be trained in parallel, but the GBDT cannot, therefore, the GBDT was superior to RF in many challenges [18,19].

3.3.3. XGBoost

XGBoost [19] is an optimized distributed gradient boosting library designed to be highly efficient, flexible and portable. XGBoost provides a parallel tree boosting that solves many data sciences in a fast and accurate way. The advantages of this method over the traditional GBDT are as follows:
(1)
Additional regularization terms can be added to help smooth the final learnt weights to avoid over-fitting.
(2)
Second-order approximation of the loss function can be used to quickly optimize the objective in the general settings.
(3)
Shrinkage (introduced by Friedman [34]), can be used to reduce the influence of each individual tree and leave more space for future trees.
(4)
The column technique in RF can be used for reference, in order to prevent over-fitting and to speed up computations of parallel algorithms.
(5)
Some techniques, including blocks for out-of-core computation and cache-aware access can be used to make tree learning faster.

3.3.4. K-Fold Cross-Validation

The K-Fold Cross-Validation is a technique to derive a more accurate f model and to help tune parameters. The original samples are randomly partitioned into k equal parts; of the k parts, a single part is used as the validation dataset, and the remaining k-1 subparts are used as the training dataset to construct the model. The same procedure is repeated k times, with a different validation dataset is chosen each time, before the final accuracy of the model is equal to the average accuracy obtained each time. The advantage of this technique over repeated random sub-sampling is that all samples are used for both training and validation, and each sample is used for validation exactly once.

3.3.5. Evaluations

In order to evaluate the performance of different classification models, several methods were used in this study, including precision, recall, F-score, confusion matrix, Receiver Operating Characteristic curve and Area Under The Curve. For classification problems, samples were divided into four categories: true positives; true negatives; false positives; and false negatives (Table 3).
Precision is the fraction of retrieved documents that are relevant to the query, and recall is the fraction of the documents that are relevant to the query that are successfully retrieved. Precision and recall are thus defined as:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
In order to measure the overall performance of the model, the F-score was considered, which contains both the precision and the recall.
2 F = 1 P r e c i s i o n + 1 R e c a l l
The confusion matrix is a specific table layout that allows the visualization of the performance of an algorithm, which shows the error rates, precision and recall values. The ROC curve illustrates the performance of a classifier system as its discrimination threshold is varied, and is created by plotting the true positive rate (TPR) against the false positive rate (FPR) at various threshold settings, thus providing tools to select possible optimal models. We can use the areas under the curve to measure the performance when the two ROC-curves were cross, which is called the AUC.
T P R = T P T P + T N
F P R = F P T N + F P

4. Results and Discussion

A dataset collected by the Geolife project [12,13] by 182 users over a period of five years (from April 2007 to August 2012) was used to validate the performance of the ensemble model. This dataset is widely distributed in over 30 cities of China, as well as cities located in the USA and Europe and contains 17,621 trajectories with a total distance of 51 km and a total duration of 50,176 h. Up to 91.5% of the trajectories are logged in a dense representation, e.g., every 1–5 s or every 5–10 m per point. In the dataset, 73 users labeled their trajectories with their starting transportation mode or when their mode changed. Figure 4 presents one example of the trajectories in the Geolife dataset, and these labeled trajectories were used to conduct the experiment.
After data preprocessing, 7985 trajectories were extracted. The dataset was separated into a training set (consisting of 70% of the data) and a testing set (30% of the data). The distribution of the number of different categories is shown in Table 4. As the distribution of the different categories was unbalanced, it was unreasonable to base our evaluation on accuracy alone, despite this being commonly done in the previous literature. Furthermore, it was difficult to evaluate the performance of the model based on a single measure as these measures are sometimes the same. Thus, in our study, different evaluation techniques were used, including Precision, Recall, F-Score, ROC cure and Confusion matrix. We used overall misclassification rate which is obtained from five-fold Cross-Validation to tune the parameters. The detailed experiments are shown below:

4.1. K Nearest Neighbor

The KNN model was implemented in the Python software along with “sklearn” package. The tuning parameter in the KNN model was the number of neighbors. Figure 5a shows the change of misclassification error with the number of neighbors, which was obtained using five-fold Cross-Validation. From Figure 5, we can see that the highest accuracy was achieved when K was 7, and the overall accuracy was 74.44%. The ROC curve is plotted in Figure 5b, and the Confusion matrix is presented in Table 5. Figure 5b and Table 5 show the poor performance of bus, bike, car, subway and train, which led to the conclusion that the KNN model found it difficult to differentiate the different transportation modes.

4.2. Decision Tree

The decision tree model was implemented in the Python software along with “sklearn” package. The tuning parameter was the maximum depth of the tree and was used to control over-fitting. Figure 6a shows how the depth of the tree impacts the performance of the model. The best performance was 86.05%, and was achieved when maximum depth was 7. The Confusion matrix is presented in Table 6, and the ROC curve is plotted in Figure 6b. In Figure 6b and Table 6, we can see that the performance of the DT model was better than the KNN base in most of the evaluation measures; however, it is still unable to effectively identify transportation modes including car, subway and train.

4.3. Support Vector Machine

The SVM model was implemented in the Python software along with “sklearn” package. The tuning parameters contained regularization parameter (c) and Gaussian parameter (σ). Figure 7a provides a heat plot that demonstrates how the regularization parameter (c) and the Gaussian parameter (σ) impacted the performance of the SVM model. The optimal parameters (c, σ) of the model were eight and five, respectively, and are circled in red. The highest accuracy was 87.69%, the Confusion matrix for the best SVM model is presented in Table 7 and the ROC curve is plotted in Figure 7b. In Figure 7b and Table 7, we can see that although the SVM model performed better than both the KNN and DT models, it still cannot effectively identify transportation modes including car, subway and train.

4.4. Random Forest

The Random Forest model was implemented in the Python software along with “sklearn” package. Figure 8a shows the change of Test-Error and the Out of Bag Error with number of the tree, and Figure 8b shows the influence of the number of features. From these two figures, 200 trees and 16 features generated the best performance. Table 8 shows the Confusion matrix for the best RF model, and the ROC curve is plotted in Figure 8c. The best five-fold Cross-Validation accuracy was 90.28%, and shows that it has remarkably outperformed the KNN, DT and SVM models. Furthermore, compared to the other models listed above, we found that the RF model could effectively identify the transportation modes that other models could not, such as car, subway and train. The detection accuracy of these models showed a significant improvement over the above models.

4.5. Gradient Boosting Decision Tree

The GBDT method was implemented in the Python software along with “sklearn” package. The tuning parameters in the GBDT method contained the number of trees, the number of features, and max depth. Figure 9a shows the change of Test-Error and CV (five-fold Cross-Validation) Error with number of the tree, Figure 9b shows the influence of number of features used in each tree. The highest accuracy (90.46%) was achieved when 500 trees, 16 features, and a max depth of six were used. The Confusion matrix for the best GBDT models is shown in Table 9 and the ROC curve are plotted in Figure 9c. The results imply a small improvement over the RF model, but GBDT is more time-consuming than RF model as it cannot process in parallel (which can be seen in the part model comparison).

4.6. XGBoost

The XGBoost method was implemented in the Python software along with “XGBoost” package. The tuning parameters in XGBoost included “num_round”, “eta”, “max_depth”, “scale_pos_weight” and “min_child_weight”. We used the five-fold Cross-Validation to tune the parameters which included the following steps: First, we selected a relatively high learning rate and determined the optimum number of trees for this learning rate. Second, we tuned tree-specific parameters (max_depth, min_child_weight and scale_pose_wight) for the decided learning rate and the number of trees. Third, we tuned the regularization parameters (lambda, alpha) for XGBoost, which can help reduce model complexity and enhance performance. Fourth, we lowered the learning rate and decided on the optimal parameters. We have provided two example of the parameters tuning process, other process can be found in Reference [38]. Figure 10a shows the influence of the number of iterations on the misclassification error and Figure 10b shows the influence of the number of features used in each iteration. Using this method, the highest accuracy was 90.77% with the parameters set as follows: num_round = 500, eta = 0.05, max_depth = 6, scale_pose_weight = 5, and min_child_weight = 1. The Confusion matrix for the best XGBoost models is shown in Table 10 and the ROC curve is plotted in Figure 10c, and shows an improvement greater than that of the GBDT model, and more discriminating in many transportation modes, with the train speed was faster than the GBDT model which will be proved in the next section.

4.7. Model Comparison

Based on the above results, the F-score measure was calculated and can be seen in Figure 11a. It is obvious that the tree-based ensemble models achieved a higher F-score than the DT, SVM and KNN models. Figure 11b shows the ROC curve and the AUC measure of different classification models. In Figure 11b, it can be clearly seen that tree-based ensemble models achieved superior results to the DT, SVM and KNN models based on ROC and AUC measures. Meanwhile, Figure 11c shows the computation time of the different models where the XGBoost and RF model consume less time than others, and can handle large volumes of data in the future. Considering the different evaluation measures, we finally concluded that the tree ensemble methods performed better than the traditional machine learning methods (K-Nearest Neighbor, Support Vector Machines and Decision Tree), and out of the tree ensemble methods, the XGBoost model produced the best performance, with the highest accuracy 90.77%, which was a nearly 3% improvement compared to the SVM model.
These results also suggest that the classification errors mainly occurred between the car and bus classifiers, as these two transportation modes always perform similarly in the GPS trajectory. One solution to reduce the error is to add more accurate sensors to improve the classification performance in feature work, such as gyroscopes and rotation vectors. Another solution is the use of GIS information to create more distinguished features; however, these two methods have several limitations, which are mentioned earlier. Therefore, the use of these two solutions depends on the detailed application. Furthermore, tree-based ensemble methods can still efficiently identify different transportation modes by combining these features into the classification procedure.

4.8. Feature Importance and Feature Reduction

As mentioned earlier, 111 features were extracted from the original GPS trajectory. To evaluate the validity of our feature extraction method, we compared our features with the following baseline features: (1) Global features: 72 dimensional features mentioned above; (2) Local features: 39 dimensional features mentioned above; (3) Zheng’s features: 12 dimensional features mentioned by Zheng [15]; (4) Dodge’s features: 48 dimensional features mentioned by Dodge [23]; and (5) Global + Local: the combination of global features and local features. We performed the experiments with the same dataset that used the same method but with different features. Table 11 shows the misclassification error with different features in the same model. We still found that tree-based ensemble methods perform better than other methods in different features, and we concluded that our feature extraction method worked better in recognizing different transportation modes than other feature extraction methods. The combination of Global features and Local Features scores the lowest error rates.
In most of the previous literature, the principal component analysis (PCA) is predominantly used to conduct feature selection prior to the classification stage. In our study, we used the tree-based ensemble method to carry out feature selection to reduce the model complexity after the classification stage, which is different from the feature selection method used in most of the literature. We removed the unimportant features one by one, and then recalculated the overall accuracy, where the importance of features was accessed based on the measure “weight”, which represents the number of times a feature is used to split the data across all trees. Therefore, the important features are features that are used frequently to split the data during the process of model construction. It was also able to discriminate between different transportation modes. In Figure 12, the red line shows how the number of features removed by the ensemble methods impacted on overall accuracy, while the green line shows how the number of features reduced by the PCA method, In Figure 12, it can be obviously seen that feature reduction method based on ensemble methods performed better the PCA method. Ensemble methods can still achieve highly predictive ability even when nearly 60 unimportant features are eliminated. Thus, 51 features can meet the algorithm to identifying the transportation modes. Figure 13 shows the 25 most important features with different classification tasks. Items in red represent local features and blue items represent global features. We found that important features contained both global and local features, which proves that both kinds of features play an important role in transportation mode recognition task. In Figure 13, we also found that important features contained a percentage of acceleration decomposition class which was labeled “high sinuosity and low deviation”, and follows upper quartiles of the speed and minimum value of acceleration.

5. Conclusions

An approach based on ensemble learning was developed to infer hybrid transportation modes that included walking, cycling, driving a car and taking a subway using only GPS data. To obtain a better performance, we first combined the global features generated by Statistics method and the local features extracted by trajectory segmentation. Second, we used tree-based ensemble models to classify the different transportation modes. These experiments revealed that a combination of global and local features made sense when detecting modes of transport. The experiments also proved that tree-based ensemble models performed better than the traditional methods (KNN, DT and SVM), using different evaluation techniques such as the F-Score, ROC curve, AUC and Confusion Matrix. Finally, the XGBoost model was found to the best out of the evaluation methods, with the highest accuracy from five-fold Cross-Validation of 90.77%. Furthermore, we used tree-based ensemble methods to conduct feature selection, which performed better than the PCA method.
In this paper, overall misclassification rate, which is obtained from five-fold Cross-Validation, is used for parameters tuning, and other measures are available as well, but due to the limitation of space, they are not introduced in detail. In order to process the imbalanced data, several methods can be applied in the future such as sampling methods (oversampling and undersampling) and weighted-learning. Future work will investigate how to explain the ensemble model more intuitively, as the ensemble method is like a black box and is not easy to intuitively explain. Another investigation needs to pay attention to how to integrate the data collected from different sensors or how to use GIS acquired information if necessary. Furthermore, approaches on how to segment the trajectory with only one transportation mode before the classify stage, as trajectories may contain several transportation modes at the same time, should be developed. Some deep learning algorithms could support feasible solution in terms of identifying transportation modes.

Acknowledgments

This paper is supported by National Natural Science Foundation of China (No. 41501485).

Author Contributions

Zhibin Xiao and Kun Fu conceived and designed the experiments; Zhibin Xiao and Fan Wu performed the experiments; Zhibin Xiao and Yang Wang analyzed the data; Zhibin Xiao wrote the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xiao, Y.; Low, D.; Bandara, T.; Pathak, P.; Lim, H.B.; Goyal, D.; Santos, J.; Cottrill, C.; Pereira, F.; Zegras, C. Transportation activity analysis using smartphones. In Proceedings of the 2012 IEEE Consumer Communications and Networking Conference (CCNC), Las Vegas, NV, USA, 7–10 January 2012; pp. 60–61.
  2. Zhang, F.; Zhu, X.; Guo, W.; Ye, X.; Hu, T.; Huang, L. Analyzing urban human mobility patterns through a thematic model at a finer scale. ISPRS Int. J. Geo-Inf. 2016, 5, 78. [Google Scholar] [CrossRef]
  3. Bohte, W.; Maat, K. Deriving and validating trip purposes and travel modes for multi-day GPS-based travel surveys: A large-scale application in The Netherlands. Transp. Res. Part C: Emerg. Technol. 2009, 17, 285–297. [Google Scholar] [CrossRef]
  4. Asakura, Y.; Tanabe, J.; Lee, Y.H. Characteristics of positioning data for monitoring travel behaviour. In Proceedings of the 7th World Congress on Intelligent Systems, Turin, Italy, 6–9 November 2000.
  5. GPS Track Log Route Exchange Forum. Available online: http://www.gpsxchange.com (accessed on 21 February 2017).
  6. Susi, M.; Renaudin, V.; Lachapelle, G. Motion mode recognition and step detection algorithms for mobile phone users. Sensors 2013, 13, 1539–1562. [Google Scholar] [CrossRef] [PubMed]
  7. Elhoushi, M.; Georgy, J.; Noureldin, A.; Korenberg, M. Online motion mode recognition for portable navigation using low-cost sensors. Navigation 2015, 62, 273–290. [Google Scholar] [CrossRef]
  8. Parkka, J.; Ermes, M.; Korpipaa, P.; Mantyjarvi, J.; Peltola, J.; Korhonen, I. Activity classification using realistic data from wearable sensors. IEEE Trans. Inf. Technol. Biomed. 2006, 10, 119–128. [Google Scholar] [CrossRef] [PubMed]
  9. Sohn, T.; Varshavsky, A.; LaMarca, A.; Chen, M.Y.; Choudhury, T.; Smith, I.; Consolvo, S.; Hightower, J.; Griswold, W.G.; De Lara, E. Mobility detection using everyday GSM traces. In Proceedings of the International Conference on Ubiquitous Computing, Orange County, CA, USA, 17–21 September 2006; pp. 212–224.
  10. Shah, R.C.; Wan, C.-Y.; Lu, H.; Nachman, L. Classifying the mode of transportation on mobile phones using GIS information. In Proceedings of the 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing, Seattle, WA, USA, 13–17 September 2014; pp. 225–229.
  11. Gong, H.; Chen, C.; Bialostozky, E.; Lawson, C.T. A GPS/GIS method for travel mode detection in new york city. Comput. Environ. Urban Syst. 2012, 36, 131–139. [Google Scholar] [CrossRef]
  12. Jahangiri, A.; Rakha, H. Developing a support vector machine (SVM) classifier for transportation mode identification by using mobile phone sensor data. In Proceeding of the Transportation Research Board 93rd Annual Meeting, Washington, DC, USA, 12–16 January 2014.
  13. Jahangiri, A.; Rakha, H.A. Applying machine learning techniques to transportation mode recognition using mobile phone sensor data. IEEE Trans. Intell. Transp. Syst. 2015, 16, 2406–2417. [Google Scholar] [CrossRef]
  14. Zheng, Y.; Liu, L.; Wang, L.; Xie, X. Learning transportation mode from raw GPS data for geographic applications on the web. In Proceedings of the 17th International Conference on World Wide Web, Beijing, China, 21–25 April 2008; pp. 247–256.
  15. Zheng, Y.; Li, Q.; Chen, Y.; Xie, X.; Ma, W.-Y. Understanding mobility based on GPS data. In Proceedings of the 10th international conference on Ubiquitous computing, Seoul, Korea, 21–24 September 2008; pp. 312–321.
  16. Nick, T.; Coersmeier, E.; Geldmacher, J.; Goetze, J. Classifying means of transportation using mobile sensor data. In Proceedings of the 2010 International Joint Conference on Neural Networks (IJCNN), Barcelona, Spain, 18–23 July 2010; pp. 1–6.
  17. Bolbol, A.; Cheng, T.; Tsapakis, I.; Haworth, J. Inferring hybrid transportation modes from sparse GPS data using a moving window SVM classification. Comput. Environ. Urban Syst. 2012, 36, 526–537. [Google Scholar] [CrossRef]
  18. Bennett, J.; Lanning, S. The netflix prize. In Proceedings of KDD Cup and Workshop, San Jose, CA, USA, 12 August 2007; p. 35.
  19. Chen, T.; Guestrin, C. XGBoost: A scalable tree boosting system. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016.
  20. Yu, H.-F.; Lo, H.-Y.; Hsieh, H.-P.; Lou, J.-K.; McKenzie, T.G.; Chou, J.-W.; Chung, P.-H.; Ho, C.-H.; Chang, C.-F.; Wei, Y.-H. Feature engineering and classifier ensemble for KDD cup 2010. In Proceedings of the KDD Cup 2010 Workshop, Washington, DC, USA, 25–28 July 2010; pp. 1–16.
  21. Zhou, Z.-H. Ensemble Methods: Foundations and Algorithms; CRC Press: Boca Raton, FL, USA, 2012. [Google Scholar]
  22. Mardia, K.V.; Jupp, P.E. Directional Statistics; John Wiley & Sons: New York, NY, USA, 2009. [Google Scholar]
  23. Dodge, S.; Weibel, R.; Forootan, E. Revealing the physics of movement: Comparing the similarity of movement characteristics of different types of moving objects. Comput. Environ. Urban Syst. 2009, 33, 419–434. [Google Scholar] [CrossRef] [Green Version]
  24. Stenneth, L.; Wolfson, O.; Yu, P.S.; Xu, B. Transportation mode detection using mobile phones and GIS information. In Proceedings of the 19th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, Chicago, IL, USA, 1–4 November 2011; pp. 54–63.
  25. Reddy, S.; Mun, M.; Burke, J.; Estrin, D.; Hansen, M.; Srivastava, M. Using mobile phones to determine transportation modes. ACM Trans. Sens. Netw. (TOSN) 2010, 6, 13. [Google Scholar] [CrossRef]
  26. Gonzalez, P.A.; Weinstein, J.S.; Barbeau, S.J.; Labrador, M.A.; Winters, P.L.; Georggi, N.L.; Perez, R. Automating mode detection for travel behaviour analysis by using global positioning systemsenabled mobile phones and neural networks. IET Intell. Transp. Syst. 2010, 4, 37–49. [Google Scholar] [CrossRef]
  27. Gonzalez, P.; Weinstein, J.; Barbeau, S.; Labrador, M.; Winters, P.; Georggi, N.L.; Perez, R. Automating mode detection using neural networks and assisted GPS data collected using GPS-enabled mobile phones. In Proceedings of the 15th World Congress on Intelligent Transportation Systems, New York, NY, USA, 16–20 November 2008.
  28. Biljecki, F.; Ledoux, H.; Van Oosterom, P. Transportation mode-based segmentation and classification of movement trajectories. Int. J. Geogr. Inf. Sci. 2013, 27, 385–407. [Google Scholar] [CrossRef]
  29. Das, R.D.; Winter, S. Detecting urban transport modes using a hybrid knowledge driven framework from GPS trajectory. ISPRS Int. J. Geo-Inf. 2016, 5, 207. [Google Scholar] [CrossRef]
  30. Bashir, F.I.; Khokhar, A.A.; Schonfeld, D. Object trajectory-based activity classification and recognition using hidden markov models. IEEE Trans. Image Process. 2007, 16, 1912–1919. [Google Scholar] [CrossRef] [PubMed]
  31. Widhalm, P.; Nitsche, P.; Brändie, N. Transport mode detection with realistic smartphone sensor data. In Proceedings of the 21st International Conference on Pattern Recognition (ICPR), Tsukuba, Japan, 11–15 November 2012; pp. 573–576.
  32. Nitsche, P.; Widhalm, P.; Breuss, S.; Maurer, P. A strategy on how to utilize smartphones for automatically reconstructing trips in travel surveys. Proced.-Soc. Behav. Sci. 2012, 48, 1033–1046. [Google Scholar] [CrossRef]
  33. Endo, Y.; Toda, H.; Nishida, K.; Kawanobe, A. Deep feature extraction from trajectories for transportation mode estimation. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining, Auckland, New Zealand, 19–22 April 2016; pp. 54–66.
  34. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  35. Hansen, L.K.; Salamon, P. Neural network ensembles. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 993–1001. [Google Scholar] [CrossRef]
  36. Schapire, R.E. The strength of weak learnability. Mach. Learn. 1990, 5, 197–227. [Google Scholar] [CrossRef]
  37. Liaw, A.; Wiener, M. Classification and regression by randomforest. R News 2002, 2, 18–22. [Google Scholar]
  38. Complete Guide to Parameter Tuning in XGBoost (with codes in python). Available online: https://www.analyticsvidhya.com/blog/2016/03/complete-guide-parameter-tuning-XGBoost-with-codes-python/ (accessed on 21 February 2017).
Figure 1. Flow chart outlining the methodology for the classification of different transportation modes.
Figure 1. Flow chart outlining the methodology for the classification of different transportation modes.
Ijgi 06 00057 g001
Figure 2. The distribution of global features and local features: (a) average speed; (b) heading change rate; (c) skewness of speed; (d) autocorrelation coefficient of speed; (e) average length of class three for accelerate decompose; and (f) percentage of class one for accelerate decompose.
Figure 2. The distribution of global features and local features: (a) average speed; (b) heading change rate; (c) skewness of speed; (d) autocorrelation coefficient of speed; (e) average length of class three for accelerate decompose; and (f) percentage of class one for accelerate decompose.
Ijgi 06 00057 g002
Figure 3. A common ensemble classifier.
Figure 3. A common ensemble classifier.
Ijgi 06 00057 g003
Figure 4. Visualization of one example from the Geolife dataset.
Figure 4. Visualization of one example from the Geolife dataset.
Ijgi 06 00057 g004
Figure 5. The KNN model: (a) the impact of number of neighbors on misclassification error; and (b) the ROC curve of different classes.
Figure 5. The KNN model: (a) the impact of number of neighbors on misclassification error; and (b) the ROC curve of different classes.
Ijgi 06 00057 g005
Figure 6. The DT model: (a) the impact of the maximum depth on the misclassification error; and (b) the ROC curve of different classes.
Figure 6. The DT model: (a) the impact of the maximum depth on the misclassification error; and (b) the ROC curve of different classes.
Ijgi 06 00057 g006
Figure 7. The SVM model: (a) the impact of regularization parameter and the Gaussian parameter impact on the misclassification error; and (b) the ROC curve of different classes.
Figure 7. The SVM model: (a) the impact of regularization parameter and the Gaussian parameter impact on the misclassification error; and (b) the ROC curve of different classes.
Ijgi 06 00057 g007
Figure 8. The Random Forest model: (a) the impact of the number of tress on the misclassification error; (b) the impact of the number of features on the misclassification error; and (c) the ROC curve of different classes.
Figure 8. The Random Forest model: (a) the impact of the number of tress on the misclassification error; (b) the impact of the number of features on the misclassification error; and (c) the ROC curve of different classes.
Ijgi 06 00057 g008
Figure 9. The Gradient boosting Decision Tree model: (a) the impact of the number of tress on the misclassification error; (b) the impact of the number of features on the misclassification error; and (c) the ROC curve of different classes.
Figure 9. The Gradient boosting Decision Tree model: (a) the impact of the number of tress on the misclassification error; (b) the impact of the number of features on the misclassification error; and (c) the ROC curve of different classes.
Ijgi 06 00057 g009
Figure 10. The XGBoost model: (a) the impact of the number of tress on the misclassification error; (b) the impact of the number of features on the misclassification error; (c) the ROC curve of different classes.
Figure 10. The XGBoost model: (a) the impact of the number of tress on the misclassification error; (b) the impact of the number of features on the misclassification error; (c) the ROC curve of different classes.
Ijgi 06 00057 g010aIjgi 06 00057 g010b
Figure 11. Model comparison: (a) the F-SCORE of different models; (b) the ROC curve of different model; and (c) the train time of different models.
Figure 11. Model comparison: (a) the F-SCORE of different models; (b) the ROC curve of different model; and (c) the train time of different models.
Ijgi 06 00057 g011
Figure 12. The impact of features reduction on overall accuracy.
Figure 12. The impact of features reduction on overall accuracy.
Ijgi 06 00057 g012
Figure 13. The feature importance of tree-based ensemble models.
Figure 13. The feature importance of tree-based ensemble models.
Ijgi 06 00057 g013
Table 1. Summary of past research.
Table 1. Summary of past research.
StudiesModesData TypeMethodAccuracy
Yu Zheng [2,15]Car, Walk, Bus, BikeGPSDT, BYS, SVM, CRF76.2%
Threresa Nick [16]Car, Train, WalkAccelerometerNB, SVM97.3%
Gong, H. [11]Walk, Car, Subway, Bus, Commuter railGPS, GISRules82.6%
Leon Stenneth [24]Bus, Car, Train, Walk, Bike, StandGPS, GISNB, BN, DT, RF93.5%
S. Reddy [25]Stationary, Run, Walk, BikeAccelerometerHMM82.14%
P.A. Gonzalez [26,27]Bus, Car, WalkGPSNeural networks91.23%
Adel Bolbol [17]Walk, Car, Train, Bicycle Bus, SubwayGPSSVM88%
Widhalm, P. [31]Bus, Car, Bike, Tram, Train, Subway, Walk, MotorcycleGPS, AccelerometerHMM75.8%
Philippe Nitsche [32]Walk, Car, Train, Bus, Bike, SubwayGPS, AccelerometerMultivariate parametric78.8%
Arash Jahangiri [12,13]Car, Bike, Walk, Run, BusAccelerometer, gyroscope, rotation vectorSVM, KNN, DT, Bag, RF95.1%
Table 2. Summary of Feature Extraction.
Table 2. Summary of Feature Extraction.
DescriptorsFeature DescriptionNumber
GlobalMean, Standard deviation, Mode, Top three, Min three, Ranges, 25th percentile, 75th percentile, Interquartile range, Skewness, Kurtosis, Coefficient of variation, Autocorrelation coefficient, each movement parameter (4)17 × 4 = 68
Heading change rate, Stop rate, Velocity change rate, Trajectory length4
LocalMean and Standard deviation of segment length, each decomposition class (4), each movement parameter (3)
Number of decomposition class changes (4)
Percentage of each decomposition class (4), each movement parameter (3)
39
Table 3. The four categories of classification problems.
Table 3. The four categories of classification problems.
True ConditionPredicted Condition
PositiveNegative
PositiveTrue Positives(TP)True Negatives(FN)
NegativeFalse Positives(FP)False Negatives(TN)
Table 4. The distribution of different transportation modes.
Table 4. The distribution of different transportation modes.
Transportation ModesTrajectory Number
Walk3043
Bus and Taxi2158
Bike1413
Car717
Subway519
Train135
Table 5. The confusion matrix of the KNN models.
Table 5. The confusion matrix of the KNN models.
KNNActualPrecision
WalkBus and TaxiBikeCarSubwayTrain
PredictedWalk90.143.8219.51.494.720.0087.49
Bus and Taxi5.2078.7526.5129.833.787.5065.19
Bike4.446.7251.111.003.385.0070.06
Car0.116.422.0961.1912.1620.0061.19
Subway0.114.130.473.4844.602.5063.46
Train0.000.150.232.991.3565.0072.22
Recall90.1478.7551.1161.1944.6065.00
Table 6. The confusion matrix of the DT models.
Table 6. The confusion matrix of the DT models.
DTActualPrecision
WalkBus and TaxiBikeCarSubwayTrain
PredictedWalk92.313.846.281.008.110.0092.81
Bus and Taxi4.1286.484.6523.3814.1910.081.24
Bike2.381.0887.441.491.350.0091.71
Car0.336.411.1671.148.7815.0068.10
Subway0.872.300.002.4967.577.5076.34
Train0.000.150.470.500.0067.5087.10
Recall92.3186.4887.4471.1467.5767.50
Table 7. The confusion matrix of the SVM models.
Table 7. The confusion matrix of the SVM models.
SVMActualPrecision
WalkBus and TaxiBikeCarSubwayTrain
PredictedWalk94.804.749.071.497.430.0091.24
Bus and Taxi3.0388.535.1228.368.7815.0082.13
Bike1.840.6184.880.500.680.0094.07
Car0.113.980.4768.162.705.0079.65
Subway2.171.990.471.0079.732.50085.51
Train0.000.150.000.500.6877.5091.18
Recall94.8088.5084.8868.1279.7377.50
Table 8. The confusion matrix of the Random Forest models.
Table 8. The confusion matrix of the Random Forest models.
Random ForestActualPrecision
WalkBus and TaxiBikeCarSubwayTrain
PredictedWalk92.963.592.990.000.740.0095.99
Bus and Taxi2.6284.790.7513.375.265.2690.37
Bike2.732.3095.771.740.000.0089.53
Car0.216.890.5084.302.632.6372.14
Subway1.471.870.000.5888.152.6380.41
Train0.000.570.000.001.4889.4785.00
Recall92.9684.7995.7784.3088.1589.47
Table 9. The confusion matrix of the GBDT models.
Table 9. The confusion matrix of the GBDT models.
GBDTActualPrecision
WalkBus and TaxiBikeCarSubwayTrain
PredictedWalk93.653.552.960.000.750.0095.88
Bus and Taxi2.5484.960.7411.375.972.6391.59
Bike2.332.2796.050.600.000.0090.89
Car0.216.810.2586.832.245.2672.14
Subway1.271.840.000.6090.302.6380.41
Train0.000.570.000.600.7589.4785.00
Recall93.6584.9696.0586.8390.3089.47
Table 10. The confusion matrix of the XGBoost models.
Table 10. The confusion matrix of the XGBoost models.
XGBoostActualPrecision
WalkBus and TaxiBikeCarSubwayTrain
predictedWalk93.963.452.240.001.360.0096.21
Bus and Taxi2.5485.240.5010.785.525.5590.83
Bike2.332.1697.080.600.000.0091.10
Car0.216.900.2587.432.040.0073.00
Subway0.951.580.000.6089.712.7785.14
Train0.000.570.000.600.1491.6782.50
Recall93.9685.2497.0887.4389.7191.67
Table 11. The misclassification error of some feature combination.
Table 11. The misclassification error of some feature combination.
FeaturesXGBoostGBDTRFSVMDTKNN
Local0.25500.25110.25740.31860.33310.4381
Global0.10030.09940.12600.16320.14330.3211
Zheng [15]0.12600.14650.18340.28780.19830.3996
Dodge [23]0.11950.15100.19240.19960.19460.4335
Global+Local0.09230.09540.09720.12310.13950.2556

Share and Cite

MDPI and ACS Style

Xiao, Z.; Wang, Y.; Fu, K.; Wu, F. Identifying Different Transportation Modes from Trajectory Data Using Tree-Based Ensemble Classifiers. ISPRS Int. J. Geo-Inf. 2017, 6, 57. https://doi.org/10.3390/ijgi6020057

AMA Style

Xiao Z, Wang Y, Fu K, Wu F. Identifying Different Transportation Modes from Trajectory Data Using Tree-Based Ensemble Classifiers. ISPRS International Journal of Geo-Information. 2017; 6(2):57. https://doi.org/10.3390/ijgi6020057

Chicago/Turabian Style

Xiao, Zhibin, Yang Wang, Kun Fu, and Fan Wu. 2017. "Identifying Different Transportation Modes from Trajectory Data Using Tree-Based Ensemble Classifiers" ISPRS International Journal of Geo-Information 6, no. 2: 57. https://doi.org/10.3390/ijgi6020057

APA Style

Xiao, Z., Wang, Y., Fu, K., & Wu, F. (2017). Identifying Different Transportation Modes from Trajectory Data Using Tree-Based Ensemble Classifiers. ISPRS International Journal of Geo-Information, 6(2), 57. https://doi.org/10.3390/ijgi6020057

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop