Next Article in Journal
The Status of Environmental Electric Field Detection Technologies: Progress and Perspectives
Previous Article in Journal
Three-Dimensional Posture Estimation of Vehicle Occupants Using Depth and Infrared Images
Previous Article in Special Issue
Evaluation of Roadside LiDAR-Based and Vision-Based Multi-Model All-Traffic Trajectory Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Vehicle Localization Using Crowdsourced Data Collected on Urban Roads

Department of Mechanical Engineering, Korea University, Seoul 02841, Republic of Korea
*
Author to whom correspondence should be addressed.
Submission received: 24 July 2024 / Revised: 24 August 2024 / Accepted: 25 August 2024 / Published: 27 August 2024

Abstract

:
Vehicle localization using mounted sensors is an essential technology for various applications, including autonomous vehicles and road mapping. Achieving high positioning accuracy through the fusion of low-cost sensors is a topic of considerable interest. Recently, applications based on crowdsourced data from a large number of vehicles have received significant attention. Equipping standard vehicles with low-cost onboard sensors offers the advantage of collecting data from multiple drives over extensive road networks at a low operational cost. These vehicle trajectories and road observations can be utilized for traffic surveys, road inspections, and mapping. However, data obtained from low-cost devices are likely to be highly inaccurate. On urban roads, unlike highways, complex road structures and GNSS signal obstructions caused by buildings are common. This study proposes a reliable vehicle localization method using a large amount of crowdsourced data collected from urban roads. The proposed localization method is designed with consideration for the high inaccuracy of the data, the complexity of road structures, and the partial use of high-definition (HD) maps that account for environmental changes. The high inaccuracy of sensor data affects the reliability of localization. Therefore, the proposed method includes a reliability assessment of the localized vehicle poses. The performance of the proposed method was evaluated using data collected from buses operating in Seoul, Korea. The data used for the evaluation were collected 18 months after the creation of the HD maps.

1. Introduction

Vehicle localization is a fundamental technology essential for applications such as autonomous driving and HD map creation [1]. Localization methods vary depending on the vehicle type, sensor configuration, sensor performance, environment, and specific application [2]. Research on achieving high accuracy and precision in vehicle localization using low-cost sensors remains ongoing [3,4,5,6]. Typically, low-cost sensors are integrated into an onboard device module with embedded processors. This device usually consists of GNSS, an IMU, and cameras, while LiDAR sensors are not utilized.
Localization methods are influenced by the driving environment, which interacts with sensor observations. For example, on highways, the environment is quite open, resulting in minimal GNSS signal interference. The surrounding features are not distinctive, consisting mainly of guardrails and parallel lane markings [7]. In contrast, urban roads often suffer from GNSS signal interference due to surrounding buildings. The lane markings in urban areas reflect complex road structures and intersections, leading to a high density of traffic lights and signs. Environmental characteristics significantly influence the methods for utilizing HD maps and processing sensor data. In this work, we use sensor data from low-cost onboard modules installed in vehicles, collected on urban roads.
Recently, solutions based on crowdsourced data have received significant attention [8,9,10,11,12]. Typically, low-cost sensors are used in vehicles to collect crowdsourced data. When large amounts of data can be obtained cost-effectively, statistical analysis becomes feasible across extensive road networks. Since the statistical analysis is conducted after data acquisition, real-time processing is not required. As a result, the localization method discussed in this study benefits from back-end processing. Vehicle poses derived from crowdsourced data-based localization can be utilized for applications such as HD map updates [8,10,13], road inspections [14], vehicle trajectory analyses [15], and road surveillance [16]. All these applications are influenced by localization errors. The data collected by low-cost sensors generally have high inaccuracy, which negatively impacts localization performance. Research on crowdsourced data-based HD map updates [10] relies on real-time localization by onboard sensors to compare HD maps with observed landmarks. In this context, back-end pose correction is expected to improve map update performance [10].
Offline localization is related to the back end of Simultaneous Localization and Mapping (SLAM). A representative example is SLAM using vehicles equipped with a Mobile Mapping System (MMS) [17]. However, MMS-based methods are not suitable for the problem addressed in this paper, as they typically rely on high-precision data collected from a single vehicle. In contrast, crowdsourced data-based mapping includes a localization scheme that leverages data from multiple vehicles. For instance, graph-based SLAM using sensor data from numerous vehicles on highways was proposed in [18,19]. However, localization methods designed for highway data are challenging to apply directly to urban road scenarios. While successful localization on limited urban roads was demonstrated in [9], our work evaluates localization performance across a broader range of urban roads.
The proposed localization method is designed with consideration for the inaccuracy of crowdsourced data and the environmental characteristics of urban roads. Traffic landmarks recognized by the camera are used for map matching at lane-level HD maps. These lane-level HD maps serve as a layer within the overall HD map for landmarks such as traffic lights, signs, and lane markings. Traffic landmark maps are utilized in vehicle localization through map matching of landmark observations [20,21]. Map matching requires the map and the environment to be identical. Therefore, the proposed method uses the map selectively, based on the reliability assessment of map matching. To take advantage of back-end processing, different localization methods are strategically selected for various sections, rather than relying on sequential tracking. The localization methods in this strategy include bidirectional local tracking based on map matching and factor graph optimization. After localization, a comparison of vehicle poses with odometry measurements is conducted. Additionally, vehicle pose clustering using large amounts of data is employed to detect outliers. As a result, reliable vehicle poses are obtained through localization and diagnostics based on extensive data. The effectiveness of the proposed method was verified through a quantitative evaluation of landmark mapping errors. The evaluation data were collected from numerous buses traveling on various urban roads.
The structure of this paper is as follows. Section 2 discusses localization based on crowdsourced data. Section 3 describes diagnostic methods used for vehicle poses after localization. Section 4 validates the effectiveness of the proposed method through the application of extensive real-world driving data. Finally, Section 5 presents the conclusions.

2. Localization Based on Crowdsourced Data

2.1. Crowdsourced Data and HD Maps

This section describes the structure of crowdsourced data and the HD maps used for localization. The crowdsourced data in this study are collected from devices equipped with low-cost cameras and GNSS as part of onboard sensor modules. The devices provide refined information processed internally. The information consists of global 3D poses p , odometry related to the 3D poses u , and observations of traffic landmarks z .
The collected data are synchronized on keyframes. Keyframes are generated at regular intervals during travel. In this study, the interval is approximately 5 m. Therefore, the traversal of a single vehicle can be represented as a set of keyframes { k 1 , , k N k } , where N k is the number of keyframes for a single traversal of the vehicle and is proportional to the travel distance. The information contained in keyframe i is expressed as follows:
k i = { p i , u i , z i , Σ p i , Σ u i , Σ z i }
In Equation (1), Σ p , Σ u , and Σ z represent the uncertainty covariance matrices corresponding to p , u , and z , respectively.
Traffic landmarks are recognized by the device’s embedded system through semantic segmentation and are represented relative to the device’s local frame. The 3D positions of traffic landmarks within camera images are obtained by tracking the same landmark during travel. By tracking a single landmark across multiple keyframes, observations are stored only in the keyframe with the lowest observational uncertainty. Traffic signs and lights are represented as 3D points at the centers of features and are defined as discrete landmarks. Lane markings are represented as 3D point sets for lines and are defined as continuous landmarks. In the recognition of dashed lane markings, the point sets for continuous landmarks are modeled to connect the empty spaces continuously, as with solid lines. In summary, the observed sets of discrete and continuous landmarks are denoted as z d and z c , respectively, and the observations are expressed as z = { z d , z c } . Similarly, the uncertainty covariance terms are represented as Σ z = { Σ z d , Σ z c } . The observation uncertainty covariance of continuous landmarks is assigned individually to all points that constitute the landmark.
The accuracy and precision of p estimated in real time by onboard devices are not guaranteed. Figure 1 visually illustrates the localization errors of p across multiple traversals on a specific urban road. As shown in Figure 1a, there are buildings surrounding the roads that obstruct GNSS signals. A comparison between Figure 1b and Figure 1c reveals that p frequently deviates from the ground-truth driving lane. There is a high probability that the significant bias errors in p are due to GNSS multipath errors. Such errors are difficult to represent with Σ p . Therefore, it is important to correct p to account for data uncertainty and environmental characteristics.
The proposed method partially utilizes HD maps. To match the observed landmarks in crowdsourced data, feature maps corresponding to the types of observed landmarks in the HD map are used. The feature maps are categorized into discrete landmarks of 3D points M d and continuous landmarks of 3D point sets M c , identical to the classification of observed landmark types. Therefore, the HD map M used in the proposed method can be expressed as follows:
M = { M d , M c }
The structure of the crowdsourced data utilized in this study is atypical. The crowdsourced data are identical to those used in previous research [10] and similar to those used in [8].

2.2. Bidirectional Local Tracking with Map Matching

This section introduces map matching-based localization for a partial set of keyframes in a single traversal. The proposed localization is a process to correct p , which is subject to high inaccuracies. Pose correction relative to the map frame can utilize landmarks such as traffic signs, lights, and lane markings in the HD map or links to the driving lane. However, there are cases where the lateral error of p is larger than the interval between adjacent lanes. In such cases, it becomes difficult to estimate the driven lane due to the similar geometric characteristics of the lane markings. Therefore, instead of uniformly correcting all keyframes in the same manner, an appropriate localization method is applied to each keyframe according to observations and map-matching states.
To achieve reliable localization within the map frame, accurate matching of observed landmarks is essential. ICP-based nearest-point matching can be utilized to find correspondences for landmarks [22]. However, nearest-point matching can cause p to converge on the incorrect lane. Therefore, we begin by identifying a single keyframe where reliable map matching is possible. To estimate a stable 3D pose through map matching, we focus on a keyframe where at least one discrete landmark and two continuous landmarks are recognized. If only continuous landmarks are recognized, correcting the direction of travel becomes difficult. Since traffic signs and lights are located at specific points along the road, the number of keyframes that satisfy the landmark observation condition is limited.
To find accurate landmark correspondences based on the inaccurate p , costs are assigned and evaluated for all discrete landmark correspondences within a certain range. Considering the inaccuracy of p , HD map landmarks of the same type as the observed discrete landmark within a distance d g are selected as matching candidates. The distance d g is empirically set based on the maximum positional error of p . The matching candidates M c a n d ( z d ) for a single observed discrete landmark z d ( z d z d ) in keyframe i are expressed as follows:
M c a n d ( z d ) = { m d | | | p i z d m d | | < d g , m d M d }
In Equation (3), m d represents a discrete landmark included in the HD map. The operator ⊕ denotes pose composition. Each element of the matching candidate M c a n d ( z d ) is evaluated by a matching cost. The matching cost is calculated based on the nearest-point match for landmarks after an initial translation t 0 by candidate correspondences. The initial translation t 0 represents the positional error between z d and the candidate in the map frame. p i becomes p i through the translation by t 0 and is expressed as follows:
t 0 = p i z d M c a n d ( z d )
p i = p i [ R 0 , t 0 ] ( R 0 = I 3 )
In Equation (5), R 0 is the identity matrix for three-dimensional rotation. The correspondence of continuous landmarks and other discrete landmarks depends on the nearest-point matching at p i . From this correspondence, error vectors for discrete landmarks e d = ( p i z d ) m d and continuous landmarks e c = ( p i z c ) m c can be derived within the map frame. The matching cost of landmarks f z ( p i ) is expressed as the weighted average of the Mahalanobis distances, as follows:
f z ( p i ) = w d n d e d T ( R i Σ z d ( R i ) T ) 1 e d + w c n c e c T ( R i Σ z c ( R i ) T ) 1 e c
In Equation (6), R i , w d , w c , n d , and n c represent the rotation matrix SO ( 3 ) of p i , the weights of discrete and continuous landmarks, the number of matched discrete landmarks, and the number of points in matched continuous landmarks, respectively. The matching cost f z ( p i ) is minimized through iterative optimization based on the Levenberg–Marquardt method [23]. After optimization, the optimal vehicle pose p i c a n d is estimated from the candidate correspondences.
For all observed discrete landmarks, matching candidates are selected, and the matching costs f z ( p i c a n d ) after optimization for each candidate are derived. If errors are minimized by incorrect landmark correspondences, f z ( p i c a n d ) is typically abnormally high. To assess the validity of each candidate, the maximum value of f z ( p i c a n d ) for correct matching is empirically set as f z , v a l i d . Therefore, a candidate with an f z ( p i c a n d ) less than f z , v a l i d is considered valid. However, the number of candidates that satisfy the validity criterion varies depending on the conditions. If no valid candidates are found, it is likely due to misdetected landmarks, high measurement uncertainty, or the detection of newly created landmarks not present on the map. Conversely, if more than two valid candidates exist, it is likely due to the similarity of road structures between adjacent lanes. In such cases, the reliability of candidate matching is low. Therefore, the optimal pose p i * is determined only when a single valid matching candidate exists. p i * is expressed by the following equation, which minimizes the matching cost:
p i * = argmin p i c a n d { f z ( p i c a n d ) }
If reliable localization based on discrete landmark matching candidates is achieved at keyframe i, reliable pose estimation using odometry is possible for the adjacent areas of keyframe i. By taking advantage of the offline system, odometry-based estimation can be performed in both the forward and reverse directions of the vehicle. The predicted poses of adjacent keyframes, estimated based on odometry from keyframe i, can be expressed by the following equation:
p i + 1 = p i * u i
p i 1 = p i * u i 1 1
Equation (8) represents the estimation in the direction of the vehicle’s movement. In contrast, Equation (9) indicates the estimation in the opposite direction of the vehicle. The term u i 1 1 represents the inverse transformation of u i 1 . From the odometry-based predicted pose p , the optimal pose is derived through corrections that include map matching. The correspondences for map matching are determined by nearest-point matching with respect to the predicted pose. The cost based on the landmark matching error follows Equation (6). Additionally, the odometry-based predicted position is incorporated into the cost through point-to-point matching. The odometry-based predicted position serves as a constraint in the direction of the vehicle’s movement. The odometry-based position error vector e u is expressed as follows:
e u = ( p i + 1 Δ T ) t ( p i + 1 ) t
In Equation (10), ( · ) t refers to the selection of the translation term from either the pose vector or the covariance matrix. Δ T is the transformation that corrects the predicted pose p i + 1 in the iterative optimization process of reducing the cost function. Consequently, the local tracking cost function f L ( p i + 1 ) is expressed as follows:
f L ( p i + 1 ) = f z ( p i + 1 ) + w u e u T ( R i ( Σ u i ) t ( R i ) T ) 1 e u
In Equation (11), R i and w u represent the rotation matrix SO ( 3 ) of p i and the weight for the odometry-based predicted position, respectively. The cost function f L ( p i + 1 ) is minimized using the Levenberg–Marquardt method [23], resulting in the optimal pose p i + 1 * . For local tracking in the opposite direction of the vehicle’s movement, in Equations (10) and (11), p i + 1 and Σ u i are replaced with p i 1 and Σ u i 1 1 , respectively.
The vehicle poses of adjacent keyframes in both directions are estimated sequentially from the estimated keyframe i in Equation (7). However, localization confidence can be reduced by factors such as a lack of observed landmarks, high measurement uncertainties, and environmental changes. In these cases, confidence can be assessed by the number of observed continuous landmarks and the cost after optimization f L ( p * ) in local tracking. For stable 3D pose estimation, at least two continuous landmarks are recognized, and the cost f L ( p * ) is lower than the empirically set threshold f L , v a l i d . However, the conditions required for reliable localization may not be met under temporary occlusions by other vehicles or short-term environmental changes. To maximize the number of keyframes estimated through local tracking, temporary low-confidence pose estimations are replaced with odometry-based predictions. If localization confidence is not maintained for a certain number of consecutive keyframes, local tracking is considered a failure.
Due to the characteristics of the road structure, only parallel lane markings are observed in straight road sections without traffic signs or lights. In this case, even if f L ( p * ) remains consistently low during local tracking over successive keyframes, drift error in the direction of the vehicle’s movement accumulates. Therefore, if drift error correction is not achieved through map matching of discrete landmarks, local tracking is terminated based on the magnitude of the propagating position uncertainty covariance matrix from odometry. The odometry-based propagated position uncertainty covariance matrix Σ i + k for consecutive keyframes from i to i + k is expressed as follows:
Σ i + k = Σ i + k 1 + R i + k 1 ( Σ u i + k 1 ) t ( R i + k 1 ) T ( Σ i = O 3 × 3 )
According to Equation (12), local tracking at keyframe i includes map matching with discrete landmarks. The decision to terminate local tracking is based on the sum of the diagonal elements of the propagated uncertainty covariance matrix, t r a c e ( Σ i + k ) . Bidirectional local tracking from a single reference keyframe generates two termination keyframes. The keyframes between these two termination keyframes can be considered as having reliable localization with map matching applied. By repeatedly applying bidirectional local tracking to the remaining unestimated keyframes, keyframes with reliable map matching-based estimations can be separated from the entire keyframe set.

2.3. Local Pose Graph Optimization

This section addresses the localization of keyframes that are not estimated by the bidirectional local tracking process. Since unestimated keyframes are likely to be difficult for map matching, we applied pose graph-based optimization without map matching. The keyframes successfully estimated during bidirectional local tracking can be used as fixed nodes in the pose graph. Therefore, instead of generating a pose graph for the entire set of keyframes from a single traversal, multiple local pose graphs were individually created and optimized based on the fixed nodes.
Figure 2 shows an example of the range within the keyframe set of a single traversal where local pose graphs are generated. Two local pose graphs are created based on the termination keyframes of bidirectional local tracking. The nodes of the local pose graph include two fixed nodes and p . Given the interval of keyframes from i to i + n , the initial node set of the local pose graph is denoted as X = { x i , x i + 1 , , x i + n 1 , x i + n } . It can be expressed as follows:
X = { p i * , p i + 1 , , p i + n 1 , p i + n * }
In Equation (13), x i and x i + n correspond to the fixed nodes p i * and p i + n * , respectively. The edge constraints of the local pose graph include the error between the odometry measurement and the transformation of successive nodes. The edge constraint for frame j to j + 1 is represented as e j , j + 1 . Therefore, the error vector based on the odometry measurement e j , j + 1 and the corresponding information matrix Ω j , j + 1 is expressed as follows:
e j , j + 1 = u j 1 ( x j 1 x j + 1 ) ( i j < i + n )
Ω j , j + 1 = ( Σ u j ) 1
In Equation (14), the expression for the error vector follows that in [24]. The odometry-based constraint smooths the keyframe pose between two fixed nodes through pose graph optimization. When optimizing the pose graph using only odometry-based constraints, drift error accumulates in the keyframe poses further from the fixed nodes as the travel distance between them increases. By adding p as an edge constraint for the global position, the drift errors can be corrected. Considering the positional inaccuracy of p , constraints for the global position are generated intermittently. As shown in Equation (12), when the uncertainty covariance of odometry-based drift errors propagated from the fixed nodes becomes larger than the inaccuracy of p , a global position constraint is added. When a global position constraint is generated at keyframe k, the error vector is expressed as a unary edge as follows:
e k , k = ( x k ) t ( p k ) t
The information matrix Ω k , k , corresponding to the error vector e k , k , is determined based on the positional inaccuracy of p . The standard deviation of the positional uncertainty σ p , is determined empirically by estimating the maximum positional error between the ground truth and p . Therefore, Ω k , k is expressed as follows:
Ω k , k = ( σ p 2 ) 1 0 0 0 ( σ p 2 ) 1 0 0 0 0
Combining the constraints from Equations (14)–(17), the cost function of the pose graph is expressed as follows:
F ( X ) = j e j , j + 1 T Ω j , j + 1 e j , j + 1 + k e k , k T Ω k , k e k , k
The cost function F ( X ) is minimized using pose graph optimization based on the g2o library [25]. After optimization, the pose graph nodes X * become the optimized keyframe poses p * for the local pose graph area.

3. Localization Diagnosis

3.1. Odometry-Based Keyframe Pose Diagnosis

The proposed localization method includes a process for assessing the reliability of localization. However, due to the similarity of road structures, bidirectional local tracking can be performed in the incorrect lane. In this case, the local pose graph optimization is affected by significant errors in the fixed nodes. Therefore, we diagnose p * by comparing the transformation between the keyframe poses after localization and the odometry.
The error vector e u * between consecutive keyframes i and i + 1 after localization is identical to the odometry constraint in the pose graph, as shown in Equation (14).
e u * = u i 1 ( ( p i * ) 1 p i + 1 * )
To diagnose the localization over a certain distance within a keyframe segment, we evaluate the average error for n consecutive keyframes. The exclusion condition for the evaluated n keyframes is expressed as follows:
1 n n ( e u * ) T ( Σ u ) 1 e u * > d o d o m
The threshold value d o d o m in Equation (20) is determined empirically. Odometry-based pose diagnosis allows for the exclusion of keyframe sets that exhibit excessive differences compared to odometry-based predictions.

3.2. Pose Clustering with Outlier Filtering

In this part, keyframe outliers are excluded through the clustering of keyframe poses. If the ratio of traversals between adjacent lanes differs significantly, the data from lanes with fewer traversals are less likely to represent the majority and are more likely to be considered outliers.
Keyframe pose clustering considers both the position and heading direction of the keyframes. The proposed method utilizes an approach that adds a vehicle heading direction metric to DBSCAN-based vehicle position clustering [26]. In the clustering of positions, the distance between clusters is set to ensure that clusters from different lanes are separated, considering the standard lane interval. Clustering based on the heading direction groups vehicles traveling in the same direction and distinguishes between adjacent lanes with opposite directions of travel or different directions at intersections.
The number of keyframes included in a cluster varies depending on the number of traversals and the presence of outliers. Evaluation for a single cluster is conducted by comparing it with clusters in adjacent areas. If the number of keyframes in the cluster is significantly smaller than in surrounding clusters, the cluster is considered an outlier. To statistically detect outliers, we utilize the MAD (Median Absolute Deviation) filter [27]. The MAD filter identifies data points that deviate significantly from the median as outliers. In our problem, only outliers with a smaller number of keyframes are selected. The scope of the MAD filter should account for the varying number of traversals for each road and lane marking. For a single cluster c i , the adjacent cluster set C M A D , i for comparison is determined by the average positional and heading direction differences between clusters. C M A D , i is expressed as follows:
C M A D , i = { c | | | C p ( c ) C p ( c i ) | | < d p , | | C h ( c C h ( c i ) ) | | < d h }
In Equation (21), C p ( c ) and C h ( c ) represent the average position and average heading direction of the keyframes contained in cluster c . d p and d h are the position and heading direction thresholds of the adjacent clusters. When the MAD filter is applied to C M A D , i , the outlier status of each element is determined. To increase the reliability of outlier detection, C M A D is generated for each cluster, and the outlier status is redundantly evaluated across all clusters. Subsequently, a single cluster receives N e v a l a t i o n outlier evaluations from adjacent clusters. If the outlier detection rate is high across these multiple evaluations, the cluster is excluded. The condition for a single cluster to be classified as an outlier is as follows:
N o u t l i e r N e v a l u a t i o n > r o u t l i e r
In Equation (22), N o u t l i e r and r o u t l i e r represent the number of outlier determinations and the criterion for the cluster exclusion rate, respectively. The outlier status is evaluated for all clusters. Finally, the keyframe poses contained in the clusters that have not been excluded are utilized.

4. Evaluation

For the evaluation, we used bus driving data from roads in Seoul, Korea. The evaluation data were collected on specific sections of the roads over two days, 18 months after the HD map was created.
Figure 3a shows a satellite image of the urban area where the evaluation dataset was collected. The data were collected from eight separate roads, with a total length of approximately 35.1 km. Multiple driving datasets were obtained from these roads, with a total driving distance of 4214 km. Figure 3b,c show examples of traffic landmarks observed by sensing devices mounted on the vehicle.
To verify the validity of the proposed localization method, we assessed the consistency between the ground-truth path and the estimated vehicle positions and driving directions. A keyframe pose estimation was considered successful if the estimated keyframe position was closest to the link of the ground-truth driving path in the HD map and the driving directions were matched. When environmental changes occurred, the HD map link was updated to reflect the conditions at the time of data acquisition and then used for evaluation. The maximum valid matching distance criterion between the keyframe and the HD map link was set to 1.5 m in accordance with lane marking width standards. For quantitative evaluation, the ratio of keyframes with successfully estimated driving paths was compared based on whether the proposed method was applied.
Table 1 presents the quantitative analysis results for the eight areas shown in Figure 3a. “Rejection” in the table indicates the ratio of keyframes excluded by the proposed method. The rejection rates varied by area, ranging from 48 % to 69 % . The variation in rejection rates was due to the challenges of localization posed by different road environments. The results indicate that even in the worst-case scenario, more than 30 % of the data were still utilized. “Raw data” and “Proposed” in the table represent the valid ratios of keyframe poses for p and the proposed method, respectively. The keyframe validity ratio for the proposed method showed no significant variation across different environments, maintaining approximately 90 % . By applying the proposed method, the keyframe validity ratio increased in all areas. Overall, the ratio of invalid keyframes in the raw data was approximately three times higher compared to the proposed method. Invalid keyframe poses are likely to cause performance degradation when used in solutions such as HD map creation or environmental change detection.
Figure 4 shows the qualitative evaluation results of localization for the road without environmental changes, the road with a newly created lane marking, and the road with a deleted lane marking. Environmental changes are represented by the creation of new landmarks not present on the map or the removal of existing landmarks from the map. By comparing Figure 4b,c, it is evident that the estimated poses mostly matched the ground-truth driving path when the proposed method was applied. The proposed method partially utilized map matching and excluded unreliable localization results. Nevertheless, as shown in Figure 4c, the mapped continuous landmarks correctly reflect the environmental changes.
Localization performance is closely related to the accuracy of mapping. To quantitatively evaluate localization, we assessed the mapping accuracy and precision of landmarks. The mapping of landmarks observed from multiple driving datasets and their correspondence with the HD map was conducted using the method described in previous research [10]. In [10], the positions of mapped discrete landmarks were estimated using Maximum A Posteriori (MAP) [28], considering the observation uncertainty for the same landmarks recognized from driving datasets. Similarly, the mapped continuous landmark is a point array obtained by clustering the lane markings detected during multiple traversals and the HD map, followed by fitting based on a Bézier curve [29].
Table 2 presents the results of the mapping error evaluation for the 198 traffic lights and signs within the sensor’s field of view. By applying the proposed method, both the accuracy and precision of the mapping were improved. Precision is measured as the average distance error between the mapped landmark and the landmark measurements.
Figure 5 shows the mapping results of the discrete landmarks. In Figure 5b, the discrete landmarks mapped by multiple vehicles display high variance with respect to the driving direction. In contrast, Figure 5c shows lower variance along with fewer outliers. Region “A” in (a) indicates a newly created traffic sign that was not included in the HD map. Figure 5c shows the observed discrete landmarks corresponding to the region, and the variance is similar to that of other discrete landmarks.
The continuous landmarks were evaluated in the same manner as the discrete landmarks. As shown in Table 3, the mapping accuracy and precision improved when using the proposed method compared to the raw data. Precision is calculated as the average distance error between the mapped continuous landmark obtained by fitting and the measured continuous landmark points.
Figure 6 shows the mapping results of the continuous landmarks. Comparing Figure 6b and Figure 6c, the thickness of the clustered lane marking measurement points for a single HD map lane marking is thinner with the proposed method than with the raw data. A thinner distribution of clustered continuous landmark points indicates a smaller lateral error in the keyframe poses. Therefore, using the vehicle pose estimated by the proposed method improves the accuracy and precision of landmark mapping. In other words, the mapping results validate the effectiveness of the proposed localization method.

5. Conclusions

In this paper, we address offline vehicle localization using a large amount of crowdsourced data. The proposed localization method is designed with consideration for the inaccuracies in crowdsourced data and the structural characteristics of urban road traffic landmarks. To account for these characteristics, we strategically apply map matching and factor graph optimization in the back end based on specific conditions. Leveraging the advantage of large datasets, inaccurate localized keyframes are excluded through a diagnostic process. This diagnostic method involves comparing the estimated keyframe poses with odometry and filtering outliers through keyframe pose clustering using a large amount of driving data. Consequently, reliable localized keyframe poses can be obtained from data collected by multiple vehicles.
The proposed method was validated using a large amount of crowdsourced data collected on urban roads. Localization was evaluated by verifying the driving path and comparing landmark mapping errors. The reliably estimated poses from multiple vehicles are expected to be applicable for map updates, vehicle trajectory estimation, mapping, and road infrastructure management.

Author Contributions

Conceptualization, S.C. and W.C.; methodology, S.C.; software, S.C.; validation, S.C. and W.C.; formal analysis, S.C.; investigation, S.C.; writing—original draft preparation, S.C.; writing—review and editing, S.C. and W.C.; visualization, S.C.; supervision, W.C.; project administration, W.C.; funding acquisition, W.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by SK Telecom Company, Ltd., and was also funded by the National Research Foundation of Korea (NRF) through the Ministry of Science and ICT (MSIT) under Grant Number NRF-2021R1A2C2007908.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, and further inquiries can be directed to the corresponding author.

Acknowledgments

The authors acknowledge Seung-Mo Cho at SK Telecom for fruitful discussions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lu, Y.; Ma, H.; Smart, E.; Yu, H. Real-time Performance-focused Localization Techniques for Autonomous Vehicle: A Review. IEEE Trans. Intell. Transp. Syst. 2021, 23, 6082–6100. [Google Scholar] [CrossRef]
  2. Chalvatzaras, A.; Pratikakis, I.; Amanatiadis, A.A. A Survey on Map-Based Localization Techniques for Autonomous Vehicles. IEEE Trans. Intell. Veh. 2023, 8, 1574–1596. [Google Scholar] [CrossRef]
  3. Suhr, J.K.; Jang, J.; Min, D.; Jung, H.G. Sensor Fusion-Based Low-Cost Vehicle Localization System for Complex Urban Environments. IEEE Trans. Intell. Transp. Syst. 2017, 18, 1078–1086. [Google Scholar] [CrossRef]
  4. Vivacqua, R.; Vassallo, R.; Martins, F. A Low Cost Sensors Approach for Accurate Vehicle Localization and Autonomous Driving Application. Sensors 2017, 17, 2359. [Google Scholar] [CrossRef] [PubMed]
  5. Pauls, J.-H.; Petek, K.; Poggenhans, F.; Stiller, C. Monocular Localization in HD Maps by Combining Semantic Segmentation and Distance Transform. In Proceedings of the 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 24 October 2020–24 January 2021; pp. 4595–4601. [Google Scholar]
  6. Wen, T.; Xiao, Z.; Wijaya, B.; Jiang, K.; Yang, M.; Yang, D. High precision vehicle localization based on tightly-coupled visual odometry and vector hd map. In Proceedings of the 2020 IEEE Intelligent Vehicles Symposium (IV), Las Vegas, NV, USA, 19 October–13 November 2020; pp. 672–679. [Google Scholar]
  7. Li, L.; Yang, M.; Li, H.; Wang, C.; Wang, B. Robust Localization for Intelligent Vehicles Based on Compressed Road Scene Map in Urban Environments. IEEE Trans. Intell. Veh. 2017, 8, 252–262. [Google Scholar] [CrossRef]
  8. Pannen, D.; Liebner, M.; Hempel, W.; Burgard, W. How to keep HD maps for automated driving up to date. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 2288–2294. [Google Scholar]
  9. Kim, C.; Cho, S.; Sunwoo, M.; Jo, K. Crowd-Sourced Mapping of New Feature Layer for High-Definition Map. Sensors 2023, 18, 4172. [Google Scholar] [CrossRef] [PubMed]
  10. Kim, K.; Cho, S.; Chung, W. HD map update for autonomous driving with crowdsourced data. IEEE Robot. Autom. Lett. 2021, 6, 1895–1901. [Google Scholar] [CrossRef]
  11. Pannen, D.; Liebner, M.; Burgard, W. Lane marking learning based on crowdsourced data. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Macau, China, 3–8 November 2019; pp. 7040–7046. [Google Scholar]
  12. Zhang, P.; Zhang, M.; Liu, J. Real-time HD map change detection for crowdsourcing update based on mid-to-high-end sensors. Sensors 2021, 21, 2477. [Google Scholar] [CrossRef] [PubMed]
  13. Zhanabatyrova, A.; Souza Leite, C.F.; Xiao, Y. Automatic Map Update Using Dashcam Videos. IEEE Internet Things J. 2023, 10, 11825–11843. [Google Scholar] [CrossRef]
  14. Mahlberg, J.A.; Li, H.; Zachrisson, B.; Leslie, D.K.; Bullock, D.M. Pavement Quality Evaluation Using Connected Vehicle Data. Sensors 2022, 22, 9109. [Google Scholar] [CrossRef] [PubMed]
  15. Dey, S.; Tomko, M.; Winter, S.; Ganguly, N. Traffic count estimation using crowd-sourced trajectory data in the absence of dedicated infrastructure. Pervasive Mob. Comput. 2024, 102, 101935. [Google Scholar] [CrossRef]
  16. Zheng, W.; Xu, H.; Li, P.; Wang, R.; Shao, X. SAC-RSM: A High-Performance UAV-Side Road Surveillance Model Based on Super-Resolution Assisted Learning. IEEE Internet Things J. 2024; early access. [Google Scholar] [CrossRef]
  17. Elhashash, M.; Albanwan, H.; Qin, R. A Review of Mobile Mapping Systems: From Sensors to Applications. Sensors 2022, 22, 4262. [Google Scholar] [CrossRef] [PubMed]
  18. Doer, C.; Henzler, M.; Messner, H.; Trommer, G.F. HD Map Generation from Vehicle Fleet Data for Highly Automated Driving on Highways. In Proceedings of the 2020 IEEE Intelligent Vehicles Symposium (IV), Las Vegas, NV, USA, 19 October–13 November 2020; pp. 2014–2020. [Google Scholar]
  19. Liebner, M.; Jain, D.; Schauseil, J.; Pannen, D.; Hackeloer, A. Crowdsourced HD map patches based on road model inference and graph-based slam. In Proceedings of the 2019 IEEE Intelligent Vehicles Symposium (IV), Paris, France, 9–12 June 2019; pp. 1211–1218. [Google Scholar]
  20. Elfring, J.; Dani, S.; Shakeri, S.; Mesri, H.; van den Brand, J.W. Vehicle localization using a traffic sign map. IEEE Trans. Intell. Transp. Syst. 2017, 1–6. [Google Scholar] [CrossRef]
  21. Choi, M.; Suhr, J.; Choi, K.; Jung, H. Low-Cost Precise Vehicle Localization Using Lane Endpoints and Road Signs for Highway Situations. IEEE Access. 2019, 7, 149846–149856. [Google Scholar] [CrossRef]
  22. Segal, A.; Hhnel, D.; Thrun, S. Generalized-ICP. In Proceedings of the Robotics: Science and Systems V (RSS), Seattle, WA, USA, 28 June–1 July 2009. [Google Scholar]
  23. Moré, J.J. The Levenberg-Marquardt Algorithm: Implementation and Theory, in Numerical Analysis; Springer: Berlin/Heidelberg, Germany, 1978; pp. 105–116. [Google Scholar]
  24. Grisetti, G.; Kummerle, R.; Stachniss, C.; Burgard, W. A tutorial on graph-based SLAM. IEEE Intell. Trans. Syst. Mag. 2010, 2, 31–43. [Google Scholar] [CrossRef]
  25. Kummerle, R.; Grisetti, G.; Strasdat, H.; Konolige, K.; Burgard, W. G2O: A general framework for graph optimization. In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011. [Google Scholar]
  26. Ester, M.; Kriegel, H.P.; Sander, J.; Xu, X. A Density-Based Algorithm for Discovering Clusters in Large Spatial Databases with Noise. In Proceedings of the Second International Conference on Knowledge Discovery and Data Mining, KDD’96, Portland, OR, USA, 2–4 August 1996; AAAI Press: Palo Alto, CA, USA, 1996; pp. 226–231. [Google Scholar]
  27. Leys, C.; Ley, C.; Klein, O.; Bernard, P.; Licata, L. Detecting outliers: Do not use standard deviation around the mean, use absolute deviation around the median. J. Exp. Soc. Psychol. 2013, 49, 764–766. [Google Scholar] [CrossRef]
  28. Kumar, M.; Garg, D.P.; Zachery, R.A. A generalized approach for inconsistency detection in data fusion from multiple sensors. In Proceedings of the 2006 American Control Conference, Minneapolis, MN, USA, 14–16 June 2006. [Google Scholar]
  29. Gonzalez, T.; Diaz-Herrera, J.; Tucker, A. Computing Handbook: Computer Science and Software Engineering, 3rd ed.; CRC Press: Boca Raton, FL, USA, 2014. [Google Scholar]
Figure 1. Localization error in p across multiple traversals. (a) Urban road image. (b) HD map with ground-truth driving lane corresponding to the driving lane in (a). (c) Trajectory of p for 172 traversals.
Figure 1. Localization error in p across multiple traversals. (a) Urban road image. (b) HD map with ground-truth driving lane corresponding to the driving lane in (a). (c) Trajectory of p for 172 traversals.
Sensors 24 05531 g001
Figure 2. Localization method for keyframes in a single traversal.
Figure 2. Localization method for keyframes in a single traversal.
Sensors 24 05531 g002
Figure 3. Visualization of data acquisition roads and sensor observations. (a) Satellite view of data acquisition roads. (b) The sensing device mounted on the vehicle and the observed traffic landmarks. (c) HD map and landmark measurements associated with a single keyframe.
Figure 3. Visualization of data acquisition roads and sensor observations. (a) Satellite view of data acquisition roads. (b) The sensing device mounted on the vehicle and the observed traffic landmarks. (c) HD map and landmark measurements associated with a single keyframe.
Sensors 24 05531 g003
Figure 4. Qualitative results of localization and mapping with environmental changes. (a) Ground-truth driving path with HD map lane markings. (b) Keyframe positions and continuous landmarks from raw data across multiple traversals. (c) The corresponding figure for (b) after applying the proposed method. Black dots represent keyframe positions, and purple dots represent continuous landmarks based on the keyframe pose.
Figure 4. Qualitative results of localization and mapping with environmental changes. (a) Ground-truth driving path with HD map lane markings. (b) Keyframe positions and continuous landmarks from raw data across multiple traversals. (c) The corresponding figure for (b) after applying the proposed method. Black dots represent keyframe positions, and purple dots represent continuous landmarks based on the keyframe pose.
Sensors 24 05531 g004
Figure 5. Comparison of discrete landmark mapping results between the raw data and the proposed method. Filled red and blue circles represent traffic signs and traffic lights on the HD map, respectively. Correspondingly, the empty red and blue circles represent the mapped observed landmarks.
Figure 5. Comparison of discrete landmark mapping results between the raw data and the proposed method. Filled red and blue circles represent traffic signs and traffic lights on the HD map, respectively. Correspondingly, the empty red and blue circles represent the mapped observed landmarks.
Sensors 24 05531 g005
Figure 6. Comparison of continuous landmark mapping results between the raw data and the proposed method. Points of the same color represent continuous landmarks clustered in the same HD map lane marking.
Figure 6. Comparison of continuous landmark mapping results between the raw data and the proposed method. Points of the same color represent continuous landmarks clustered in the same HD map lane marking.
Sensors 24 05531 g006
Table 1. Comparison of valid keyframe ratios resulting from the use of the proposed method in each evaluation area.
Table 1. Comparison of valid keyframe ratios resulting from the use of the proposed method in each evaluation area.
AreaLength of Link (km)Driving Distance (km)Rejection (%)Raw Data (%)Proposed (%)
15.8340.863.872.094.5
24.3338.467.563.990.2
310.61289.852.874.491.8
42.2945.748.475.088.3
56.0763.369.454.489.6
63.3386.051.175.692.7
72.271.857.572.298.3
80.777.762.875.191.8
Overall35.14213.557.069.990.9
Table 2. Evaluation of discrete landmark mapping: 198 traffic lights and signs were used.
Table 2. Evaluation of discrete landmark mapping: 198 traffic lights and signs were used.
MethodAccuracy (m)Precision (m)
Raw data (p)1.051.46
Proposed0.530.78
Table 3. Evaluation of continuous landmark mapping: The total length of lane markings is 122.6 km.
Table 3. Evaluation of continuous landmark mapping: The total length of lane markings is 122.6 km.
MethodAccuracy (m)Precision (m)
Raw data (p)0.290.39
Proposed0.210.21
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cho, S.; Chung, W. Vehicle Localization Using Crowdsourced Data Collected on Urban Roads. Sensors 2024, 24, 5531. https://doi.org/10.3390/s24175531

AMA Style

Cho S, Chung W. Vehicle Localization Using Crowdsourced Data Collected on Urban Roads. Sensors. 2024; 24(17):5531. https://doi.org/10.3390/s24175531

Chicago/Turabian Style

Cho, Soohyun, and Woojin Chung. 2024. "Vehicle Localization Using Crowdsourced Data Collected on Urban Roads" Sensors 24, no. 17: 5531. https://doi.org/10.3390/s24175531

APA Style

Cho, S., & Chung, W. (2024). Vehicle Localization Using Crowdsourced Data Collected on Urban Roads. Sensors, 24(17), 5531. https://doi.org/10.3390/s24175531

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop