Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (698)

Search Parameters:
Keywords = pedestrian detection

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
27 pages, 7600 KiB  
Article
Spiking Neural Networks for Real-Time Pedestrian Street-Crossing Detection Using Dynamic Vision Sensors in Simulated Adverse Weather Conditions
by Mustafa Sakhai, Szymon Mazurek, Jakub Caputa, Jan K. Argasiński and Maciej Wielgosz
Electronics 2024, 13(21), 4280; https://doi.org/10.3390/electronics13214280 - 31 Oct 2024
Viewed by 472
Abstract
This study explores the integration of Spiking Neural Networks (SNNs) with Dynamic Vision Sensors (DVSs) to enhance pedestrian street-crossing detection in adverse weather conditions—a critical challenge for autonomous vehicle systems. Utilizing the high temporal resolution and low latency of DVSs, which excel in [...] Read more.
This study explores the integration of Spiking Neural Networks (SNNs) with Dynamic Vision Sensors (DVSs) to enhance pedestrian street-crossing detection in adverse weather conditions—a critical challenge for autonomous vehicle systems. Utilizing the high temporal resolution and low latency of DVSs, which excel in dynamic, low-light, and high-contrast environments, this research evaluates the effectiveness of SNNs compared to traditional Convolutional Neural Networks (CNNs). The experimental setup involved a custom dataset from the CARLA simulator, designed to mimic real-world variability, including rain, fog, and varying lighting conditions. Additionally, the JAAD dataset was adopted to allow for evaluations using real-world data. The SNN models were optimized using Temporally Effective Batch Normalization (TEBN) and benchmarked against well-established deep learning models, concerning their accuracy, computational efficiency, and energy efficiency in complex weather conditions. This study also conducted a comprehensive analysis of energy consumption, highlighting the significant reduction in energy usage achieved by SNNs when processing DVS data. The results indicate that SNNs, when integrated with DVSs, not only reduce computational overhead but also dramatically lower energy consumption, making them a highly efficient choice for real-time applications in autonomous vehicles (AVs). Full article
(This article belongs to the Special Issue Autonomous and Connected Vehicles)
Show Figures

Figure 1

19 pages, 1689 KiB  
Article
PE-MCAT: Leveraging Image Sensor Fusion and Adaptive Thresholds for Semi-Supervised 3D Object Detection
by Bohao Li, Shaojing Song and Luxia Ai
Sensors 2024, 24(21), 6940; https://doi.org/10.3390/s24216940 - 29 Oct 2024
Viewed by 430
Abstract
Existing 3D object detection frameworks in sensor-based applications heavily rely on large-scale annotated data to achieve optimal performance. However, obtaining such annotations from sensor data—like LiDAR or image sensors—is both time-consuming and costly. Semi-supervised learning offers an efficient solution to this challenge and [...] Read more.
Existing 3D object detection frameworks in sensor-based applications heavily rely on large-scale annotated data to achieve optimal performance. However, obtaining such annotations from sensor data—like LiDAR or image sensors—is both time-consuming and costly. Semi-supervised learning offers an efficient solution to this challenge and holds significant potential for sensor-driven artificial intelligence (AI) applications. While it reduces the need for labeled data, semi-supervised learning still depends on a small amount of labeled samples for training. In the initial stages, relying on such limited samples can adversely affect the effective training of student–teacher networks. In this paper, we propose PE-MCAT, a semi-supervised 3D object detection method that generates high-precision pseudo-labels. First, to address the challenges of insufficient local feature capture and poor robustness in point cloud data, we introduce a point enrichment module. This module incorporates information from image sensors and combines multiple feature fusion methods of local and self-features to directly enhance the quality of point clouds and pseudo-labels, compensating for the limitations posed by using only a few labeled samples. Second, we explore the relationship between the teacher network and the pseudo-labels it generates. We propose a multi-class adaptive threshold strategy to initially filter and create a high-quality pseudo-label set. Furthermore, a joint variable threshold strategy is introduced to refine this set further, enhancing the selection of superior pseudo-labels.Extensive experiments demonstrate that PE-MCAT consistently outperforms recent state-of-the-art methods across different datasets. Specifically, on the KITTI dataset and using only 2% of labeled samples, our method improved the mean Average Precision (mAP) by 0.7% for cars, 3.7% for pedestrians, and 3.0% for cyclists. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

26 pages, 7432 KiB  
Article
Research on Deep Learning Detection Model for Pedestrian Objects in Complex Scenes Based on Improved YOLOv7
by Jun Hu, Yongqi Zhou, Hao Wang, Peng Qiao and Wenwei Wan
Sensors 2024, 24(21), 6922; https://doi.org/10.3390/s24216922 - 29 Oct 2024
Viewed by 473
Abstract
Objective: Pedestrian detection is very important for the environment perception and safety action of intelligent robots and autonomous driving, and is the key to ensuring the safe action of intelligent robots and auto assisted driving. Methods: In response to the characteristics of pedestrian [...] Read more.
Objective: Pedestrian detection is very important for the environment perception and safety action of intelligent robots and autonomous driving, and is the key to ensuring the safe action of intelligent robots and auto assisted driving. Methods: In response to the characteristics of pedestrian objects occupying a small image area, diverse poses, complex scenes and severe occlusion, this paper proposes an improved pedestrian object detection method based on the YOLOv7 model, which adopts the Convolutional Block Attention Module (CBAM) attention mechanism and Deformable ConvNets v2 (DCNv2) in the two Efficient Layer Aggregation Network (ELAN) modules of the backbone feature extraction network. In addition, the detection head is replaced with a Dynamic Head (DyHead) detector head with an attention mechanism; unnecessary background information around the pedestrian object is also effectively excluded, making the model learn more concentrated feature representations. Results: Compared with the original model, the log-average miss rate of the improved YOLOv7 model is significantly reduced in both the Citypersons dataset and the INRIA dataset. Conclusions: The improved YOLOv7 model proposed in this paper achieved good performance improvement in different pedestrian detection problems. The research in this paper has important reference significance for pedestrian detection in complex scenes such as small, occluded and overlapping objects. Full article
Show Figures

Figure 1

15 pages, 2191 KiB  
Article
YOLO-ESL: An Enhanced Pedestrian Recognition Network Based on YOLO
by Feilong Wang, Xiaobing Yang and Juan Wei
Appl. Sci. 2024, 14(20), 9588; https://doi.org/10.3390/app14209588 - 21 Oct 2024
Viewed by 611
Abstract
Pedestrian detection is a critical task in computer vision; however, mainstream algorithms often struggle to achieve high detection accuracy in complex scenarios, particularly due to target occlusion and the presence of small objects. This paper introduces a novel pedestrian detection algorithm, YOLO-ESL, based [...] Read more.
Pedestrian detection is a critical task in computer vision; however, mainstream algorithms often struggle to achieve high detection accuracy in complex scenarios, particularly due to target occlusion and the presence of small objects. This paper introduces a novel pedestrian detection algorithm, YOLO-ESL, based on the YOLOv7 framework. YOLO-ESL integrates the ELAN-SA module, designed to enhance feature extraction, with the LGA module, which improves feature fusion. The ELAN-SA module optimizes the flexibility and efficiency of small object feature extraction, while the LGA module effectively integrates multi-scale features through local and global attention mechanisms. Additionally, the CIOUNMS algorithm addresses the issue of target loss in cases of high overlap, improving boundary box filtering. Evaluated on the VOC2012 pedestrian dataset, YOLO-ESL achieved an accuracy of 93.7%, surpassing the baseline model by 3.0%. Compared to existing methods, this model not only demonstrates strong performance in handling occluded and small object detection but also remarkable robustness and efficiency. Full article
(This article belongs to the Special Issue Applications of Artificial Intelligence in Industrial Engineering)
Show Figures

Figure 1

23 pages, 7971 KiB  
Article
Three-Dimensional Outdoor Object Detection in Quadrupedal Robots for Surveillance Navigations
by Muhammad Hassan Tanveer, Zainab Fatima, Hira Mariam, Tanazzah Rehman and Razvan Cristian Voicu
Actuators 2024, 13(10), 422; https://doi.org/10.3390/act13100422 - 16 Oct 2024
Viewed by 722
Abstract
Quadrupedal robots are confronted with the intricate challenge of navigating dynamic environments fraught with diverse and unpredictable scenarios. Effectively identifying and responding to obstacles is paramount for ensuring safe and reliable navigation. This paper introduces a pioneering method for 3D object detection, termed [...] Read more.
Quadrupedal robots are confronted with the intricate challenge of navigating dynamic environments fraught with diverse and unpredictable scenarios. Effectively identifying and responding to obstacles is paramount for ensuring safe and reliable navigation. This paper introduces a pioneering method for 3D object detection, termed viewpoint feature histograms, which leverages the established paradigm of 2D detection in projection. By translating 2D bounding boxes into 3D object proposals, this approach not only enables the reuse of existing 2D detectors but also significantly increases the performance with less computation required, allowing for real-time detection. Our method is versatile, targeting both bird’s eye view objects (e.g., cars) and frontal view objects (e.g., pedestrians), accommodating various types of 2D object detectors. We showcase the efficacy of our approach through the integration of YOLO3D, utilizing LiDAR point clouds on the KITTI dataset, to achieve real-time efficiency aligned with the demands of autonomous vehicle navigation. Our model selection process, tailored to the specific needs of quadrupedal robots, emphasizes considerations such as model complexity, inference speed, and customization flexibility, achieving an accuracy of up to 99.93%. This research represents a significant advancement in enabling quadrupedal robots to navigate complex and dynamic environments with heightened precision and safety. Full article
(This article belongs to the Section Actuators for Robotics)
Show Figures

Figure 1

15 pages, 6894 KiB  
Article
A Novel Approach to Pedestrian Re-Identification in Low-Light and Zero-Shot Scenarios: Exploring Transposed Convolutional Reflectance Decoders
by Zhenghao Li and Jiping Xiong
Electronics 2024, 13(20), 4069; https://doi.org/10.3390/electronics13204069 - 16 Oct 2024
Viewed by 596
Abstract
In recent years, pedestrian re-identification technology has made significant progress, with various neural network models performing well under normal conditions, such as good weather and adequate lighting. However, most research has overlooked extreme environments, such as rainy weather and nighttime. Additionally, the existing [...] Read more.
In recent years, pedestrian re-identification technology has made significant progress, with various neural network models performing well under normal conditions, such as good weather and adequate lighting. However, most research has overlooked extreme environments, such as rainy weather and nighttime. Additionally, the existing pedestrian re-identification datasets predominantly consist of well-lit images. Although some studies have started to address these issues by proposing methods for enhancing low-light images to restore their original features, the effectiveness of these approaches remains limited. We noted that a method based on Retinex theory designed a reflectance representation learning module aimed at restoring image features as much as possible. However, this method has so far only been applied in object detection networks. In response to this, we improved the method and applied it to pedestrian re-identification, proposing a transposed convolution reflectance decoder (TransConvRefDecoder) to better restore details in low-light images. Extensive experiments on the Market1501, CUHK03, and MSMT17 datasets demonstrated that our approach delivered superior performance. Full article
Show Figures

Figure 1

25 pages, 6736 KiB  
Article
LFIR-YOLO: Lightweight Model for Infrared Vehicle and Pedestrian Detection
by Quan Wang, Fengyuan Liu, Yi Cao, Farhan Ullah and Muxiong Zhou
Sensors 2024, 24(20), 6609; https://doi.org/10.3390/s24206609 - 14 Oct 2024
Viewed by 955
Abstract
The complexity of urban road scenes at night and the inadequacy of visible light imaging in such conditions pose significant challenges. To address the issues of insufficient color information, texture detail, and low spatial resolution in infrared imagery, we propose an enhanced infrared [...] Read more.
The complexity of urban road scenes at night and the inadequacy of visible light imaging in such conditions pose significant challenges. To address the issues of insufficient color information, texture detail, and low spatial resolution in infrared imagery, we propose an enhanced infrared detection model called LFIR-YOLO, which is built upon the YOLOv8 architecture. The primary goal is to improve the accuracy of infrared target detection in nighttime traffic scenarios while meeting practical deployment requirements. First, to address challenges such as limited contrast and occlusion noise in infrared images, the C2f module in the high-level backbone network is augmented with a Dilation-wise Residual (DWR) module, incorporating multi-scale infrared contextual information to enhance feature extraction capabilities. Secondly, at the neck of the network, a Content-guided Attention (CGA) mechanism is applied to fuse features and re-modulate both initial and advanced features, catering to the low signal-to-noise ratio and sparse detail features characteristic of infrared images. Third, a shared convolution strategy is employed in the detection head, replacing the decoupled head strategy and utilizing shared Detail Enhancement Convolution (DEConv) and Group Norm (GN) operations to achieve lightweight yet precise improvements. Finally, loss functions, PIoU v2 and Adaptive Threshold Focal Loss (ATFL), are integrated into the model to better decouple infrared targets from the background and to enhance convergence speed. The experimental results on the FLIR and multispectral datasets show that the proposed LFIR-YOLO model achieves an improvement in detection accuracy of 4.3% and 2.6%, respectively, compared to the YOLOv8 model. Furthermore, the model demonstrates a reduction in parameters and computational complexity by 15.5% and 34%, respectively, enhancing its suitability for real-time deployment on resource-constrained edge devices. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

20 pages, 3290 KiB  
Article
Vehicle and Pedestrian Detection Based on Improved YOLOv7-Tiny
by Zhen Liang, Wei Wang, Ruifeng Meng, Hongyu Yang, Jinlei Wang, He Gao, Biao Li and Jungeng Fan
Electronics 2024, 13(20), 4010; https://doi.org/10.3390/electronics13204010 - 12 Oct 2024
Viewed by 672
Abstract
To improve the detection accuracy of vehicles and pedestrians in traffic scenes using object detection algorithms, this paper presents modifications, compression, and deployment of the single-stage typical algorithm YOLOv7-tiny. In the model improvement section: firstly, to address the problem of small object missed [...] Read more.
To improve the detection accuracy of vehicles and pedestrians in traffic scenes using object detection algorithms, this paper presents modifications, compression, and deployment of the single-stage typical algorithm YOLOv7-tiny. In the model improvement section: firstly, to address the problem of small object missed detection, shallower feature layer information is incorporated into the original feature fusion branch, forming a four-scale detection head; secondly, a Multi-Stage Feature Fusion (MSFF) module is proposed to fully integrate shallow, middle, and deep feature information to extract more comprehensive small object information. In the model compression section: the Layer-Adaptive Magnitude-based Pruning (LAMP) algorithm and the Torch-Pruning library are combined, setting different pruning rates for the improved model. In the model deployment section: the V7-tiny-P2-MSFF model, pruned by 45% using LAMP, is deployed on the embedded platform NVIDIA Jetson AGX Xavier. Experimental results show that the improved and pruned model achieves a 12.3% increase in [email protected] compared to the original model, with parameter volume, computation volume, and model size reduced by 76.74%, 7.57%, and 70.94%, respectively. Moreover, the inference speed of a single image for the pruned and quantized model deployed on Xavier is 9.5 ms. Full article
Show Figures

Figure 1

31 pages, 50761 KiB  
Article
Intelligent Structural Health Monitoring and Noncontact Measurement Method of Small Reservoir Dams Using UAV Photogrammetry and Anomaly Detection
by Sizeng Zhao, Fei Kang, Lina He, Junjie Li, Yiqing Si and Yiping Xu
Appl. Sci. 2024, 14(20), 9156; https://doi.org/10.3390/app14209156 - 10 Oct 2024
Viewed by 589
Abstract
This study proposes a UAV-based remote measurement method for accurately locating pedestrians and other small targets within small reservoir dams. To address the imprecise coordinate information in reservoir areas after prolonged operations, a transformation method for converting UAV coordinates into the local coordinate [...] Read more.
This study proposes a UAV-based remote measurement method for accurately locating pedestrians and other small targets within small reservoir dams. To address the imprecise coordinate information in reservoir areas after prolonged operations, a transformation method for converting UAV coordinates into the local coordinate system without relying on preset parameters is introduced, accomplished by integrating the Structure from Motion (SfM) algorithm to calculate the transformation parameters. An improved YOLOv8 network is introduced for the high-precision detection of small pedestrian targets, complemented by a laser rangefinder to facilitate accurate 3D locating of targets from varying postures and positions. Furthermore, the integration of a thermal infrared camera facilitates the detection and localization of potential seepage. The experimental validation and application across two real small reservoir dams confirm the accuracy and applicability of the proposed approach, demonstrating the efficiency of the proposed routine UAV surveillance strategy and proving its potential to establish electronic fences and enhance maintenance operations. Full article
Show Figures

Figure 1

20 pages, 29723 KiB  
Article
Optimized Right-Turn Pedestrian Collision Avoidance System Using Intersection LiDAR
by Soo-Yong Park and Seok-Cheol Kee
World Electr. Veh. J. 2024, 15(10), 452; https://doi.org/10.3390/wevj15100452 - 6 Oct 2024
Viewed by 447
Abstract
The incidence of right-turning pedestrian accidents is increasing in South Korea. Most of the accidents occur when a large vehicle is turning right, and the main cause of the accidents was found to be the driver’s limited field of vision. After these accidents, [...] Read more.
The incidence of right-turning pedestrian accidents is increasing in South Korea. Most of the accidents occur when a large vehicle is turning right, and the main cause of the accidents was found to be the driver’s limited field of vision. After these accidents, the government implemented a series of institutional measures with the objective of preventing such accidents. However, despite the institutional arrangements in place, pedestrian accidents continue to occur. We focused on the many limitations that autonomous vehicles, like humans, can face in such situations. To address this issue, we propose a right-turn pedestrian collision avoidance system by installing a LiDAR sensor in the center of the intersection to facilitate pedestrian detection. Furthermore, the urban road environment is considered, as this provides the optimal conditions for the model to perform at its best. During this research, we collected data on right-turn accidents using the CARLA simulator and ROS interface and demonstrated the effectiveness of our approach in preventing such incidents. Our results suggest that the implementation of this method can effectively reduce the incidence of right-turn accidents in autonomous vehicles. Full article
Show Figures

Figure 1

21 pages, 10718 KiB  
Review
Pedestrian Fall Detection Methods for Public Traffic Areas: A Literature Review
by Rongyong Zhao, Wenjie Zhu, Chuanfeng Han, Bingyu Wei, Hao Zhang, Arifur Rahman and Cuiling Li
Appl. Sci. 2024, 14(19), 8934; https://doi.org/10.3390/app14198934 - 4 Oct 2024
Viewed by 727
Abstract
Crowd accident surveys have shown that regardless of the initial triggering factors, pedestrian fall behavior is the most critical factor causing and aggravating crowd accidents in public traffic areas (PTAs). The application of pedestrian fall behavior detection methods in PTAs is significant. Once [...] Read more.
Crowd accident surveys have shown that regardless of the initial triggering factors, pedestrian fall behavior is the most critical factor causing and aggravating crowd accidents in public traffic areas (PTAs). The application of pedestrian fall behavior detection methods in PTAs is significant. Once deployed, they would prevent many pedestrians from losing life in crowded traffic area accidents. However, most existing methods are still focused on medical assistance for the elderly. Therefore, this paper conducted bibliometric and content analyses, combining fall detection-related keywords from internationally recognized literature databases and benchmark pedestrian behavior datasets. Based on the analysis of the state-of-the-art (SOTA) achievements in fall detection methods, the fall detection methods were classified into different categories according to the research approach. This study undertakes a comprehensive analysis of five predominant methods, namely, computer vision, Internet of Things, smartphone, kinematic, and wearable device-based methods. Furthermore, the benchmark datasets, including fall scenarios, were introduced and compared. Finally, this study provides a detailed discussion of existing fall detection methods, and possible future directions are identified considering the application requirements in PTAs. This overview may help researchers understand the SOTA fall detection methods and devise new methodologies by improving and synthesizing the highlighted issues in PTAs. Full article
Show Figures

Figure 1

18 pages, 5232 KiB  
Article
Vehicle and Pedestrian Traffic Signal Performance Measures Using LiDAR-Derived Trajectory Data
by Enrique D. Saldivar-Carranza, Jairaj Desai, Andrew Thompson, Mark Taylor, James Sturdevant and Darcy M. Bullock
Sensors 2024, 24(19), 6410; https://doi.org/10.3390/s24196410 - 3 Oct 2024
Viewed by 560
Abstract
Light Detection and Ranging (LiDAR) sensors at signalized intersections can accurately track the movement of virtually all objects passing through at high sampling rates. This study presents methodologies to estimate vehicle and pedestrian traffic signal performance measures using LiDAR trajectory data. Over 15,000,000 [...] Read more.
Light Detection and Ranging (LiDAR) sensors at signalized intersections can accurately track the movement of virtually all objects passing through at high sampling rates. This study presents methodologies to estimate vehicle and pedestrian traffic signal performance measures using LiDAR trajectory data. Over 15,000,000 vehicle and 170,000 pedestrian waypoints detected during a 24 h period at an intersection in Utah are analyzed to describe the proposed techniques. Sampled trajectories are linear referenced to generate Purdue Probe Diagrams (PPDs). Vehicle-based PPDs are used to estimate movement level turning counts, 85th percentile queue lengths (85QL), arrivals on green (AOG), highway capacity manual (HCM) level of service (LOS), split failures (SF), and downstream blockage (DSB) by time of day (TOD). Pedestrian-based PPDs are used to estimate wait times and the proportion of people that traverse multiple crosswalks. Although vehicle signal performance can be estimated from several days of aggregated connected vehicle (CV) data, LiDAR data provides the ability to measure performance in real time. Furthermore, LiDAR can measure pedestrian speeds. At the studied location, the 15th percentile pedestrian walking speed was estimated to be 3.9 ft/s. The ability to directly measure these pedestrian speeds allows agencies to consider alternative crossing times than those suggested by the Manual on Uniform Traffic Control Devices (MUTCD). Full article
(This article belongs to the Section Radar Sensors)
Show Figures

Figure 1

17 pages, 5388 KiB  
Article
Research on Pedestrian and Cyclist Classification Method Based on Micro-Doppler Effect
by Xinyu Chen, Xiao Luo, Zeyu Xie, Defang Zhao, Zhen Zheng and Xiaodong Sun
Sensors 2024, 24(19), 6398; https://doi.org/10.3390/s24196398 - 2 Oct 2024
Viewed by 503
Abstract
In the field of autonomous driving, it is important to protect vulnerable road users (VRUs) and ensure the safety of autonomous driving effectively by improving the detection accuracy of VRUs in the driver’s field of vision. However, due to the strong temporal similarity [...] Read more.
In the field of autonomous driving, it is important to protect vulnerable road users (VRUs) and ensure the safety of autonomous driving effectively by improving the detection accuracy of VRUs in the driver’s field of vision. However, due to the strong temporal similarity between pedestrians and cyclists, the insensitivity of the traditional least squares method to their differences results in its suboptimal classification performance. In response to this issue, this paper proposes an algorithm for classifying pedestrian and cyclist targets based on the micro-Doppler effect. Firstly, distinct from conventional time-frequency fusion methods, a preprocessing module was developed to solely perform frequency-domain fitting on radar echo data of pedestrians and cyclists in forward motion, with the purpose of generating fitting coefficients for the classification task. Herein, wavelet threshold processing, short-time Fourier transform, and periodogram methods are employed to process radar echo data. Then, for the heightened sensitivity to inter-class differences, a fractional polynomial is introduced into the extraction of micro-Doppler characteristics of VRU targets to enhance extraction precision. Subsequently, the support vector machine technique is embedded for precise feature classification. Finally, subjective comparisons, objective explanations, and ablation experiments demonstrate the superior performance of our algorithm in the field of VRU target classification. Full article
(This article belongs to the Section Vehicular Sensing)
Show Figures

Figure 1

21 pages, 3267 KiB  
Article
Attention-Guided Sample-Based Feature Enhancement Network for Crowded Pedestrian Detection Using Vision Sensors
by Shuyuan Tang, Yiqing Zhou, Jintao Li, Chang Liu and Jinglin Shi
Sensors 2024, 24(19), 6350; https://doi.org/10.3390/s24196350 - 30 Sep 2024
Viewed by 495
Abstract
Occlusion presents a major obstacle in the development of pedestrian detection technologies utilizing computer vision. This challenge includes both inter-class occlusion caused by environmental objects obscuring pedestrians, and intra-class occlusion resulting from interactions between pedestrians. In complex and variable urban settings, these compounded [...] Read more.
Occlusion presents a major obstacle in the development of pedestrian detection technologies utilizing computer vision. This challenge includes both inter-class occlusion caused by environmental objects obscuring pedestrians, and intra-class occlusion resulting from interactions between pedestrians. In complex and variable urban settings, these compounded occlusion patterns critically limit the efficacy of both one-stage and two-stage pedestrian detectors, leading to suboptimal detection performance. To address this, we introduce a novel architecture termed the Attention-Guided Feature Enhancement Network (AGFEN), designed within the deep convolutional neural network framework. AGFEN improves the semantic information of high-level features by mapping it onto low-level feature details through sampling, creating an effect comparable to mask modulation. This technique enhances both channel-level and spatial-level features concurrently without incurring additional annotation costs. Furthermore, we transition from a traditional one-to-one correspondence between proposals and predictions to a one-to-multiple paradigm, facilitating non-maximum suppression using the prediction set as the fundamental unit. Additionally, we integrate these methodologies by aggregating local features between regions of interest (RoI) through the reuse of classification weights, effectively mitigating false positives. Our experimental evaluations on three widely used datasets demonstrate that AGFEN achieves a 2.38% improvement over the baseline detector on the CrowdHuman dataset, underscoring its effectiveness and potential for advancing pedestrian detection technologies. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

15 pages, 5543 KiB  
Article
Reflective Adversarial Attacks against Pedestrian Detection Systems for Vehicles at Night
by Yuanwan Chen, Yalun Wu, Xiaoshu Cui, Qiong Li, Jiqiang Liu and Wenjia Niu
Symmetry 2024, 16(10), 1262; https://doi.org/10.3390/sym16101262 - 25 Sep 2024
Viewed by 524
Abstract
The advancements in deep learning have significantly enhanced the accuracy and robustness of pedestrian detection. However, recent studies reveal that adversarial attacks can exploit the vulnerabilities of deep learning models to mislead detection systems. These attacks are effective not only in digital environments [...] Read more.
The advancements in deep learning have significantly enhanced the accuracy and robustness of pedestrian detection. However, recent studies reveal that adversarial attacks can exploit the vulnerabilities of deep learning models to mislead detection systems. These attacks are effective not only in digital environments but also pose significant threats to the reliability of pedestrian detection systems in the physical world. Existing adversarial attacks targeting pedestrian detection primarily focus on daytime scenarios and are easily noticeable by road observers. In this paper, we propose a novel adversarial attack method against vehicle–pedestrian detection systems at night. Our approach utilizes reflective optical materials that can effectively reflect light back to its source. We optimize the placement of these reflective patches using the particle swarm optimization (PSO) algorithm and deploy patches that blend with the color of pedestrian clothing in real-world scenarios. These patches remain inconspicuous during the day or under low-light conditions, but at night, the reflected light from vehicle headlights effectively disrupts the vehicle’s pedestrian detection systems. Considering that real-world detection models are often black-box systems, we propose a “symmetry” strategy, which involves using the behavior of an alternative model to simulate the response of the target model to adversarial patches. We generate adversarial examples using YOLOv5 and apply our attack to various types of pedestrian detection models. Experiments demonstrate that our approach is both effective and broadly applicable. Full article
(This article belongs to the Special Issue Advanced Studies of Symmetry/Asymmetry in Cybersecurity)
Show Figures

Figure 1

Back to TopTop