Next Article in Journal
Reconstructing the Three-Dimensional Thermohaline Structure of Mesoscale Eddies in the South China Sea Using In Situ Measurements and Multi-Sensor Satellites
Previous Article in Journal
YOLO-IRS: Infrared Ship Detection Algorithm Based on Self-Attention Mechanism and KAN in Complex Marine Background
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SAR-PATT: A Physical Adversarial Attack for SAR Image Automatic Target Recognition

School of Geosciences and Info-Physics, Central South University, South Lushan Road, Changsha 410083, China
*
Author to whom correspondence should be addressed.
Submission received: 5 November 2024 / Revised: 17 December 2024 / Accepted: 23 December 2024 / Published: 25 December 2024
(This article belongs to the Section AI Remote Sensing)

Abstract

:
Deep neural network-based synthetic aperture radar (SAR) automatic target recognition (ATR) systems are susceptible to attack by adversarial examples, which leads to misclassification by the SAR ATR system, resulting in theoretical model robustness problems and security problems in practice. Inspired by optical images, current SAR ATR adversarial example generation is performed in the image domain. However, the imaging principle of SAR images is based on the imaging of the echo signals interacting between the SAR and objects. Generating adversarial examples only in the image domain cannot change the physical world to achieve adversarial attacks. To solve these problems, this article proposes a framework for generating SAR adversarial examples in a 3D physical scene. First, adversarial attacks are implemented in the 2D image space, and the perturbation in the image space is converted into simulated rays that constitute SAR images through backpropagation optimization methods. The mapping between the simulated rays constituting SAR images and the 3D model is established through coordinate transformation, and point correspondence to triangular faces and intensity values to texture parameters are established. Thus, the simulated rays constituting SAR images are mapped to the 3D model, and the perturbation in the 2D image space is converted back to the 3D physical space to obtain the position and intensity of the perturbation in the 3D physical space, thereby achieving physical adversarial attacks. The experimental results show that our attack method can effectively perform SAR adversarial attacks in the physical world. In the digital world, we achieved an average fooling rate of up to 99.02% for three objects in six classification networks. In the physical world, we achieved an average fooling rate of up to 97.87% for these objects, with a certain degree of transferability across the six different network architectures. To the best of our knowledge, this is the first work to implement physical attacks in a full physical simulation condition. Our research establishes a theoretical foundation for the future concealment of SAR targets in practical settings and offers valuable insights for enhancing the attack and defense capabilities of subsequent DNNs in SAR ATR systems.

1. Introduction

Deep neural networks (DNNs) have been widely used in the field of remote sensing image recognition, including optical remote sensing image target recognition and synthetic aperture radar (SAR) automatic target recognition (ATR) [1,2,3]. However, the existence of adversarial examples [4] limits the application of SAR ATR. Through carefully designed, imperceptible perturbations that are difficult for the human eye to detect, it is possible to mislead DNNs into generating incorrect predictions. The issue of adversarial examples has received widespread attention from the academic and industrial communities, becoming a focal point in the field of artificial intelligence security [5,6,7].
Currently, in the remote sensing field, there are many studies of adversarial attacks [8,9]. In the SAR ATR field, researchers are also conducting research on adversarial attacks, with the majority being digital attacks, i.e., directly adding perturbations to the pixels of the image, with the perturbations thus existing in digital space. For example, attackers can directly apply attack methods and principles from optical remote sensing images to adversarial attacks in SAR images [10,11,12,13] or use a sparse attack strategy to confine perturbations to the target area in the image [14,15]. These attack methods generate perturbations that are stealthy, making them difficult for observers to detect and posing a threat to the widespread application of SAR ATR.
However, most of these digital attack methods are based on the principles of optical image attacks and do not consider the differences between the imaging mechanisms of SAR and optical images. The pixel information in SAR images is actually related to backscatter information associated with factors such as radar system parameters and the surface specificities of objects, while visible light imaging is influenced mainly by the color and surface specificities of objects, with different colors and materials absorbing and reflecting light of different wavelengths. Therefore, images generated based on the principles of optical image attacks may not conform to SAR imaging characteristics. Considering the potential for adversarial attacks to be implemented in the real world, in this article, we focus on physical attacks, i.e., adding perturbations in the real-world environment or physical simulators so that the perturbations exist in physical space. Compared to digital attacks, physical attacks pose greater potential threats to society. Moreover, we believe that adversarial examples generated through simulators based on SAR imaging mechanisms are more in line with the SAR imaging mechanisms, making the attacks stealthier.
Physical adversarial attacks have received widespread attention in the optical image recognition field [16,17]. Examples include patch-based attacks [18] and camouflage-based attacks [19]. These methods mount attacks by modifying the patterns on the surface of objects. However, these methods cannot be directly transferred to the SAR field because of the different imaging mechanisms. SAR images record the backscatter information of terrestrial targets, so simply modifying the pattern on the surface of an object cannot change the pixel values of the images. It is necessary to change the surface texture properties of the object, such as the scattering coefficient, surface roughness, and other parameters, to cause changes in the radar echo signal and thereby change the pixel values of the SAR images.
To address the above issues, we propose a pipeline for SAR physical adversarial attack (SAR-PATT). First, to connect the digital world with the physical world, our attack includes both digital and physical parts. Our digital adversarial perturbations act on the output data in SAR geometry, i.e., the scattering points and scattering intensity information generated by the ray tracer. Our physical adversarial perturbations act on the texture parameters of 3D objects, and we use the SAR simulation program RaySAR [20], which is based on ray tracing, to generate 2D SAR images. In previous research, RaySAR has been widely used as a simulator for SAR imaging [21]. Since RaySAR is a nondifferentiable system and physical adversarial perturbations cannot be obtained directly through optimization, we first design an effective loss function and update the digital perturbations through backpropagation. Finally, the final digital perturbations are mapped back into the texture information of the 3D model through a mapping module, thereby updating the physical perturbations. To test our attack method, we add three categories to the Moving and Stationary Target Acquisition and Recognition (MSTAR) dataset [22], namely, CAR, JEEP, and VAN, thus forming a classification dataset with a total of 13 categories, and conduct a series of experiments to prove the effectiveness of our proposed attack method. Our main contributions are summarized as follows:
  • To the best of our knowledge, we are the first to design an end-to-end physical adversarial attack pipeline for SAR images that achieves adversarial attacks by changing the surface texture parameters of 3D targets in the physical world, thereby converting adversarial perturbations from the digital space to the physical space.
  • We propose a new adversarial attack method for SAR ATRs based on the SAR imaging mechanism, which can generate adversarial examples that conform to SAR imaging characteristics. In the digital part, we use an optimization method to generate adversarial perturbations in the SAR geometric output data. In the physical part, we establish a digital–physical mapping module so that the perturbations ultimately act in physical space. Through the SAR imaging simulation process, the perturbations are converted into perturbations in the SAR images, making the attack stealthy.
  • Our method has the potential to be implemented in the real world. Experiments show that in the physical world, our method achieves an average fooling rate of up to 97.87%. In addition, the digital and physical attacks we propose have certain transferability to different network architectures. The adversarial examples generated based on VGG-16 have an average fooling rate of 77.5% for the other five black-box models. Moreover, the physical attack is robust under different azimuths.

2. Related Works

To study physical adversarial attacks on SAR ATR, we investigate the current state of research in three related fields, namely, SAR automatic target recognition, adversarial attacks on SAR ATR, and SAR simulation.

2.1. SAR Automatic Target Recognition

Traditional feature extraction methods rely on human experience and have poor generalizability. However, deep learning (especially convolutional neural networks, (CNNs) can automatically learn features from data, extract features, and perform classification. Therefore, deep learning-based methods have strong high-level feature learning ability and high classification accuracy [23,24,25,26,27]. Because of these advantages, SAR ATR has also gradually incorporated deep learning, and in recent years, deep learning-based SAR ATR has developed rapidly.
Shao et al. [28] were the first to apply several representative CNN models to SAR ATR. They analyzed the performances of different CNN models on the MSTAR dataset and their results show that compared with traditional algorithms, most CNNs demonstrate superior performance, achieving an accuracy rate of over 99% on the MSTAR dataset. Soldin et al. [29] used ResNet-18 on the MSTAR dataset to verify the effectiveness of SAR ATR based on deep learning, achieving an accuracy rate of up to 99% for 10 types of targets. The aforementioned research uses ready-made CNN models to demonstrate the effectiveness of incorporating deep learning into SAR ATR. In addition, some researchers have designed specialized CNNs. Xu et al. [30] proposed the deep CNN model SARNet for SAR ATR, which has two convolution-pooling layers and two fully connected layers. It achieved an accuracy rate of 95.68% on the MSTAR dataset. Zhai et al. [31] proposed MF-SARNet for SAR ATR. The Fire module in the network is used to extract features and requires fewer parameters. MF-SARNet consists of 18 convolutional layers, 2 fully connected layers, and 8 Fire modules. Using data enhancement based on clockwise rotation, the dataset was expanded 360 times, achieving an accuracy rate of 98.53% on the MSATR dataset. Shang et al. [32] proposed M-Net to solve the overfitting problem caused by insufficient samples. M-Net uses an information recorder to store spatial features and uses spatial similarity to predict the labels of unknown samples. To better optimize M-Net, parameter transfer training was adopted. The first step involved training the CNN on M-Net and initializing the parameters. The second step involved using the initialized parameters and training the model with the MSTAR dataset. The model achieved an accuracy rate of 99.71% on the MSTAR dataset. However, given the widespread application of SAR ATR, it is also necessary to study adversarial attacks against SAR ATR to help enhance the security of SAR ATR and thus expand the application field of SAR ATR.

2.2. Adversarial Attacks on SAR ATR

Since Szegedy et al. [4] first discovered adversarial examples, a large amount of research has been conducted on adversarial examples in the computer vision field. Currently, attacks on SAR ATR can be divided into two main categories: attacks in the computer vision field and attacks designed for SAR ATR.

2.2.1. Attacks in the Computer Vision Field

These methods verify the vulnerability of SAR ATR models by drawing on attack algorithms in the computer vision field. Initially, Li et al. [10] used two classic white-box attack methods, FGSM and BIM, to generate adversarial examples from the MSTAR dataset and the SENSAR dataset and evaluated the vulnerability of six commonly used CNNs. The results show that adversarial examples are quite effective in fooling SAR ATR. Huang et al. [11] used three mainstream algorithms to generate adversarial examples to attack three typical CNNs of SAR ATR using the MSTAR dataset for experiments and proved the vulnerability of SAR ATR. Pang et al. [33] designed a CNN for SAR target classification and conducted an adversarial example analysis on it using FGSM and BIM. Wang et al. [13] attacked SAR ATR using universal adversarial perturbations (UAPs). Thereafter, more researchers began to enter this field. Zhou et al. [34] compared several mainstream adversarial attack methods and evaluated the security of the used DNN from the perspective of attention. Peng et al. [35] proposed a completely black-box universal attack framework FBUA for creating UAPs for SAR ATR. FBUA can be divided into three main stages: (1) SAR image simulation, (2) substitute model training, and (3) UAP generation. Comprehensive evaluations were conducted on the MSTAR and SARSIM datasets to analyze the performance of the FBUA.

2.2.2. Attacks Designed for SAR ATR

These methods usually consider the distinct mechanism of SAR imaging and are highly relevant for SAR imaging. Meng et al. [14] proposed a new target-area-based adversarial attack method for SAR ATR. First, they extracted the mask of the target in the SAR image and then introduced the mask parameters into the loss function of the perturbation generator to aggregate the perturbations into the target area. This method provides a theoretical basis for the camouflaging of SAR targets in the physical world. Zhang et al. [15] proposed a new adversarial attack algorithm for SAR ATR that has a higher deception success rate, higher recognition confidence, and smaller perturbation coverage than other advanced SAR attack methods. Peng et al. [36] proposed a speckle-variant attack (SVA), which achieves speckle variant transformation, continuously reconstructs the speckle noise pattern in each iteration process to achieve strong transferability and ensures the feasibility of adversarial perturbations in actual scenarios by limiting the area of perturbation. Therefore, the SVA can generate adversarial examples with greater transferability and physical feasibility. At the same time, researchers have begun to focus on the practical feasibility of adversarial attacks. Xia et al. [37] proposed SAR-PeGA, a perturbation generation algorithm for SAR ATR, which adjusts the phase characteristics of the reflected signal with variable phase sequences, thereby generating adversarial perturbations on the SAR echo signal. Peng et al. [38] explored the domain knowledge of the SAR imaging process and proposed a novel scattering model-guided adversarial attack (SMGAA) algorithm, which can generate adversarial perturbations in the form of electromagnetic scattering responses (called adversarial scatterers). Zhou et al. [39] believe that the adversarial attack algorithm for SAR ATR should consider scattering features such as the attribute scattering center (ASC) and proposed an attack algorithm called ASC-STA. This method can achieve a high success rate without complex parameter settings and can generate high-quality perturbed images. Ma et al. [40] argue that only partially occluding the target region, such as by coating it with absorbing materials, can effectively interfere with the SAR target recognition results and thus, proposed a method for generating adversarial examples of partially obscured SAR targets. Cui et al. [41] proposed a physically oriented adversarial attack for SAR image target recognition. They first identified the prominent areas of the SAR target that the classifier focused on and then continuously refined the adversarial scatterer attribute parameters based on dynamic step optimization of the differential equation. Xie et al. [42] discussed the physical realizability of SAR adversarial examples. They proposed a metasurface interference-guided attack to craft metasurface-based adversarial perturbations. Zhang et al. [43] proposed a SAR adversarial attack method that includes target perturbation generation and background perturbation generation. They use attention mechanisms to extract target regions and scattering models to generate scatterer images as perturbations. Ma et al. [44] integrated electromagnetic computation and a differential evolution algorithm and proposed an adversarial attack against SAR ATR that only uses simple scatterers near the target, such as corner reflectors.
Because attacks in the computer vision field lack prior information and domain knowledge of SAR imaging and because the attacks designed for SAR ATR have not involved research on physical domain attacks, these attack methods are difficult to implement in real scenarios. We therefore consider physical domain attacks, hoping to inspire further research in this field.

2.3. SAR Simulation

To obtain SAR images under the constraints of physical conditions, various types of SAR simulation techniques have been proposed that can simulate the effects of changes in radar system parameters and target electromagnetic scattering characteristics on imaging. RaySAR is a SAR simulator based on ray-tracing methods developed by Auer et al. [45] It extends the open-source ray-tracing rendering software POV-Ray, which is aimed at simulating radar signals in 3D coordinate systems, including azimuth, range, and elevation. RaySAR uses an approximate physical optics model to simulate the specular reflection component and diffuse reflection component of radar signals and employs ray-tracing methods to record the reflection direction and reflection components, with the ray paths determining the positions of echoes in the image. CohRaS®, developed by Hammer and Schulz [46], is also based on ray tracing and is used to simulate high-resolution small scenes, providing training data for classifiers and sample data for the training of image analysis personnel. SARViz, developed by Balz and Stilla [47], is based on rasterization methods for real-time single-bounce simulations. It allows for very fast simulations but has certain limitations in terms of geometric and radiometric accuracy.
Based on RaySAR, Tao et al. [48] proposed a method to support automatic interpretation of meter-level resolution SAR images in complex urban scenes. They compared simulated images with TerraSAR-X Spotlight images, demonstrating good geocoding accuracy, reasonable mask layers, and precise overlapping outlines of individual buildings. Mason et al. [49] developed a method to detect urban flooding using near real-time SAR data. This method utilizes RaySAR combined with LiDAR data of urban areas to predict radar shadows and stagnation areas caused by buildings and tall vegetation in images. Niu et al. [50] proposed a framework to train a DNN to establish a relationship between SAR images and RaySAR simulation parameters, further improving the accuracy of SAR images. Yu and Takeuchi [21] used RaySAR to generate reflectivity maps corresponding to different types of radar backscatter signals for 3D typical targets and a collapse model. By combining the geometric and physical information of the targets, they analyzed the correspondence between point and line features, different scattering signals, and specific target structures in the simulated SAR images. Jia et al. [51] conducted a comprehensive comparative study on SAR imaging in large coastal scenes based on raw Sentinel-1 data, SAR imaging simulations, and Google Maps. They validated the effectiveness of the parallel Range-Doppler algorithm and the RaySAR simulator based on backwards ray tracing. Boyoğlu et al. [52] used RaySAR to simulate TerraSAR-X data for looting holes research. They excavated simulated looting holes and constructed photogrammetric representations for RaySAR to simulate, which shows the benefits of SAR simulation in saving cultural historical objects. Therefore, we use RaySAR to simulate SAR imaging in this study.

3. Method

This section introduces the preliminaries. Then, the proposed method is described in detail, including how to generate adversarial examples in digital and physical worlds.

3.1. Preliminaries

First, we define the forward process of generating SAR images from a 3D model as follows: A 3D model has mesh information M and texture information T . The mesh information M includes vertex coordinates and triangle indices, and the texture information T includes the reflection coefficient, scattering coefficient, surface roughness, etc. The imaging parameter θ c includes the radar signal source coordinates, receiving antenna coordinates, pointing position, etc. We use a ray-tracing renderer R , adopting an approximate physical optics model to simulate the specular and diffuse reflections of radar signals, and we record the reflection direction and ray path C that determine the echo position in the image. This process can be expressed by C = R ( M , T ; θ c ) . To generate a 2D SAR image I , reprocessing F needs to be performed on C , which can be expressed by I = F ( C ; θ p ) , I R H × W × 1 , where θ p is the reprocessing parameter, including the maximum/minimum values of the azimuth, the maximum/minimum values and spacing of the range, the maximum number of ray bounces, etc. Then, we input the image I into the SAR ATR classification network C and obtain the predicted label y = C ( I ) .
Our attack is implemented in the digital space and physical space sequentially. First, adversarial perturbations are generated in the digital space for the SAR geometry data C , and then the perturbations are further mapped to the texture information T in the physical space. Therefore, we define the perturbation of C as a digital attack, and we regard the generation of perturbation as an optimization problem. The objective function is
C a d v * = arg max L ( C ( F ( C a d v , θ p ) ) , y g t )
where C a d v * is the final adversarial SAR geometry data, L ( · , · ) is the loss function, and  y g t is the ground truth label. By solving the optimization problem described in Equation (1), we obtain the adversarial SAR geometry data.
After implementing the perturbation to C , we further map it to the physical space through a mapping M . We define this process as physical attack, which can be represented as
M , T a d v * = M ( M , T , C a d v * )
where T a d v * is the final adversarial texture information. We can obtain the final physical adversarial perturbation via the mapping process. The notations used in the text are summarized in Table 1.

3.2. The SAR-PATT Framework

To generate SAR physical adversarial perturbations, we propose a framework for SAR physical adversarial attacks, which includes a differentiable SAR geometry data reprocessing module. Our framework is shown in Figure 1.
We refer to the process of generating perturbations in SAR geometry data as digital attacks because these perturbations still exist in the digital space. In this section, we use optimization methods to generate adversarial perturbations through gradient backpropagation by calculating the loss. Therefore, the loss function is crucial in this process because it determines our final attack effectiveness. We will discuss our loss function design in detail.
Analogously, we refer to the process of mapping perturbations generated in SAR geometry data to a 3D model as physical attacks because this step maps perturbations from the digital space to the physical space. In this section, we discuss in detail the design of the mapping module and how to implement this mapping. The procedure of our attack is represented by Algorithm 1.
Algorithm 1 SAR-PATT
Input: 3D model ( M , T ) , camera parameter θ c , ground truth label y g t
Output: Adversarial texture T a d v *
  1:
C = R ( M , T ; θ c )
  2:
Initial C a d v * with C
  3:
for the max iteration do
  4:
    update m a s k
  5:
     C a d v * = C + ( ( C a d v * C ) m a s k )
  6:
     I a d v = F ( C a d v * )
  7:
     y = C ( I a d v )
  8:
    calculate L by Equation (3)
  9:
    update C a d v * with gradient backpropagation
10:
end for
11:
calculate the nearest meshes M for each scattering points in C a d v *
12:
update T a d v * for selected meshes M
13:
return T a d v *

3.3. Generating Adversarial Perturbations in the Digital Domain (GaPD)

To achieve the digital attack, we use an optimization method to generate adversarial perturbations in the SAR geometry data C . When designing the loss function, we consider the following three factors:
(1) Objective function. Our strategy is to use a nontarget attack method to automatically find the false target label. Our optimization goal is to make the predicted value of the adversarial example continuously approach this target label, ultimately causing model misclassification. Specifically, we use the model’s classification loss function, i.e., cross-entropy loss, as the objective function and select the label with the highest predicted probability other than the correct label as the target label to minimize the loss function and thus find an adversarial example. However, using a single loss function is insufficient to ensure the effectiveness of the adversarial example because there is no limit to the range of perturbation optimization, which may make the adversarial example deviate from the imaging principles of SAR or conflict with certain characteristics inherent in SAR images.
(2) Intensity range limitation. Since our optimization target is the intensity values contributed by scattering points in SAR geometry data, we constrain the range of intensity values between [ 0 , 1 ] to generate effective 2D SAR images. Let the perturbation vector be δ ; our goal is to ensure that 0 x i + δ i 1 . As the range of the sigmoid function is ( 0 , 1 ) , we can let δ i = s i g m o i d ( ω i ) x i so that 0 s i g m o i d ( ω i ) = x i + δ i 1 . Therefore, we use a sigmoid function to limit the range of the optimization parameters, which can ensure that the generated perturbations are reasonable. When initializing parameter ω i , we use the inverse function of sigmoid l o g i t ( x ) = ln ( x 1 x ) to ensure the similarity of the perturbation vector to the original intensity vector.
(3) Perturbation amplitude limitation. To increase the authenticity of the generated adversarial examples, and because in subsequent physical attacks, we need to map the perturbation of the intensity values to the texture parameters of the 3D model, the smaller the amplitude of the perturbation is, the more conducive it is to the subsequent scattering coefficient matching process. We use the L 2 norm to limit the perturbation vector δ , i.e., to minimize δ 2 . Our final loss function is
L = c · s i g m o i d ( ω ) x 2 2 + H ( y , y a d v )
where ω is the parameter we need to optimize, x is the intensity of the SAR geometry data, c is a constant that can balance the classification loss and perturbation amplitude limitation loss, and H is the cross-entropy loss function.
In addition, in the next step of the physical attack, we need to modify the texture information of the triangular faces of the 3D model corresponding to each scattering point. Therefore, we use a sparse attack strategy to limit the coverage of the perturbed scattering points, i.e., the number of perturbed scattering points. We use the iterative algorithm in the L 0 attack proposed by Carlini and Wagner [53] to implement a sparse attack, fixing some pixels that do not significantly affect the classifier output in each iteration until a minimum subset of perturbed pixels is determined. Specifically, we first initialize a mask vector of the same shape as the perturbation vector, with all elements being 1. The generated perturbation is elementwise-multiplied by the mask vector (Hadamard product), so to limit the coverage of the perturbation, we need only to limit the number of 1 elements in the mask vector. The elements in the perturbation vector are ordered by size. If an element is smaller than a certain threshold, the element of the corresponding index in the mask vector is set to 0, i.e., perturbations smaller than a certain threshold in the perturbation vector are eliminated, and the number of 0 elements in the mask vector increases in the iteration.

3.4. Mapping Adversarial Perturbations from the Digital Domain to the Physical Domain (MaPP)

To map the perturbation of SAR geometry data C in the digital space to the physical space, we define a mapping module M to generate the final adversarial texture information. We refer to this process as a physical attack. In this mapping module, we identify three transformation relationships:
  • Coordinate transformation: Transformation is established from the coordinate system of the scattering points in C to the coordinate system of the 3D model.
  • Point correspondence to triangular faces: The basic units of the intensity values in C are scattering points, while the basic units affected by the texture parameters of the 3D model are triangular faces. Thus, a correspondence needs to be established.
  • Intensity values to texture parameters: The changes in intensity values in SAR geometry data correspond to changes in texture parameters in the 3D model.
For coordinate transformation, we record the coordinates of each scattering point in the 3D model’s coordinate system through a ray-tracing renderer R , thus achieving a transformation from the scattering point’s coordinate system to the 3D model’s coordinate system. For point correspondence to triangular faces, since the areas of the triangular faces composing the 3D model are not uniform, based on the different area sizes of triangular faces, our experiments show that smaller triangular faces may correspond to a single scattering point, while larger triangular faces may correspond to multiple scattering points, as shown in Figure 2. Our goal is to establish a one-to-one correspondence between the basic unit affected by the texture parameters of the 3D model (triangular face) and the basic unit of intensity values in C (scattering point) so that changes in the texture parameters of the triangular faces can simulate changes in the intensity values of the scattering points. Therefore, considering that multiple scattering points may have different impacts on the same triangular face, we select the scattering point with the largest change in intensity values before and after perturbation as the reference point to calculate the texture parameters of the triangular face.
Since the calculation of intensity values in the ray-tracing renderer R needs to consider many factors, such as the angle between the triangular element and the light source, it is difficult to directly obtain the texture parameters of the corresponding triangular face from a determined intensity value. In addition, in SAR imaging, radar signals are usually diffusely rather than specularly reflected on the surface of objects, so the scattering coefficient dominates the impact on intensity values among all texture parameters. We use a single-variable optimizer to optimize the bounds of the scattering coefficient of each triangular face corresponding to each scattering point. Specifically, we define a function that inputs the scattering coefficient and uses the same imaging parameters as before to output the absolute value of the difference between the intensity value output of the ray-tracing renderer R and the target intensity value. We use the Brent algorithm [54] to find the local minimum of this function in the interval [ 0 , 1 ] , thereby obtaining the scattering coefficient corresponding to the target intensity value on this triangular face.
There are two potential ways to implement this physical attack in the real world. First, to reduce the scattering coefficient in the real world, as shown by research into microwave absorption theory, there are radar absorbing materials like carbon fibers, nanotubes, graphene, etc. [55]. Second, to increase the scattering coefficient in the real world, radar reflectors like corner reflectors could be used to amplifying the radar signal. When radar waves hit the reflector, they bounce off the surfaces and return to the radar source. By arranging these materials and radar reflectors in the position of perturbed triangular faces, the SAR echo signals could be perturbed like in the simulation.

4. Experiments and Results

4.1. Dataset

To the best of our knowledge, there is currently no physical dataset available for SAR simulation research (i.e., a dataset containing various 3D models and scenes). The commonly used SAR dataset for researchers, MSTAR, has 10 categories, and it is costly to construct a physical dataset with 10 categories from scratch. Considering the cost and richness of dataset categories, we chose to construct a new dataset by expanding the MSTAR dataset. The expansion method is as follows: we used three different 3D models of vehicles, namely, car, jeep, and van. The rendered images of these 3D models are shown in Figure 3, and the numbers of vertices and triangles for each model are shown in Table 2. Specifically, we used parameters similar to those of the MSTAR dataset at the same azimuth to generate SAR images of size 172 × 172 pixels at 1-degree intervals over 360 degrees. We utilized these 3D models to generate 2D SAR images and expanded the MSTAR dataset accordingly. The final dataset has 13 categories: 2S1, BMP2, BRDM_2, BTR60, D7, SLICY, T62, T72, ZIL131, ZSU_23_4, CAR, JEEP, and VAN. The number of images per category ranges from 428 to 720, as detailed in Table 3. Figure 4 displays SAR images of some targets, along with a comparison between our simulated SAR images and real SAR images in the MSTAR dataset.

4.2. Experimental Setup

Due to the gap in the dataset, our attack principle differs from the principles of existing attacks, as shown in Table 4 and Figure 5. Regardless of whether the attack is digital or physical, our adversarial perturbations are not directly applied in the image space, so we do not select a baseline for our experiments. Therefore, our experiments verify the feasibility of our proposed attack method rather than whether the attack is powerful. We first verify the effects of the proposed digital attack and physical attack, then verify their transferability in different network architectures, and finally, verify the impact of different azimuths on the attack effects.
We choose six different classifier networks to test our attack, namely, three mainstream networks and three lightweight networks: ResNet-50 [56], VGG-16 [57], DenseNet-121 [58], Mobilenet V2 [59], SqueezeNet [60], and ShuffleNet V2 [61]. All networks are from the official implementation of PyTorch. We train these classifier networks separately using ImageNet pretrained models with the following hyperparameter settings: an SGD with momentum optimizer, a learning rate of 0.001, a momentum of 0.9, a batch size of 32, and a maximum number of epochs of 80. The classification accuracy of the classifier networks is shown in Table 5.
For our adversarial attack, the hyperparameter settings are as follows: an Adam optimizer, a learning rate of 0.05, and a maximum number of iterations of 50. The value of constant mentioned in Equation (3) is 1 and the threshold of perturbation size is 0.0001. We run the experiments on a workstation with an NVIDIA GTX 1080 Ti 11 GB GPU.
We choose the fooling rate as the metric to evaluate the effectiveness of our attack. The definition of the fooling rate is
f r = i N t o t a l ( F ( x i a d v ) y i g t ) N t o t a l
where N t o t a l is the total number of samples that are attacked, F is the classifier network, x a d v is the generated adversarial example, and y g t is the ground truth label. The fooling rate is the ratio of the number of samples misclassified to the total number of samples that are attacked.

4.3. Results

We conducted attack experiments on the test set and ensured that all the images were correctly classified before the attack. According to our attack framework, we first conduct digital domain attacks and then map the digital attacks to the physical domain. We divide the results into two parts: the effect of attacks in the digital domain and their corresponding effects in the physical domain.

4.3.1. Digital Attack

In the digital domain, we analyze the fooling rates of the adversarial examples generated on three target objects for six network architectures, as shown in Table 6. The results show that our attack has a certain fooling rate for general and lightweight network architectures, with the highest fooling rate being that with ShuffleNetV2 and the lowest being that with ResNet-50. Among the different attack targets, the fooling rate is highest for the CAR target, averaging 99.02%, while the lowest average fooling rate is 77.14% for the JEEP target. These results indicate that our attack method is effective in fooling SAR target automatic recognition networks. We provide some examples of generated adversarial examples, as shown in Figure 6. After applying adversarial perturbations, the targets are all misclassified into other categories. The left side of the figure is the original image, and the right side is the image with the applied perturbation. Our attack can deceive the classifier with high confidence while maintaining a level of stealth that is difficult for the human eye to detect.

4.3.2. Physical Attack

By mapping the adversarial perturbations generated in the digital domain to the physical domain, we calculate the corresponding fooling rates of the adversarial examples, as shown in Table 7. Compared to the digital world, the fooling rates in the physical world decreased, with ShuffleNetV2 still having the highest fooling rate and VGG-16 having the lowest. Among the different attack targets, the average fooling rate is still highest for the CAR target, at 97.87%, while the lowest average fooling rate is 67.02% for the JEEP target. We show the generated adversarial perturbations corresponding to the target in Figure 7, with the triangular faces with adversarial texture parameters marked in red. We find that in the physical world, the fooling rate is related to the number and area of the triangular faces of the 3D model. For example, for the JEEP target, the number of triangular faces is the lowest (Table 2), and our attack yields the lowest average fooling rate for this target. In the VAN target, as shown in Figure 7, the perturbed triangular faces have a larger area, resulting in the most significant decrease in the average fooling rate for this target. In the CAR target, the number of triangular faces is the highest, and their area is relatively small, leading our attack to yield the highest average fooling rate for this target.

4.3.3. Transferability of the SAR-PATT

In this section, we first test the transferability of digital attacks between different network architectures, as shown in Table 8. We generate adversarial examples on the surrogate model and use them to attack the black-box target model. The black-box attack has an average fooling rate of 80.11% on ResNet-50, while the adversarial examples based on VGG-16 have an average fooling rate of 77.5% on the other five black-box models. Table 8 shows that the transferability of adversarial attacks varies between different networks, which means that their robustness to black-box adversarial attacks is different. For example, the VGG-16 and ShuffleNetV2 networks have lower fooling rates against attacks from all sources, indicating that they have greater robustness to black-box adversarial attacks. In addition, we found that network characteristics affect transferability, and different network architectures have different sensitivities to attacks. For example, ResNet-50, as the source network, has a fooling rate of 72.73% for attacks on SqueezeNet, while the fooling rate for ShuffleNetV2 is only 3.79%. In summary, our results show that the threat of physical adversarial attacks still exists even in black-box network architectures.
We also test the transferability of physical attacks, as shown in Table 9. The results show that the transferability of physical attacks is similar to that of digital attacks. Since physical attacks are essentially mappings of digital attacks in the physical world, the results are as expected.

4.3.4. Robust Attack on Different Azimuths

Under real-world conditions, the SAR image perspective is related to changes in the SAR satellite imaging azimuth. To simulate the impact of this change on attacks, we alter the positions of the camera and objects in the simulated scene to achieve variations in the imaging azimuth. This alteration is performed to demonstrate the robustness of our adversarial perturbations at different azimuths. We test the fooling rates at different azimuths ranging from 19 to 27 degrees with a 2-degree interval under the ResNet-50 network, as shown in Table 10. In each azimuth, the objects are also rotated from angle 0 to 360 degrees with a interval of 1 degree. These setups almost cover the imaging angles that match the real-world scenarios. The results show that the changes in the fooling rates at different azimuths for the three targets are not significant. The average fooling rates for the CAR and VAN targets are higher at different azimuths, while the average fooling rate for the JEEP target is the lowest. These results suggest that our attack method is robust at different azimuths, and our attack is not limited to a particular perspective.

5. Conclusions

In this article, we propose a novel physical adversarial attack for SAR ATR, namely, SAR-PATT, which can map adversarial perturbations that are difficult for the human eye to recognize in the digital space to the physical space, leading to misclassification of the SAR ATR system. We achieve transformation from the physical world to the digital world through a SAR simulation renderer, RaySAR, and update digital perturbations through backpropagation by designing effective loss functions. Finally, the resulting digital perturbations are mapped back to the texture information of the 3D model through a mapping module to realize the final physical perturbation. The results show that our attack can effectively achieve physical–world SAR adversarial attacks, achieving a highest average fooling rate of 99.02% in the digital world and a highest average fooling rate of 97.87% in the physical world. Compared to existing methods, our approach bridges the gap between digital and physical attacks, showing the potential for real-world implementation.
In future work, we hope to increase the aggregation of adversarial triangular faces, thereby reducing the difficulty of real-world implementation. In addition, we hope to propose a differentiable SAR simulation renderer to directly generate physical adversarial perturbations through optimization, thereby improving attack efficiency and implementing universal physical adversarial attacks from multiple views to enhance the robustness of our attacks.

Author Contributions

Conceptualization, B.L. and H.C.; methodology, B.L.; software, B.L.; validation, B.L., H.C., J.C. and X.L.; formal analysis, B.L.; investigation, B.L.; resources, H.L.; data curation, B.L.; writing—original draft preparation, B.L.; writing—review and editing, H.L., C.P. and J.H.; visualization, B.L.; supervision, H.L., C.P. and J.H.; project administration, H.L. All authors have read and agreed to the published version of the manuscript.

Funding

The work was supported by the National Natural Science Foundation of China under Grant no. 42301433 and the Natural Science Foundation of Hunan Province of China under Grant no. 2024JJ6496.

Data Availability Statement

The data associated with this research are available online. The MSTAR dataset is available at https://www.kaggle.com/datasets/atreyamajumdar/mstar-dataset-8-classes (accessed on 20 April 2024).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhu, X.; Montazeri, S.; Ali, M.; Hua, Y.; Wang, Y.; Mou, L.; Shi, Y.; Xu, F.; Bamler, R. Deep Learning Meets SAR: Concepts, Models, Pitfalls, and Perspectives. IEEE Geosci. Remote Sens. Mag. 2021, 9, 143–172. [Google Scholar] [CrossRef]
  2. Kechagias-Stamatis, O.; Aouf, N. Automatic Target Recognition on Synthetic Aperture Radar Imagery: A Survey. IEEE Aerosp. Electron. Syst. Mag. 2021, 36, 56–81. [Google Scholar] [CrossRef]
  3. Li, J.; Yu, Z.; Yu, L.; Cheng, P.; Chen, J.; Chi, C. A Comprehensive Survey on SAR ATR in Deep-Learning Era. Remote Sens. 2023, 15, 1454. [Google Scholar] [CrossRef]
  4. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing Properties of Neural Networks. arXiv 2013, arXiv:1312.6199. [Google Scholar]
  5. Yuan, X.; He, P.; Zhu, Q.; Li, X. Adversarial Examples: Attacks and Defenses for Deep Learning. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 2805–2824. [Google Scholar] [CrossRef]
  6. Zhang, J.; Li, C. Adversarial Examples: Opportunities and Challenges. IEEE Trans. Neural Netw. Learn. Syst. 2019, 31, 2578–2593. [Google Scholar] [CrossRef]
  7. Akhtar, N.; Mian, A. Threat of Adversarial Attacks on Deep Learning in Computer Vision: A Survey. IEEE Access 2018, 6, 14410–14430. [Google Scholar] [CrossRef]
  8. Czaja, W.; Fendley, N.; Pekala, M.; Ratto, C.; Wang, I.J. Adversarial Examples in Remote Sensing. In Proceedings of the 26th ACM SIGSPATIAL International Conference on Advances in Geographic Information Systems, Seattle, WA, USA, 6–9 November 2018; pp. 408–411. [Google Scholar] [CrossRef]
  9. Chen, L.; Xu, Z.; Li, Q.; Peng, J.; Wang, S.; Li, H. An Empirical Study of Adversarial Examples on Remote Sensing Image Scene Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7419–7433. [Google Scholar] [CrossRef]
  10. Li, H.; Huang, H.; Chen, L.; Peng, J.; Huang, H.; Cui, Z.; Mei, X.; Wu, G. Adversarial Examples for CNN-Based SAR Image Classification: An Experience Study. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 1333–1347. [Google Scholar] [CrossRef]
  11. Huang, T.; Zhang, Q.; Liu, J.; Hou, R.; Wang, X.; Li, Y. Adversarial Attacks on Deep-Learning-Based SAR Image Target Recognition. J. Netw. Comput. Appl. 2020, 162, 102632. [Google Scholar] [CrossRef]
  12. Sun, H.; Xu, Y.; Kuang, G.; Chen, J. Adversarial Robustness Evaluation of Deep Convolutional Neural Network Based SAR ATR Algorithm. In Proceedings of the 2021 IEEE International Geoscience and Remote Sensing Symposium IGARSS, Brussels, Belgium, 11–16 July 2021; pp. 5263–5266. [Google Scholar]
  13. Wang, L.; Wang, X.; Ma, S.; Zhang, Y. Universal Adversarial Perturbation of SAR Images for Deep Learning Based Target Classification. In Proceedings of the 2021 IEEE 4th International Conference on Electronics Technology (ICET), Chengdu, China, 7–10 May 2021; pp. 1272–1276. [Google Scholar] [CrossRef]
  14. Meng, T.; Zhang, F.; Ma, F. A Target-region-based SAR ATR Adversarial Deception Method. In Proceedings of the 2022 7th International Conference on Signal and Image Processing (ICSIP), Suzhou, China, 20–22 July 2022; pp. 142–146. [Google Scholar]
  15. Zhang, F.; Meng, T.; Xiang, D.; Ma, F.; Sun, X.; Zhou, Y. Adversarial Deception against SAR Target Recognition Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 4507–4520. [Google Scholar] [CrossRef]
  16. Wei, H.; Tang, H.; Jia, X.; Yu, H.; Li, Z.; Wang, Z.; Satoh, S.; Wang, Z. Physical Adversarial Attack Meets Computer Vision: A Decade Survey. arXiv 2022, arXiv:2209.15179. [Google Scholar] [CrossRef] [PubMed]
  17. Wang, D.; Yao, W.; Jiang, T.; Tang, G.; Chen, X. A Survey on Physical Adversarial Attack in Computer Vision. arXiv 2023, arXiv:2209.14262. [Google Scholar]
  18. Brown, T.B.; Mané, D.; Roy, A.; Abadi, M.; Gilmer, J. Adversarial Patch. arXiv 2018, arXiv:1712.09665. [Google Scholar]
  19. Zeng, X.; Liu, C.; Wang, Y.S.; Qiu, W.; Xie, L.; Tai, Y.W.; Tang, C.K.; Yuille, A.L. Adversarial Attacks beyond the Image Space. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2019; pp. 4302–4311. [Google Scholar]
  20. Auer, S.J. 3D Synthetic Aperture Radar Simulation for Interpreting Complex Urban Reflection Scenarios. Ph.D. Thesis, Technische Universität München, München, Germany, 2011. [Google Scholar]
  21. Yu, Y.; Takeuchi, W. Analysis of SAR Backscatter Intensity Characteristics for Inverse Estimation of Earthquake-Damaged Buildings. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 350–353. [Google Scholar] [CrossRef]
  22. Keydel, E.R.; Lee, S.W.; Moore, J.T. MSTAR Extended Operating Conditions: A Tutorial. Algorithms Synth. Aperture Radar Imag. III 1996, 2757, 228–242. [Google Scholar]
  23. Li, H.; Cao, J.; Zhu, J.; Liu, Y.; Zhu, Q.; Wu, G. Curvature graph neural network. Inf. Sci. 2022, 592, 50–66. [Google Scholar] [CrossRef]
  24. Zhu, J.; Han, X.; Deng, H.; Tao, C.; Zhao, L.; Wang, P. KST-GCN: A Knowledge-Driven Spatial-Temporal Graph Convolutional Network for Traffic Forecasting. IEEE Trans. Intell. Transp. Syst. 2022, 23, 15055–15065. [Google Scholar] [CrossRef]
  25. Peng, J.; Ye, D.; Tang, B.; Lei, Y.; Liu, Y.; Li, H. Lifelong Learning With Cycle Memory Networks. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 16439–16452. [Google Scholar] [CrossRef]
  26. Zhang, Z.; Ren, Z.; Tao, c.; Zhang, Y.; Peng, C.; Li, H. GraSS: Contrastive Learning With Gradient-Guided Sampling Strategy for Remote Sensing Image Semantic Segmentation. IEEE Trans. Geosci. Remote. Sens. 2023, 61, 5626814. [Google Scholar] [CrossRef]
  27. Li, H.; Cao, J.; Zhu, J.; Luo, Q.; He, S.; Wang, X. Augmentation-Free Graph Contrastive Learning of Invariant-Discriminative Representations. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 11157–11167. [Google Scholar] [CrossRef]
  28. Shao, J.; Qu, C.; Li, J. A Performance Analysis of Convolutional Neural Network Models in SAR Target Recognition. In Proceedings of the 2017 SAR in Big Data Era: Models, Methods and Applications (BIGSARDATA), Beijing, China, 13–19 November 2017; pp. 1–6. [Google Scholar] [CrossRef]
  29. Soldin, R.J. SAR Target Recognition with Deep Learning. In Proceedings of the 2018 IEEE Applied Imagery Pattern Recognition Workshop (AIPR), Washington, DC, USA, 9–11 October 2018; pp. 1–8. [Google Scholar]
  30. Xu, Y.; Liu, K.; Ying, Z.; Shang, L.; Liu, J.; Zhai, Y.; Piuri, V.; Scotti, F. SAR Automatic Target Recognition Based on Deep Convolutional Neural Network. In Proceedings of the Image and Graphics: 9th International Conference, ICIG 2017, Shanghai, China, 13–15 September 2017, Revised Selected Papers, Part III 9; Zhao, Y., Kong, X., Taubman, D., Eds.; Springer: Cham, Switzerland, 2017; Volume 10668, pp. 656–667. [Google Scholar] [CrossRef]
  31. Zhai, Y.; Ma, H.; Cao, H.; Deng, W.; Liu, J.; Zhang, Z.; Guan, H.; Zhi, Y.; Wang, J.; Zhou, J. MF-SarNet: Effective CNN with Data Augmentation for SAR Automatic Target Recognition. J. Eng. 2019, 2019, 5813–5818. [Google Scholar] [CrossRef]
  32. Shang, R.; Wang, J.; Jiao, L.; Stolkin, R.; Hou, B.; Li, Y. SAR Targets Classification Based on Deep Memory Convolution Neural Networks and Transfer Parameters. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 2834–2846. [Google Scholar] [CrossRef]
  33. Pang, L.; Wang, L.; Zhang, Y.; Li, H. Adversarial Examples of SAR Images for Deep Learning Based Automatic Target Recognition. In Proceedings of the 2021 IEEE 6th International Conference on Signal and Image Processing (ICSIP), Nanjing, China, 22–24 October 2021; pp. 24–27. [Google Scholar] [CrossRef]
  34. Zhou, J.; Peng, B.; Peng, B. Adversarial Attacks on Radar Target Recognition Based on Deep Learning. In Proceedings of the IGARSS 2022—2022 IEEE International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 17–22 July 2022; pp. 2646–2649. [Google Scholar] [CrossRef]
  35. Peng, B.; Peng, B.; Yong, S.; Liu, L. An Empirical Study of Fully Black-Box and Universal Adversarial Attack for SAR Target Recognition. Remote Sens. 2022, 14, 4017. [Google Scholar] [CrossRef]
  36. Peng, B.; Peng, B.; Zhou, J.; Xia, J.; Liu, L. Speckle-Variant Attack: Toward Transferable Adversarial Attack to SAR Target Recognition. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  37. Xia, W.; Liu, Z.; Li, Y. SAR-PeGA: A Generation Method of Adversarial Examples for SAR Image Target Recognition Network. IEEE Trans. Aerosp. Electron. Syst. 2022, 52, 1910–1920. [Google Scholar] [CrossRef]
  38. Peng, B.; Peng, B.; Zhou, J.; Xie, J.; Liu, L. Scattering Model Guided Adversarial Examples for SAR Target Recognition: Attack and Defense. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–17. [Google Scholar] [CrossRef]
  39. Zhou, J.; Feng, S.; Sun, H.; Zhang, L.; Kuang, G. Attributed Scattering Center Guided Adversarial Attack for DCNN SAR Target Recognition. IEEE Geosci. Remote Sens. Lett. 2023, 20, 1–5. [Google Scholar] [CrossRef]
  40. Ma, Y.; Xu, L.; Pei, J.; Huo, W.; Zhang, Y.; Huang, Y.; Yang, J. Target Partial-Occlusion: An Adversarial Examples Generation Approach Against SAR Target Recognition Networks. In Proceedings of the IGARSS 2023—2023 IEEE International Geoscience and Remote Sensing Symposium, Pasadena, CA, USA, 16–21 July 2023; pp. 7054–7057. [Google Scholar] [CrossRef]
  41. Cui, J.; Shao, R.; Li, H. Physics-Oriented Adversarial Attacks on SAR Image Target Recognition. In Proceedings of the The Second Workshop on New Frontiers in Adversarial Machine Learning, Honolulu, HI, USA, 23–29 July 2023. [Google Scholar]
  42. Xie, J.; Peng, B.; Lu, Z.; Zhou, J.; Peng, B. MIGAA: A Physical Adversarial Attack Method against SAR Recognition Models. In Proceedings of the 2024 9th International Conference on Computer and Communication Systems (ICCCS), Xi’an, China, 19–22 July 2024; pp. 309–314. [Google Scholar] [CrossRef]
  43. Zhang, F.; Yu, Y.; Ma, F.; Zhou, Y. A Physically Realizable Adversarial Attack Method Against SAR Target Recognition Model. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 11943–11957. [Google Scholar] [CrossRef]
  44. Ma, Y.; Pei, J.; Huo, W.; Zhang, Y.; Huang, Y.; Chen, K.; Yang, J. SAR-PAA: A Physically Adversarial Attack Approach Against SAR Intelligent Target Recognition. IEEE Trans. Aerosp. Electron. Syst. 2024, 1–18. [Google Scholar] [CrossRef]
  45. Auer, S.; Hinz, S.; Bamler, R. Ray-Tracing Simulation Techniques for Understanding High-Resolution SAR Images. IEEE Trans. Geosci. Remote Sens. 2010, 48, 1445–1456. [Google Scholar] [CrossRef]
  46. Hammer, H.; Schulz, K. Coherent Simulation of SAR Images. In Proceedings of the Image and Signal Processing for Remote Sensing XV, Berlin, Germany, 31 August–2 September 2009; Volume 7477, pp. 406–414. [Google Scholar] [CrossRef]
  47. Balz, T.; Stilla, U. Hybrid GPU-based Single-and Double-Bounce SAR Simulation. IEEE Trans. Geosci. Remote Sens. 2009, 47, 3519–3529. [Google Scholar] [CrossRef]
  48. Tao, J.; Auer, S.; Palubinskas, G.; Reinartz, P.; Bamler, R. Automatic SAR Simulation Technique for Object Identification in Complex Urban Scenarios. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2013, 7, 994–1003. [Google Scholar] [CrossRef]
  49. Mason, D.C.; Dance, S.L.; Vetra-Carvalho, S.; Cloke, H.L. Robust Algorithm for Detecting Floodwater in Urban Areas Using Synthetic Aperture Radar Images. J. Appl. Remote Sens. 2018, 12, 04501. [Google Scholar] [CrossRef]
  50. Niu, S.; Qiu, X.; Lei, B.; Ding, C.; Fu, K. Parameter Extraction Based on Deep Neural Network for SAR Target Simulation. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4901–4914. [Google Scholar] [CrossRef]
  51. Jia, H.; Yang, P.; Wu, R. Comparative Study of Sentinel-1-Focused and Simulated SAR Images Using LiDAR Point Cloud Modeling for Coastal Areas. Electronics 2023, 12, 4312. [Google Scholar] [CrossRef]
  52. Boyoğlu, C.S.; Balz, T.; Sultanbekova, A. SAR Simulation of Looting Holes for Better Image Understanding. In Proceedings of the IGARSS 2024—2024 IEEE International Geoscience and Remote Sensing Symposium, Athens, Greece, 7–12 July 2024; pp. 3337–3340. [Google Scholar] [CrossRef]
  53. Carlini, N.; Wagner, D. Towards Evaluating the Robustness of Neural Networks. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (Sp), San Jose, CA, USA, 22–24 May 2017; pp. 39–57. [Google Scholar]
  54. Brent, R.P. Algorithms for Minimization Without Derivatives; Courier Corporation: San Francisco, CA, USA, 2013. [Google Scholar]
  55. Ruiz-Perez, F.; López-Estrada, S.M.; Tolentino-Hernández, R.V.; Caballero-Briones, F. Carbon-Based Radar Absorbing Materials: A Critical Review. J. Sci. Adv. Mater. Devices 2022, 7, 100454. [Google Scholar] [CrossRef]
  56. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  57. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  58. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  59. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  60. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level Accuracy with 50x Fewer Parameters and <0.5MB Model Size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  61. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. Shufflenet v2: Practical Guidelines for Efficient Cnn Architecture Design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 116–131. [Google Scholar]
Figure 1. In the SAR-PATT framework, we first generate adversarial perturbations in the SAR geometry data via an optimization method. We design a loss function to update the perturbations through backpropagation, thereby obtaining the final perturbations. Finally, these perturbations are mapped to the texture information of the 3D model through the mapping module.
Figure 1. In the SAR-PATT framework, we first generate adversarial perturbations in the SAR geometry data via an optimization method. We design a loss function to update the perturbations through backpropagation, thereby obtaining the final perturbations. Finally, these perturbations are mapped to the texture information of the 3D model through the mapping module.
Remotesensing 17 00021 g001
Figure 2. The triangular faces that compose the 3D model. When mapping the scattering points to the triangular faces, due to the different areas of the triangular faces, a larger triangular face can correspond to multiple scattering points. Nonetheless, we can select only one scattering point as a reference to calculate the texture parameters of the triangular face.
Figure 2. The triangular faces that compose the 3D model. When mapping the scattering points to the triangular faces, due to the different areas of the triangular faces, a larger triangular face can correspond to multiple scattering points. Nonetheless, we can select only one scattering point as a reference to calculate the texture parameters of the triangular face.
Remotesensing 17 00021 g002
Figure 3. The rendered images of the 3D models we used.
Figure 3. The rendered images of the 3D models we used.
Remotesensing 17 00021 g003
Figure 4. Comparison of the generated SAR simulation images and real scene SAR images in the MSTAR dataset, with the real SAR images at the top and the SAR simulation images at the bottom.
Figure 4. Comparison of the generated SAR simulation images and real scene SAR images in the MSTAR dataset, with the real SAR images at the top and the SAR simulation images at the bottom.
Remotesensing 17 00021 g004
Figure 5. The difference between our attack and other attacks.
Figure 5. The difference between our attack and other attacks.
Remotesensing 17 00021 g005
Figure 6. Generated adversarial examples and their original labels and misclassified labels. The confidence levels are listed. These adversarial examples are generated on ResNet-50.
Figure 6. Generated adversarial examples and their original labels and misclassified labels. The confidence levels are listed. These adversarial examples are generated on ResNet-50.
Remotesensing 17 00021 g006
Figure 7. The positions of the physical adversarial perturbations generated on three types of 3D models. From top to bottom, they are generated by ResNet-50, VGG-16, DenseNet-121, MobileNetV2, SqueezeNet, and ShuffleNetV2. The triangular faces with adversarial texture parameters are marked in red, and the number of triangular faces is labeled below the image. It can be observed that the perturbed triangular faces have different sizes.
Figure 7. The positions of the physical adversarial perturbations generated on three types of 3D models. From top to bottom, they are generated by ResNet-50, VGG-16, DenseNet-121, MobileNetV2, SqueezeNet, and ShuffleNetV2. The triangular faces with adversarial texture parameters are marked in red, and the number of triangular faces is labeled below the image. It can be observed that the perturbed triangular faces have different sizes.
Remotesensing 17 00021 g007aRemotesensing 17 00021 g007b
Table 1. Notations.
Table 1. Notations.
NotationsDescriptionNotationsDescription
M Mesh information T Texture information
θ c Imaging parameters R ( · , · ; · ) Ray-tracing renderer
C SAR geometry data I SAR image
F ( · ; · ) Reprocessing module θ p Reprocessing parameters
C ( · ) SAR ATR classifieryLabel of prediction
C a d v * Adversarial SAR geometry data L ( · , · ) Loss function
y g t Ground truth label M ( · , · , · ) Mapping module
T a d v * Adversarial texture information
Table 2. Comparison of 3D models and their vertices and triangles.
Table 2. Comparison of 3D models and their vertices and triangles.
3D ModelCARJEEPVAN
vertices116,92950,65176,214
triangles207,83093,927139,097
Table 3. Categories of training and test set after expansion of the MSTAR dataset.
Table 3. Categories of training and test set after expansion of the MSTAR dataset.
ClassTraining SetTest Set
2S1233274
BMP2233195
BRDM_2233274
BTR60233195
D7233274
SLICY233274
T62233273
T72232196
ZIL131233274
ZSU_23_4233274
CAR360360
JEEP360360
VAN360360
Table 4. Comparison of SAR-PATT and existing methods.
Table 4. Comparison of SAR-PATT and existing methods.
MethodFull Physical SimulationBased onPerturbation at
SVA [36]×GradientImage pixel
SAR-PeGA [37]×OptimizationSAR echo signal
SMGAA [38]×GradientAdversarial scatterer
SAR-PATT(Ours)Optimization3D model texture
Table 5. The accuracy of all the classifier networks on the test set.
Table 5. The accuracy of all the classifier networks on the test set.
ResNet-50VGG-16DenseNet-121MobileNetV2SqueezeNetShuffleNetV2
test accuracy98.6603%98.2138%92.4365%98.4092%93.8878%97.2370%
Table 6. The fooling rates of digital attacks generated in six classification networks and their average fooling rates.
Table 6. The fooling rates of digital attacks generated in six classification networks and their average fooling rates.
ResNet-50VGG-16DenseNet-121MobileNetV2SqueezeNetShuffleNetV2Avg.
CAR100%94.12%100%100%100%100%99.02%
JEEP63.89%85.24%100%59.17%57.50%97.02%77.14%
VAN77.42%95.52%83.61%100%99.58%99.37%92.58%
Table 7. The fooling rates of physical attacks on different targets under six classification networks and their average fooling rates, with levels of variation indicated in the table.
Table 7. The fooling rates of physical attacks on different targets under six classification networks and their average fooling rates, with levels of variation indicated in the table.
ResNet-50VGG-16DenseNet-121MobileNetV2SqueezeNetShuffleNetV2Avg.
CAR100% ↓ 0%88.39% ↓ 5.73%100% ↓ 0%100% ↓ 0%100% ↓ 0%98.85% ↓ 1.15%97.87% ↓ 1.15%
JEEP59.57% ↓ 4.32%66.34% ↓ 18.90%100% ↓ 0%41.18% ↓ 17.99%44.25% ↓ 13.25%90.80% ↓ 6.22%67.02% ↓ 10.12%
VAN81.67% ↑ 4.25%26.39% ↓ 69.13%78.64% ↓ 4.97%96.31% ↓ 3.69%84.81% ↓ 14.77%91.77% ↓ 7.60%76.60% ↓ 15.98%
Table 8. The transferability of adversarial examples generated in 6 different classifier networks in the digital domain among six different network architectures. The corresponding fooling rates are listed.
Table 8. The transferability of adversarial examples generated in 6 different classifier networks in the digital domain among six different network architectures. The corresponding fooling rates are listed.
Source\TargetResNet-50VGG-16DenseNet-121MobileNetV2SqueezeNetShuffleNetV2Avg.
ResNet-500%62.12%59.85%72.73%3.79%39.70%
VGG-1696.13%86.61%92.56%83.04%29.17%77.50%
DenseNet-12173.44%0%69.53%46.09%10.94%40.00%
MobileNetV256.03%0%62.07%70.69%5.17%38.79%
SqueezeNet90.78%2.13%68.79%86.52%22.70%54.18%
ShuffleNetV284.15%3.46%88.18%84.44%74.93%67.03%
Avg.80.11%1.12%73.55%78.58%69.50%14.35%
Table 9. The transferability of adversarial examples generated in six different classifier networks in the physical domain among six different network architectures. The corresponding fooling rates are listed.
Table 9. The transferability of adversarial examples generated in six different classifier networks in the physical domain among six different network architectures. The corresponding fooling rates are listed.
Source\TargetResNet-50VGG-16DenseNet-121MobileNetV2SqueezeNetShuffleNetV2Avg.
ResNet-500%61.36%62.88%72.73%3.79%40.15%
VGG-1696.13%87.50%93.15%78.87%28.57%76.84%
DenseNet-12173.44%0.78%71.09%47.66%10.94%40.78%
MobileNetV260.34%0%64.66%72.41%5.17%40.52%
SqueezeNet92.20%2.13%72.34%88.65%22.70%55.60%
ShuffleNetV287.32%4.90%87.90%84.44%75.79%68.07%
Avg.81.89%1.56%74.75%80.04%69.49%14.23%
Table 10. The fooling rates of adversarial examples generated at different azimuths in the ResNet-50 network.
Table 10. The fooling rates of adversarial examples generated at different azimuths in the ResNet-50 network.
Azimuth 19 21 23 25 27
CAR100%100%100%100%99.32%
JEEP51.94%53.33%58.77%55.43%57.50%
VAN98.91%100%99.39%100%100%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Luo, B.; Cao, H.; Cui, J.; Lv, X.; He, J.; Li, H.; Peng, C. SAR-PATT: A Physical Adversarial Attack for SAR Image Automatic Target Recognition. Remote Sens. 2025, 17, 21. https://doi.org/10.3390/rs17010021

AMA Style

Luo B, Cao H, Cui J, Lv X, He J, Li H, Peng C. SAR-PATT: A Physical Adversarial Attack for SAR Image Automatic Target Recognition. Remote Sensing. 2025; 17(1):21. https://doi.org/10.3390/rs17010021

Chicago/Turabian Style

Luo, Binyan, Hang Cao, Jiahao Cui, Xun Lv, Jinqiang He, Haifeng Li, and Chengli Peng. 2025. "SAR-PATT: A Physical Adversarial Attack for SAR Image Automatic Target Recognition" Remote Sensing 17, no. 1: 21. https://doi.org/10.3390/rs17010021

APA Style

Luo, B., Cao, H., Cui, J., Lv, X., He, J., Li, H., & Peng, C. (2025). SAR-PATT: A Physical Adversarial Attack for SAR Image Automatic Target Recognition. Remote Sensing, 17(1), 21. https://doi.org/10.3390/rs17010021

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop