1. Introduction
Synthetic aperture radar (SAR) is widely used in military and civilian fields for its ability to image targets with high resolution under all-time and all-weather conditions [
1,
2,
3]. However, unlike natural images, it is difficult for humans to intuitively understand SAR images without resorting to interpretation techniques. The most popular interpretation method at present is the SAR automatic target recognition (SAR-ATR) technology based on deep neural networks (DNNs) [
4,
5,
6,
7,
8]. With their powerful representation capabilities, DNNs outperform traditional supervised methods in image classification tasks. Yet, some researchers have proved that DNN-based SAR target recognition models are vulnerable to adversarial examples [
9].
Szegedy et al. [
10] first proposed the concept of adversarial examples, that is, a well-designed tiny perturbation can lead to the misclassification of a well-trained recognition model. This discovery makes adversarial attacks become one of the biggest threats to artificial intelligence (AI) security. Thus far, researchers have proposed a series of adversarial attack methods, which can be divided into two categories from the perspective of prior knowledge: white-box attacks and black-box attacks. In white-box conditions, the attacker has high access to the victim model, which means that the attacker can utilize lots of prior information to craft adversarial examples. The typical white-box methods are gradient-based attacks [
11,
12], boundary-based attacks [
13], saliency map-based attacks [
14], etc. Conversely, in black-box conditions, the biggest challenge for attackers is that they can only access the output information of the victim model or even less. The representative black-box methods are probability label-based attacks [
15,
16], decision-based attacks [
17], and transferability-based attacks [
18]. While the above methods achieve fantastic attack performance, they all fool DNNs with data-dependent perturbations, i.e., each input corresponds to a different adversarial perturbation, which is hard to satisfy in real-world deployments. Moosavi et al. [
19] first proposed a universal adversarial perturbation (UAP) that can deceive DNNs independently of the input data. Subsequently, the work in [
20] designed a universal adversarial network to learn the mapping from noise to UAPs and demonstrated the transferability of UAPs across different network structures. Mopuri et al. [
21] argue that it is difficult for attackers to obtain the training dataset of the victim model, so to reduce the dependence on the dataset, they proposed a data-free method to generate UAPs by destroying the features extracted by convolutional layers. Another data-free work [
22] used class impressions to simulate a real data distribution, generating UAPs with high transferability. In the field of remote sensing, Xu et al. [
23] were the first to investigate the adversarial attack and defense in safety-critical remote sensing tasks. Meanwhile, they also proposed the mixup attack [
24] to craft universal adversarial examples for remote sensing data. Furthermore, researchers [
25] have successfully attacked an advanced YOLOv2 detector in the real world with just a printed patch. Thus, a further study on adversarial examples, especially UAPs, is necessary for both attackers and defenders.
With the wide application of DNNs in the field of SAR-ATR, researchers have embarked on investigating the adversarial examples of SAR images. In terms of data-dependent perturbations, Li et al. [
26] used the FGSM and BIM algorithms to produce abundant adversarial examples for a CNN-based SAR image classification model and comprehensively analyzed various factors affecting the attack success rate. The work in [
27] presented a Fast C&W algorithm for real-time attacks that introduces an encoder network to generate adversarial examples through the one-step forward mapping of SAR images. To enhance the universality of adversarial perturbations, Wang et al. [
28] utilized the method proposed in [
19] to craft UAPs for SAR images and achieved high attack success rates. In addition, the latest research [
29] has broken through the limitations of the digital domain and implemented the UAP of SAR images in the signal domain by transmitting a two-dimensional jamming signal.
Although the above methods perform well in fooling SAR target recognition models, they are vulnerable and inefficient in practical applications. Specifically, existing attack methods work on the assumption that the adversarial perturbations can be fully fed to the victim model, while this is not always true in practice, i.e., in many cases the perturbations fed to the model are incomplete, resulting in the failure of the adversarial attacks. We attribute the failure to the vulnerability of adversarial attacks and call this situation perturbation offset. For ease of understanding, we detail a specific example in
Figure 1. On the other hand, we calculate the model’s attention heatmaps [
30] through layer-wise relevance propagation (LRP) [
31], which is used to analyze the relevance of each pixel in the SAR image to the recognition results. The pixel-wise attention heatmaps can be found in
Section 4.3. The fact is that the background regions of SAR images have little relevance to the model’s outputs, and the features that greatly impact the recognition results are mainly concentrated in the target regions. However, existing attack methods fool DNN models by global perturbations so that massive time and computing resources are allocated to design perturbations for low-relevance background regions, which is undoubtedly inefficient. Therefore, the vulnerability and inefficiency of adversarial attacks are pending to be solved in real-world implementations.
In this paper, we propose a semi-whitebox [
32] attack network called Universal Local Adversarial Network (ULAN) to generate UAPs for target regions of SAR images. Specifically, we first calculate the model’s attention heatmaps through LRP to locate the target regions in SAR images that have high relevance to the recognition results. Then, we utilize U-Net [
33] to learn the mapping from noise to UAPs and craft the adversarial examples by adding the generated local perturbations to the target regions. In this way, attackers can focus perturbations on the high-relevance target regions, which significantly improves the efficiency of adversarial attacks. Meanwhile, the proposed method also can make the adversarial perturbations be fed to the victim model as completely as possible, preventing perturbation offset to the greatest extent.
The main contributions of this paper are summarized as follows.
- (1)
We are the first to evaluate the adversarial attacks against DNN-based SAR-ATR models in the case of perturbation offset and analyze the relevance of each pixel in SAR images to the recognition results. Our research reveals the vulnerability and inefficiency of existing adversarial attacks in SAR target recognition tasks.
- (2)
This paper designs a generative network to craft UAPs for the target regions of SAR images under semi-white-box conditions. The proposed method requires model information only during the training phase. Once the network is trained and given inputs, it can generate adversarial examples in real time for the victim model through one-step forward mapping without requiring access to the model itself anymore. Thus, our method possesses higher application potential than traditional iterative methods.
- (3)
Experiments indicate that the proposed method not only prevents perturbation offset effectively but also achieves comparable attack performance to the conventional global UAPs by perturbing only a quarter or less of the SAR image area. Furthermore, we evaluate the attack performance of ULAN under small sample conditions. The results show that given five images per class, our method can cause a misclassification rate of over 70% in non-targeted attacks and make the probability of victim models outputting specified results in targeted attacks close to 80%.
The rest of this paper is organized as follows.
Section 2 introduces the relevant preparation knowledge. In
Section 3, we describe the proposed method in detail. The experimental results are shown in
Section 4. The discussions and conclusions are given in
Section 5 and
Section 6, respectively.
3. The Proposed Universal Local Adversarial Network (ULAN)
The framework of the Universal Local Adversarial Network (ULAN) is shown in
Figure 2. To describe the training process of ULAN more clearly, we divide it into four steps. The first step uses a generator to learn the mapping from normal distribution noise into universal adversarial perturbations (UAPs). Next, the second step calculates the pixel-wise attention heatmaps of the surrogate model through layer-wise relevance propagation (LRP). Then, the third step utilizes UAPs and attention heatmaps to craft adversarial examples of SAR images. Finally, the fourth step computes the training loss and updates the generator’s parameters through backward propagation. Note that the victim model is a white box in the training phase, but in the testing phase it is a black box, and thus, we calculate the heatmap of the surrogate model as an alternative to the victim network’s heatmap. This chapter will introduce each of the above steps in detail.
3.1. Structure of Generative Network
In order to craft UAPs independently of the input data, this paper trains a generative network
to transform the normal distribution noise
into a UAP
as follows:
where
and
have the same size, denoted as
. Meanwhile, we set the size of SAR images to
. Since the generated
is a local perturbation, the relationship between
and
is
.
The characteristics of SAR images should be taken into account when choosing the generative network. First of all, a SAR image mainly consists of the target and background clutter. Yet, the features that have great impact on the recognition results are mainly concentrated in the target region, which only occupies a tiny part of the SAR image. Second, compared to natural images, the professionalism and confidentiality of SAR images make them challenging to access. This means that we need to consider adversarial attacks under small sample conditions, so a lightweight generator is necessary to prevent network overfitting. In summary, this paper takes U-Net as the generator to craft UAPs.
Figure 3 shows the detailed U-Net structure. It was first proposed to segment biomedical images [
33] and mainly consists of an encoder and a decoder. The encoder extracts features by down-sampling the input data, while the decoder recovers the data by up-sampling feature maps. The biggest difference between the U-Net and other common encoder-decoder models is that the former introduces a skip connection operation to fuse features from different layers. Specifically, both the encoder and the decoder consist of four sub-blocks. The encoder block contains two
convolutional layers and a
max-pooling layer, while the decoder block contains a
transposed convolutional layer and two
convolutional layers. Note that the last layer of the decoder utilizes a
convolutional layer to make the number of input and output channels identical. The network parameters are given in
Table 1.
3.2. Layer-Wise Relevance Propagation (LRP)
To analyze the relevance of each pixel in SAR images to the recognition results, we must obtain the DNN model’s attention heatmaps first. In this paper, we apply layer-wise relevance propagation (LRP) [
31], which takes as input the model’s logits outputs and outputs the pixel-wise attention heatmaps of the surrogate model
. For an easy explanation, we suppose
is an
l-layer DNN without the softmax output layer.
Figure 4 illustrates the network’s forward propagation and LRP.
The left of
Figure 4 shows a standard forward propagation, which takes a SAR image
x as input and outputs a logits vector
. A common mapping from one layer to the next one can be expressed as follows:
where
and
denote the pre-activation and post-activation of the corresponding node (the superscript and subscript denote layer and node indices, respectively),
is an activation function,
and
can be understood as the weight and local pre-activation between nodes
and
, and
is a bias term. The activation function
is usually nonlinear, such as the hyperbolic tangent
or the rectification function
, which can enhance the network’s representation capacity. Note that the input and output layers typically do not include activation functions, and the output
is a logits vector without softmax operations.
As for LRP, given a target class output
as input, its output is a pixel-wise attention heatmap reflecting the image regions most relevant to
. Specifically, we sequentially decompose the relevance of each node for the target class output
from the neural network’s output layer to the input layer. Meanwhile, the backward propagation of the relevance must satisfy the following conservation property:
A common decomposition is to allocate the relevance according to the ratio of local to global pre-activations in the forward propagation, as follows:
where
denotes the relevance assigned from node
to node
. This decomposition can approximately satisfy the conservation property in (
8):
Additionally, considering that if
goes to zero, then
will close to infinity, so (
9) can be modified by introducing a stable term
as follows:
In summary, we can calculate the relevance of each node for the target class output through the following recursion formula and backward-pass the relevance until reaching the input layer.
3.3. Adversarial Examples of SAR Images
To add the local perturbations generated in
Section 3.1 to the target regions of SAR images, we determine the perturbation location through the attention heatmaps calculated by
Section 3.2. Therefore, we take the attention heatmap centroid as the perturbation center and design a perturbation function to craft the adversarial examples.
First of all, the coordinates of the image centroid can be calculated by the following formula [
37]:
where
is the zero-order moment of the image, and
and
are the first-order moments of the image. This involves the calculation of higher-order moments, which are generally defined as:
For a digital image, we regard the coordinates of the pixel as a two-dimensional random variable
, and the value of each pixel is regarded as the density of the point. Thus, a gray-scale image can be represented by a two-dimensional gray-scale density function
, and its higher-order moments can be expressed as:
Note that the premise here is a two-dimensional gray-scale image, so we convert the attention heatmap
to a single-channel gray-scale image first and then preprocess it with Gaussian blur and binarization algorithms [
38].
Then, we take the attention heatmap centroid as the perturbation center, so the pixel coordinates corresponding to
, i.e., the perturbation origin, can be derived as:
where
w and
h are the width and height of
,
and
represent the displacement difference between the perturbation center and the perturbation origin in the horizontal and vertical directions, and
means rounding down. Meanwhile, this paper adds a two-dimensional random noise
on the centroid coordinates to improve the generalization of our attack.
Next, we add the UAP
to the perturbed region through the following perturbation function. Let
be a function that takes as input the perturbation origin coordinates
, a UAP
, and the size of SAR images
and outputs an adversarial perturbation
of the same size as SAR images, defined as:
In brief, the adversarial perturbation equals zero at all pixels except the pixels in the perturbed region.
Finally, the adversarial example
can be expressed as:
The clipping operation restricts the pixel values of to the interval of , ensuring that is still an 8-bit gray-scale image.
3.4. Design of Loss Functions
To effectively fool the DNN model with a minor perturbation, we design a loss function consisting of an attack loss and a norm loss . This section will introduce them in detail.
For the non-targeted attack: In this paper, we design an attack loss
on the basis of the following standard cross-entropy loss.
where
is the logits output of the victim model. The above formula actually contains the following softmax operation:
Obviously, the cross-entropy loss in (
19) has been widely used in network training to improve the DNN model’s classification accuracy by increasing the confidence of true classes. Instead, according to (
2), the non-targeted attack can minimize the classification accuracy by decreasing the confidence of true classes, i.e., increasing the confidence of others, and thus, the attack loss
can be expressed as:
Meanwhile, a norm loss
is introduced to limit the perturbation magnitude. We use the traditional
-norm to measure the degree of image distortion as follows:
Then, we apply the linear weighted sum method to balance the relationship between
and
, so the total loss
can be represented as:
where
is a constant that measures the relative importance of the attack’s effectiveness and the attack’s stealthiness.
For the targeted attack: According to (
3), a targeted attack aims to maximize the probability that the victim model recognizes samples as target classes. In other words, we need to increase the confidence of target classes. Thus, the attack loss
of targeted attacks can be expressed as:
The norm loss
is the same as (
22), so the total loss
of the targeted attack can be derived as follows:
4. Experiments
4.1. Dataset and Implementation Details
4.1.1. Dataset
The moving and stationary target acquisition and recognition (MSTAR) dataset [
39] published by the U.S. Defense Advanced Research Projects Agency (DARPA) is employed in our experiments. MSTAR is collected by the high-resolution spotlight SAR and contains SAR images of Soviet military vehicle targets at different azimuth and depression angles. All the experiments were performed under standard operating conditions (SOC), which included ten ground target classes, such as self-propelled howitzers (2S1); infantry fighting vehicles (BMP2); armored reconnaissance vehicles (BRDM2); wheeled armored transport vehicles (BTR60, BTR70); bulldozers (D7); main battle tanks (T62, T72); cargo trucks (ZIL131); and self-propelled artillery (ZSU234). The training dataset contains 2747 images collected at a 17° depression angle, and the testing dataset contains 2426 images captured at a 15° depression angle. More details about the dataset are shown in
Table A1, and
Figure A2 shows the optical images and corresponding SAR images of ten ground target classes.
4.1.2. Implementation Details
Due to the different sizes of SAR images in MSTAR, we first center-cropped the images to
. In practice, however, the target is not necessarily located in the center of the SAR image. Thus, we randomly cropped the cropped images to
again and finally normalized them to
. For the victim models, we adopted six common DNNs, A-ConvNets-BN [
40], VGG16-BN [
41], GoogLeNet [
42], InceptionV3 [
43], ResNet50 [
44], and ResNeXt50 [
45], which were trained on the MSTAR dataset and had a classification accuracy of over
. The surrogate model employed a well-trained VGG16-BN network to approximate the pixel-wise attention heatmap of the victim model. During the training phase, we formed the validation dataset by uniformly sampling
of data from the training dataset and used the Adam optimizer [
46] with a learning rate of
, a training epoch of 15, and a training batch size of 32. The size of UAPs defaults to
, the norm type defaults to the
-norm, and the weight coefficient
defaults to
. The above parameter settings have been experimentally proven to achieve excellent attack performance. We will discuss the influence of parameters on UAPs in
Section 4.7.
Considering that most of the current research aims to craft global adversarial perturbations for SAR images, few scholars have focused on universal or local perturbations. Therefore, in the comparative experiments, we took the methods proposed in [
20,
47] as baselines to compare with ULAN. Note that baseline methods generate global UAPs for SAR images, while our method only needs to perturb local regions. All codes were written in Pytorch, and the experimental environment consisted of Windows 10 with an NVIDIA GeForce RTX 2080 Ti GPU and a
GHz Intel Core i9-9900K CPU).
4.2. Evaluation Metrics
This paper takes into account two factors to comprehensively evaluate the performance of adversarial attacks: the attack’s effectiveness and the attack’s stealthiness. In the experiments, we crafted adversarial examples for all samples in the SAR image dataset, so the victim model’s classification accuracy directly reflects the attack effectiveness of UAPs:
where
and
represent the true and target classes of the input data,
k is the number of target classes, and
is a discriminant function. In non-targeted attacks, the Acc metric reflects the probability that victim models correctly recognize adversarial examples. The lower the classification accuracy of the victim model on adversarial examples, the better the non-targeted attacks. In targeted attacks, the Acc metric represents the probability of victim models identifying adversarial examples as target classes. The higher the Acc metric, the stronger the targeted attacks. In conclusion, the non-targeted attack’s effectiveness is inversely proportional to the Acc metric, and the targeted attack’s effectiveness is proportional to this metric. Moreover, to verify the reliability of attacks, we also compared the confidence level of target classes before and after the attack.
When evaluating the attack stealthiness, in addition to using the
-norm to measure the degree of image distortion, we also introduced the structural similarity (SSIM) [
48], a metric more in line with human visual perception, for a more objective evaluation, defined as:
where
is the adversarial example of
,
,
and
,
are the mean and standard deviation of the corresponding image,
is the covariance, and
,
are the constants used to keep the metric stable. Equation (
27) calculates the mean of the SSIM value between all the samples in the dataset and the corresponding adversarial examples, which ranges from
to 1. The higher the SSIM, the more imperceptible the UAPs, and the better the attack’s stealthiness is.
4.3. Attention Heatmaps for DNN-Based SAR Target Recognition Models
For the six victim models mentioned in
Section 4.1.2, given ten SAR images from different target classes as input, they all correctly classified the targets with high confidence. Then, we calculated pixel-wise attention heatmaps for the victim models by LRP, as shown in
Figure A3. The results are similar to the natural image in
Figure A1, i.e., the pixels that have a great impact on the SAR image classifiers are mainly concentrated in the target regions. Furthermore, we found that the attention heatmaps of different models have similar structures, which proves the feasibility of our method. Specifically, since the victim model is a black box in the testing phase, attackers are unable to directly obtain its attention heatmaps through LRP. However, due to the similarity of attention heatmaps between different DNN models, we can calculate a white-box surrogate model’s attention heatmap as an alternative. Meanwhile, since the attention heatmap of VGG16-BN best matches the target shape and has the clearest boundary, the surrogate model adopts a well-trained VGG16-BN network to approximate the attention heatmap of the victim model.
4.4. Adversarial Attacks without Perturbation Offset
In this experiment, we evaluated the non-targeted and targeted attack performance of each method without perturbation offset. Specifically, we first cropped the SAR images to
, as mentioned in
Section 4.1.2, and then crafted adversarial examples by adding well-designed perturbations to the cropped images, which ensured that the perturbations could be fully fed to the victim model. Note that the structures and parameters of the model were known in the training phase, while these details were unavailable in the testing phase. Moreover, we emphasize that the UAPs generated by baseline methods cover the global SAR images, but our method only needs to perturb target regions. The results of the non-targeted and targeted attacks are shown in
Table 2 and
Table 3, respectively. There are four metrics in the table to evaluate the attack performance: the classification accuracy and target class confidence before and after the attack, the
-norm of image distortion, and the SSIM between clean and adversarial examples.
In the non-targeted attack, the classification accuracy of each DNN model on the testing dataset exceeds
, and the true class confidence is over
. However, after the attack, the average decrease in the classification accuracy exceeds
, and the maximum drop in the true class confidence reaches
. From the perspective of attack effectiveness, the UAN performs the best, followed by ULAN and U-Net, and the worst is the ResNet Generator. Yet, the biggest drawback of baseline methods is that they need to perturb the global regions of size
, but our method perturbs the target regions of size
. Even though ULAN only perturbs a quarter of the SAR image area, it achieves comparable attack performance to the global UAPs. We speculate the reason is that the features within target regions have stronger relevance with the recognition results than others, so a focused perturbation on the target region is more efficient than a global perturbation. In terms of the attack’s stealthiness,
Table 2 lists the
-norm value of image distortion caused by each method and the SSIM between the adversarial examples and clean SAR images. An interesting phenomenon is that sometimes ULAN causes a larger image distortion but still performs better on the SSIM metric than baseline methods. We attribute this to the fact that the human eye is more sensitive to large-range minor perturbations than small-range focused ones, resulting in the superior performance of our method on the SSIM metric. It also illustrates that local perturbations can enhance the imperceptibility of adversarial attacks.
In the targeted attack, we regard the target category as the correct class, so the classification accuracy of DNN models on the testing dataset reflects the data distribution, i.e., each category accounts for about one-tenth of the total dataset. According to
Table 3, adversarial examples lead to a sharp rise in the Acc metric, the average increase reaches
, and the maximum rise of the true class confidence exceeds
. This means that the generated UAPs can induce DNN models to output specified results with high confidence. In general, ULAN is slightly inferior to UAN and U-Net regarding the attack’s effectiveness but performs much better than baseline methods on the attack’s stealthiness. Thus, we believe that given a fixed SSIM value, ULAN can achieve the best attack performance.
To visualize the adversarial examples generated by different methods, we take the VGG16-BN-based SAR-ATR model as the victim network and display the adversarial examples for the non-targeted and targeted attacks in
Figure 5 and
Figure 6, respectively. We list the prediction and confidence output by the victim model at the top of each adversarial example, and the bottom of each figure shows the sizes of the corresponding image and perturbation. As we can see, the UAPs generated by baseline methods fully cover the SAR images fed to the model, while ULAN can locate and perturb the target (green box) region effectively. Meanwhile, according to
Figure 5 and
Figure 6, there are apparent shadow and texture traces in the adversarial examples crafted by baseline methods, which also suggests that the global perturbations are more perceptible than the local ones. In summary, compared to baseline methods, our method can achieve good attack performance with smaller perturbed regions and lower perceptions.
4.5. Adversarial Attacks with Perturbation Offset
We now evaluate the adversarial attacks in the case of perturbation offset. Specifically, we first recover the adversarial examples generated in
Section 4.4 to
and next obtain the input data by randomly cropping the recovered images to
again. In this way, we cause a mismatch between the input and perturbed regions. As shown in
Figure 1, the input and perturbed regions correspond to the red and green box regions such that the adversarial perturbations cannot be fed to the victim model completely, and thus, the perturbation offset condition is constructed. The results of non-targeted and targeted attacks in the case of perturbation offset are shown in
Table 4 and
Table 5, respectively.
The experimental results suggest that perturbation offset severely impacts the attack performance of baseline methods. In non-targeted attacks, the Acc metric of baseline methods deteriorates rapidly, the average increase exceeds
, and the maximum increase in true class confidence reaches
. A similar situation also occurs in targeted attacks, where the UAPs generated by baseline methods are likely to be ineffective in the case of perturbation offset. The average decrease of the Acc metric exceeds
, and the maximum drop in the target class confidence reaches
. In contrast, the attack performance of our method is hardly affected under the same experiment condition. Detailed experimental data are displayed in
Table 4 and
Table 5.
In summary, the global UAPs generated by baseline methods are vulnerable to perturbation offset. They might be ineffective unless the victim model accurately takes the perturbed region as input. However, the local perturbations generated by ULAN only cover the target regions of SAR images so that they can be fed to the model as completely as possible regardless of the input regions, which effectively prevents perturbation offset.
4.6. Adversarial Attacks under Small Sample Conditions
Thus far, we have assumed attackers share full access to any images used to train the victim model. However, the professionalism and confidentiality of SAR images make them challenging to access in practice. In other words, it is difficult for attackers to obtain sufficient data to support the training of attack networks. Therefore, we now evaluate the adversarial attacks under stronger assumptions of attacker access to training data.
We consider an extreme situation where attack networks are trained on a subset containing only 50 samples (5 per class). Specifically, we uniformly sample 50 images from the full training dataset to form the subset and compare the attack performance of attack networks trained on the subset and full training dataset against different DNN models. The results of non-targeted and targeted attacks based on different size datasets are shown in
Table 6 and
Table 7, respectively.
As we can see, the reduction in training data seriously impacts the attack performance of the UAN and ResNet Genenrator. Although a slight deterioration in the Acc metric can be tolerated, the average decrease in the SSIM metric is nearly . This means that the above methods severely sacrifice the attack’s stealthiness for better attack effectiveness, which makes the generated adversarial examples easily detected by defenders. However, ULAN and U-Net still maintain good attack effectiveness and stealthiness under small sample conditions. The average change in the Acc metric in both attack modes is less than , and the mean decrease in the SSIM metric is within .
The reasons for the above results might be due to the skip connection structure of the network and the fixation structure of the SAR image. The decoder of ULAN and U-Net fuses the features from different layers through the skip connection structure, which can help the generator learn the data distribution sufficiently. Moreover, the low dependence on the training data can also be attributed to the fixation structure of the SAR image itself such that its semantic features are more easily extracted and represented than natural images. Thus, our approach can work well in the situation where attackers have difficulty obtaining sufficient training data.
4.7. Influence of Parameters
This section evaluates the attack performance of ULAN trained on different parameter settings, providing guidance for attackers to achieve superior attack performance. The parameters mainly include the perturbation size , the weight coefficient , and the type of -norm.
4.7.1. Perturbation Size
To investigate the influence of the perturbation size
on the attack’s performance, we trained ULAN on seven different size settings:
,
,
,
,
,
, and
. Then, we evaluate the attack performance on the testing dataset, and the results are shown in
Figure 7. As expected, for both non-targeted and targeted attacks, a larger perturbation size improves the attack effectiveness, while the attack stealthiness becomes worse. Meanwhile, we find that when the perturbation size exceeds
, the SSIM metric of each DNN model shown in
Figure 7b,d continuously decreases, while the corresponding Acc metric shown in
Figure 7a,c tends to a stable value. We speculate the reason is that perturbation offset will inevitably occur as the perturbation size increases, resulting in the fact that only partial perturbations can be fed to the victim model such that the attack effectiveness is no longer improved. Therefore, the advised perturbation size in this paper is between
and
.
Furthermore, ULAN has superior attack performance even in the case of perturbation offset, which is quite different from baseline methods. Specifically, according to
Table 4 and
Table 5, a large number of global UAPs generated by baseline methods fail to attack the victim model in the case of perturbation offset. Yet, when the perturbation size reaches
, more than
of the adversarial examples generated by ULAN still work well. This is because the perturbation size is too large to prevent perturbation offset during the training phase. In other words, ULAN itself is trained in the case of perturbation offset. Thus, there is no doubt that a well-trained ULAN has already been equipped with the ability to fool models effectively in the case of perturbation offset.
4.7.2. Weight Coefficient
The weight coefficient
is a constant measuring the relative importance of attack effectiveness and stealthiness, which has a great impact on the attack performance. We now train ULAN on nine different weight coefficients,
,
,
,
,
,
,
,
, and
, and report attack results on the testing dataset in
Figure 8. We can see that for non-targeted attacks, the Acc and SSIM metrics increase as
becomes larger. In targeted attacks, the Acc metric declines as
grows, while the SSIM metric is still increasing. Since the non-targeted attack effectiveness is inversely proportional to the Acc metric, and the targeted attack effectiveness is proportional to this metric, the effectiveness of adversarial attacks becomes worse as
increases. However, in both attack modes, the SSIM metric is always proportional to the attack stealthiness such that UAPs become more imperceptible as
gets larger. Meanwhile,
Figure 8a,c suggests that the Acc metric of each DNN model cannot converge to a stable value, and the corresponding SSIM metric shown in
Figure 8b,d is also constantly changing. Thus, for superior attack performance, attackers are supposed to choose an appropriate weight as needed in the training phase of ULAN.
4.7.3. Type of -Norm
Thus far, we have adopted the -norm to measure the image distortion caused by adversarial attacks. However, in addition to the -norm, there are many distance metrics, such as the -norm and the -norm, etc.
In this section, we evaluate the attack performance of ULAN trained on different distance metrics: the
-norm and the
-norm. Note that the values of image distortion calculated by the two metrics differ by several orders of magnitude, so we set the weight
of the
-norm to
and 10 for the
-norm. The results of non-targeted and targeted attacks are shown in
Table 8. We can find that ULAN trained on the
-norm has better performance on both the attack effectiveness and stealthiness. Therefore, to obtain a more threatening attack network, the advised distance metric in this paper is the
-norm.
5. Discussion
The above research demonstrates that our method can efficiently attack DNN models on the MSTAR dataset. To further investigate the generality of the proposed method in SAR target recognition tasks, we also apply experiments on the FUSAR-Ship dataset [
49]. Specifically, we select four kinds of sub-class targets for the experiments, and the details of the dataset are displayed in
Table A2. Considering the size of SAR images is
, we set the input size of models to
, the perturbation size to
, and the weight coefficient
to
. For the victim models, we adopt four common DNNs, GoogLeNet [
42], InceptionV3 [
43], ResNet50 [
44], and ResNeXt50 [
45]. The attack results of ULAN against DNN models on the FUSAR-Ship dataset are shown in
Table A3. Experiments suggest that our method can fool DNN models on the FUSAR-Ship dataset by perturbing the target regions of SAR images. Meanwhile, the results in
Table A4 indicate that the adversarial examples generated by ULAN prevent perturbation offset effectively. In summary, the method proposed in this paper has a promising application in adversarial attacks against DNN-based SAR-ATR models.