Next Article in Journal
Spatial-Convolution Spectral-Transformer Interactive Network for Large-Scale Fast Refined Land Cover Classification and Mapping Based on ZY1-02D Satellite Hyperspectral Imagery
Next Article in Special Issue
Neural-Network-Based Equalization and Detection for Underwater Acoustic Orthogonal Frequency Division Multiplexing Communications: A Low-Complexity Approach
Previous Article in Journal
Dim and Small Target Detection Based on Energy Sensing of Local Multi-Directional Gradient Information
Previous Article in Special Issue
Research on High Robustness Underwater Target Estimation Method Based on Variational Sparse Bayesian Inference
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Joint Detection and Reconstruction of Weak Spectral Lines under Non-Gaussian Impulsive Noise with Deep Learning

1
Acoustic Science and Technology Laboratory, Harbin Engineering University, Harbin 150001, China
2
Key Laboratory of Marine Information Acquisition and Security (Harbin Engineering University), Ministry of Industry and Information Technology, Harbin 150001, China
3
College of Underwater Acoustic Engineering, Harbin Engineering University, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Submission received: 15 March 2023 / Revised: 20 June 2023 / Accepted: 22 June 2023 / Published: 25 June 2023
(This article belongs to the Special Issue Advanced Array Signal Processing for Target Imaging and Detection)

Abstract

:
Non-Gaussian impulsive noise in marine environments strongly influences the detection of weak spectral lines. However, existing detection algorithms based on the Gaussian noise model are futile under non-Gaussian impulsive noise. Therefore, a deep-learning method called AINP+LR-DRNet is proposed for joint detection and the reconstruction of weak spectral lines. First, non-Gaussian impulsive noise suppression was performed by an impulsive noise preprocessor (AINP). Second, a special detection and reconstruction network (DRNet) was proposed. An end-to-end training application learns to detect and reconstruct weak spectral lines by adding into an adaptive weighted loss function based on dual classification. Finally, a spectral line-detection algorithm based on DRNet (LR-DRNet) was proposed to improve the detection performance. The simulation indicated that the proposed AINP+LR-DRNet can detect and reconstruct weak spectral line features under non-Gaussian impulsive noise, even for a mixed signal-to-noise ratio as low as −26 dB. The performance of the proposed method was validated using experimental data. The proposed AINP+LR-DRNet detects and reconstructs spectral lines under strong background noise and interference with better reliability than other algorithms.

1. Introduction

The single-frequency detection of underwater radiation noise with abundant single-frequency components is crucial for detecting quiet targets [1]. The time-frequency analysis is projected on the time and frequency planes to form a three-dimensional stereogram (lofargram). It presents the abundant features of underwater radiation noise [2]. Therefore, the lofargram is regularly employed to analyze its features for passive sonar signals. However, for low signal-to-noise ratios (SNRs), frequency fluctuations caused by a moving target [3] and a high amount of background noise may weaken spectral-line detection.
The detection of weak spectral lines using a lofargram has long been an attractive research topic. Image-processing methods, neural networks, and statistical models are applied to detect weak spectral lines in a lofargram. Image-processing and neural-network methods obtain spectral-line traces from complex image semantic features; however, their performance is usually unsatisfactory for low SNRs [4,5,6,7]. Some deep-learning methods [8,9,10,11] achieve good line-spectrum estimation, but the SNR requirement is relatively high. To overcome this limitation, deepLofargram was proposed to recover invisible and irregularly fluctuating frequency lines at low SNRs [12]. Furthermore, in lofargrams, when the weak spectral lines are far beyond the perceptual range of human vision, this is referred to as low SNR. A statistical model such as the hidden Markov model (HMM) can track the optimal spectral-line trajectory from multi-frame power-spectrum data [13,14]. Most of the aforementioned studies were applied to marine ambient noise following a Gaussian distribution. In particular, marine ambient noise presents strong impulsive characteristics owing to the superposition of seawater thermal noise, hydrodynamic noise, under-ice noise, biological noise, and other noises [15]. Existing underwater-acoustic-signal-processing methods may be invalidated under such non-Gaussian impulsive noise. To overcome this problem, several studies [16,17,18] have performed statistical analyses and models of non-Gaussian marine ambient noise. It was found that the generation and propagation of underwater impulsive noise are in accordance with the “heavy tail” statistical characteristics of the symmetric α-stable (SαS) distribution [19]. Furthermore, SNR and mixed signal-to-noise ratio (MSNR) have been used to characterize the energies of Gaussian noise and non-Gaussian impulsive noise [19]. Various preprocessors have been proposed to suppress non-Gaussian impulsive noise, which can be described by the SαS distribution, including the standard median filter (SMF) [20,21,22] and the memoryless analog nonlinear preprocessor (MANP) [23]. Nevertheless, weak spectral-line detection is unreliable at low MSNRs.
In recent years, with the introduction and development of deep convolutional structures such as UNet [24], SegNet [25], and LinkNet [26], image-semantic-segmentation technology based on deep learning has developed rapidly. In deep-learning semantic segmentation, the semantic features in an image are captured by finding semantic correlations between pixel points from global or local contextual information. In passive sonar-signal processing, weak spectral lines have time-frequency correlations, making them relatively continuous in a lofargram, even though they cannot be observed. Therefore, we argue that when combined with a preprocessor and a deep convolution structure, a lofargram would be able to handle the detection and reconstruction of weak spectral lines under non-Gaussian impulsive noise. Moreover, by “reconstruction,” we mean that potential spectral-line features are recovered to output a lofargram with significant spectral lines.
In this study, we propose a novel method, called AINP+LR-DRNet, which is suitable for the detection and reconstruction of weak spectral lines under non-Gaussian impulsive noise. The spectral-line detection-and-reconstruction problem is redefined as a binary classification problem. First, an impulsive noise preprocessor (AINP) was applied to suppress the non-Gaussian impulsive noise. Second, a specially constructed DRNet was built to detect and reconstruct weak spectral lines. Third, a dual classification adaptive weighted loss was applied to obtain the optimal DRNet during training iterations. Fourth, the detection performance was further improved by the proposed LR-DRNet algorithm. Finally, we validated the ability of the proposed method to detect and reconstruct weak spectral lines under non-Gaussian-impulse noise using simulated and measured data sets.

2. Proposed Framework and Training

Considering deep learning (DL) techniques, we formulate the spectral-line detection-and-reconstruction problem in a lofargram as a binary classification problem. Thus, binary hypothesis testing can be performed, which is defined as follows:
H 1 : i , j i = T , j = F [ s ( t i , f j ) + u ( t i , f j ) ] H 0 : i , j i = T , j = F u ( t i , f j ) ,
where H 1 and H 0 indicate the presence of spectral-line pixels and noise pixels in a lofargram, respectively. i , j i = T , j = F s ( t i , f j ) describes the set of spectral-line pixels, and i , j i = T , j = F u ( t i , f j ) describes the set of noise pixels.
Thus, the spectral-line detection-and-reconstruction framework are proposed to solve Equation (1). As shown in Figure 1, the proposed framework, with the sampling, detection, and reconstruction algorithm, is illustrated. In the sampling stage, the passive SONAR system collects the acoustic signals and noises. The received data are preprocessed by AINP to construct the dataset. Subsequently, a specially designed LR-DRNet is pre-trained to obtain the optimal model parameters in offline training by adding into an adaptive weighted loss function based on dual classification. The well-trained LR-DRNet is utilized to fine-tune the parameters to detect and reconstruct the measured unlabeled samples in online detection and reconstruction. More details are described below:

2.1. Detection-and-Reconstruction Algorithm

2.1.1. Data Preprocessing

The heavy impulsive noise causes broadband interference in a lofargram. Therefore, appropriate preprocessing is required. In this study, the AINP method [24] is used to nonlinearly suppress the abnormal amplitude in the input signal s ( t ) , which is more prominent than the amplitude threshold θ ( t ) . The influence function for the AINP is as follows:
e ( t ) = s ( t ) 1 , s ( t ) θ ( t ) ( θ ( t ) s ( t ) ) 2 , s ( t ) > θ ( t ) ,
where θ ( t ) can be obtained from Equation (3)
θ ( t ) = ( 1 + 2 θ 0 ) Q 2 ( t ) .
In Equation (3), Q 2 ( t ) represents the second quartile of the absolute value of the input signal g ( t ) , and θ 0 is a coefficient, which is set to 1.5, as in [23].

2.1.2. DRNet Structure

The proposed DRNet is derived from LinkNet [26], including the shared encoder, detection decoder, and reconstruction decoder. One part of the shared encoder, illustrated in Figure 2a, is stacked with a series of residual convolution structures to extract spectral-line features. For spectral-line detection, the detection decoder is added to output result of spectral-line detection (for H 1 and H 0 , respectively). As illustrated in Figure 2b, plugging into the squeeze-and-excitation (SE) blocks, the reconstruction decoder can perform channel enhancement by obtaining the importance of each channel through “squeeze” and “excitation“ operations [27]. The reconstruction decoder outputs spectral line-reconstruction results through FinalConv, as shown in Figure 2c. Moreover, a reconstruction decoder is enabled when the detection decoder announces that H 1 holds during the online detection-and-reconstruction stage. The complete network structure is depicted in Figure 3. As shown in Figure 3, the proposed DRNet involves multi-task learning (MTL). For MTL, the model relies on the relative weighting between each task’s loss, and manually adjusting these weights is difficult and time-consuming. Hence, inspired by [28], the adaptive weighted dual loss function is considered by modeling reconstruction-and-detection-task uncertainty. The details are described in Section 2.1.3.

2.1.3. Adaptive Weighted Loss Function Based on Dual Classification

For MTL, the loss function is weighted for each task loss. Thereafter, the MTL loss can be expressed as:
L = i = 1 I w i L i ,
where w i and L i denote the weight and loss of the i-th task, respectively. The I indicates the number of tasks. In this study, there are spectral-line detection and reconstruction tasks with different loss scales. A basic approach to overcoming the large loss difference between the detection and reconstruction tasks involves the model adaptively adjusting the weights w i according to uncertainty of each task.
As the spectral-line detection-and-reconstruction problem is treated as a binary classification task, an adaptive weighted loss function based on a dual classification is used to train the model. Following the derivation in [28], when f W ( x ) is a sufficient statistic, the following multi-mask likelihood is expressed as
p ( y d , y r | f W ( x ) ) = p ( y d | f W ( x ) ) p ( y r | f W ( x ) ) ,
where f W ( x ) represents the detection-and-reconstruction-prediction results of model with parameters W on input x . y d and y r are the ground-truth labels for detection and reconstruction tasks.
Under random noise, the log-likelihood of the detection and reconstruction task is output through a Softmax function, which can be expressed as
log p ( y d | f W ( x ) ) = log ( Softmax ( 1 σ d 2 f d W ( x ) ) ) = 1 σ d 2 f d W ( x ) log c 1 exp ( 1 σ d 2 f c 1 W ( x ) ) ,
log p ( y r | f W ( x ) ) = log ( Softmax ( 1 σ r 2 f r W ( x ) ) ) = 1 σ r 2 f r W ( x ) log c 2 exp ( 1 σ r 2 f c 2 W ( x ) ) ,
where c 1 and c 2 denote the categories of the detection and reconstruction tasks, respectively. The σ d 2 and σ r 2 denote the observation-noise parameters of the model for the detection and reconstruction tasks, respectively.
When the Softmax likelihood is modeled for the detection and reconstruction, the joint loss L ( W , σ d , σ r ) is
L ( W , σ d , σ r ) = log p ( y d , y r | f W ( x ) ) = log [ Softmax ( 1 σ d 2 f d W ( x ) ) Softmax ( 1 σ r 2 f r W ( x ) ) ] = 1 σ d 2 f d W ( x ) log c 1 exp ( 1 σ d 2 f c 1 W ( x ) ) + 1 σ r 2 f r W ( x ) log c 2 exp ( 1 σ r 2 f c 2 W ( x ) ) = 1 σ d 2 [ f d W ( x ) log c 1 exp ( 1 σ d 2 f c 1 W ( x ) ) ] + 1 σ r 2 [ f r W ( x ) log c 2 exp ( 1 σ r 2 f c 2 W ( x ) ) ] + log c 1 exp ( 1 σ d 2 f c 1 W ( x ) ) / ( c 1 exp ( f c 1 W ( x ) ) ) 1 σ d 2 + log c 2 exp ( 1 σ r 2 f c 2 W ( x ) ) / ( c 2 exp ( f c 2 W ( x ) ) ) 1 σ r 2 1 σ d 2 [ log ( Softmax ( y d , f d W ( x ) ) ) ] + 1 σ r 2 [ log ( Softmax ( y r , f r W ( x ) ) ) ] + log σ d σ r = 1 σ d 2 L d ( W ) + 1 σ r 2 L r ( W ) + log σ d σ r ,
where f d W ( x ) and f r W ( x ) represent the outputs of detection and reconstruction in f c 1 W ( x ) and f c 2 W ( x ) , respectively. Equation (8) can be applied for approximation, as follows:
c 1 exp ( 1 σ d 2 f c 1 W ( x ) ) / c 1 exp ( f c 1 W ( x ) ) 1 σ d 2 σ d , c 2 exp ( 1 σ r 2 f c 2 W ( x ) ) / c 2 exp ( f c 2 W ( x ) ) 1 σ r 2 σ r ,
where Equation (9) becomes equal when σ d , σ r 1 .
Referring to the suggestion in [28], we set s d = log σ d 2 , s e = log σ e 2 . Accordingly, L ( W , σ d , σ r ) can be rewritten as
L ( W , σ d , σ r ) = 2 exp ( s d ) L d ( W ) + 2 exp ( s r ) L r ( W ) + s d + s e .
For the detection and reconstruction loss functions L d ( W ) and L r ( W ) , we adopt the two-class cross-entropy and the class-balanced cross-entropy loss functions in [12], as follows:
L d ( W ) = i = 1 N 1 h log p + ( 1 h ) log ( 1 p ) ,
where N 1 indicates the number of samples in the batch size. Moreover, h { 0 , 1 } represents the H 0 and H 1 hypotheses, and p indicates the probability of the output sample class when using a Softmax function.
L r ( W ) = i = 1 N 2 λ [ f , t G + log p f , t + ( 1 λ ) f , t G log ( 1 p f , t ) ] ,
where λ = G / G and 1 λ = G + / G . The G + and G represent the spectral-line and noise ground-truth label sets, respectively. The p f , t indicates the predicted value of the H 1 samples at the ( f , t ) position by a sigmoid function.
According to Equations (10)–(12), the joint-loss form of the multi-task can be obtained. Simultaneously, two weight parameters, σ d and σ r , are adaptively adjusted during the training process. Thus, the purpose of adaptive loss weighting is achieved.

2.1.4. DRNet-Based Spectral-Line-Detection Algorithm

Inspired by the application of the CNN-based spectrum sensing algorithm [29] in narrowband spectrum sensing, which provides a path for detecting spectral lines in a lofargram, a LR-DRNet algorithm is proposed by considering only a single receiver hydrophone. In the proposed algorithm, we use DRNet for offline training and adopt a threshold-based mechanism for online detection.

Offline Training

In offline training, the dataset of the lofargram is constructed under H 0 and H 1 after applying AINP and labeled as follows:
( O l , Z ) = { ( l ( 1 ) , z ( 1 ) ) , ( l ( 2 ) , z ( 2 ) ) , ... , ( l ( M ) , z ( M ) ) } ,
where O l denotes the set of lofargrams l, and Z is its label. The ( l ( m ) , z ( m ) ) represents the m-th sample in the training set.
For the test statistic, the proposed LR-DRNet can extract weak spectral-line features in a lofargram. The output node of the detection decoder was set to 2 by converting spectral line detection into an image binary classification. After a series of convolutional layers, pooling layers, and activation functions, the probability that the lofargram belongs to H 0 or H 1 can be obtained. For the detection task, Equation (1) can be rewritten as follows:
H 1 : P ( z ( m ) = 1 | l ( m ) ; ϑ ) = h ϑ | H 1 ( l ( m ) ) H 0 : P ( z ( m ) = 0 | l ( m ) ; ϑ ) = h ϑ | H 0 ( l ( m ) ) ,
where h ϑ ( ) represents a nonlinear expression of the model with parameters ϑ . After a Softmax function, the network’s output layer has:
h ϑ | H 1 ( l ( m ) ) + h ϑ | H 0 ( l ( m ) ) = 1 .
Next, Equation (11), as a training-error loss function, can be rewritten as:
J L R - D R N e t ( ϑ ) = 1 K k = 1 K { z ( m ) log h ϑ | H 1 ( l ( m ) ) + ( 1 z ( m ) ) log h ϑ | H 0 ( l ( m ) ) } .
Training LR-DRNet minimizes the error loss in Equation (16) and maximizes the posterior probability of the parameter set ϑ . The optimal parameter set ϑ can be obtained as follows:
ϑ = arg max P ( Z | L ; ϑ ) .
Based on the loss function in Equation (10), the backpropagation algorithm is employed to gradually update the parameters of LR-DRNet. Hence, the well-trained LR-DRNet can be illustrated as follows:
h ϑ | H i ( l ( m ) ) = h ϑ | H 1 ( l ( m ) ) h ϑ | H 0 ( l ( m ) ) .
Considering the Bayesian and Neyman–Pearson (NP) criterion, and assuming that P ( H 0 ) = P ( H 1 ) , the test statistics under the proposed LR-DRNet can be acquired as:
Λ L R - D R N e t = h ϑ | H 1 ( l ( m ) ) h ϑ | H 0 ( l ( m ) ) η ,
where η denotes the detection threshold. The presence or absence of spectral lines in the lofargram can be adjudicated by comparing the test statistic and detection threshold.
Next, the detection threshold should be determined. First, M, noise-sample data sets composed of H 0 lofargrams after applying AINP, are constructed. M lofargrams under H 0 after applying AINP are costructed.
O n = { n ( 1 ) , n ( 2 ) , ... , n ( M ) } .
where O n denotes the set of lofargrams n under H 0 .
The probability of detection ( P D ) and the false-alarm probability ( P f ) are defined as follows:
P D = P [ Λ L R - D R N e t | H 1 > η ] ,
P f = P [ Λ L R - D R N e t | H 0 > η ] .
Subsequently, the data set O n is fed as samples into the pre-trained DRNet and the test statistics of all lofargrams under the H 0 hypothesis are obtained.
Λ L R - D R N e t | H 0 = h ϑ | H 1 ( n ( m ) ) h ϑ | H 0 ( n ( m ) ) .
By arranging these values in descending order to form a sequence T ( m ) , the detection threshold of the artificially set false-alarm-probability value P f can be acquired [29].
Λ L R - D R N e t | H i ( z ˜ ) = h γ ^ | H 1 ( l ( m ) ) h γ ^ | H 0 ( l ( m ) ) H 1 H 0 η .
where is the nearest smaller integer. The Λ L R - D R N e t | H 0 ( m ) denotes the m-th sample value of T ( m ) in descending order.

Online Detection

According to Equation (24), a detection threshold η is set. The unlabeled lofargrams, denoted as z ˜ , are input into the well-trained LR-DRNet. Subsequently, online detection, based on LR-DRNet, is performed, that is,
Λ L R - D R N e t | H i ( z ˜ ) = h γ ^ | H 1 ( l ( m ) ) h γ ^ | H 0 ( l ( m ) ) H 1 H 0 η .
When the test statistic is obtained, we can rapidly decide whether there are spectral lines in a lofargram by comparing it to the preset threshold.

2.2. Training Process

Training was optimized for the loss function in Equation (10) using the mini-batch gradient of the Adam optimizer [30], and by setting s d and s e to log2. The batch size was 128. Xavier weight initialization was performed [31]. As expressed in [12], the network’s loss function is ineffective at converging at low SNRs. Hence, we first pre-trained the model with a learning rate of 10−4 for lofargrams with MSNR ranging from −19 dB to −22 dB. The model was then retrained with a learning rate of 10−5 for lofargrams with MSNR ranging from −23 dB to −26 dB. The learning rate was not fixed and was adjusted according to the cosine annealing warm restart [32] and gradual warmup [33]. Here, the gradual warmup was up to the 10th epoch, the initial restart epoch was set to 15, and the restart factor was set to 2. To prevent network overfitting and the problem of insufficient data, data augmentation was performed during training using methods such as horizontal and vertical flipping of images, random cropping, and grayscale maps. Both the above training procedures were terminated after approximately 300 epochs.

3. Simulation Analysis

This section first introduces the synthesis of the datasets and the network performance evaluation metrics. Subsequently, we illustrate the effectiveness of the proposed method by analyzing the effect of the network structure on performance. Finally, the performances of some existing methods are compared and analyzed through simulations.

3.1. Datasets

Non-Gaussian impulsive noise can be described by an α -stable distribution, whose characteristic function can be expressed as in [34].
φ ( t ) = exp { j b t γ t α [ 1 + j β sgn ( t ) ω ( t , α ) ] } ,
ω ( t , α ) = tan ( π α 2 ) ,   α 1 ( 2 π ) log t ,   α = 1 ,
where 0 < α 2 , 1 β 1 , γ > 0 , and < b < . The characteristic exponent α determines the impulse intensity of the distribution; the higher the value of α , the lower the intensity. The position parameter b determines the center of the distribution. The dispersion coefficient γ measures the sample’s degree of deviation by taking values relative to the mean, which is similar to the variance in a Gaussian distribution. The symmetry parameter β is used to describe the skewness of the distribution. When β = 0 , the distribution is named the SαS distribution.
In this case of 0 < α 2 , only the first order is presented in α -stable distributed noise. Therefore, the SNR defined under traditional Gaussian noise is inapplicable. The mixed signal-to-noise ratio (MSNR) is defined as follows:
M S N R = 10 log 10 ( v s 2 γ ) ,
where v s 2 denotes the signal variance.
The α -stable distribution degenerates into a Gaussian distribution when α = 2 . A conventional SNR measure of the relationship between the signal and noise power can be obtained as follows:
S N R = 10 log 10 ( a 2 v n 2 ) ,
where α and v n 2 denote the signal amplitude and the noise variance, respectively.
A low-frequency spectral line, radiated by underwater and surface vehicles, under Gaussian/non-Gaussian impulsive noise, is discussed in this study. Owing to the motion of vehicles (variable speed or steering), the spectral lines fluctuate even at low frequencies. The fluctuating spectral lines can be simulated using a series of sinusoidal signals. The fluctuating spectral lines observed during the k-th time interval are described as:
s ( t k ) = i = 1 I a i ( t k ) sin ( 2 π f i ( t k ) + φ k ) + n ( t k ) , k = 0 , 1 , ... , T 1 ,
where I represents the number of spectral lines. The f i ( t k ) represents the frequency that subsequently varies t k , meaning that the spectral line has unpredictable fluctuations. The φ k [ 0   , 2 π ] is the initial phase, and n ( t k ) represents the sampling point of α -stable distribution noise in t k -th.
The underwater acoustic channel contains Gaussian and non-Gaussian impulse noises. Before establishing the dataset, the modeling and statistical analysis of the measured marine environmental noise were performed. We first modeled three typical marine ambient noises with normal and α -stable distributions. Figure 4 shows that the α -stable distribution is approximate to the marine ambient noise, particularly at high pulse intensities, which conforms with the results reported in [14]. Subsequently, the characteristic function method [35] was applied to estimate the α -stable distribution parameters and statistically acquire the parameter-distribution regularities. Figure 5 presents the statistical conclusions for the four parameters estimated by the α -stable distribution. The α is distributed between 1.7 and 2.0, indicating that the analyzed marine-ambient-noise data contain weak pulse characteristics. The β is distributed at approximately 0, indicating that the S α S distribution can model the noise. The γ values are relatively low, ranging from 0 to 0.01. The data amplitudes are relatively concentrated, which is consistent with the weak pulse characteristics. The δ is distributed at approximately 0, indicating that the measured noise data are concentrated around the zero value. Therefore, the simulation dataset was synthesized according to the distribution regularities of the parameters above and Equation (30).
In the simulation of the SαS distribution noise, α was randomly selected in the range of [1.3, 2], β is set to 0, and γ , δ were set to 1 and 0, respectively. The fluctuating spectral lines within 100 Hz and MSNR in the range of [−26, −19] dB were considered. The sampling rate f s was 1000 Hz. Our synthetic dataset contained one to five fluctuating spectral lines, and multiple spectral lines had harmonic relations. The SαS distribution noise was added to the time-domain amplitude of sinusoids in the form of Equations (28) and (29) with the MSNR and SNR. Figure 6 presents the H 1 lofargrams of multiple sinusoidal signals of different MSNRs and H 0 lofargrams. The presence of spectral lines in lofargrams is not perceived through the visual senses below −22 dB. For MSNR in the range of −22 dB to −26 dB, we repeated the Monte Carlo simulation 1200 times to simulate the scenario under various parameters, splitting the dataset into 85% for training and 15% for testing. Therefore, our training datasets comprised 9600 H 1 lofargrams and 6800 H 0 lofargrams, while the test set had 1440 H 1 lofargrams and 1200 H 0 lofargrams.

3.2. Evaluation Metrics

The following assessment metrics were utilized to analyze the detection and reconstruction performance.
First, the receiver operating characteristic (ROC) curve was used to evaluate the detection performance. Using Equations (22)–(24), we set various P f to obtain thresholds in the offline training stage. A serial set of P f and P D representing the points of the curve could be obtained, and these points together formed the ROCs.
Second, to evaluate the quality of the reconstruction lofargrams, the mIoU [36] and line-location accuracy (LLA) [37] were employed.
mIOU = 1 k + 1 i = 0 k T P F N + F P + T P .
where T P , F N , F P , and T N denote the true positives, false negatives, false positives, and true negatives, respectively.
LLA = 1 max ( B 1 , B 2 ) ( m , n ) G 1 1 1 + λ min ( i , j ) G 2 ( [ m , n ] [ i , j ] 2 ) .
where B 1 and B 2 denote the accumulation of non-zero elements in the predicted lofargram map G 1 and actual lofargram map G 2 , respectively. The [ m , n ] [ i , j ] 2 indicates the Euclidean distance between the detected spectral lines and the actual spectral lines. We set λ = 1 , as in [37].

3.3. Performance Analysis and Discussion

3.3.1. Necessity of AINP

Figure 7 and Figure 8 compare the performances of the AINP under different intensity levels of impulse noise. As shown in Figure 7, heavy SαS noise creates broadband interference in lofargrams. At MSNR = −22 dB, the interference gradually increased as the  value decreased. After the preprocessing with the AINP method, the broadband interference in the lofargrams was largely suppressed. However, the spectral-line pixels of the lofargram were still mixed with the low-amplitude impulse-noise pixels and were not visually distinguishable. The following LR-DRNet further processed lofargrams containing a significant amount of low-amplitude impulse-noise pixels. To further indicate the necessity of the AINP in the proposed method, Figure 8 presents a comparison of the performances of LR-DRNet and AINP+LR-DRNet. As shown in Figure 8, when α = 1.9 , for the cases of −22 and −23 dB, LR-DRNet and AINP+LR-DRNet exhibited comparable performances. As the MSNRs were further reduced to −25 and −26 dB, the performance of AINP+LR-DRNet was better than that of the LR-DRNet. This implies that LR-DRNet has the ability to adapt to weak impulse noise. However, when α decreased to 1.6, the performance of LR-DRNet degraded dynamically. Thus, it can be concluded that the AINP can effectively suppress the broadband interference caused by heavy SαS noise in lofargrams and is necessary for our method to detect and reconstruct weak spectral lines under non-Gaussian impulsive noise.

3.3.2. Network-Structure Analysis

Specific tasks may require suitable network structures. The simulation analyzed the appropriate network structure for spectral-line detection and reconstruction. The two network structures chosen for this analysis were LR-DRNet18 and LR-DRNet34. As shown in Figure 9, the impact of LR-DRNet depth on performance varies across all MSNRs. Compared with LR-DRNet18, the deeper LR-DRNet exhibited comparable performances in the cases of −22 and −23 dB, and exhibited better performances with −24, −25, and −26 dB, respectively. This shows that increasing the network depth improves network performance. Therefore, the coding layer of the LR-DRNet was set to 34 layers.
Theoretically, the relevance of each feature channel can be automatically determined by the SE structure through learning. This learning of the SE structure determines the significance of each feature channel, which consequently strengthens the desirable features. Therefore, the SE structure needed to be analyzed to determine the performance of the proposed LR-DRNet. As shown in Figure 10, compared with LR-DRNet without SE, LR-DRNet with SE had a significant improvement in detection performance, especially at −25 and −26 dB, along with a slight improvement in reconstruction performance. The parameters of the SE structure participated in the end-to-end network parameter optimization process and optimized the encoder and decoder. Thus, we conclude that the SE structure can significantly enhance detection and reconstruction.

3.3.3. Detection and Reconstruction Performance Evaluation

With AINP used as a preprocessor, the outcomes of the proposed AINP+LR-DRNet were compared under various α values and MSNRs. As shown in Figure 11, the performance gradually decreased with decreases in alpha and MSNR, especially at α = 1.3 and 1.5. The performances were comparable at high MSNR and α values, presenting more advantages at lower MSNR and α values. At an MSNR of −24 dB and an α of 1.7, the proposed AINP+LR-DRNet still had a P D of approximately 78%, a mIOU of 0.59, and a LLA of 0.42. In particular, the stronger impulsive noise intensity and lower MSNR affected the feature-extraction ability of the network, encumbering the detection and reconstruction. Nevertheless, the proposed AINP+LR-DRNet is adaptable to low MSNR and strong impulse-noise intensity.
To verify the feasibility of the proposed LR-DRNet under Gaussian noise, its performance under Gaussian noise was compared with those of other methods. The LR-DRNet34 under a single detection task (LR-DNet34), HMM [14], UNet [24], SegNet [25], ResNet18 [38], ResNet34 [38], and LR-DRNet34 under a single reconstruction task (RNet34) were introduced for performance comparison. To ensure that this comparison was fair, ResNet used the same spectral-line-detection algorithm as LR-DRNet.
Figure 12 compares the detection performances of the proposed LR-DRNet with that of several deep learning methods under Gaussian noise. The proposed LR-DRNet achieved a higher detection rate, particularly at SNR values of −24 dB to −26 dB. Figure 13 presents the differences in the reconstruction performances of the five methods. The reconstruction performance of the proposed LR-DRNet was slightly better than that of RNet34, and better than that of the HMM and other deep-learning methods. In terms of reconstruction, as shown in Figure 14, the proposed LR-DRNet reconstructed weak spectral lines more accurately than the other methods at -25 and −26 dB, while exhibiting comparable performances at −22 and −23 dB. This was consistent with the analysis shown in Figure 13. The excellent detection and reconstruction performance of the proposed LR-DRNet and the superiority of MTL over single-task learning (STL) are illustrated. Thus, the feasibility of the proposed LR-DRNet method under Gaussian noise is illustrated.

3.3.4. Comparison with Existing Methods

The detection and reconstruction performances of the proposed AINP+LR-DRNet were compared with those of other methods. Deep classification networks, such as ResNet34 [38] and DNet34, and a detector based on a Gaussian function (GF) [39] were introduced for the detection. Semantic segmentation structures, such as UNet [24], SegNet [25], RNet34, and HMM [14] were introduced for the reconstruction. In GF, the scale parameter c was set to 2.0, and the impulse intensity α and the dispersion coefficient γ were considered in plotting the ROC curve. The number of search times of the spectral line was set to four in the HMM. For a fair comparison, AINP and the algorithm in Section 2.1.4 were used for all the comparison algorithms, except GF.
First, we compared the detection performances of various methods with that of the proposed AINP+LR-DRNet. Figure 15 presents the ROCs of the four methods for MSNR from −22 dB to −26 dB. The AINP+LR-ResNet34, AINP+LR-DNet34, and the proposed AINP+LR-DRNet exhibited discrepancies, particularly at low MSNR values. Furthermore, the GF and the proposed AINP+LR-DRNet at the same P f were compared. The GF detector filtered out impulse noise with large amplitudes via a nonlinear transformation, which suggests it had the worst performance. The superiority of the proposed AINP+LR-DRNet in detection is attributable to its specially designed network, which matches the spectral-line-detection algorithm, which is highly capable of feature extraction. The structures of other advanced networks and the disadvantages of the features of the traditional detection algorithm at a low MSNR may hinder detection.
Subsequently, we compared the different reconstruction methods. The SegNet, UNet, LinkNet34, and the proposed AINP+LR-DRNet are encoding and decoding networks, which segment features with different scales and complex boundaries by extracting the features of the encoding layer and reconstructing the decoding layer. As indicated in Table 1 and Table 2, the proposed AINP+LR-DRNet outperformed the other methods by a considerable margin. Specifically, the mIOU and LLA of the AINP+HMM among the five MSNRs ranged from 0.4841 to 0.5566 and 0.2336 to 0.3985, respectively. The mIOU and LLA of AINP+SegNet and AINP+UNet in the five MSNRs were approximately 0.4917 to 0.6719 and 0.0246 to 0.5757, respectively. Accordingly, AINP+RNet34 was superior to the previous three methods, ranging from 0.5305 to 0.6881 and 0.2118 to 0.5859, respectively; however, the proposed AINP+LR-DRNet achieved impressive performances, ranging from 0.5387 to 0.6932 and from 0.2777 to 0.5950, respectively. Figure 16 presents the lofargram reconstruction of the five methods. The lofargrams reconstructed by the AINP+HMM appeared as false spectral-line pixels after −23 dB, and the line profile became cluttered. The AINP+UNet and AINP+SegNet still worked at −22 dB, but the spectral line broke at varying degrees after −23 dB, and their integrity was reduced. The proposed AINP+LR-DRNet had a prominent spectral-line profile and a higher integrity at −22, −23, and −24 dB, respectively. At −25 and −26 dB, the spectral line could not be reconstructed in some positions because of the excessive background noise. The unique design of the network structure is more suited to reconstruction than those of other segmentation structures.
Finally, we trained AINP+DNet34, AINP+RNet34, and the proposed AINP+LR-DRNet model separately to examine the validity of the MSL. As displayed in Figure 15 and Figure 16, the proposed AINP+LR-DRNet improved the detection and reconstruction performances after utilizing an adaptive weighted loss function based on dual classification. Owing to the multi-task loss function, detection and reconstruction tasks complement each other by sharing valuable information.

4. Experimental Data Analysis

The detection and reconstruction of the proposed AINP+LR-DRNet in Gaussian/non-Gaussian impulsive noise were verified by the aforementioned simulation analysis. At this point, the weights of the pre-trained model in the simulation were fine-tuned using experimental data. The ability of the proposed AINP+LR-DRNet to detect and reconstruct single and multiple weak spectral lines was analyzed by employing two different experimental datasets, and the performances were compared with those of other methods.

4.1. Reconstruction of Weak Single Spectral Line from Strong Background Noise

The data for single-spectral-line detection and reconstruction were received from an experiment conducted in the South China Sea in July 2021. A vertical line array (VLA) composed of 32 hydrophones with an interval of 2 m was employed at a depth of 275–337 m. The sampling rate of the acoustic collector was 10 kHz. During the experiment, the sound source transmitted a single-frequency signal of 71 Hz and was towed 1.5 to 11 km away from the receiving array at a depth of approximately 20 m. Figure 17 displays the hydrophone arrays used in our experiment. We intercepted 2k signal and noise samples from VLA-1 to VLA-32, which formed the measured sample set.
In Figure 18a, two unreconstructed lofargrams are displayed in the experimental data, with a relatively weak spectral line. The AINP+HMM, AINP+RNet34, and the proposed AINP+LR-DRNet effectively reconstructed the regions with obvious spectral lines, as shown to the left of Figure 18b–d. As the MSNR was low, the HMM reconstructed some false spectral-line pixels, and AINP+RNet34 reconstructed a few spectral-line pixels. The weak spectral line was reconstructed using the proposed AINP+LR-DRNet, despite the strong background noise in the two cases. Meanwhile, the experimental data show that MTL outperformed STL. Consequently, the proposed AINP+LR-DRNet is suitable for extracting weak single spectral lines from noise-dominated lofargrams.

4.2. Weak Multiple-Spectral-Line Reconstruction against Strong Interference Background

Another experiment conducted in the South China Sea in September 2021 was used to detect and reconstruct multiple spectral lines. An ocean-bottom seismometer (OBS) was deployed every 5 km, with a line length of more than 100 km. The entire seabed was initially relatively flat, and it gradually became inclined near the destination. The sampling rate of the OBS was 100 Hz. As shown in Figure 19, the ship sailed along a straight line at a certain speed for the deployment and recovery of the OBS. Therefore, the OBS can collect ship-noise samples at low SNRs, as well as marine-ambient-noise samples. The test data set was formed with 5000 more signal and noise samples from OBS-1 to OBS-25.
As shown in Figure 20a, the spectral lines of the ship were affected by the strong interference. In one of the cases on the left, the spectral lines at 25 Hz and 33 Hz were blurred on the original lofargram. Furthermore, in another case, the original lofargram did not present a spectral line at 16, 25, or 33 Hz. Figure 20b indicates that AINP+HMM can only reconstruct spectral lines with higher SNR, but becomes ineffective under strong interference. In addition, more spectral-line pixels were reconstructed using AINP+RNet34, as shown in Figure 20c. The proposed AINP+LR-DRNet is more appropriate for spectral line reconstruction than HMM and RNet34, thereby highlighting the spectral lines and suppressing noise. Hence, the proposed AINP+LR-DRNet is applicable for multiple weak ship spectral line reconstruction under intense interference.

4.3. Detection Performances with Two Real-World Data

Finally, the detection performances of the GF, AINP+ResNet18, and AINP+DNet34 were compared to evaluate the proposed AINP+LR-DRNet. The P f was certain for a fixed test set. For a fair comparison, the GF was compared with the detection rate under the false alarm rate obtained by the proposed AINP+LR-DRNet.
As summarized in Table 3, GF displayed the lowest values at low SNR. Compared with GF, the P D and P f of AINP+LR-DNet34 were higher. The proposed AINP+LR-DRNet exhibited the highest P D for the two measured datasets, reaching 94.73% and 94.49%, respectively. Values of P f of 2.21% and 5.93% were also obtained, which was the best performance of all the methods. This analysis indicates that the proposed AINP+LR-DRNet has the advantage of detection at low SNR under MTL.

5. Conclusions

In this study, the joint detection and reconstruction of weak spectral lines under non-Gaussian impulsive noise using DL was investigated. First, with DL, the detection and reconstruction of spectral lines were formulated as a binary classification problem. Subsequently, a framework for weak-line-spectrum detection and reconstruction based on AINP and DRNet was developed. Under the developed framework, a LR-DRNet detection algorithm was designed, and the lofargrams after the AINP were used as the input of the LR-DRNet. In particular, LR-DRNet was trained by the dual classification adaptive loss to output high detection results and lofargrams with significant spectral lines. Finally, simulated data and real data from the South China Sea were used to verify the performance of AINP+LR-DRNet. The results show that the proposed AINP+LR-DRNet can effectively detect and reconstruct weak spectral lines under non-Gaussian impulsive noise.
In the future, various underwater acoustic signals and marine ambient noises following other distributions will be examined. Furthermore, weak-spectral-line detection based on unsupervised learning will be considered to alleviate the lack of underwater acoustic data and labeling requirements.

Author Contributions

Conceptualization, Z.L.; methodology, Z.L.; investigation, Z.L., X.W. and J.G.; data curation, X.W. and J.G.; writing—original draft preparation, Z.L.; writing—review and editing, X.W. and J.G.; visualization, Z.L.; simulations, Z.L.; supervision, X.W. and J.G. All authors have read and agreed to the published version of the manuscript.

Funding

We thank the staff of the South China Sea experiment in 2021 for their assistance in providing us with valuable data. This work was supported by the National Natural Science Foundation of China under grant no. 12174078.

Data Availability Statement

The data presented in this paper are available upon reasonable request to the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, Q.; Li, M.; Yang, X. The detection of single frequency component of underwater radiated noise of target: Theoretical analysis. Acta Acust. 2008, 33, 193–196. [Google Scholar]
  2. Cohen, L. Time-frequency distributions—A review. Proc. IEEE 1989, 77, 941–981. [Google Scholar] [CrossRef] [Green Version]
  3. Yu, G.; Yang, T.C.; Piao, S. Estimating the delay-Doppler of target echo in a high clutter underwater environment using wideband linear chirp signals: Evaluation of performance with experimental data. J. Acoust. Soc. Am. 2017, 142, 2047–2057. [Google Scholar] [CrossRef] [PubMed]
  4. Abel, J.S.; Lee, H.J.; Lowell, A.P. An image processing approach to frequency tracking (application to sonar data). In Proceedings of the ICASSP-92: 1992 IEEE International Conference on Acoustics, Speech, and Signal Processing, San Francisco, CA, USA, 23–26 March 1992; IEEE: Piscataway, NJ, USA, 1992; Volume 2, pp. 561–564. [Google Scholar]
  5. Gillespie, D. Detection and classification of right whale calls using an ‘edge’ detector operating on a smoothed spectrogram. Can. Acoust. 2004, 32, 39–47. [Google Scholar]
  6. Khotanzad, A.; Lu, J.H.; Srinath, M.D. Target detection using a neural network based passive sonar system. In Proceedings of the International Joint Conference on Neural Networks, Washington, DC, USA, 18–22 June 1989; Volume 1, pp. 335–440. [Google Scholar]
  7. Leeming, N. Artificial neural nets to detect lines in noise. In Proceedings of the International Conference on Acoustic Sensing and Imaging, London, UK, 29–30 March 1993; IET: London, UK, 1993; pp. 147–152. [Google Scholar]
  8. Izacard, G.; Bernstein, B.; Fernandez-Granda, C. A learning-based framework for line-spectra super-resolution. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 3632–3636. [Google Scholar]
  9. Izacard, G.; Mohan, S.; Fernandez-Granda, C. Data-driven estimation of sinusoid frequencies. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2019; Volume 32. [Google Scholar]
  10. Jiang, Y.; Li, H.; Rangaswamy, M. Deep learning denoising based line spectral estimation. IEEE Signal Process. Lett. 2019, 26, 1573–1577. [Google Scholar] [CrossRef]
  11. Jiang, Y.; Zhang, T.; Zhang, W. Model-Based Neural Network and Its Application to Line Spectral Estimation. arXiv 2022, arXiv:2202.06485. [Google Scholar]
  12. Han, Y.; Li, Y.; Liu, Q. DeepLofargram: A deep learning based fluctuating dim frequency line detection and recovery. J. Acoust. Soc. Am. 2020, 148, 2182–2194. [Google Scholar] [CrossRef]
  13. Paris, S.; Jauffret, C. Frequency line tracking using hmm-based schemes [passive sonar]. IEEE Trans. Aerosp. Electron. Syst. 2003, 39, 439–449. [Google Scholar] [CrossRef]
  14. Luo, X.; Shen, Z. A sensing and tracking algorithm for multiple frequency line components in underwater acoustic signals. Sensors 2019, 19, 4866. [Google Scholar] [CrossRef] [Green Version]
  15. Nikias, C.L.; Shao, M. Signal Processing with Alpha-Stable Distributions and Applications; Wiley-Interscience: Hoboken, NJ, USA, 1995. [Google Scholar]
  16. Webster, R.J. A random number generator for ocean noise statistics. IEEE J. Ocean. Eng. 1994, 19, 134–137. [Google Scholar] [CrossRef]
  17. Traverso, F.; Vernazza, G.; Trucco, A. Simulation of non-white and non-Gaussian underwater ambient noise. In Proceedings of the 2012 Oceans-Yeosu, Yeosu, Republic of Korea, 21–24 May 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 1–10. [Google Scholar]
  18. Song, G.; Guo, X.; Li, H. The α stable distribution model in ocean ambient noise. Chin. J. Acoust. 2021, 40, 63–79. [Google Scholar]
  19. Wang, J.; Li, J.; Yan, S. A novel underwater acoustic signal denoising algorithm for Gaussian/non-Gaussian impulsive noise. IEEE Trans. Veh. Technol. 2020, 70, 429–445. [Google Scholar] [CrossRef]
  20. Vijaykumar, V.R.; Mari, G.S.; Ebenezer, D. Fast switching based median–mean filter for high density salt and pepper noise removal. AEU Int. J. Electron. Commun. 2014, 68, 1145–1155. [Google Scholar] [CrossRef]
  21. Sheela, C.J.J.; Suganthi, G. An efficient denoising of impulse noise from MRI using adaptive switching modified decision based unsymmetric trimmed median filter. Biomed. Signal Process. Control 2020, 55, 101657. [Google Scholar] [CrossRef]
  22. Chanu, P.R.; Singh, K.M. A two-stage switching vector median filter based on quaternion for removing impulse noise in color images. Multimed. Tools Appl. 2019, 78, 15375–15401. [Google Scholar] [CrossRef]
  23. Barazideh, R.; Sun, W.; Natarajan, B.; Nikitin, A.V.; Wang, Z. Impulsive noise mitigation in underwater acoustic communication systems: Experimental studies. In Proceedings of the 2019 IEEE 9th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 7–9 January 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 880–885. [Google Scholar]
  24. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015, Proceedings of the 18th International Conference, Munich, Germany, 5–9 October 2015; Part III 18; Springer International Publishing: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  25. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
  26. Chaurasia, A.; Culurciello, E. Linknet: Exploiting encoder representations for efficient semantic segmentation. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–4. [Google Scholar]
  27. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  28. Kendall, A.; Gal, Y.; Cipolla, R. Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7482–7491. [Google Scholar]
  29. Liu, C.; Wang, J.; Liu, X. Deep CM-CNN for spectrum sensing in cognitive radio. IEEE J. Sel. Areas Commun. 2019, 37, 2306–2321. [Google Scholar] [CrossRef]
  30. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  31. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, Sardinia, Italy, 13–15 May 2010; JMLR Workshop and Conference Proceedings. pp. 249–256. [Google Scholar]
  32. Goyal, P.; Dollár, P.; Girshick, R. Accurate, large minibatch sgd: Training imagenet in 1 hour. arXiv 2017, arXiv:1706.02677. [Google Scholar]
  33. Loshchilov, I.; Hutter, F. Sgdr: Stochastic gradient descent with warm restarts. arXiv 2016, arXiv:1608.03983. [Google Scholar]
  34. Samorodnitsky, G.; Taqqu, M.S.; Linde, R.W. Stable non-gaussian random processes: Stochastic models with infinite variance. Bull. Lond. Math. Soc. 1996, 28, 554–555. [Google Scholar]
  35. Koutrouvelis, I.A. An iterative procedure for the estimation of the parameters of stable laws: An iterative procedure for the estimation. Commun. Stat-Simul. Comput. 1981, 10, 17–28. [Google Scholar] [CrossRef]
  36. Garcia-Garcia, A.; Orts-Escolano, S.; Oprea, S. A review on deep learning techniques applied to semantic segmentation. arXiv 2017, arXiv:1704.06857. [Google Scholar]
  37. Lampert, T.A.; O’Keefe, S.E.M. A survey of spectrogram track detection algorithms. Appl. Acoust. 2010, 71, 87–100. [Google Scholar] [CrossRef] [Green Version]
  38. He, K.; Zhang, X.; Ren, S. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  39. Luo, J.; Wang, S.; Zhang, E. Signal detection based on a decreasing exponential function in alpha-stable distributed noise. KSII Trans. Internet Inf. Syst. TIIS 2018, 12, 269–286. [Google Scholar]
Figure 1. Proposed spectral-line detection-and-reconstruction framework.
Figure 1. Proposed spectral-line detection-and-reconstruction framework.
Remotesensing 15 03268 g001
Figure 2. Structural diagram of each part in DRNet. (a) Structure of convolutional modules in Encoder (n). (b) Structure of the decoding layer. (c) Structure of the FinalConv layer.
Figure 2. Structural diagram of each part in DRNet. (a) Structure of convolutional modules in Encoder (n). (b) Structure of the decoding layer. (c) Structure of the FinalConv layer.
Remotesensing 15 03268 g002
Figure 3. Architecture of DRNet.
Figure 3. Architecture of DRNet.
Remotesensing 15 03268 g003
Figure 4. Comparison of the modeling of the normal distribution and α -stable distribution under various disturbances. (a) In a quiet environment; (b) in a ship-interference environment; (c) under airgun interference.
Figure 4. Comparison of the modeling of the normal distribution and α -stable distribution under various disturbances. (a) In a quiet environment; (b) in a ship-interference environment; (c) under airgun interference.
Remotesensing 15 03268 g004
Figure 5. Estimation results of α-stable distribution parameters. (a) α -value distribution statistics; (b) β -value distribution statistics; (c) γ -value distribution statistics; (d) δ -value distribution statistics.
Figure 5. Estimation results of α-stable distribution parameters. (a) α -value distribution statistics; (b) β -value distribution statistics; (c) γ -value distribution statistics; (d) δ -value distribution statistics.
Remotesensing 15 03268 g005
Figure 6. Lofargrams: (a) only the SαS distributed noise lofargram under H 0 ; (b) lofargram of the signal at MSNR = −5 dB under H 1 ; (c) lofargram of the same signal at MSNR = −15 dB under H 1 ; (d) lofargram of the same signal at MSNR = −22 dB under H 1 .
Figure 6. Lofargrams: (a) only the SαS distributed noise lofargram under H 0 ; (b) lofargram of the signal at MSNR = −5 dB under H 1 ; (c) lofargram of the same signal at MSNR = −15 dB under H 1 ; (d) lofargram of the same signal at MSNR = −22 dB under H 1 .
Remotesensing 15 03268 g006
Figure 7. Comparison of original and AINP outcomes with different values of α at MSNR = −22 dB. (a) α = 1.6 ; (b) α = 1.9 .
Figure 7. Comparison of original and AINP outcomes with different values of α at MSNR = −22 dB. (a) α = 1.6 ; (b) α = 1.9 .
Remotesensing 15 03268 g007
Figure 8. Performances of the LR-DRNet and the AINP+LR-DRNet under α = 1.6 and α = 1.9 . (a) ROC; (b) mIOU; (c) LLA.
Figure 8. Performances of the LR-DRNet and the AINP+LR-DRNet under α = 1.6 and α = 1.9 . (a) ROC; (b) mIOU; (c) LLA.
Remotesensing 15 03268 g008
Figure 9. Performances of the AINP+LR-DRNet with 18 and 34 depths in different MSNRs. (a) ROC; (b) mIOU; (c) LLA.
Figure 9. Performances of the AINP+LR-DRNet with 18 and 34 depths in different MSNRs. (a) ROC; (b) mIOU; (c) LLA.
Remotesensing 15 03268 g009
Figure 10. Performances of the AINP+LR-DRNet with and without SE in different MSNRs. (a) ROC; (b) mIOU; (c) LLA.
Figure 10. Performances of the AINP+LR-DRNet with and without SE in different MSNRs. (a) ROC; (b) mIOU; (c) LLA.
Remotesensing 15 03268 g010
Figure 11. Detection and reconstruction with the SαS distribution noise, with α values of 1.3, 1.5, 1.7, and 1.9. (a) ROC; (b) mIOU; (c) LLA.
Figure 11. Detection and reconstruction with the SαS distribution noise, with α values of 1.3, 1.5, 1.7, and 1.9. (a) ROC; (b) mIOU; (c) LLA.
Remotesensing 15 03268 g011
Figure 12. Comparison of ROCs of different methods under Gaussian noise in different SNRs. (a) SNR = −22 dB; (b) SNR = −23 dB; (c) SNR = −24 dB; (d) SNR = −25 dB; (e) SNR = −26 dB.
Figure 12. Comparison of ROCs of different methods under Gaussian noise in different SNRs. (a) SNR = −22 dB; (b) SNR = −23 dB; (c) SNR = −24 dB; (d) SNR = −25 dB; (e) SNR = −26 dB.
Remotesensing 15 03268 g012
Figure 13. Comparison of the reconstruction performance of LR-DRNet, RNet34, SegNet, UNet, and HMM under different SNRs. (a) mIOU; (b) LLA.
Figure 13. Comparison of the reconstruction performance of LR-DRNet, RNet34, SegNet, UNet, and HMM under different SNRs. (a) mIOU; (b) LLA.
Remotesensing 15 03268 g013
Figure 14. Reconstruction of five methods in different SNRs. The original lofargrams with SNR in the range of [−26, −22] dB are shown in (a). The same samples were reconstructed by HMM, SegNet, UNet, RNet34, and LR-DRNet34 in different SNRs, as shown in (bf).
Figure 14. Reconstruction of five methods in different SNRs. The original lofargrams with SNR in the range of [−26, −22] dB are shown in (a). The same samples were reconstructed by HMM, SegNet, UNet, RNet34, and LR-DRNet34 in different SNRs, as shown in (bf).
Remotesensing 15 03268 g014
Figure 15. Comparison of ROCs of four methods under the SαS distribution noise for different MSNRs. (a) MSNR = −22 dB, (b) MSNR = −23 dB, (c) MSNR = −24 dB, (d) MSNR = −25 dB, and (e) MSNR = −26 dB.
Figure 15. Comparison of ROCs of four methods under the SαS distribution noise for different MSNRs. (a) MSNR = −22 dB, (b) MSNR = −23 dB, (c) MSNR = −24 dB, (d) MSNR = −25 dB, and (e) MSNR = −26 dB.
Remotesensing 15 03268 g015
Figure 16. Reconstruction results of different methods in different MSNRs. The original lofargrams as shown in (a). Reconstructed by AINP+HMM, AINP+SegNet, AINP+UNet, AINP+RNet34, and the proposed AINP+LR-DRNet with MSNR in the range of [−26, −22] dB, as shown in (bf).
Figure 16. Reconstruction results of different methods in different MSNRs. The original lofargrams as shown in (a). Reconstructed by AINP+HMM, AINP+SegNet, AINP+UNet, AINP+RNet34, and the proposed AINP+LR-DRNet with MSNR in the range of [−26, −22] dB, as shown in (bf).
Remotesensing 15 03268 g016
Figure 17. Schematic of ship movement and VLA deployment.
Figure 17. Schematic of ship movement and VLA deployment.
Remotesensing 15 03268 g017
Figure 18. Lofargram reconstruction results of experimental data-1 by AINP+HMM, AINP+RNet34, and the proposed AINP+LR-DRNet, shown in (bd), respectively. (a) Original lofargrams, which were not reconstructed.
Figure 18. Lofargram reconstruction results of experimental data-1 by AINP+HMM, AINP+RNet34, and the proposed AINP+LR-DRNet, shown in (bd), respectively. (a) Original lofargrams, which were not reconstructed.
Remotesensing 15 03268 g018
Figure 19. Schematic of OBS deployment and recovery in the experiment.
Figure 19. Schematic of OBS deployment and recovery in the experiment.
Remotesensing 15 03268 g019
Figure 20. Lofargram reconstruction results of experimental data-2 by AINP+HMM, AINP+RNet34, and the proposed AINP+LR-DRNet in (bd), respectively. (a) Two different original lofargrams.
Figure 20. Lofargram reconstruction results of experimental data-2 by AINP+HMM, AINP+RNet34, and the proposed AINP+LR-DRNet in (bd), respectively. (a) Two different original lofargrams.
Remotesensing 15 03268 g020
Table 1. mIOU values of different methods for different MSNRs.
Table 1. mIOU values of different methods for different MSNRs.
MethodsMSNR/dB
−22−23−24−25−26
AINP+HMM0.55660.53830.52360.50470.4841
AINP+SegNet0.65840.59090.53010.50270.4917
AINP+UNet0.67190.62050.56600.52050.4991
AINP+RNet340.68810.66550.63000.58330.5305
AINP+LR-DRNet340.69320.66880.63160.59050.5387
Table 2. LLA values of different methods for different MSNRs.
Table 2. LLA values of different methods for different MSNRs.
MethodsMSNR/dB
−22−23−24−25−26
AINP+HMM0.39850.35990.32770.28400.2336
AINP+SegNet0.55270.41820.19160.07340.0246
AINP+UNet0.57570.51690.34060.14160.0499
AINP+RNet340.58590.57740.52210.43280.2118
AINP+LR-DRNet340.59500.57830.53730.44240.2777
Table 3. Detection performances on practical data.
Table 3. Detection performances on practical data.
MethodsAn Experiment in July 2021An Experiment in September 2021
P f P D P f P D
GF2.21%62.03%5.93%22.79%
AINP+LR-DNet3411.0%89.47%14.83%76.47%
AINP+LR-DRNet342.21%94.73%5.93%94.79%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Z.; Guo, J.; Wang, X. Joint Detection and Reconstruction of Weak Spectral Lines under Non-Gaussian Impulsive Noise with Deep Learning. Remote Sens. 2023, 15, 3268. https://doi.org/10.3390/rs15133268

AMA Style

Li Z, Guo J, Wang X. Joint Detection and Reconstruction of Weak Spectral Lines under Non-Gaussian Impulsive Noise with Deep Learning. Remote Sensing. 2023; 15(13):3268. https://doi.org/10.3390/rs15133268

Chicago/Turabian Style

Li, Zhen, Junyuan Guo, and Xiaohan Wang. 2023. "Joint Detection and Reconstruction of Weak Spectral Lines under Non-Gaussian Impulsive Noise with Deep Learning" Remote Sensing 15, no. 13: 3268. https://doi.org/10.3390/rs15133268

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop