Next Article in Journal
The Contact Phase in Vehicle–Pedestrian Accident Reconstruction
Next Article in Special Issue
MFFNet: A Building Extraction Network for Multi-Source High-Resolution Remote Sensing Data
Previous Article in Journal
An Energy-Efficient Optimization Method for High-Speed Rail Communication Systems Assisted by Intelligent Reflecting Surfaces (IRS)
Previous Article in Special Issue
GLFFNet: A Global and Local Features Fusion Network with Biencoder for Remote Sensing Image Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved Sea Ice Image Segmentation Using U2-Net and Dataset Augmentation

Institute of Geospatial Information, Information Engineering University, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(16), 9402; https://doi.org/10.3390/app13169402
Submission received: 13 July 2023 / Revised: 12 August 2023 / Accepted: 17 August 2023 / Published: 18 August 2023
(This article belongs to the Special Issue Deep Learning in Satellite Remote Sensing Applications)

Abstract

:
Sea ice extraction and segmentation of remote sensing images is the basis for sea ice monitoring. Traditional image segmentation methods rely on manual sampling and require complex feature extraction. Deep-learning-based semantic segmentation methods have the advantages of high efficiency, intelligence, and automation. Sea ice segmentation using deep learning methods faces the following problems: in terms of datasets, the high cost of sea ice image label production leads to fewer datasets for sea ice segmentation; in terms of image quality, remote sensing image noise and severe weather conditions affect image quality, which affects the accuracy of sea ice extraction. To address the quantity and quality of the dataset, this study used multiple data augmentation methods for data expansion. To improve the semantic segmentation accuracy, the SC-U2-Net network was constructed using multiscale inflation convolution and a multilayer convolutional block attention module (CBAM) attention mechanism for the U2-Net network. The experiments showed that (1) data augmentation solved the problem of an insufficient number of training samples to a certain extent and improved the accuracy of image segmentation; (2) this study designed a multilevel Gaussian noise data augmentation scheme to improve the network’s ability to resist noise interference and achieve a more accurate segmentation of images with different degrees of noise pollution; (3) the inclusion of a multiscale inflation perceptron and multilayer CBAM attention mechanism improved the ability of U2-Net network feature extraction and enhanced the model accuracy and generalization ability.

1. Introduction

Sea ice has a substantial impact on global climate change, geophysical activities such as ocean surface physical properties and currents, and economic and social activities such as maritime shipping and transportation [1]. The freezing and thawing of sea ice and sea ice drift in winter interfere with sea-related engineering, marine trade, and various offshore industrial production activities to varying degrees. Internationally, the melting of Arctic sea ice is intensifying due to global warming, and the Arctic shipping route connecting the Atlantic Ocean and the Pacific Ocean will soon open, shortening the voyage from the eastern coast of China to the east coast of the United States (instead of through the Panama Canal) by at least two thousand nautical miles, and reducing the distance from the northern port of China to the western and northern coastal ports of Europe by 25–55% [2]. Therefore, sea ice monitoring is important for maritime shipping, environmental changes, and disaster prevention. There are many satellites used for sea ice remote sensing in the world, such as Landsat series, SPOT series, Worldview series, GeoEye series, Sentinel series and so on. China’s sea ice remote sensing satellites mainly include GF series satellites, FY series meteorological satellites and HY series satellites [3]. Therefore, the use of sea ice remote sensing images to automatically identify and extract sea ice has a broad application prospect.
Currently, sea ice segmentation methods can be roughly divided into three types: threshold segmentation, machine learning, and deep learning. For example, Wang used multiple thresholding and random forest methods to invert FY-4A Bohai Sea regional images and introduced a seed-filling algorithm to revise the results, which improved the accuracy of sea ice inversion under non-clear sky conditions [4]. Li proposed turbid seawater end elements for Bohai Sea ice and water classification, used multifeature binomial tree classification to solve the problem of the difficult distinction between turbid seawater and sea ice, and used mixed-image element decomposition to enter the pixel interior to analyze its category [5]. Li et al. proposed an AL-TSVM sea ice image classification method from the perspective of combining active learning with semisupervised learning by combining a small amount of labeled, known training data and the overall data feature and spatial distribution patterns [6]. Han et al. obtained a combination of informative and low-similarity superior bands using the mutual information similarity metric and classified them using a support vector machine [7]. Zhou et al. improved the OSSP algorithm in three aspects: training set composition, classification result output, and tilted image geometry correction, to improve the classification accuracy of ship-based images [8]. Yu proposed a pixel-level domain relationship context classification method for sea ice spatial neighborhood relations [9].
In the field of deep learning, Dowden et al. evaluated SegNet and PSPNet101 neural networks based on self-built training and testing sets. The sea ice classification dataset consisted of 1090 images with labels; for the test set of 104 images, the classification accuracy was 98.3% or better for both, validating the applicability of deep learning methods for sea ice detection [10]. Han et al. proposed a multilevel, feature-fusion image classification method based on a residual network PCA method to extract the first principal component of the original image, used a residual network to deepen the number of network- layer FPN, PAN, and SPP modules to increase the mining between layer and layer features, merged the features between different layers, and used the hyperspectral image of Bohai Bay for validation; the method improved the sea ice classification accuracy [11].
Shi proposed using the PCANet network to select adaptive convolutional filter banks to mine sea ice depth features, adding hash binarization mapping and chunked histograms to enhance feature separation and reduce feature dimensionality. The author designed a two-branch, multisource, remote sensing, deep learning model for optical and SAR images to obtain good classification results with fewer training samples [12]. The improved SIS-U-Net network outperformed the classical U-Net network by adding a residual structure and void space pyramidal pooling structure to the U-Net network [13]. Cui et al. used the convolutional neural network (CNN) model for image segmentation and selected the appropriate cost and activation functions according to the principle of migration learning. They examined the HJ-1A/B Bohai Sea ice images as the experimental data source labeled samples and achieved better experimental results [14]. Han et al. proposed a spectral–spatial joint feature concept for hyperspectral sea ice image classification and designed a three-dimensional (3D-CNN) model to extract the deep spectral–spatial features of sea ice by conducting sea ice classification experiments using two hyperspectral datasets, Baffin Bay and Bohai Bay, with experimental results based on a single-feature CNN algorithm [15]. Han et al. used the advantage of a CNN in deep feature extraction to design a deep learning network structure for SAR and optical images to achieve sea ice image classification by feature extraction and feature-level fusion of heterogeneous data; the effectiveness of the method was verified using two sets of heterogeneous satellite data in the Hudson Bay area [16]. Zhang et al. classified the Beaufort Sea and Severnaya Zemlya based on a Micro Sea Ice Residual Convolution Network (MSI-ResNet); the MSI-ResNet method performed better than the traditional support vector machine (SVM) classifier for identifying sea ice [17]. Cheng Wen et al. proposed an automatic LFSI extraction method for the Laptev Sea in the eastern Arctic Ocean based on the conditional generative adversarial network Pix2Pix and validated it experimentally using true color images from the Moderate Resolution Imaging Spectroradiometer (MODIS) [18].
In terms of data augmentation, Liu et al. proposed a data augmentation method based on image gradients, which can freely choose the number of expansions and image sizes to effectively expand the dataset, and demonstrated through comparison experiments that the accuracy of the network model was improved after expanding the dataset by this method; the improvement was more obvious when the dataset was small, and the model accuracy could be improved by effectively reducing overfitting. The improvement is more obvious in the case of small datasets and can effectively reduce overfitting to improve the model accuracy [19]. Ziqi et al. fused random probability resampling with adaptive scale equalization, added the fusion expansion algorithm to different target detection algorithms for experiments, and verified that the expansion algorithm can effectively reduce misdetection and false detection of small targets in road scenes [20]. By improving the generator and the discriminator of the Wasserstein–Generative Adversarial Network (W-GAN) and introducing reconstruction and perceptual style loss to enhance the ability of generating remote sensing images by ship, Yang et al. used the remote sensing images generated by ship-WGAN to train the image recognition model, and the recognition accuracy was substantially improved by sample expansion of the generated samples to achieve the effect of data augmentation [21]. The recognition accuracy was substantially improved by expanding the generated samples, and data enhancement was achieved.
Sea ice dataset labeling requires manual interpretation and mapping, and relies on ship-based and shore-based observations, which are costly. In this context, to obtain the segmentation of sea ice remote sensing images under limited sample conditions and achieve good generalization of the semantic segmentation model, this study improved two aspects of data augmentation and network structure, and used a test set to check the accuracy of the improved model. The experimental area is located at lower latitudes (where data are easier to access and process). We hope to build richer training sets to train lightweight neural networks in the future to obtain more robust deep learning models, which will later realize automated sea ice extraction and operational release of satellite-based remote sensing data. In the next step, we will realize a more detailed sea ice segmentation, which classifies sea ice into new ice, one-year ice, multiyear ice, etc. In addition to this, we will fuse more remote sensing data (e.g., synthetic aperture radar (SAR)) and use this method for the study of sea ice in the polar regions.

2. Materials and Methods

2.1. Data Augmentation

Difficulty in acquiring optical images of sea ice, the high cost of labeling, and an insufficient number of samples are obstacles to sea ice segmentation, classification, and detection. Owing to complex sensor factors and special weather conditions, image noise and cloud occlusion are major obstacles to the inversion of optical sea ice images. Image augmentation is an effective method for solving data limitations in deep learning model training. Data augmentation not only increases the number of samples in the dataset but also improves the generalization ability of neural networks. The main data augmentation methods used in this study were affine transform, fuzzy, mirror image, noise, and optical spatial transform augmentations [22].
In noisy data augmentation experiments, the Gaussian random noise is representative of noise type in image processing. In this study, we used the Box–Muller transform method to generate Gaussian random noise and studied the effects of noise with different parameters on the model accuracy and generalization ability. The principle is as follows: the joint two-dimensional distribution of two mutually independent Gaussian random numbers with zero mean and the same variance is radially symmetrical, and the Gaussian random number output by the algorithm can be considered to be the coordinates of a random point in the two-dimensional plane, the amplitude of which is transformed from a random number obeying a uniform distribution on the interval. Its phase is obtained by multiplying a uniform random number on the interval, and the random point is mapped onto the Cartesian coordinate axis. The corresponding coordinate point is a random number that follows a Gaussian distribution.
X and Y are assumed to obey normal distributions and the random variables; X and Y , are transformed as (Equations (1) and (2)):
X = R cos ( θ )
Y = R sin ( θ )
The distribution functions are described by Equations (3)–(5):
P R ( R r ) = 0 2 π 0 r 1 2 π e R 2 2 R d θ d R = 1 e r 2 2
P θ ( θ ϕ ) = 0 ϕ 0 1 2 π e R 2 2 R d θ d R = ϕ 2 π
Retrieved from P R inverse function
R = F R 1 ( 1 z ) = 2 ln ( 1 z )
When z follows a uniform distribution in [0, 1], the distribution function of R is P R . Thus, two random variables, U1 and U2, which obey a uniform distribution on [0, 1], can be selected such that
θ = 2 π U 1 , 1 z = U 2 , R = 2 ln U 2
By substituting this into Equations (1) and (2), the normally distributed random quantities, X and Y, were constructed. In this study, based on this principle, Gaussian random numbers were generated from uniformly distributed, pseudorandom numbers to approximately obey a normal distribution [23].
To intuitively understand the strength of the noise, this study used the following strategy to generate the noise: first, the image grayscale value was divided by 255 to normalize to the interval [0, 1]. Then, a noisy image was generated with a mean value of 0 and a variance of a given value of σ . The noise and image were superimposed and set to 1 if the pixel value was greater than 1, and set to 0 if the pixel value was less than 0. Finally, the value was multiplied by 255 to map the pixel grayscale value back to 0–255. The “0.1 noise” mentioned in this paper indicates that “0.1” is the value of the parameter, σ , in the noise generation process.
The study area was the Bohai Sea ice monitoring dataset, and the dataset was the sea ice target monitoring dataset from the visible image of HY-1. The sea ice tag images were obtained from manual image interpretation. The pixel depth of the images was 24 bits, the original images were in red, green, and blue channels, the tag images were 8-bit grayscale images, the sea ice area was marked as 255, and the non-sea ice area was marked as 0. Data were from the following website: https://cg.cs.tsinghua.edu.cn/jittor/news/2021-09-02-00-00-jdet2/ (accessed on 5 July 2022).
The original training set had a total of 1200 images, and the 1200 images were used as the basic unit for augmentation using image rotation, brightness variation, and noise injection, respectively, before the experiment. The training set numbers and compositions are listed in Table 1. To more accurately measure the accuracy of each model, the test set (300 frames) was expanded (90-degree rotation, 180-degree rotation, blurring, and brightening) to obtain an expanded test set (1500 frames).

2.2. U2-Net Retrofit

2.2.1. U2-Net Network and Convolutional Block Attention Module (CBAM)

Semantic segmentation networks have focused on extracting deep features using existing backbones, such as Alexnet, VGG, ResNet, ResNeXt, DenseNet, etc. The above networks have the following problems: first, the network backbones are designed for image classification tasks, and the features they extract represent semantic rather than local details and global contrast information, which are crucial for semantic segmentation; second, these networks need to be pretrained on the ImageNet dataset; third, these networks add feature aggregation modules in order to extract multilayer salient features leading to complex network structure; and fourth, the deeper architecture is achieved by sacrificing the high-resolution feature mapping. The U2-Net network has the following advantages: firstly, the network is a two-layer, nested, U-shaped structure that does not use a pretrained backbone model for image classification, and the model weights are trained from the training set; secondly, the new architecture allows the network to go deeper and obtain high-resolution features without substantially increasing memory and computational costs. In the bottom layer, a new RSU is designed to extract intrastage, multiscale features without reducing the feature mapping resolution. In the top layer, there is a U-Net-like structure, where each stage is populated by a residual U-block. The feature map is downsampled twice after each Encoder and upsampled before passing through each Decoder twice before passing through each decoder. The residual U-blocks in the network can be divided into two categories: Encoder1–Encoder4 and Decoder1–Decoder4. They use the same modular structure of RSU-L, only the depth (L) is different. Taking L = 7 as an example, the maximum downsampling time was 32 times. When Encoder5–6 and Decoder6 used the module RSU-F, the main difference between the RSU-L and RSU-F structures was that there was no more downsampling (purple part). Because the image size has been sufficiently downsampled, much contextual information will be lost when the image size is decreased again, which affects image segmentation. Therefore, RSU-4F is used in the deep layer, and the module decreases the downsampling part and uses multilayer expansion convolution to enlarge the receptive field. The right side shows the output of the fusion of the features of each layer after upsampling to obtain the segmented image as large as the original image, and the feature map of each layer is then convolved by channel number 1 to obtain the final output [24].
The CBAM contains the channel attention mechanism (channel attention module, CAM) and spatial attention mechanism (spatial attention module, SAM), with two sub-modules (Figure 1 and Figure 2, respectively). The channel attention mechanism is a one-dimensional vector obtained by compressing a feature map in the spatial dimension. Mean and maximum pooling operations are used to compress the spatial and channel dimensions. The mean and maximum pooling aggregate the spatial information of the feature map, which is then mapped to the weights of each channel through the convolutional layer or fully connected layer. The original features are multiplied with this vector in the channel dimension to obtain the weighted feature map, and the size of the weights reflects the degree of relevance and importance of the features in the layer (channel) for the key information. CBAM contains a spatial attention mechanism and a channel attention mechanism, and the structure is shown in Figure 3. The CBAM is a simple yet effective attention module for feedforward CNNs. The module sequentially infers attention maps along two separate dimensions, channel and spatial, and then multiplies the attention maps with the input feature map for adaptive feature refinement. It can be seamlessly integrated into any CNN architecture with negligible overhead and is end-to-end trainable along with the base CNNs to improve the accuracy of the CNN [25].
Li et al. proposed an improved YOLOv4-based pavement damage detection model. The model improves the saliency of pavement damage by introducing a convolutional block attention module (CBAM) to suppress background noise and explores the influence of the embedding position of the CBAM module in the YOLOv4 model on detection accuracy. The results indicate that embedding CBAM into the neck and head modules can effectively improve the detection accuracy of the YOLOv4 model [26].
Sun et al. proposed an attention-based feature pyramid module (AFPM), which integrates the attention mechanism based on a multilevel feature pyramid network to efficiently and pertinently extract high-level semantic features and low-level spatial structure features, thus improving the accuracy of instance segmentation [27]. According to the above findings, the convolutional block attention module (CBAM) can independently learn the importance of each channel and space feature, recalibrate the channel and space features, and improve image classification performance [28]. Therefore, in this study, CBAM was added to multiple locations of U2-Net.

2.2.2. SCM-RSU and SC-U2-Net Network

SC-U2-Net refers to U2-Net with Skip connect structure and CBAM improvements. The SC-U2-Net improvement was divided into two parts. First, the SCM-RSU was used instead of the RSU (Figure 4). SCM-RSU is the RSU that incorporates Skip connect structure, CBAM, multiscale dilated convolution. In the downsampling stage, SCM-RSU used the residual structure, and the output of the upper layer skipped the intermediate convolution as the input of the lower convolution directly to reduce the loss of features in the downsampling process. The U2-Net network used multiscale expansion convolution only in RSU-4F, which was located in the deep layer of the network, to extract richer features at different scales. The SCM-RSU deep perception module was changed to multiscale expansion convolution, where the input feature maps were subjected to 1 × 1 convolution with an expansion factor (d) of 1, 3 × 3 convolution with a d of 1, and 3 × 3 expansion with a d of 2, respectively. Subsequently, the outputs of the three convolutions were stitched in the channel dimension, and, finally, channel fusion was performed using a 1 × 1 convolution. In the decoding stage, the output of the same depth encoding and that of the previous decoding depth were processed by the CBAM attention mechanism and inputted.
Secondly, in the overall framework, multiple attention mechanisms were added (Figure 5). The feature map of the encoder output feature map decoder upsampling after using CBAM processing and then input to the corresponding decoder; the decoder output feature map enables CBAM processing and then convolution operation to obtain the feature map with channel number 1; each feature map upsampled to the input image. After upsampling each feature map to the size of the input image, stitching was performed to obtain an output with a channel number of 6 and the same size as the original image, and then CBAM was used to process the output. Finally, the output with a channel number of 1 was obtained using 1 × 1 convolution.

2.3. Sea Ice Image Segmentation Experimental Setup

In order to improve the accuracy of sea ice image segmentation, this paper improves two aspects of data augmentation and the U2-Net network structure. In terms of data augmentation, five datasets were constructed using various data augmentation methods (such as noisy data augmentation), the model weights were obtained by training the U2-Net network with the five datasets, respectively, and the accuracy was checked using the test set. In terms of network structure improvement, a multilayer CBAM attention mechanism and multiscale expanded convolution were added to U2-Net to enhance the network feature extraction ability, and the accuracy and generalization ability of the improved model were checked using the test set. The SC-U2-Net network model was constructed by adding a multilayer, CBAM attention mechanism and multiscale inflation convolution to enhance the network feature extraction ability. The accuracy and generalization ability of the improved model were tested using test data (Figure 6).
The initial learning rate was set to 0.001, and this study used the learning rate warmup. The learning rate warmup [29] and cosine annealing [30] were used to combine the learning rate change strategy. The weights of the neural network were randomly initialized at the beginning of training, and the warmup gradually increased the learning rate from low to high to ensure a good convergence of the network. When the gradient descent algorithm was used to optimize the objective function, cosine annealing reduced the learning rate of the cosine function as it approached the global minimum of the loss value, making the model as close to the optimal value as possible. The loss function is a binary cross-entropy loss function (Equation (7)), where y is the result of the model prediction sample and y is the sample label.
The loss function of U2-Net is calculated in Equation (8), which contains two parts: m = 1 M w s i d e ( m ) l s i d e ( m ) represents the sum of the cross-entropy of the output results of different depth decoder and GT images, and w f u s e l f u s e is the cross-entropy loss of the final output and GT images after multichannel fusion.
L o s s = ( y log ( y ) + ( 1 y ) log ( 1 y )
L = m = 1 M w s i d e ( m ) l s i d e ( m ) + w f u s e l f u s e
The validation accuracy of the model increased with an increase in epochs, and the model started to converge when the epoch reached 360 when the U2-Net network was trained using training set 1. To balance model accuracy and training efficiency, the epoch was set to 360 in this study. Five test sets were used for training, in which the epochs of training datasets 1 and 2 were set to 360, the epochs of datasets 3, 4, and 5 were set to 90, and the epochs of datasets 3, 4, and 5 were set to 90.
The accuracy evaluation metrics selected in this study are the intersection over union (IoU, F1-score, and recall. If sea ice is called a positive case (positive), non-sea ice is called a negative case (negative), and the classifier predicting correctly is noted as true (true) and incorrectly predicting as false (false), and the four basic terms are combined with each other to form the four basic elements of the confusion matrix: true case (TP), false positive case (FP), false negative case (FN), and true negative case (TN). Then IoU, F1-score, and recall are calculated by Equations (9)–(11).
IoU = TP TP + FP + FN
F 1 = 2 TP 2 TP + FN + FP
Re call = TP TP + FN
To distinguish the network model from the dataset used for training, the naming rules for the network and model weights were as follows: The U2-Net network was trained using dataset i, and the names of the model weights were obtained as U2-Net (Training set i). In the illustration, to save space, we use U2-Net-1 to equivalently represent U2-Net (Training set 1).

3. Results

3.1. Data Augmentation Experiments

The average cross-merge ratio of U2-Net (Training set 1) tested on the test set was 0.842, the average recall was 0.897, and the average F1-score was 0.889. U2-Net (Training set 1) predicted noise-free sea ice images well, but there was overfitting of the network weights, and weak noise interfered with the segmentation, as shown in Figure 7d. The recall distribution curves of U2-Net (Training set 1) predicting noise-free images and weakly noisy test set images are shown in Figure 8.
To further study the effect of noise, we trained U2-Net using different training sets to obtain the corresponding model weights (U2-Net (Training set 1) was the weight obtained by training U2-Net on training set 1). We performed a segmentation of the test set containing different noise levels using U2-Net trained from different training sets. The average IoU, average recall, and average F1-score of the test set with different noise levels were counted (Table 2). The curves of average IoU, average recall, and average F1-score with noise level of the test set were made according to Table 2, as shown in Figure 9, Figure 10 and Figure 11, respectively.
Figure 12 and Figure 13 show the results of U2-Net (Training set 2) and model U2-Net (Training set 3) predicting different levels of test set. Model 2 could predict images with low noise, but the predicted noise limit was approximately 0.15; U2-Net (Training set 3) had a stronger generalization ability than model U2-Net (Training set 2) and was able to segment images with more severe noise pollution.
To further verify the usefulness of multiple data augmentation methods, the transformations (e.g., rotation and mirroring) were performed on the test set simultaneously to obtain a test set with different geometric and radiometric characteristics from the original test set in this study. The experimental accuracies are shown in Table 3 using the transformed test sets for U2-Net (Training set 3), U2-Net (Training set 4), and U2-Net (Training set 5).
The accuracy evaluation indices of U2-Net (Training set 4) were higher than those of U2-Net (Training set 1) when predicting the noiseless test set (Figure 14). The accuracy evaluation indices of U2-Net (Training set 5) were higher than those of U2-Net (Training set 3) when predicting the test set with 0.2 noise level (Figure 15).

3.2. SC-U2-Net Network

The SC-U2-Net (Training set 1) and U2-Net (Training set 1) network was trained using the same dataset (Training dataset 1). The accuracies of SC-U2-Net (Training set 1) and U2-Net (Training set 1) were tested using the test set, and the accuracy of SC-U2-Net (Training set 1) and U2-Net (Training set 1) tests are shown in Table 4.
Figure 16 shows some experimental results, where figures (a) to (h) show the sea ice images of different regions, and (1) to (4) show the original image, labeled image, U2-Net segmentation result, and SC-U2-Net segmentation result, respectively. U2-Net was less effective in segmenting some sea ice, such as the broken ice area at the land edge. SC-U2-Net performed better and could extract the outline and some details of sea ice, as shown in (a) of Figure 16.
The accuracy of SC-U2-Net (Training set 5) was examined using an extended test set (1500 images) and compared to U2-Net (Training set 5). The IoU, F1-score, and recall of U2-Net (Training set 5) and SC-U2-Net (Training set 5) on the extended test set are shown in Table 5.
Using the transformed test sets to test U2-Net (Training set 1) and SC-U2-Net (Training set 5), the statistics of IoU, F1-score, and recall per image for each test set were calculated (Table 3; Figure 17). SC-U2-Net (Training set 5) had a much better segmentation effect than U2-Net (Training set 1) on each test set (Table 6). The results in Table 6 show that the simultaneous use of data augmentation and network improvements can improve the accuracy and generalization of the model.

4. Discussion

4.1. Data Augmentation Experiments

U2-Net (Training set 1) is very sensitive to noise, and the low intensity of Gaussian noise makes it difficult for Model 1 to achieve the semantic segmentation of sea ice images. When the test set did not contain noise, the recall was mainly concentrated at 0.8–1.0, and when the noise level was 0.05, the recall was bimodal (partly concentrated in 0.8–1.0 and partly concentrated in 0–0.2, which indicates that the weak noise interfered with the prediction of the model and can only achieve more accurate segmentation for a part of the images, while the segmentation accuracy of another part of the images was close to 0; the noise level was 0.1, and the recall of all predicted images was concentrated around 0. This shows that noise reduces the accuracy of U2-Net semantic segmentation. There are two main reasons why U2-Net (Training set 1) is sensitive to noise. On the one hand, it is the convolutional neural network itself, and there is an overfitting phenomenon in the network training; on the other hand, the training dataset is small, and the single feature of the sea ice image makes it difficult for the network to recognize more complex ice conditions.
To enhance the generalization ability of U2-Net, different levels of noise are added to the training set for augmentation. The training set after adding noise has a minimal loss of accuracy, but the generalization ability of the model is enhanced. The comparative analysis led to the following conclusions: first, for the noise-free dataset, U2-Net (Training set 1) had the best segmentation effect; second, in terms of noise resistance, U2-Net (Training set 3) was better than U2-Net (Training set 2), which was better than U2-Net (Training set 1). Especially in U2-Net (Training set 1) (no noise was added to the training set), the noise made the prediction accuracy decay rapidly. Third, in terms of generalization, the model with the noisy training set resisted the interference of noise; the richer the noise level was, the stronger its noise resistance was. U2-Net (Training set 2) could predict images with low noise, but the predicted noise limit was approximately 0.15; U2-Net (Training set 3) had a stronger generalization ability than model U2-Net (Training set 2) and was able to segment images with more severe noise pollution.
U-Net (Training set 4) was obtained using Training set 4, and the accuracy evaluation indices of U2-Net (Training set 4) were higher than those of U2-Net (Training set 1) when predicting the noiseless test set. U2-Net (Training set 5) was obtained by training U2-Net network with dataset 5, and the accuracy evaluation indices of U2-Net (Training set 5) were higher than those of U2-Net (Training set 3) when predicting the test set with 0.2 noise level. Using multiple data augmentation methods not only improves the model’s ability to cope with complex scene transformations but also improves the accuracy of semantic segmentation. We also constructed additional test sets with affine transformation, mirror flip, and blurring. The augmented training set images in the training of U2-Net (Training set 4) and U2-Net (Training set 5) enabled the network to learn multiperspective and multiscale semantic features, making U2-Net (Training set 4) and U2-Net (Training set 5) cope well with complex scenarios. We also constructed additional test sets with affine transformation, mirror flip, and blurring. U2-Net (Training set 4) and U2-Net (Training set 5) performed better on these test sets.
Data augmentation experiments showed that U2-Net (Training set 1), which was trained using only the original data, was very sensitive to noise and that adding a small perturbation to the grayscale values of the original images caused U2-Net (Training set 1) to fail. This was because the number of images in the training set was small, and the scene was single, which results in U2-Net overfitting. Noisy data augmentation expanded the sample size and improved the generalization ability of the model. U2-Net (Training set 2) and U2-Net (Training set 3) showed a more stable prediction ability when predicting images with noise. The semantic segmentation accuracies of U2-Net (Training set 4) and U2-Net (Training set 5) were close to that of U2-Net (Training set 3) tested on the test set, and the new test set was obtained by subjecting the test set to affine transformation, mirror flipping, and blurring. The semantic segmentation accuracy of U2-Net (Training set 4) and U2-Net (Training set 5) was much better than that of U2-Net (Training set 3) on the new test set because the training sets of U2-Net (Training set 4) and U2-Net (Training set 5) used a variety of data augmentation in the training. The augmented training set images in the training of U2-Net (Training set 4) and U2-Net (Training set 5) enabled the network to learn multiperspective and multiscale semantic features, making U2-Net (Training set 4) and U2-Net (Training set 5) cope well with complex scenarios.

4.2. SC-U2-Net Network

In the comparison experiments of U2-Net and SC-U2-Net, we used the same training set to train U2-Net and SC-U2-Net, so as to exclude the accuracy improvement brought by the increase of dataset, and the experiments show that when using the same training and test sets, the IoU, F1-score, and recall of SC-U2-Net were higher than those of U2-Net. SC-U2-Net is more effective for sea ice segmentation on remote sensing images. We also compared U2-Net (Training set 1) (U2-Net trained without data augmentation) and SC-U2-Net (Training set 5) (SC-U2-Net trained with data augmentation), and the results show that the simultaneous use of data augmentation and network improvement can improve the accuracy and generalization of the model. SC-U2-Net was able to segment the narrowly shaped sea ice with a smaller area, and the segmentation results are basically consistent with the labeled images. U2-Net is effective in segmenting sea ice over a large area on the sea surface, while narrowly shaped sea ice extending into the land interior is ignored. Compared with U2-Net, the improved SC-U2-Net was able to pay more attention to the details of the target and segmented the discontinuous sea ice better.

5. Conclusions

Based on the U2-Net semantic segmentation network, this study expanded the training set using a data augmentation method and investigated the effect of data augmentation on the accuracy and generalization ability of U2-Net. In the case of poor segmentation of some sea ice images, the SC-U2-Net network was constructed by adding a multiscale, inflation convolution and multilayer CBAM attention mechanism on top of U2-Net, and its accuracy was compared with that of the U2-Net network. The study concluded the following: (1) U2-Net could segment the original test set images well, but the model generalization was poor. (2) The multilevel Gaussian noise data enhancement scheme designed in this study improved the noise interference resistance of the network, considered the generalization performance and accuracy of the model, and achieved more accurate segmentation of images with different degrees of noise pollution. (3) In SC-U2-Net, the residual structure reduced the loss of features during downsampling, multiscale inflation convolution increased the perceptual field of deep convolution, and the multilayer CBAM attention mechanism improved the recognition ability of the network for local features. SC-U2-Net had a higher average IoU, average F1-score, and average recall rate than U2-Net for each test set, especially for fragmented sea ice regions.
The limitations of the experiments were as follows: (1) From the experimental data, the amount of training and test data were relatively small, which affects the reliability of the network training effect and test accuracy. (2) In the experimental setup, only U2-Net and SC-U2-Net were compared, and the other networks were not used as references in the accuracy assessment. (3) The experimental results indicated that, although both data augmentation and network improvement could improve the accuracy of semantic segmentation, the improvement was not substantial enough.

Author Contributions

Conceptualization, Y.L., H.L. and S.J.; methodology, Y.L.; software, S.J.; validation, Z.L. and D.F.; formal analysis, S.J.; investigation, S.J.; resources, H.L.; data curation, H.L.; writing—original draft preparation, Y.L.; writing—review and editing, S.J.; visualization, Y.L.; supervision, S.J.; project administration, S.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by National Natural Science Foundation of China (No. 41971427). Songshan Laboratory Project (No. 221100211000-5). National Science Foundation of Henan Province under Grant (No. 222300420592). The High Resolution Remote Sensing, Surveying and Mapping Application Demonstration System (Phase II) (No. 42-Y30B04-9001-19/21).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, M. Research on the Sea Ice Classification and Thickness Detection with High-Resolution and Polarimetric SAR Data. Ph.D. Thesis, College of Information and Control Engineering, China University of Petroleum (East China), Dongying, China, 2016. [Google Scholar]
  2. Liang, J. Maritime Strategic Game in Arctic Region: China and India. S. Asian Stud. Q. 2019, 11, 24–33. [Google Scholar]
  3. Cheng, X.; Chen, Z.; Hui, F. Current Status and Outlook of Space-Based Remote Sensing Observation in Polar Regions of China. Sci. Technol. Foresight 2022, 1, 183–197. [Google Scholar]
  4. Wang, L. Research on Sea Ice Inversion Algorithm Based on Satellite Remote Sensing Data. Master’s Thesis, Nanjing University of Information Engineering, Nanjing, China, 2021. [Google Scholar]
  5. Li, Y. Research on Sea Ice Detection Method Based on the Decomposition of Mixed Pixels. Master’s Thesis, Wuhan University, Wuhan, China, 2020. [Google Scholar]
  6. Li, P. Combining Active Learning and Semi-supervised Learning for Sea Ice Image Classification. Master’s Thesis, Shanghai Ocean University, Shanghai, China, 2018. [Google Scholar]
  7. Han, Y.; Li, J.; Zhang, Y.; Hong, Z. Hyperspectral Sea Ice Detection Using Improved Similarity Metric. Remote Sens. Inf. 2018, 33, 76–85. [Google Scholar]
  8. Zhou, J.; Lu, P.; Wang, Q.; Xie, F.; Li, R. Research on Automatic Detection Algorithm Based on Video Image Acquisition for Ice Surface Features. Hydro Sci. Cold Zone Eng. 2021, 4, 60–65. [Google Scholar]
  9. Yu, Z. Sea Ice Classification of Remote Sensing Image Based on Neighborhood Relationships. Master’s Thesis, CUPB (East China), Dongying, China, 2019. [Google Scholar]
  10. Dowden, B.; De Silva, O.; Huang, W.; Oldford, D. Sea Ice Classification via Deep Neural Network Semantic Segmentation. IEEE Sens. J. 2021, 21, 11879–11888. [Google Scholar] [CrossRef]
  11. Han, Y.; Cui, P.; Zhang, Y.; Zhou, R.; Yang, S.; Wang, J. Remote Sensing Sea Ice Image Classification Based on Multilevel Feature Fusion and Residual Network. Math. Probl. Eng. 2021, 2021, 9928351. [Google Scholar] [CrossRef]
  12. Shi, Q. Homologous and Heterologous Remote Sensing Sea Ice Classification Based on Deep Learning. Master’s Thesis, Shanghai Ocean University, Shanghai, China, 2022. [Google Scholar]
  13. Fang, Y. Research on Sea Ice Area Identification Based on MODIS Satellite Remote Sensing Images. Master’s Thesis, Qingdao University of Science and Technology, Qingdao, China, 2021. [Google Scholar]
  14. Cui, Y.; Zou, B.; Han, Z.; Shi, L.; Liu, S. Application of Convolutional Neural Networks in Satellite Remote Sensing Sea Ice Image Classification: A Case Study of Sea Ice in the Bohai Sea. Acta Oceanol. Sin. 2020, 42, 100–109. [Google Scholar] [CrossRef]
  15. Han, Y.; Gao, Y.; Zhang, Y.; Wang, J.; Yang, S. Hyperspectral Sea Ice Image Classification Based on the Spectral-Spatial-Joint Feature with Deep Learning. Remote Sens. 2019, 11, 2170. [Google Scholar] [CrossRef]
  16. Han, Y.; Liu, Y.; Hong, Z.; Zhang, Y.; Yang, S.; Wang, J. Sea Ice Image Classification Based on Heterogeneous Data Fusion and Deep Learning. Remote Sens. 2021, 13, 592. [Google Scholar] [CrossRef]
  17. Zhang, T.; Yang, Y.; Shokr, M.; Mi, C.; Li, X.; Cheng, X.; Hui, F. Deep Learning Based Sea Ice Classification with Gaofen-3 Fully Polarimetric SAR Data. Remote Sens. 2019, 13, 1452. [Google Scholar] [CrossRef]
  18. Wen, C.; Zhai, M.; Lei, R.; Xie, T.; Zhu, J. Automated Identification of Landfast Sea Ice in the Laptev Sea from the True-Color MODIS Images Using the Method of Deep Learning. Remote Sens. 2023, 15, 1610. [Google Scholar] [CrossRef]
  19. Liu, Z.; Zhang, S.; Liu, Y.; Luo, C.; Li, M. Data Augmentation Method Based on Image Gradient. J. Appl. Sci. 2023, 39, 302–311. [Google Scholar]
  20. Huang, Z.; Liu, X.; Shi, Y.; Lin, C. Small Object Detection in Road Science Base on Data Augmenttation. J. Wuhan Univ. Technol. 2022, 44, 79–87. [Google Scholar]
  21. Yang, Z.; Yang, Y.; Cang, S.; Li, Y.; Hao, Y.; Zhang, F.; Wu, G. Data Augmentation Method of Ship Remote Sensing Images Based on GAN. Appl. Sci. Technol. 2022, 49, 8. [Google Scholar]
  22. Lin, C.; Shan, C.; Zhao, G.; Yang, Z.; Peng, J.; Chen, S.; Huang, R.; Li, Z.; Yi, X.; Du, J.; et al. Review of Image Data Augmentation in Computer Vision. J. Front. Comput. Sci. Technol. 2021, 15, 583–611. [Google Scholar]
  23. Lee, D.-U.; Villasenor, J.D.; Luk, W.; Leong, P.H.W. A Hardware Gaussian Noise Generator Using the Box-Muller Method and Its Error Analysis. IEEE Trans. Comput. 2006, 55, 659–671. [Google Scholar] [CrossRef]
  24. Qin, X.; Zhang, Z.; Huang, C.; Dehghan, M.; Zaiane, O.R.; Jagersand, M. U2-Net: Going Deeper with Nested U-Structure for Salient Object Detection. Pattern Recognit. 2020, 106, 107404. [Google Scholar] [CrossRef]
  25. Woo, S.; Park, J.; Lee, J.; Kweon, I.S. Cbam: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar] [CrossRef]
  26. Li, L.; Fang, B.; Zhu, J. Performance Analysis of the YOLOv4 Algorithm for Pavement Damage Image Detection with Different Embedding Positions of CBAM Modules. Appl. Sci. 2022, 12, 10180. [Google Scholar] [CrossRef]
  27. Sun, Y.; Gao, W.; Pan, S.; Zhao, T.; Peng, Y. An Efficient Module for Instance Segmentation Based on Multi-level Features and Attention Mechanisms. Appl. Sci. 2021, 11, 968. [Google Scholar] [CrossRef]
  28. Zhang, L.; Duan, L. Cross-Scenario Transfer Diagnosis of Reciprocating Compressor Based on CBAM and ResNet. J. Intell. Fuzzy Syst. 2022, 43, 5929–5943. [Google Scholar] [CrossRef]
  29. Xiong, R.; Yang, Y.; He, D.; Zheng, K.; Zheng, S.; Xing, C.; Zhang, H.; Lan, Y.; Wang, L.; Liu, T. On Layer Normalization in the Transformer Architecture. In Proceedings of the International Conference on Machine Learning, Virtual Event, 13–18 July 2020. [Google Scholar]
  30. Loshchilov, I.; Hutter, F. SGDR: Stochastic Gradient Descent with Warm Restarts. arXiv 2016, arXiv:1608.03983. [Google Scholar]
Figure 1. CAM structure.
Figure 1. CAM structure.
Applsci 13 09402 g001
Figure 2. SAM network structure.
Figure 2. SAM network structure.
Applsci 13 09402 g002
Figure 3. CABM network structure.
Figure 3. CABM network structure.
Applsci 13 09402 g003
Figure 4. SCM-RSU structure.
Figure 4. SCM-RSU structure.
Applsci 13 09402 g004
Figure 5. Improving the U2-Net network.
Figure 5. Improving the U2-Net network.
Applsci 13 09402 g005
Figure 6. Flow chart of sea ice segmentation.
Figure 6. Flow chart of sea ice segmentation.
Applsci 13 09402 g006
Figure 7. U2-Net (Training set 1) predicted noise image. (a) Original image; (b) sea ice GT value; (c) noiseless image prediction; (d) 0.05 noise image prediction.
Figure 7. U2-Net (Training set 1) predicted noise image. (a) Original image; (b) sea ice GT value; (c) noiseless image prediction; (d) 0.05 noise image prediction.
Applsci 13 09402 g007
Figure 8. U2-Net (Training set 1) recall distribution probability curve.
Figure 8. U2-Net (Training set 1) recall distribution probability curve.
Applsci 13 09402 g008
Figure 9. Mean value of IoU of U2-Net versus noise level.
Figure 9. Mean value of IoU of U2-Net versus noise level.
Applsci 13 09402 g009
Figure 10. Mean value of F1-score of U2-Net versus noise level.
Figure 10. Mean value of F1-score of U2-Net versus noise level.
Applsci 13 09402 g010
Figure 11. Mean value of recall of U2-Net versus noise level.
Figure 11. Mean value of recall of U2-Net versus noise level.
Applsci 13 09402 g011
Figure 12. U2-Net (Training set 2) predicted noise images. (a) Original image; (b) sea ice GT value; (c) Noiseless image segmentation; (d) 0.1 noise image segmentation; (e) 0.11 noise image segmentation; (f) 0.15 noise image segmentation; (g) 0.16 noise image segmentation; (h) 0.17 noise image segmentation.
Figure 12. U2-Net (Training set 2) predicted noise images. (a) Original image; (b) sea ice GT value; (c) Noiseless image segmentation; (d) 0.1 noise image segmentation; (e) 0.11 noise image segmentation; (f) 0.15 noise image segmentation; (g) 0.16 noise image segmentation; (h) 0.17 noise image segmentation.
Applsci 13 09402 g012
Figure 13. U2-Net (Training set 3) predicted noise images. (a) Original image; (b) sea ice GT value; (c) noiseless image segmentation; (d) 0.15 noise image segmentation; (e) 0.25 noise image segmentation; (f) 0.30 noise image segmentation; (g) 0.40 noise image segmentation; (h) 0.55 noise image segmentation.
Figure 13. U2-Net (Training set 3) predicted noise images. (a) Original image; (b) sea ice GT value; (c) noiseless image segmentation; (d) 0.15 noise image segmentation; (e) 0.25 noise image segmentation; (f) 0.30 noise image segmentation; (g) 0.40 noise image segmentation; (h) 0.55 noise image segmentation.
Applsci 13 09402 g013aApplsci 13 09402 g013b
Figure 14. U2-Net (Training set 1) and U2-Net (Training set 4) predicted noise-free test sets.
Figure 14. U2-Net (Training set 1) and U2-Net (Training set 4) predicted noise-free test sets.
Applsci 13 09402 g014
Figure 15. U2-Net (Training set 3) and U2-Net (Training set 5) predicted 0.2 noise test set.
Figure 15. U2-Net (Training set 3) and U2-Net (Training set 5) predicted 0.2 noise test set.
Applsci 13 09402 g015
Figure 16. Comparison of SC-U2-Net and U2-Net segmentation results.
Figure 16. Comparison of SC-U2-Net and U2-Net segmentation results.
Applsci 13 09402 g016
Figure 17. U2-Net (Training set 1) and SC-U2-Net (Training set 5) accuracy rating box line diagram.
Figure 17. U2-Net (Training set 1) and SC-U2-Net (Training set 5) accuracy rating box line diagram.
Applsci 13 09402 g017
Table 1. Training set composition.
Table 1. Training set composition.
Training Set NumberTraining Set 1Training Set 2Training Set 3Training Set 4Training Set 5
Training Set Composition
Original training set (1200)
90° rotation (1200)
180° rotation (1200)
Horizontal mirroring (1200)
Brightness enhancement (1200)
Brightness reduction (1200)
Random noise (1200)
Gaussian blur (1200)
0.1 (1200)
0.15 noise (1200)
0.20 noise (1200)
Total number of images120012004800960013,200
Note: A checkmark indicates that the dataset used a type of augmentation; “0.1 noise” means 1200 images of the training set with 0.1 level of noise added.
Table 2. U2-Net predicts the average metrics of the test set with different noise levels.
Table 2. U2-Net predicts the average metrics of the test set with different noise levels.
Noise LevelU2-Net (Training Set 1)U2-Net (Training Set 2)U2-Net (Training Set 3)U2-Net (Training Set 4)U2-Net (Training Set 5)
ImouF1RecallImouF1RecallImouF1RecallImouF1RecallImouF1Recall
00.8420.8970.8890.8020.870.8820.8110.8770.8890.8790.9260.9180.8490.9030.9
0.050.4210.490.4420.8020.8710.8950.80.8680.8840.8560.9090.90.850.9060.906
0.100.1460.1720.1540.7920.860.8620.8110.8770.8890.8120.8780.8720.8340.8950.899
0.11 0.7860.8530.8640.7940.8640.8850.760.8310.8220.8230.8870.895
0.13 0.7760.8470.8560.790.8610.8860.7220.7920.7870.8210.8850.897
0.15 0.770.8440.8490.7840.8570.8830.5530.6190.5990.7990.8690.877
0.16 0.7140.8010.7690.7770.8510.877 0.7650.8370.85
0.17 0.570.6570.6130.7740.8490.878 0.7200.7960.804
0.20 0.3460.4270.3570.7710.850.885 0.6280.7010.697
0.25 0.1310.1580.1440.7320.8110.85 0.4840.5530.535
0.30 0.7130.7950.841 0.3050.3490.337
0.45 0.6150.7060.735 0.1250.1450.134
0.55 0.5100.5990.612
0.60 0.4620.5440.553
Table 3. Accuracy of U2-Net on different test sets.
Table 3. Accuracy of U2-Net on different test sets.
ModelsU2-Net (Training Set 3)U2-Net (Training Set 4)U2-Net (Training Set 5)
IndicatorsImouF1RecallImouF1RecallImouF1Recall
Test Set
Original test set0.8110.8770.8890.8590.9160.910.8490.9030.9
90° rotation0.6930.7760.7640.8280.8960.8940.8180.8840.878
180° rotation0.6650.7490.7310.8430.9080.910.8310.8960.882
Blur0.8620.9020.9330.9230.9510.9540.9110.9370.941
Brighten0.6900.7700.7890.8360.8930.9110.8310.8920.897
Table 4. SC-U2-Net (Training set 1) vs. U2-Net (Training set 1) test accuracy.
Table 4. SC-U2-Net (Training set 1) vs. U2-Net (Training set 1) test accuracy.
IndicatorsImouF1Recall
U2-Net (Training set 1)0.8420.8970.889
SC-U2-Net (Training set 1)0.8570.9130.920
Table 5. Accuracy of U2-Net and SC-U2-Net on augmented test set.
Table 5. Accuracy of U2-Net and SC-U2-Net on augmented test set.
IndicatorsImouF1Recall
U2-Net (Training set 5)0.8340.8860.884
SC-U2-Net (Training set 5)0.8360.8970.898
Table 6. Accuracy of U2-Net (Training set 1) and SC-U2-Net (Training set 5) on the augmented test set.
Table 6. Accuracy of U2-Net (Training set 1) and SC-U2-Net (Training set 5) on the augmented test set.
ModelsU2-Net (Training Set 1)SC-U2-Net (Training Set 5)
IndicatorsIoUF1RecallIoUF1Recall
Test Set
Original image0.8120.8650.8570.8470.9070.911
90° rotation0.7930.8570.8450.8270.8940.898
180° rotation0.7860.8550.8400.8170.8870.886
Blur0.8070.8660.8530.8380.9000.901
Brighten0.7970.8560.8430.8360.8970.898
Mean of all test sets0.7990.8600.8480.8320.8960.897
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Li, H.; Fan, D.; Li, Z.; Ji, S. Improved Sea Ice Image Segmentation Using U2-Net and Dataset Augmentation. Appl. Sci. 2023, 13, 9402. https://doi.org/10.3390/app13169402

AMA Style

Li Y, Li H, Fan D, Li Z, Ji S. Improved Sea Ice Image Segmentation Using U2-Net and Dataset Augmentation. Applied Sciences. 2023; 13(16):9402. https://doi.org/10.3390/app13169402

Chicago/Turabian Style

Li, Yongjian, He Li, Dazhao Fan, Zhixin Li, and Song Ji. 2023. "Improved Sea Ice Image Segmentation Using U2-Net and Dataset Augmentation" Applied Sciences 13, no. 16: 9402. https://doi.org/10.3390/app13169402

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop