Next Article in Journal
Prediction of Particle Settling Velocity in Newtonian and Power-Law Fluids Using Artificial Neural Network Model
Next Article in Special Issue
Advancing Biological Research: New Automated Analysis of Immunofluorescence Signals
Previous Article in Journal
Uniaxial Compressive Damage Characteristics of Rock-like Materials with Prefabricated Conjugate Cracks
Previous Article in Special Issue
SIPNet & SAHI: Multiscale Sunspot Extraction for High-Resolution Full Solar Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Super-Resolution Reconstruction Method for Infrared Polarization Images with Sparse Representation of Over-Complete Basis Sets

1
Shanghai Institute of Technical Physics, Chinese Academy of Sciences, Shanghai 200083, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
Key Laboratory of Infrared Detection and Imaging Technology, Chinese Academy of Sciences, Shanghai 200083, China
4
Key Laboratory of Electro-Optical Countermeasures Test & Evaluation Technology, Luoyang 471003, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(2), 825; https://doi.org/10.3390/app14020825
Submission received: 15 December 2023 / Revised: 14 January 2024 / Accepted: 16 January 2024 / Published: 18 January 2024
(This article belongs to the Special Issue Advanced Image Analysis and Processing Technologies and Applications)

Abstract

:
The spatial resolution of an infrared focal plane polarization detection system is limited by the structure of the detector, resulting in lower resolution than the actual array size. To overcome this limitation and improve imaging resolution, we propose an infrared polarization super-resolution reconstruction model based on sparse representation, optimized using Stokes vector images. This model forms the basis for our method aimed at achieving super-resolution reconstruction of infrared polarization images. In this method, we utilize the proposed model to initially reconstruct low-resolution images in blocks. Subsequently, we perform a division by weight, followed by iterative back projection to enhance details and achieve high-resolution reconstruction results. As a supplement, we establish a near-real-time short-wave infrared time-sharing polarization system for data collection. The dataset was acquired to gather prior knowledge of the over-complete basis set and to generate a series of simulated focal plane images. Simulation experimental results demonstrate the superiority of our method over several advanced methods in objective evaluation indexes, exhibiting strong noise robustness in quantitative experiments. Finally, to validate the practical application of our method, we establish a split-focal plane polarization short-wave infrared system for scene testing. Experimental results confirm the effective processing of actual captured data by our method.

1. Introduction

With advancements in infrared technology, infrared polarization imaging has found growing applications in various domains, such as industrial inspection [1,2,3], medical diagnosis [4,5,6], and astronomical remote sensing [7,8,9]. This technology facilitates the augmentation of information acquired through infrared imaging and enables further exploration of post-image processing algorithms. The infrared focal plane polarization detector has garnered significant interest as the central component of infrared polarization imaging. It consists of multiple groups of micro-linear polarization units and infrared detectors [10]. The micro-linear polarization units are arranged in an ordered structure on the surface of the infrared detector, with each group containing four pixels. Each group records four different necessary intensity response values [11]. However, this design structure decreases the spatial resolution of the infrared focal plane polarization imaging system, resulting in a significant impact on the measurement accuracy of polarization information. To address the issue of limited sampling by the detector and the inability to obtain a full-resolution polarization image, super-resolution reconstruction of the polarization image is necessary. This involves utilizing an effective reconstruction algorithm to enhance image details and clarity, compensating for the loss of resolution caused by the system structure.
Super-resolution reconstruction techniques can be categorized based on their implementation methods into interpolation methods [12,13,14,15,16], reconstruction methods [17,18,19], and learning methods [20,21,22,23,24,25,26,27,28]. The interpolation methods include bilinear and bicubic interpolation methods [12], which were the initial algorithms utilized for super-resolution tasks. Subsequently, new algorithms were proposed, such as gradient-based interpolation [13], Newton polynomial-based interpolation (NP) [14], Edge-Aware Residual Interpolation (EARI) [15], and polarization image demosaicking using Polarization Channel Difference Prior (PCDP) [16], which leverage image geometric information and channel correlation. These interpolation techniques offer the advantage of quick reconstruction and effective edge preservation. However, since the degradation information of the actual pixel is not always accurately known, the models employed are not fixed.
The primary objective of super-resolution algorithms in reconstruction methods is to reverse-engineer the imaging process of an image. This involves establishing an observation model that links the low-resolution image with the high-resolution image. To solve the model, image-specific prior knowledge is introduced, either at a local or global level. As an example, Chen et al. combined the 3D block-matching algorithm with the concept of non-local filtering and introduced the Block-matching 3D Projectile On Convex Sets (BPOCS) algorithm [17]. Zhang et al. proposed the Frequency domain Phase-based Projection On Convex Sets (FPPOCS) algorithm, which combines the Wiener filter with the phase spectrum features of high-resolution images [18]. Similarly, in our previous work, we utilized principal component analysis and soft threshold denoising, using micro-scanning to capture sub-pixel micro-displacement frame sequences without altering the optical system’s structure. We proposed the Micro-scanning Polarization Projection On Convex Sets (MPPOCS) algorithm, designed specifically for infrared focal plane polarization detection systems [19]. The primary focus of the reconstruction method’s super-resolution algorithm is the degradation process of the image. It aims to narrow down the range of potential solutions using an observational model, thereby ultimately improving the preservation of the image’s geometry. However, due to the complex degradation factors encountered during the imaging process, artificial modeling cannot fully establish this process comprehensively.
The underlying principle of learning methods involves establishing correspondence between low-resolution and high-resolution images, utilizing the complementary information within similar blocks to preserve signal characteristics. In 2018, Zhang et al. introduced the Polarization Demosaicing Convolutional Neural Network (PDCNN) [20] to address the challenge of polarization super resolution, which lacked a specialized network for focal plane polarizing images. Subsequently, to mitigate accumulated errors caused by formula computation steps, Zeng et al. proposed an end-to-end fully convolutional neural network, FORK-NET [21]. This network takes the focal plane image as input and generates output intensity, polarization degree, and polarization angle images. An objective evaluation index indicated the superior performance of FORK-NET compared to PDCNN. Consequently, researchers explored alternative network models, such as the Multi-Scale Adaptive Weighted Network (MSAWN) [22], Deep Compressed Sensing (DCS) [23], and the sparsely polarimetric image demosaicing model (Sparse-PDM) [24]. These networks primarily target visible light polarization and necessitate extensive training data, typically sourced from Sony IMX250 series visible light polarization cameras. Articles [24,25] describe the publicly available dataset created by these researchers. The advantage of deep learning methods lies in their ability to achieve excellent reconstruction effects. Nevertheless, these methods also possess certain drawbacks that limit their suitability for specific tasks. Super-resolution performance becomes heavily reliant on computational power and data volume, limiting their deployment in computationally constrained and data-limited environments. Furthermore, the lack of clear interpretation in the structural units is evident. Conversely, shallow learning methods derived from traditional machine learning algorithms offer advantages such as adaptability to small datasets, relatively lower computational costs, and interpretability.
In this paper, we propose an infrared polarization super-resolution reconstruction model based on sparse representation. Our model optimizes the Stokes vector images, eliminating the need for mechanical reconstruction of a multi-channel polarization image. To enhance efficiency and accuracy, we construct three over-complete sets of Stokes vector image dictionaries. These dictionaries streamline the calculation process of the polarization degree and polarization angle while minimizing the potential for error.
Furthermore, we introduce a reconstruction method based on our model. This method utilizes the proposed model to initially reconstruct the low-resolution image in blocks. The image is then divided by the weight coefficients of the overlaps, and the details are refined using the Iterative Back Projection (IBP) technique to achieve high-resolution reconstruction results.
To build the dictionary, we designed a near-real-time short-wave infrared time-sharing polarization system prototype, leveraging our previous error analysis work [29]. This prototype enabled us to acquire the necessary dataset for simulation experiments. Additionally, we established a short-wave infrared focal plane polarization system for capturing real data. The experimental results demonstrate the superiority of our proposed method in both objective and subjective evaluation indexes.
The rest of this paper is organized as follows: Section 2 introduces the reconstruction model and algorithm, including the establishment of an over-complete dictionary set, noise analysis, and the iterative solution method. Section 3 presents the simulation and real experiment results, provides the objective evaluation index, and explains the subjective index. The anti-noise performance of the proposed algorithm is analyzed, and the running time of the algorithm is provided. Next, we discuss the results and introduce future research directions. Finally, Section 4 discusses the advantages of the proposed method, summarizes the novelty of the method and our contributions, and describes future research direction.

2. Super-Resolution Methods

An infrared focal plane detection system can capture essential intensity images from four different angles in a single exposure. Due to limitations imposed by the detector structure, the effective resolution of the four declination angle images is reduced by half of the original pixel size, expressed mathematically as follows:
I θ i = 1 , 2 , 3 , 4 M / 2 × N / 2 = I out M × N M ask θ
where I out M × N represents the cumulative light intensity recorded by the detector, analogous to an image obtained from a single exposure at a specific angle. M ask θ is an intensity extraction mask that biases the detection process to isolate the individual angle components within the overall light intensity. This mask is dependent on the pixel arrangement employed by the self-developed detector. Each superpixel within this arrangement consists of a tuple of four pixels. The resulting separated images for each declination angle, denoted as I θ i = 1 , 2 , 3 , 4 M / 2 × N / 2 , exhibit a reduced size of half the original dimensions. The Stokes notation is a widely utilized method for representing polarization states in engineering, and its formula is provided below:
S = [ S 0 S 1 S 2 S 3 ]
where S 0 , S 1 , S 2 , and S 3 are parameters that describe different aspects of the incident light. S 0 represents the total incident light intensity. S 1 is the difference in intensity between the horizontal and vertical directions. S 2 is the intensity difference between polarized light at 45 degrees and 90 degrees concerning a reference direction. S 3 specifically relates to the detection of left-handed versus right-handed circular polarization. In typical projects, the detection of linear polarized light is generally carried out without considering the S 3 component. By using the above formula, it is possible to solve for the Stokes vector of the incident light inversely. This method allows for the retrieval of information about the polarization degree and polarization angle of the current scene.
D P = S 1 2 + S 2 2 S 0 = ( I 0 I 90 ) 2 + ( I 45 I 135 ) 2 ( I 0 + I 45 + I 90 + I 135 ) × 1 2
A P = 1 2 arctan ( S 2 S 1 ) = 1 2 arctan ( I 45 I 135 I 0 I 90 )
When I θ i = 1 , 2 , 3 , 4 M / 2 × N / 2 is directly included in the aforementioned calculation formula, it becomes apparent that the dimensions of D P and A P calculated in this manner are M / 2 × N / 2 . Furthermore, due to the unique arrangement structure of the superpixel, only a single declining-angle light intensity of a given pixel can be obtained in a single exposure. This leads to an instantaneous field error in the calculated polarization signal. Hence, it is imperative to conduct super-resolution reconstruction of I θ i = 1 , 2 , 3 , 4 M / 2 × N / 2 .

2.1. Infrared Polarization Super-Resolution Reconstruction Model

To obtain the high-resolution image I S R , the low-resolution image I L R is used. It is assumed that I S R can be linearly represented as α using the over-complete basis vector space D S R . According to the theory of root sparse coding primitives, we can infer that the following formula holds when α is sufficiently sparse:
I ˜ S R = D S R α α R k   w i t h   α 0 < < k
The over-complete basis vector D S R , which can also be referred to as an over-complete dictionary set, is utilized. The objective is to minimize the Euclidean distance between the reconstructed image I ˜ S R and the actual high-resolution image, while ensuring that α satisfies the sparsity condition. This can be formulated as an optimization problem:
min a α 0 s . t . D S R α I S R 2 2 < ε
Formula (6) satisfies the condition that the square of the 2-norm is smaller than the error ε , rather than just the 2-norm being smaller than the error ε . This choice is justified by previous research which has shown that in the field of super-resolution reconstruction, penalizing high-frequency errors by using the square of the 2-norm yields better results [26,27,28]. When performing super-resolution tasks with the input of image I L R , it is evident that Formula (6) does not have enough known optimization solutions. In this context, the existence of dictionary D L R is assumed. This dictionary shares the sparse representation α with D S R , and solving Formula (6) can be equivalently stated as the following equation:
min a α 0 s . t . G D L R α G I L R 2 2 < ε
In the equation above, G denotes an edge extraction operator, responsible for generating edge features for Formula (7). We observe that the pseudo-polarization information is primarily concentrated in the high-frequency portion. Precisely reconstructing the missing high-frequency content in the targeted high-resolution image holds the utmost significance. Extracting the high-frequency components from low-resolution images proves advantageous for the reconstruction of high-resolution images [30]. Formula (7) corresponds to an L0 problem, which can be approximated through an L1 problem using the Lagrange multiplier method:
min a G D L R α G I L R 2 2 + λ α 1
In Formula (8), λ is an adjustable parameter that is associated with the noise variance of the input image I L R . The derivation process will be conducted later on. With the given image I L R and over-complete dictionary sets D S R and D S R , a set of optimal sparse representations α is obtained by solving Formula (8). Subsequently, the reconstructed image I ˜ S R can be obtained by applying it to Formula (5). The resulting image satisfies the constraint stated in Formula (6).
We introduce adjustable parameter λ into Formula (8), which is derived and analyzed using the Maximum A Posteriori (MAP) theory. According to the MAP theory, the squared term of the L2 norm in Formula (8) can be interpreted as the maximum likelihood estimate of α . Additionally, part of the L1 norm can be considered as the Laplace prior of α . Without taking into account the edge extraction operator G , this prior can be expressed as follows:
α = ar g max α     p ( I L R     |     D L R , α ) p ( α )
The α variable in Formula (9) follows a Laplace distribution with the parameter (0, b). As a result, its probability density function can be expressed as follows:
p ( α ) = 1 2 b e α 1 b
The variable I L R in Formula (9) is distributed according to a Gaussian distribution with the parameter ( D L a , δ 2 ) . Thus, its probability density function can be expressed as follows:
p ( I L R     |     D L R , α ) = 1 2 π δ e 1 2 δ 2 D L R α I L R 2 2
By incorporating Formulas (10) and (11) into Formula (9), we derive the resulting Formula (12):
α * = arg max α log p ( I L R     |     D L R , α ) + log p ( α ) = arg max α 1 2 δ 2 D L R α I L R 2 2 + 1 2 b α 1 = arg min α D L R α I L R 2 2 + δ 2 b α 1
Equation λ = δ 2 / b can be observed from the aforementioned formula. Given the sparsity of α , it is assumed that the scale parameter b remains relatively constant, while λ should be increased in the presence of higher input image noise. For the experiment, a value of 0.2 was assigned to λ .

2.2. Reconstruction Method

I θ i = 1 , 2 , 3 , 4 M / 2 × N / 2 is a low-resolution image. If I θ i = 1 , 2 , 3 , 4 M × N in the same scene is solved, the D P and A P of full-resolution images can be solved using Formulas (3) and (4). Consequently, the Stokes vector needs to be calculated. The calculation process involves finding the difference between two groups of orthogonal light intensities. It is important to note that the different image is significantly influenced by the displacement of the target. Based on the experimental findings in [19], displacement-induced pseudo-polarization has a substantial impact on subjective perception. Therefore, our reconstruction method is expressed in the following equation:
α * = arg min a α 0 S . T . G D L R α G I L R 2 2 < ε I L R = { S 0 L R , ( S 1 / S 0 ) L R , ( S 2 / S 0 ) L R } I S R = D S R α *
S 0 L R is reconstructed as S 0 S R through the use of super resolution. Subsequently, ( S 1 / S 0 ) L R and ( S 2 / S 0 ) L R are reconstructed as ( S 1 / S 0 ) S R and ( S 2 / S 0 ) S R , respectively. Finally, the D P and A P images at full resolution are calculated. This reconstruction approach offers several benefits. First, it eliminates the need to calculate the difference between each angular component, thereby avoiding any potential angle errors. Additionally, the reconstruction of ( S 1 / S 0 ) S R and ( S 2 / S 0 ) S R does not rely on the light intensity of the scene. The light intensity captured by the detector encompasses both natural and fully polarized light components, which can be derived using Malus’ law:
I θ = 1 2 I N + I × P × cos 2 ( θ A )
In the formula above, I θ represents the intensity of partially polarized light at a given declination angle θ . This can also be interpreted as the light intensity received by the detector at that angle. I N represents natural light, while I N represents the total light intensity. I N denotes the degree of polarization, while A represents the polarization angle. The orthogonal difference of polarization is defined [31] as:
I θ = I θ I θ + π 2 = I ( 1 P ) × ( cos 2 ( θ A ) cos 2 ( π 2 + θ A ) )
In the formula above, when θ = 0 ° , I θ = S 1 , and θ = 45 ° , I θ = S 2 . The Stokes representation of the total light intensity I is denoted as S 0 . Therefore, ( S 1 / S 0 ) L R and ( S 2 / S 0 ) L R are influenced by the polarization degree and polarization angle. By reconstructing them, it is possible to achieve better restoration of the D P and A P images at full resolution, instead of directly reconstructing the D P and A P low-resolution images from the high-resolution S 0 , D P , and D P images. The proposed method does not take the entire image as input. Instead, a sliding window reconstruction is performed on a low-resolution image. The window size is set at 5 × 5 , with a step length of 1. During the reconstruction process, the number of overlapping pixels is calculated, and the reconstructed I ˜ S R is divided by the overlap coefficient W . The operational details are presented in the figure below.
As shown in Figure 1, the red segment illustrates a pixel-level reconstruction process where each pixel is reconstructed once, using a weight factor of W = 1 . The yellow part represents another reconstruction process. It is important to clarify that this specific depiction of traversing during the second step is not obligatory, as any traversal with a step size of 1 covering the entire image would be sufficient. It is worth noting that there is an overlap between the yellow and red segments. In terms of pixel-level analysis, the overlapping region carries a weight coefficient of W = 2 . The resulting reconstructed I S R is then divided by the overlap coefficient W . Additionally, the Iterative Back Projection (IBP) algorithm is employed to enhance image details and minimize the reconstruction error of the reconstructed I ˜ S R . The input for the IBP algorithm is composed of the reconstructed image I ˜ S R and the reference image obtained through the sampling of I L R using the bicubic interpolation method.
The overall method consists of the following steps:
(1)
The acquired low-resolution image is initially divided into sliding window blocks with a step size of 1 and window size of 5 × 5 .
(2)
The α of each component is solved using Formula (8), and in this process, the feature symbol search algorithm proposed in [32] is employed.
(3)
The reconstructed image I ˜ S R of each component is then divided by the overlap coefficient W .
(4)
The Iterative Back Projection (IBP) algorithm is used to further reconstruct each component’s W .
(5)
Finally, the Stokes vector image, polarization degree image, and polarization angle image are calculated based on the reconstructed results.

2.3. Establishment of Dictionary Set

Section 2.1 discusses the dictionary, called D L R , which is mentioned as having a shared sparse representation with D S R . To achieve our desired solution, we need to construct a joint dictionary. The optimization problem for this joint dictionary is represented in the downward form:
min D L R , D S R , α U 1 C 1 G D L R α U G T L R 2 2 + 1 C 2 D S R α U T S R 2 2 + λ ( 1 C 1 + 1 C 2 ) α U 1
The variables C 1 and C 2 in the formula above are used to balance the costs of the Dictionary-based Super-Resolution D S R and Dictionary-based Low-Resolution D L R methods. Since the reconstruction method requires a sliding window, the dictionary set is also built using a sliding window approach with a window size of 5 and step size of 1. However, this practice leads to inconsistent matrix sizes, so we up-sample the low-resolution image before building the dictionary set. The values of C 1 and C 2 are set to the number of atoms corresponding to D L R and D S R , respectively. In our case, C 2 is set to 25 and C 1 is set to 100, as our edge extraction operator G is defined as the first and second derivatives in the x and y directions. Additionally, we further simplify Formula (16):
min D U , α U D U α U T U 2 2 + λ α U 1 D U = 1 10 G D L R 1 5 D S R T U = 1 10 G T L R 1 5 T S R
According to [33], to optimize the problem of solving Formula (17), it is necessary to constrain D U . This ensures the existence of both d U , i rank   ( D U ) × 1 and α U atomic sets within D U , which leads to consistent results for d U , i rank   ( D U ) × 1 α U and approaching 0 for α U 1 . As a result, it is crucial to normalize the atomic sets of D U and impose constraints before training. Representation of the entire dictionary set can be expressed using the following formula, with K being the total number of atomic sets in the dictionary set, set to 512 during training:
D U * = argmin D U , α U D U α U T U 2 2 + λ α U 1 s . t . d u . i r a n k ( α U ) 1 2 2 1 , i = 1 , 2 , 3 , K
The formula is not convex on both D U and α U . However, it is convex on one of them. To address this, we follow the approach proposed in [32]. We first initialize D U and solve α U using the feature symbol search algorithm. Then, we fix α U and solve the duality problem of α U using the conjugate gradient method. Solving the duality problem offers the advantage of generating fewer variables during the optimization process, thus making it faster.

3. Results

To validate the proposed method, we created a prototype near-real-time short-wave infrared time-sharing polarization system for collecting training and simulation data, building upon our previous work. As shown in Figure 2. The system utilized a custom short-wave detector that operated within the 0.95~1.65 μm wavelength range. Additionally, a self-designed zoom lens with a resolution of 120 lp/mm was employed. Real-time communication with the host optical signal was facilitated by the fully electronic design of the rotating target wheel component. The system offered a polarization time resolution of 25 frames. A total of 60 scenes, encompassing indoor and outdoor settings, were captured. Amongst these scenes, 54 were used to form the training dictionary set, while 6 scenes were reserved for simulation testing. The S 0 images from the simulation test are depicted in Figure A1 in Appendix A, showcasing cars, a window, a wall, a building, a tower, and a tank. Each image has a resolution of 640 × 512 pixels.
In this paper, four interpolation methods were utilized: the Newton polynomial-based interpolation method (NP) [14], the Edge Sensing Residual Interpolation method (EARI) [15], a polarization difference priority-based interpolation algorithm (PCDP) [16], and the Micro-scanning Polarization Projection On Convex Sets (MPPOCS) algorithm [19], which we previously proposed. As the MPPOCS algorithm is a multi-frame super-resolution algorithm, we modified its input. In the simulation experiment, we simulated the micro-scan image input using a single frame image. For the real image test, we directly employed a multi-frame image as the input.
In terms of objective evaluation indicators, three commonly used methods in super-resolution reconstruction tasks were selected: a structural similarity coefficient (SSIM), the Peak Signal-to-Noise Ratio (PSNR), and Root Mean Square Error (RMSE). These evaluation criteria were applied to the normalized images of S 0 , S 1 , S 2 , D P , and A P , respectively. A higher SSIM indicated that the reconstructed image closely resembled the reference image in terms of intensity, contrast, and structure. The accuracy of each polarization component’s reconstructed image reflected the true polarization characteristics of the object. The noise in the polarization image mainly manifested as false polarization. PSNR, which measures the ratio of peak signal energy to average noise energy, was used to assess the pseudo-polarization level in each component of the reconstructed polarization image. RMSE was used to gauge the degree of similarity between the reconstructed polarization image and the true polarization image in terms of pixel position. A smaller RMSE value indicated a closer match between the reconstructed and true polarization values for the same pixel position. By combining these three objective evaluation measures with subjective criteria, a comprehensive analysis of the proposed reconstruction method was conducted.

3.1. Super-Resolution Performance Test

The experimental setup consisted of a PC host with an i7-4720HQ CPU (Intel Corporation, Santa Clara, CA, USA) and GTX960 GPU (Nvidia Corporation, Santa Clara, CA, USA). The specific parameter settings for both the reconstruction and training parts of the dictionary set are provided and explained in Section 2. It is important to note that dictionary set learning takes place after the image segmentation process. For training purposes, we randomly selected 100,000 windows from 54 training scenes, each with dimensions of 5 × 5. No noise reduction algorithm was employed following the non-uniformity correction and blind element repair of the captured images. Consequently, the resulting images may have contained electronic noise, dark current noise [34], stray light response, and some uncorrected blind elements. The objective evaluation indicators for the simulation results are depicted in Table 1.
The table above clearly demonstrates the excellent reconstruction performance achieved by the proposed method. All five algorithms yielded good results for the S 1 and S 2 images. The majority of pixels in the S 1 and S 2 images represented the low-frequency component. Despite this, our approach still stood out and achieved the best results. Unlike the other algorithms, our method did not obtain the S 1 and S 2 images by detecting the difference in declination angle. On the S 0 index, the MPPOCS method exhibited a low reconstruction effect. This is because the filtering method used in this algorithm sacrifices some details of the decline-detecting image, reduces the noise of the orthogonal difference image, and neglects the quality of the weighted image. In our reproductions of the other algorithms, some did not align with the original data. We analyzed these replications in our experiments using the publicly available source code provided by the original authors. The primary difference lay in the down-sampling technique. The authors used interlaced pixel extraction for the simulation, resulting in simulated images with minimal noise. In contrast, our simulated image contained significantly more noise. Additionally, our degradation process employs bicubic interpolation to obtain the subsampled image, where the pixels in the neighborhood are double cubic-weighted. This is because we considered that one of the imaging units of the detector does not correspond one-to-one with the real scene, as it contains a weighted component from the neighborhood due to the system’s point spread function. The impact of calculation errors was evident in the D P and A P indexes. If one of the Stokes vector images exhibited unsatisfactory reconstruction, the objective indexes of D P and A P were greatly diminished. In other words, the D P and A P images were more susceptible to noise. Therefore, it was crucial to analyze the noise robustness of the method, and a detailed experimental analysis will be presented in the next section. Figure 3 showcases the simulation reconstruction results of the different algorithms for a tower, where it is evident that the proposed method yielded results closer to the real image. Additionally, Table A1 in Appendix A displays the objective index of S 0 after the reconstruction of other test images, while Table A2 presents the objective indicators of S 1 , and Table A3 exhibits the objective indicators of S 2 . Furthermore, Table A4 provides the objective index of D P , and Table A5 presents the objective index of A P after the reconstruction of other test images.

3.2. Noise Robustness

Limited by the manufacturing process of the detector, the noise of a focal plane polarization system is generally higher than that of a time-sharing polarization system. Therefore, it was crucial to analyze the noise robustness of the algorithm. Based on the comparison results presented in the previous section, both the proposed method and MPPOCS outperformed other methods. Consequently, this section compares the noise robustness between these two methods. The test image again was a tower. Gaussian white noise with mean 0 and standard deviation ranging from 0 to 3 was added to the focal plane image synthesized by each bias detection component of the tower. The value of λ , introduced in this paper, was set to 0.02. The objective indexes for the reconstructed S 0 and D P images of the two algorithms are depicted in Figure 4, while the reconstructed indexes for the S 1 , S 2 , and A P images are shown in the attached Figure A2 and Figure A3a,b. SSIM is represented on the black axis, PSNR on the intermediate axis, and RMSE on the blue axis. The + line represents the algorithm proposed in this paper, and the × line represents MPPOCS.
In Figure 4a, the objective evaluation index of the proposed algorithm gradually decreased as the level of noise increased, with a decline rate higher than that of MPPOCS. In Figure 4b, when the mean noise was 0.9, the SSIM index of our proposed method started to become lower than that of MPPOCS. When the mean noise values were 2.7 and 2.8, respectively, both PSNR and RMSE were weaker than those of the MPPOCS algorithm. Similar results were observed for the S 1 , S 2 , and A P images. In this set of experiments, we kept the λ values unchanged. These findings proved that the proposed algorithm possessed inherent noise resistance, as the reconstruction results were superior to MPPOCS when the noise level was low. Additionally, our algorithm demonstrated better performance in the case of low noise, as shown in Figure 4a. This can be attributed to the fact that the S 0 image was less sensitive to the noise in each channel compared to the other components, resulting in consistently optimal reconstruction results. We analyzed the results to ascertain the reasons behind the MPPOCS algorithm’s high noise robustness. The inclusion of wavelet threshold noise reduction in the algorithm grants it a natural adaptability to noise, especially in the presence of high levels of noise. However, in the context of weak noise, this can lead to excessively smooth image results, as illustrated in Figure 3.

3.3. Real Image Testing

We established a focal plane polarization detection system operating in the short-wave infrared range to collect real images to test our algorithm. The size of the focal plane images was 320 × 256 pixels. The obtained test results are shown in Figure 5.
The selected test scene had cloudy weather conditions. The drone used as the target was the DJI Spirit 4Pro (DJI, Shenzhen, China). All display results were normalized to the same quantization level. No additional image enhancement algorithms were applied to the presentation results. Figure 5 shows the full-field S 0 image, revealing a dim target. Among the various algorithms tested, the EARI algorithm produced the fuzziest S 0 image. In several other channels, it was not possible to identify the target as a drone. However, some areas of the D P image allowed for the calculation of the degree of polarization. The NP and PCDP algorithms did not clearly show the target in the S 1 image. In the D P image, neither of these two algorithms accurately calculated the polarization degree of the UAV. Man-made objects, being more polarized than the natural background, did not exhibit this phenomenon well in the results. These two algorithms introduced significant errors in the A P images. Due to the drone’s orientation towards the lens, the polarization angles of its left and right sides did not significantly differ. As for MPPOCS, it can be observed that the S 0 image was overly blurry. The filter utilized in this algorithm reduced noise in the entire image, but it also sacrificed some high-frequency information. The S 1 and S 2 images performed better than with the NP and PCDP algorithms. The D P image provided a clearer representation of the target’s polarization information compared to the EARI algorithm. The A P image was slightly more distinct than with the first three algorithms. Ultimately, our proposed S 0 algorithm yielded the best results for the image. The brackets on the left and right holding the rotors were separated from the drone’s fuselage. The S 1 and S 2 images outperformed other algorithms, particularly in the A P image, where our proposed algorithm accurately reconstructed the entire shape of the UAV. With a certain level of prior knowledge, the UAV could be identified as the DJI Genie 4Pro drone even without specialized equipment. Additionally, in the D P image, our reconstruction results displayed the target’s polarization degree. The target’s edges were sharp, and the difference between the target and background was highly noticeable, providing a strong foundation for the recognition of small polarized targets. Finally, the processing time of the proposed algorithm was statistically analyzed and is shown in Table 2.
The running time of the method proposed in this paper appeared to be relatively long, as shown in Table 2. To analyze the source code, we employed MATLAB’s runtime analysis tool. Our analysis revealed that the feature symbol search algorithm accounted for a significant portion of the computing time, specifically 80.1%. Considering this finding, we are convinced that there is ample scope for improvement. Thus, one of our forthcoming research areas will focus on optimizing the feature symbol search algorithm.

4. Conclusions

This article presents a sparse representation-based model for infrared polarization super-resolution reconstruction. We propose a method to overcome the limitations of the detector’s sampling on the focal plane and achieve satisfactory results without requiring a large training set. Consequently, this approach is suitable for tasks where obtaining a vast amount of data is not feasible. Specifically, instead of independently solving each polarization angle, we solve the model on a designed Stokes vector field. This strategy effectively avoids errors caused by differential calculations during the polarization image-solving process. The simulation results depicted in Figure 3 demonstrate that our method effectively suppresses partial pseudo-polarization phenomena and generates more realistic polarization information compared to other existing methods. Furthermore, the results from a real test scene illustrated in Figure 5 reveal the significant advantages of our reconstruction results in object detection and classification tasks. Our algorithm delivers clearer edges of the targets and a noticeable contrast between the target and background, marking an improvement over the other algorithms discussed in this paper. In addition, to construct the dictionary set, we employ a self-developed detector to create a near-real-time short-wave infrared time-sharing polarization system. This system is equipped with a high-precision grating disk, enabling the continuous acquisition of polarization images. The main findings of our research are described in this article as follows:
(1) We propose a super-resolution reconstruction model for infrared polarization. (2) We design a reconstruction method based on the model. (3) We produce a prototype of a near-real-time short-wave infrared time-sharing polarization system and collect an infrared polarization dataset. (4) We build a short-wave infrared focal plane polarization system and conduct experiments to demonstrate the excellent results achieved by the proposed method in terms of quantitative measurement and visual quality.
In our future research, we intend to optimize the algorithms mentioned in the paper to enhance their speed in a software environment, with a primary focus on accelerating the feature symbol search algorithm. Moreover, we plan to integrate the proposed algorithm into our developed short-wave focal plane polarization detection system. Additionally, our future work involves incorporating convex optimization algorithms and deep learning methods to further improve image quality. Furthermore, we aim to explore the combination of metasurface design and its application in achieving full optical image reconstruction.

Author Contributions

Methodology, Y.M.; software, Y.M. and T.L.; validation, Y.M., S.W. and Z.Y.; formal analysis, S.W.; resources, F.L., S.W. and W.Q.; data curation, Y.M. and L.L.; writing—original draft preparation, Y.M.; writing—review and editing, F.L.; project administration, F.L. and L.L.; funding acquisition, F.L. and W.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Pre-research Program during the 14th Five-Year Plan, grant number No. 514010405.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Our data are derived from a self-developed time-sharing and focal-plane short-wave polarized infrared system. The system has been thoroughly refined and is now available for purchase. It encompasses imaging equipment, acquisition cards, an industrial computer, and software. Due to the inclusion of sensitive information, the dataset is not publicly accessible. Please reach out to the first author for the source code if necessary.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Figure A1. Simulation data strength diagram used for testing: (a) Cars, (b) Window, (c) Wall, (d) Building, (e) Tower, (f) Tank.
Figure A1. Simulation data strength diagram used for testing: (a) Cars, (b) Window, (c) Wall, (d) Building, (e) Tower, (f) Tank.
Applsci 14 00825 g0a1
Table A1. Objective evaluation indexes of S0 results of different algorithms on different test images.
Table A1. Objective evaluation indexes of S0 results of different algorithms on different test images.
MethodCarsWindowWallBuildingTank
EARISSIM = 0.4032SSIM = 0.7799SSIM = 0.4203SSIM = 0.5723SSIM = 0.3943
PSNR = 13.4202PSNR = 14.5269PSNR = 13.4207PSNR = 21.7906PSNR = 14.1210
RMSE = 0.2133RMSE = 0.1878RMSE = 0.2133RMSE = 0.0814RMSE = 0.1968
NPSSIM = 0.4374SSIM = 0.7858SSIM = 0.4502SSIM = 0.5519SSIM = 0.4114
PSNR = 13.8023PSNR = 14.6402PSNR = 12.6379PSNR = 21.5929PSNR = 14.6783
RMSE = 0.2041RMSE = 0.1854RMSE = 0.2334RMSE = 0.0832RMSE = 0.1845
PCDPSSIM = 0.4481SSIM = 0.7918SSIM = 0.4629SSIM = 0.5441SSIM = 0.4148
PSNR = 13.6557PSNR = 14.7660PSNR = 12.8692PSNR = 19.9623PSNR = 14.4895
RMSE = 0.2076RMSE = 0.1827RMSE = 0.2273RMSE = 0.1004RMSE = 0.1886
MPPOCSSSIM = 0.6442SSIM = 0.8353SSIM = 0.6514SSIM = 0.8971SSIM = 0.6470
PSNR = 17.5827PSNR = 15.7587PSNR = 20.3389PSNR = 36.8555PSNR = 16.9773
RMSE = 0.1321RMSE = 0.1630RMSE = 0.0962RMSE = 0.0144RMSE = 0.1416
This WorkSSIM = 0.6966SSIM = 0.8318SSIM = 0.7093SSIM = 0.9210SSIM = 0.6933
PSNR = 18.0517PSNR = 15.7436PSNR = 20.8368PSNR = 38.7827PSNR = 17.3267
RMSE = 0.1252RMSE = 0.1632RMSE = 0.0908RMSE = 0.0115RMSE = 0.1360
Table A2. Objective evaluation indexes of S1 results of different algorithms on different test images.
Table A2. Objective evaluation indexes of S1 results of different algorithms on different test images.
MethodCarsWindowWallBuildingTank
EARISSIM = 0.9637SSIM = 0.9922SSIM = 0.9195SSIM = 0.8338SSIM = 0.9886
PSNR = 39.6600PSNR = 39.6600PSNR = 37.5078PSNR = 32.5971PSNR = 46.5985
RMSE = 0.0104RMSE = 0.0045RMSE = 0.0133RMSE = 0.0235RMSE = 0.0047
NPSSIM = 0.9806SSIM = 0.9958SSIM = 0.9639SSIM = 0.8393SSIM = 0.9931
PSNR = 42.4280PSNR = 48.6937PSNR = 40.3364PSNR = 33.1142PSNR = 48.0276
RMSE = 0.0076RMSE = 0.0037RMSE = 0.0096RMSE = 0.0221RMSE = 0.0040
PCDPSSIM = 0.9781SSIM = 0.9947SSIM = 0.9498SSIM = 0.8365SSIM = 0.9916
PSNR = 41.7149PSNR = 47.9808PSNR = 39.1224PSNR = 33.0923PSNR = 47.3843
RMSE = 0.0082RMSE = 0.0040RMSE = 0.0111RMSE = 0.0222RMSE = 0.0042
MPSSIM = 0.9938SSIM = 0.9989SSIM = 0.9984SSIM = 0.9882SSIM = 0.9986
PSNR = 48.9114PSNR = 59.5889PSNR = 58.3427PSNR = 48.7383PSNR = 58.9306
RMSE = 0.0036RMSE = 0.0011RMSE = 0.0012RMSE = 0.0037RMSE = 0.0011
This WorkSSIM = 0.9959SSIM = 0.9991SSIM = 0.9986SSIM = 0.9926SSIM = 0.9989
PSNR = 51.2608PSNR = 60.5056PSNR = 59.0080PSNR = 51.1248PSNR = 60.0059
RMSE = 0.0027RMSE = 0.0010RMSE = 0.0011RMSE = 0.0028RMSE = 0.0010
Table A3. Objective evaluation indexes of S2 results of different algorithms on different test images.
Table A3. Objective evaluation indexes of S2 results of different algorithms on different test images.
MethodCarsWindowWallBuildingTank
EARISSIM = 0.9466SSIM = 0.9909SSIM = 0.9196SSIM = 0.9356SSIM = 0.9894
PSNR = 36.0382PSNR = 45.3721PSNR = 36.9036PSNR = 37.7624PSNR = 46.8909
RMSE = 0.0158RMSE = 0.0055RMSE = 0.0143RMSE = 0.0129RMSE = 0.0045
NPSSIM = 0.9582SSIM = 0.9933SSIM = 0.9614SSIM = 0.9518SSIM = 0.9927
PSNR = 37.4198PSNR = 46.3064PSNR = 39.6867PSNR = 39.1404PSNR = 48.1039
RMSE = 0.01346RMSE = 0.0048RMSE = 0.0104RMSE = 0.0110RMSE = 0.0039
PCDPSSIM = 0.9554SSIM = 0.9930SSIM = 0.9435SSIM = 0.9518SSIM = 0.9920
PSNR = 36.9204PSNR = 46.1596PSNR = 38.2077PSNR = 38.9414PSNR = 47.6478
RMSE = 0.01426RMSE = 0.0049RMSE = 0.0123RMSE = 0.0113RMSE = 0.0042
MPSSIM = 0.9918SSIM = 0.9983SSIM = 0.9974SSIM = 0.9706SSIM = 0.9977
PSNR = 47.1006PSNR = 56.4553PSNR = 55.7543PSNR = 42.0427PSNR = 56.1393
RMSE = 0.0044RMSE = 0.0015RMSE = 0.0016RMSE = 0.0079RMSE = 0.0016
This WorkSSIM = 0.9948SSIM = 0.9888SSIM = 0.9981SSIM = 0.9857SSIM = 0.9965
PSNR = 49.6021PSNR = 45.8931PSNR = 57.6130PSNR = 45.9872PSNR = 53.0169
RMSE = 0.0033RMSE = 0.0051RMSE = 0.0013RMSE = 0.0050RMSE = 0.0022
Table A4. Objective evaluation indexes of DP results of different algorithms on different test images.
Table A4. Objective evaluation indexes of DP results of different algorithms on different test images.
MethodCarsWindowWallBuildingTank
EARISSIM = 0.8003SSIM = 0.9442SSIM = 0.6495SSIM = 0.7586SSIM = 0.6876
PSNR = 27.8272PSNR = 30.8589PSNR = 21.5937PSNR = 26.9068PSNR = 27.2853
RMSE = 0.0406RMSE = 0.0287RMSE = 0.0832RMSE = 0.0452RMSE = 0.0432
NPSSIM = 0.8284SSIM = 0.9419SSIM = 0.6942SSIM = 0.7590SSIM = 0.7227
PSNR = 29.3589PSNR = 30.6841PSNR = 24.7070PSNR = 27.8181PSNR = 28.3058
RMSE = 0.0340RMSE = 0.0292RMSE = 0.0582RMSE = 0.0407RMSE = 0.0384
PCDPSSIM = 0.8216SSIM = 0.9484SSIM = 0.6709SSIM = 0.7798SSIM = 0.7113
PSNR = 28.9635PSNR = 31.2784PSNR = 22.9016PSNR = 28.0442PSNR = 27.7868
RMSE = 0.0356RMSE = 0.0273RMSE = 0.0716RMSE = 0.0396RMSE = 0.0408
MPSSIM = 0.9496SSIM = 0.9747SSIM = 0.9682SSIM = 0.9293SSIM = 0.9213
PSNR = 36.8699PSNR = 40.9748PSNR = 44.0480PSNR = 36.1728PSNR = 37.4231
RMSE = 0.0143RMSE = 0.0089RMSE = 0.0063RMSE = 0.0155RMSE = 0.0135
This WorkSSIM = 0.9653SSIM = 0.9496SSIM = 0.9775SSIM = 0.9599SSIM = 0.9338
PSNR = 40.6423PSNR = 31.4468PSNR = 45.8423PSNR = 39.6173PSNR = 35.5246
RMSE = 0.0093RMSE = 0.0268RMSE = 0.0051RMSE = 0.0105RMSE = 0.0167
Table A5. Objective evaluation indexes of AP results of different algorithms on different test images.
Table A5. Objective evaluation indexes of AP results of different algorithms on different test images.
MethodCarsWindowWallBuildingTank
EARISSIM = 0.9665SSIM = 0.9834SSIM = 0.9616SSIM = 0.9223SSIM = 0.9788
PSNR = 31.4232PSNR = 41.2151PSNR = 31.9594PSNR = 27.5346PSNR = 41.9182
RMSE = 0.0268RMSE = 0.0087RMSE = 0.0252RMSE = 0.0420RMSE = 0.0080
NPSSIM = 0.9740SSIM = 0.9848SSIM = 0.9665SSIM = 0.9477SSIM = 0.9834
PSNR = 32.7683PSNR = 42.4306PSNR = 33.1721PSNR = 29.4241PSNR = 43.2909
RMSE = 0.0230RMSE = 0.0076RMSE = 0.0220RMSE = 0.0338RMSE = 0.0069
PCDPSSIM = 0.9752SSIM = 0.9859SSIM = 0.9696SSIM = 0.9484SSIM = 0.9848
PSNR = 32.8784PSNR = 42.6716PSNR = 33.6649PSNR = 29.4233PSNR = 43.6527
RMSE = 0.0227RMSE = 0.0074RMSE = 0.0207RMSE = 0.0338RMSE = 0.0066
MPSSIM = 0.9725SSIM = 0.9852SSIM = 0.9675SSIM = 0.9335SSIM = 0.9818
PSNR = 32.2656PSNR = 41.3038PSNR = 32.7580PSNR = 27.6061PSNR = 42.0923
RMSE = 0.0244RMSE = 0.0086RMSE = 0.0230RMSE = 0.0417RMSE = 0.0079
This WorkSSIM = 0.9845SSIM = 0.9892SSIM = 0.9810SSIM = 0.9757SSIM = 0.9918
PSNR = 34.7240PSNR = 46.0793PSNR = 37.4299PSNR = 33.1091PSNR = 46.8903
RMSE = 0.0184RMSE = 0.0050RMSE = 0.0134RMSE = 0.0221RMSE = 0.0045
Figure A2. The objective evaluation indexes of our algorithm and MPPOCS in the tower dataset on AP images under different noise conditions.
Figure A2. The objective evaluation indexes of our algorithm and MPPOCS in the tower dataset on AP images under different noise conditions.
Applsci 14 00825 g0a2
Figure A3. The objective evaluation indexes of our algorithm and MPPOCS for a tower, as measured under varying noise conditions: (a) Objective evaluation indexes of S1, (b) Objective evaluation indexes of S2.
Figure A3. The objective evaluation indexes of our algorithm and MPPOCS for a tower, as measured under varying noise conditions: (a) Objective evaluation indexes of S1, (b) Objective evaluation indexes of S2.
Applsci 14 00825 g0a3

References

  1. Rosenberger, M.; Horn, R.; Golomoz, A.; Dittrich, P.-G.; Illmann, R.; Fütterer, R.; Notni, G. Investigation on surface inspection using polarizing image sensors. In Proceedings of the Image Sensing Technologies: Materials, Devices, Systems, and Applications IX, SPIE, Orlando, FL, USA, 30 May 2022; Volume 12091, pp. 113–125. [Google Scholar]
  2. Schommer, D.; Duhovic, M.; Hoffmann, T.; Ernst, J.; Schladitz, K.; Moghiseh, A.; Gortner, F.; Hausmann, J.; Mitschang, P.; Steiner, K. Polarization imaging for surface fiber orientation measurements of carbon fiber sheet molding compounds. Compos. Commun. 2023, 37, 101456. [Google Scholar]
  3. Xu, C.; Zheng, D.; Huang, Y.; Zeng, Z. Defect inspection of optical components based on dark-field polarization imaging and convolutional neural network recognition algorithms. Opt. Eng. 2023, 62, 043101. [Google Scholar]
  4. Zhao, Y.; Reda, M.; Feng, K.; Zhang, P.; Cheng, G.; Ren, Z.; Kong, S.G.; Su, S.; Huang, H.X.; Zang, J. Detecting giant cell tumor of bone lesions using Mueller matrix polarization microscopic imaging and multi-parameters fusion network. IEEE Sens. J. 2020, 20, 7208–7215. [Google Scholar] [CrossRef]
  5. Ahmad, I.; Khaliq, A.; Iqbal, M.; Khan, S. Mueller matrix polarimetry for characterization of skin tissue samples: A review. Photodiagn. Photodyn. Ther. 2020, 30, 101708. [Google Scholar] [CrossRef]
  6. Rahaman, A.; Anantharaju, A.; Jeyachandran, K.; Manideep, R.; Pal, U.M. Optical imaging for early detection of cervical cancer: State of the art and perspectives. J. Biomed. Opt. 2023, 28, 080902. [Google Scholar]
  7. Tauc, M.J.; Rehbein, E.M.; Eshelman, L.M.; Shaw, J.A. Polarization enhancement of passive SWIR cloud thermodynamic phase remote sensing. In Proceedings of the Polarization Science and Remote Sensing IX, SPIE, San Diego, CA, USA, 6 September 2019; Volume 11132, pp. 25–33. [Google Scholar]
  8. Chen, W.; Yan, L.; Chandrasekar, V. Optical polarization remote sensing. Int. J. Remote Sens. 2020, 41, 4849–4852. [Google Scholar]
  9. Wang, H.; Zhang, P.; Yin, D.; Li, Z.; Shang, H.; Xu, H.; Shang, J.; Gu, S.; Hu, X. Shortwave Infrared Multi-Angle Polarization Imager (MAPI) Onboard Fengyun-3 Precipitation Satellite for Enhanced Cloud Characterization. Remote Sens. 2022, 14, 4855. [Google Scholar] [CrossRef]
  10. Huang, H.; Yu, Y.; Li, X.; Sun, D.; Zhang, G.; Li, T.; Shao, X.; Yang, B. Shortwave Infrared InGaAs Detectors On-Chip Integrated with Subwavelength Polarization Gratings. Nanomaterials 2023, 13, 2512. [Google Scholar] [CrossRef]
  11. Zhang, J.C.; Chen, J.L.; Luo, H.B. Polarization image interpolation algorithm based on tensor non-negative sparse decomposition. Acta Opt. Sin. 2021, 41, 1411001. [Google Scholar]
  12. Gao, S.; Gruev, V. Bilinear and bicubic interpolation methods for division of focal plane polarimeters. Opt. Express 2011, 19, 26161–26173. [Google Scholar] [CrossRef]
  13. Gao, S.; Gruev, V. Gradient-based interpolation method for division-of-focal-plane polarimeters. Opt. Express 2013, 21, 1137–1151. [Google Scholar] [CrossRef]
  14. Li, N.; Zhao, Y.; Pan, Q.; Kong, S.G. Demosaicking DoFP images using Newton’s polynomial interpolation and polarization difference model. Opt. Express 2019, 27, 1376–1391. [Google Scholar] [CrossRef] [PubMed]
  15. Morimatsu, M.; Monno, Y.; Tanaka, M.; Okutomi, M. Monochrome and color polarization demosaicking based on intensity-guided residual interpolation. IEEE Sens. J. 2021, 21, 26985–26996. [Google Scholar] [CrossRef]
  16. Wu, R.; Zhao, Y.; Li, N.; Kong, S.G. Polarization image demosaicking using polarization channel difference prior. Opt. Express 2021, 29, 22066–22079. [Google Scholar] [CrossRef]
  17. Chen, J.; Wang, W.G.; Liu, T.X.; Zhang, Z.D.; Gao, H.B. A POCS super resolution restoration algorithm based on BM3D. Sci. Rep. 2017, 7, 15049. [Google Scholar] [CrossRef] [PubMed]
  18. Zhang, X.F.; Huang, W.; Xu, M.F.; Jia, S.Q.; Xu, X.R.; Li, F.B.; Zheng, Y.D. Super-resolution imaging for infrared micro-scanning optical system. Opt. Express 2019, 27, 7719–7737. [Google Scholar] [CrossRef]
  19. Ma, Y.; Wang, S.; Lei, T.; Li, B.; Li, F. Super-resolution reconstruction of focal plane infrared polarization microscan images. Opt. Precis. Eng. 2023, 31, 2418–2429. [Google Scholar] [CrossRef]
  20. Zhang, J.; Shao, J.; Luo, H.; Zhang, X.; Hui, B.; Chang, Z.; Liang, R. Learning a convolutional demosaicing network for microgrid polarimeter imagery. Opt. Lett. 2018, 43, 4534–4537. [Google Scholar]
  21. Zeng, X.; Luo, Y.; Zhao, X.; Ye, W. An end-to-end fully-convolutional neural network for division of focal plane sensors to reconstruct s 0, dolp, and aop. Opt. Express 2019, 27, 8566–8577. [Google Scholar] [CrossRef]
  22. Xu, G.; Wang, J.; Zhang, L.; Ma, J.; Wang, Y.; Liu, J. Multi-scale adaptive weighted network for polarization computational imaging super-resolution. Appl. Phys. B 2022, 128, 200. [Google Scholar] [CrossRef]
  23. Xu, M.; Wang, C.; Wang, K.; Shi, H.; Li, Y.; Jiang, H. Polarization Super-Resolution Imaging Method Based on Deep Compressed Sensing. Sensors 2022, 22, 9676. [Google Scholar] [CrossRef] [PubMed]
  24. Liu, J.; Duan, J.; Hao, Y.; Chen, G.; Zhang, H.; Zheng, Y. Polarization image demosaicing and RGB image enhancement for a color polarization sparse focal plane array. Opt. Express 2023, 31, 23475–23490. [Google Scholar] [CrossRef]
  25. Wen, S.; Zheng, Y.; Lu, F.; Zhao, Q. Convolutional demosaicing network for joint chromatic and polarimetric imagery. Opt. Lett. 2019, 44, 5646–5649. [Google Scholar] [PubMed]
  26. Wen, S.; Zheng, Y.; Lu, F. A sparse representation based joint demosaicing method for single-chip polarized color sensor. IEEE Trans. Image Process. 2021, 30, 4171–4182. [Google Scholar] [PubMed]
  27. Zhang, J.; Chen, J.; Yu, H.; Yang, D.; Liang, B.; Xing, M. Polarization image demosaicking via nonlocal sparse tensor factorization. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5607210. [Google Scholar]
  28. Zhang, J.; Luo, H.; Liang, R.; Ahmed, A.; Zhang, X.; Hui, B.; Chang, Z. Sparse representation-based demosaicing method for microgrid polarimeter imagery. Opt. Lett. 2018, 43, 3265–3268. [Google Scholar] [PubMed]
  29. Ma, Y.; Wang, S.; Lei, T.; Li, B.; Li, F. Imaging model and error analysis of quasi-real-time polarization detection system based on continuous rotation. Optoelectron. Technol. Appl. 2022, 37, 82–87+102. [Google Scholar]
  30. Chang, H.; Yeung, D.Y.; Xiong, Y. Super-resolution through neighbor embedding. In Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision and Pattern Recognition 2004, CVPR 2004, Washington, DC, USA, 27 June–2 July 2004; IEEE: Piscataway, NJ, USA, 2004; Volume 1, p. I. [Google Scholar]
  31. Zhu, Z.; Li, X.; Zhang, J.; Zhou, Y.; Zhao, Y. Specular flare suppression method for reflective materials based on the optimal polarizing angle. Appl. Opt. 2022, 61, 8034–8041. [Google Scholar] [CrossRef]
  32. Lee, H.; Battle, A.; Raina, R.; Ng, A. Efficient sparse coding algorithms. Adv. Neural Inf. Process. Syst. 2006, 19, 801–808. [Google Scholar]
  33. Olshausen, B.A.; Field, D.J. Sparse coding with an overcomplete basis set: A strategy employed by V1? Vis. Res. 1997, 37, 3311–3325. [Google Scholar] [CrossRef]
  34. Ma, Y.; Li, X.; Shao, X.; Deng, S.; Cheng, J.; Gu, Y.; Liu, Y.; Chen, Y.; Zhu, X.; Li, T.; et al. 320 × 256 Extended Wavelength InxGa1−xAs/InP Focal Plane Arrays: Dislocation Defect, Dark Signal and Noise. IEEE J. Sel. Top. Quantum Electron. 2021, 28, 3800411. [Google Scholar]
Figure 1. Change in the overlap coefficient in the sliding window super-resolution process. The red window represents one step of the rebuilding process, while the yellow window represents another step. The overlapping portion carries a weight of 2, whereas the non-overlapping section holds a weight of 1.
Figure 1. Change in the overlap coefficient in the sliding window super-resolution process. The red window represents one step of the rebuilding process, while the yellow window represents another step. The overlapping portion carries a weight of 2, whereas the non-overlapping section holds a weight of 1.
Applsci 14 00825 g001
Figure 2. Production of the near-real-time short-wave infrared time-sharing polarization system. In our study, the incident natural light is partially polarized when it reflects off the target. Subsequently, it passes through a polarizer to obtain linear polarization. The linearly polarized light then travels through the lens and reaches the shortwave infrared detector. In terms of the physical products we manufactured, the foremost component was a motor-controlled rotating target wheel. We achieved high-precision angle control using a grating ruler. The entire system was connected to the upper computer via fiber optics for transmission.
Figure 2. Production of the near-real-time short-wave infrared time-sharing polarization system. In our study, the incident natural light is partially polarized when it reflects off the target. Subsequently, it passes through a polarizer to obtain linear polarization. The linearly polarized light then travels through the lens and reaches the shortwave infrared detector. In terms of the physical products we manufactured, the foremost component was a motor-controlled rotating target wheel. We achieved high-precision angle control using a grating ruler. The entire system was connected to the upper computer via fiber optics for transmission.
Applsci 14 00825 g002
Figure 3. The simulation reconstruction results of each algorithm for a tower. The ground truth image provides a high-resolution representation of each image. The EARI algorithm exhibits a notable false bias in the AP and DP images, and the S0 image appears blurred. Additionally, the edge details show a noticeable jagged appearance. Both the NP and PCDP algorithms display clear high-frequency errors in the reconstructed S1 and S2 images, as they rely on a set of orthogonal polarization angle image differences to obtain these components. When compared with the ground truth’s S1 and S2 images, it is evident that these two methods yield larger errors. Although visually improved, the DP and AP images reconstructed by the NP and PCDP algorithms display sharp edges resulting from pseudo-polarization effects. Due to the strong noise reduction algorithm used in the MPPOCS algorithm, some loss of detail is observed, leading to a relatively fuzzy reconstruction of the S0 image. Our proposed algorithm achieves a close resemblance to the ground truth in terms of the S1 and S2 images; however, it also captures additional details in the AP and DP images compared to the MPPOCS algorithm.
Figure 3. The simulation reconstruction results of each algorithm for a tower. The ground truth image provides a high-resolution representation of each image. The EARI algorithm exhibits a notable false bias in the AP and DP images, and the S0 image appears blurred. Additionally, the edge details show a noticeable jagged appearance. Both the NP and PCDP algorithms display clear high-frequency errors in the reconstructed S1 and S2 images, as they rely on a set of orthogonal polarization angle image differences to obtain these components. When compared with the ground truth’s S1 and S2 images, it is evident that these two methods yield larger errors. Although visually improved, the DP and AP images reconstructed by the NP and PCDP algorithms display sharp edges resulting from pseudo-polarization effects. Due to the strong noise reduction algorithm used in the MPPOCS algorithm, some loss of detail is observed, leading to a relatively fuzzy reconstruction of the S0 image. Our proposed algorithm achieves a close resemblance to the ground truth in terms of the S1 and S2 images; however, it also captures additional details in the AP and DP images compared to the MPPOCS algorithm.
Applsci 14 00825 g003
Figure 4. The objective evaluation indexes of our algorithm and MPPOCS for a tower, as measured under varying noise conditions. (a) Objective evaluation indexes of S0. (b) Objective evaluation indexes of DP.
Figure 4. The objective evaluation indexes of our algorithm and MPPOCS for a tower, as measured under varying noise conditions. (a) Objective evaluation indexes of S0. (b) Objective evaluation indexes of DP.
Applsci 14 00825 g004
Figure 5. Real scene reconstruction results of each algorithm. Our algorithm achieved good performance in image processing. In the S0 image, the body of the unmanned aerial vehicle is distinguished from the fixed rotor support structures on its sides. Our algorithm effectively restored the form of the unmanned aerial vehicle. Even without specialized equipment, we were capable of visually recognizing the target. In the AP and DP images, our reconstruction results distinctly demonstrate the polarization degree of the target. The target’s edges are clearly visible, and there is a noticeable contrast between the target and background. Our algorithm provides a reliable foundation for identifying small polarized targets.
Figure 5. Real scene reconstruction results of each algorithm. Our algorithm achieved good performance in image processing. In the S0 image, the body of the unmanned aerial vehicle is distinguished from the fixed rotor support structures on its sides. Our algorithm effectively restored the form of the unmanned aerial vehicle. Even without specialized equipment, we were capable of visually recognizing the target. In the AP and DP images, our reconstruction results distinctly demonstrate the polarization degree of the target. The target’s edges are clearly visible, and there is a noticeable contrast between the target and background. Our algorithm provides a reliable foundation for identifying small polarized targets.
Applsci 14 00825 g005
Table 1. The objective evaluation indexes of each algorithm for a tower.
Table 1. The objective evaluation indexes of each algorithm for a tower.
MethodS0S1S2DPAP
EARISSIM = 0.9649SSIM = 0.9670SSIM = 0.9751SSIM = 0.8148SSIM = 0.7559
PSNR = 35.6509PSNR = 41.4362PSNR = 42.4668PSNR = 27.9892PSNR = 19.8982
RMSE = 0.0165RMSE = 0.0085RMSE = 0.0075RMSE = 0.0399RMSE = 0.1012
NPSSIM = 0.9708SSIM = 0.9807SSIM = 0.9842SSIM = 0.8599SSIM = 0.7749
PSNR = 36.8702PSNR = 43.5596PSNR = 44.0752PSNR = 29.7688PSNR = 23.2195
RMSE = 0.0143RMSE = 0.0066RMSE = 0.0063RMSE = 0.0324RMSE = 0.0690
PCDPSSIM = 0.9723SSIM = 0.9748SSIM = 0.9827SSIM = 0.8512SSIM = 0.7602
PSNR = 37.1946PSNR = 42.4574PSNR = 43.67161PSNR = 29.2385PSNR = 22.4482
RMSE = 0.0138RMSE = 0.0075RMSE = 0.0066RMSE = 0.0345RMSE = 0.0754
MPPOCSSSIM = 0.9687SSIM = 0.9972SSIM = 0.9952SSIM = 0.9610SSIM = 0.9452
PSNR = 36.1099PSNR = 54.8412PSNR = 50.9966PSNR = 38.9372PSNR = 38.6978
RMSE = 0.0156RMSE = 0.0018RMSE = 0.0028RMSE = 0.0113RMSE = 0.0116
This WorkSSIM = 0.9809SSIM = 0.9980SSIM = 0.9972SSIM = 0.9729SSIM = 0.9595
PSNR = 40.0757PSNR = 56.8644PSNR = 54.0460PSNR = 41.8172PSNR = 41.3630
RMSE = 0.0099RMSE = 0.0014RMSE = 0.0019RMSE = 0.0081RMSE = 0.0086
Table 2. The real image reconstruction results of each algorithm’s running time.
Table 2. The real image reconstruction results of each algorithm’s running time.
MethodEARINPPCDPMPPOCSThis Work
Running Time/s0.310.190.031.1818.84
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ma, Y.; Lei, T.; Wang, S.; Yang, Z.; Li, L.; Qu, W.; Li, F. A Super-Resolution Reconstruction Method for Infrared Polarization Images with Sparse Representation of Over-Complete Basis Sets. Appl. Sci. 2024, 14, 825. https://doi.org/10.3390/app14020825

AMA Style

Ma Y, Lei T, Wang S, Yang Z, Li L, Qu W, Li F. A Super-Resolution Reconstruction Method for Infrared Polarization Images with Sparse Representation of Over-Complete Basis Sets. Applied Sciences. 2024; 14(2):825. https://doi.org/10.3390/app14020825

Chicago/Turabian Style

Ma, Yizhe, Teng Lei, Shiyong Wang, Zhengye Yang, Linhan Li, Weidong Qu, and Fanming Li. 2024. "A Super-Resolution Reconstruction Method for Infrared Polarization Images with Sparse Representation of Over-Complete Basis Sets" Applied Sciences 14, no. 2: 825. https://doi.org/10.3390/app14020825

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop