PLoS ONE
Home Fusion algorithm of visible and infrared image based on anisotropic diffusion and image enhancement (capitalize only the first word in a title (or heading), the first word in a subtitle (or subheading), and any proper nouns)
Fusion algorithm of visible and infrared image based on anisotropic diffusion and image enhancement (capitalize only the first word in a title (or heading), the first word in a subtitle (or subheading), and any proper nouns)
Fusion algorithm of visible and infrared image based on anisotropic diffusion and image enhancement (capitalize only the first word in a title (or heading), the first word in a subtitle (or subheading), and any proper nouns)

Competing Interests: The authors have declared that no competing interests exist.

Article Type: research-article Article History
Abstract

Aiming at the situation that the existing visible and infrared images fusion algorithms only focus on highlighting infrared targets and neglect the performance of image details, and cannot take into account the characteristics of infrared and visible images, this paper proposes an image enhancement fusion algorithm combining Karhunen-Loeve transform and Laplacian pyramid fusion. The detail layer of the source image is obtained by anisotropic diffusion to get more abundant texture information. The infrared images adopt adaptive histogram partition and brightness correction enhancement algorithm to highlight thermal radiation targets. A novel power function enhancement algorithm that simulates illumination is proposed for visible images to improve the contrast of visible images and facilitate human observation. In order to improve the fusion quality of images, the source image and the enhanced images are transformed by Karhunen-Loeve to form new visible and infrared images. Laplacian pyramid fusion is performed on the new visible and infrared images, and superimposed with the detail layer images to obtain the fusion result. Experimental results show that the method in this paper is superior to several representative image fusion algorithms in subjective visual effects on public data sets. In terms of objective evaluation, the fusion result performed well on the 8 evaluation indicators, and its own quality was high.

Huang,Dong,Xue,Liu,Hua,and Raja: Fusion algorithm of visible and infrared image based on anisotropic diffusion and image enhancement (capitalize only the first word in a title (or heading), the first word in a subtitle (or subheading), and any proper nouns)

1 Introduction

Due to the limitation of the sensor system, the image information acquired by a single sensor cannot comprehensively describe the target scene. Therefore, it is very important to capture the detailed information of corresponding scenes several times by combining multiple sensors information to capture the target scene. Image fusion is to process the images collected by multiple source channels in the same scene to maximize the effective information in different channels, improve the utilization of each image, and make the fusion result more comprehensive and clear, which is convenient for people to observe. Image fusion is not only a kind of image enhancement technology, but also an important branch of information fusion. It is a hot spot in information fusion research. It is widely used in remote sensing [1], digital camera [2], military technology [3, 4], objects recognition [5, 6], medical imaging [7] and other fields.

Generally, image fusion is divided into three levels: data-level fusion, feature-level fusion, and decision-level fusion. Data-level fusion is also called pixel-level fusion, which refers to the process of directly processing the data collected by the sensor to obtain a fusion image, make full use of the input data information to better retain the object and background information [8]. Although pixel-level image fusion is the lowest level of fusion, it directly operates on pixels and serves as the basis for other fusion levels, enriches useful information, and the obtained images are consistent with the human visual system [9]. Infrared images are not affected by lighting and camouflage, but the resolution is low. The visible image has high spatial resolution, but it is susceptible to external interference. These two kinds of images have complementary characteristics. The fusion of them is one of the important applications of pixel-level multi-source image fusion [10, 11].

The pixel-level image fusion algorithm includes two kinds of algorithms in the spatial domain and the transform domain. The spatial domain algorithm includes many fusion rules such as gray-scale weighted average method [12], contrast modulation method [13], and Principal Component Analysis(PCA) [7], etc. Transform domain algorithms include Wavelet Transform(WT) method [14, 15], pyramid decomposition fusion algorithm [16], Curvelet Transform(CVT) [17], Non-Subsampled Contourslet Transform(NSCT) [18], Non-Subsampled Shearlet Transform(NSST) [19], etc. The algorithm based on transform domain is the current mainstream fusion algorithm of infrared image and visible image [8]. The main idea is to map the source image from the spatial domain to a sparse transform domain, and corresponding fusion is carried out according to the rules of the transform domain, and the result of image fusion is obtained after inverse transformation. In recent years, many scholars have conducted in-depth research in this area. Chao [20] proposed a wavelet-based image fusion algorithm, which uses high frequency band and low frequency band to perform fusion. Although the fusion efficiency is very high, the fusion results in some direction information are missed. On this basis, Do [21] proposed a wavelet-based contourslet conversion algorithm, which flexibly realizes multi-resolution and directional extensibility, can better preserve directional information. Adu [22, 23] proposed a coefficient selection method based on NSCT, visual and gradient features, which achieves remarkable results in target extraction, and also achieves good results in detail preservation. Ma [24] proposed a gradient transfer fusion algorithm based on gradient transfer and total variation minimization, which can effectively saves the main intensity distribution in the infrared image and the gradient change in the visible image, and it also promotes the ability to correct without pre-registration. Bavirisetti [25] proposed a fusion algorithm that combines Anisotropic Diffusion(AD) and Karhunen-Loeve(K-L) transform. This algorithm can extract detailed layer information according to AD and retain the detailed texture. Fu [26] proposed a fusion algorithm using Robust Principal Component Analysis(RPCA) and NSCT to decompose the image through RPCA to obtain the corresponding sparse matrix, and used the sparse matrix for the fusion of high and low frequency coefficients. This algorithm can effectively highlight infrared targets and retain the background information in the visible image. Huang [27] proposed an NSST algorithm based on different restricted conditions, which fuses high frequency bands through gradient constraints to ensure that the image can get more details. The low frequency bands are merged through saliency constraints to make the target more prominent. Ma [28, 29] first proposed an end-to-end infrared and visible image fusion algorithm, which generates a fusion image with main infrared intensity and additional visible gradient under the framework of Generative Adversarial Networks(GAN), and then introduced the detail loss and target edge intensity loss to further enrich the detail texture. Liu [30, 31] first proposed a Convolutional Neural Network (CNN) based deep learning model, which is effectively saves the details of the image, and then a novel fusion scheme based on Non-Subsampled Shearlet Transform (NSST), visual saliency and multi-objective artificial bee colony optimizing spiking cortical mode is proposed. It can be seen that in the fusion research of infrared and visible images, the prominence of infrared targets and the details of visible gradients are of great concern. People hope to improve the resolution and detail richness through various fusion algorithms.

This research focuses on pixel-level image fusion. In order to better integrate the characteristics of infrared images and visible images, this algorithm performs different types of enhancement and fusion on the two source image to highlight their respective advantages. The structure of this paper is shown in Fig 1 The main contributions of the proposed fusion algorithm are as follows:

The structure of this paper.
Fig 1

The structure of this paper.

The main contributions of the proposed fusion algorithm are as follows:

    The detail layers Iird and Ivisd of infrared and visible images can be obtain by AD algorithm, to better preserve clear outline and detail information.

    We propose a Power Function Enhancement (PFE) algorithm to simulate illumination for visible images to obtain visible enhanced images Ivise, and use Adaptive Histogram Partition (AHP) and Brightness Correction (BC) algorithms to enhance infrared images to obtain infrared enhanced images Iire.

    Based on the K-L transform to fuse the enhanced image and the source image, the Iiref results and Ivisf are used as new visible and infrared source image to fully retain the structural characteristics of the image and improve the fusion quality.

    In order to better preserve the characteristics of the image and improve the brightness of the image, Laplacian fusion is performed on Iiref and Ivisf to obtain the fused image Ilp, and the weighted fusion of Ilp, Iird and Iird is used to obtain the result Iresult.

The main structure of this paper is as follows: section 2 briefly introduces some basic concepts, section 3 proposes a specific image fusion scheme, section 4 provides experimental results and data analysis, and section 5 summarizes the full work.

2 Anisotropic diffusion

Anisotropic Diffusion(AD) is also known as the Perona-Malik(P-M) formula, the noise in image is eliminated by Partial Differential Equation(PDE). The image is regarded as a heat field, and each pixel is used as a heat flow. According to the relationship between the current pixel and the surrounding pixels, it is determined whether to diffuse to the surroundings. For example, when the distance between current pixel and surrounding pixel is large, the surrounding pixels may form boundary. Then the current pixel will not diffuse to the boundary, and this boundary is preserved. It overcomes the shortcomings of isotropic diffusion, isotropic diffusion may smooth the background heavily, thus the edge information is lost. The AD equation [32] is proposed:

Where It is the source image, div represents the divergence operator, c(x, y, t) represents the flux function or diffusion rate, ∇ represents the gradient operator, Δ represents the Laplacian operator, t represents the time or the number of iterations, and ⋅ represents the value range.

Convert Eq (1) into a thermal equation, use Forward-Time and Central-Space (FTCS) to solve this equation [24], the solution of the partial differential equation is as follows:

From Eq (2), we can see that the image Ix,yt+1 at the higher scale t+ 1 depends on the image Ix,yt at the previous scale t, λ is a stable constant, and satisfies 0λ14. ¯E, ¯S, ¯W and ¯N represents the difference between the nearest neighbours in the east, south, west, and north directions, respectively, defined as follows:

Similarly, cE, cS, cW and cN represent the flux function or diffusion rate in the east, south, west, and north directions, respectively, which are defined as follows:

In formula Eq (4), g(⋅) represents a monotonic decreasing function, where g(0) = 1. g(⋅) can be expressed by different functions, Perona sit [30] proposed two forms as follows:

Where K represents the thermal conductivity or gradient threshold, which is used to control the sensitivity of the edge. The functions in Eqs (5) and (6) provide a trade-off between smoothing and edge preservation. Eq (5) is suitable for images with more high-contrast edges. Eq (6) is suitable for images with large areas covering small areas. For a given image I, the AD is denoted as AD(I).

Infrared image and visible source image are represented by Iir(x, y) and Ivis(x, y). The base layer images obtained by the infrared image and the visible source image through the anisotropic diffusion algorithm are expressed as IirA(x,y) and IvisA(x,y), where IirA(x,y)=AD(Iir) and IvisA(x,y)=AD(Ivis). According to the difference between the source image and the base layer image, the detailed layer images of the infrared image Iird(x, y) and the visible image Ivisd(x, y) are defined as follows:

3 The proposed fusion method

3.1 Power function enhancement method of visible image

High visibility image details can more clearly reflect the targets in the scene, the image in the process of shooting is influenced by factors such as light and noise, the image visual quality is not satisfactory, then combined with the principle of fitting method, this study aims at dark scene under visible image, this paper presents a simple and practical power function enhancement algorithm for visible images in dark scenes. In the fitting problem, it is not necessary for the curve to pass through all given points. The aim of the fitting problem is to find a function curve that is the closest to all data points under a certain criterion, that is, the curve fitting is the best.

Power function is one of the basic elementary functions, which can be used to enhance the image. The formula of PFE is defined as follows:

Where C1 and C2 are constants, Ivis represents the input visible source image, and Ivise represents the enhanced result of visible image.

Dividing the grayscale values from 0 to 255 into 20 grayscales, the initial illumination brightness l0 is the lowest brightness of the light, the remaining five brightness stages are set as (l1, l2, l3, l4, l5), by collecting the gray value of the same color level displayed under different illumination intensity conditions, the variation of color levels under different illumination conditions is studied, the constants cx1 and cx2 in the power function at different brightness, in order to reduce the error and noise effect, calculate the average illumination brightness laverage to obtain the best enhancement effect, the final constants Caverage1 and Caverage2 are as follows:

Take the color block with gray value of 0, 20, 40 as an example, and the results acquired at different brightness stages are shown in Fig 2.

Results acquired at different brightness stages.
Fig 2

Results acquired at different brightness stages.

It can be found in Fig 2 that when the brightness reaches a certain value, the collected image reaches a stable state without significant changes.

According to the fitting results, C1 = 16.176 and C2 = 0.5339 are obtained, and the final form of power function of simulated illumination enhancement is as follows:

3.2 Infrared image enhancement base on adaptive histogram partition and brightness correction

The fusion of traditional infrared and visible images algorithms often only decomposes the image and operates on images of different frequency bands without paying attention to the image itself. Infrared images contain a lot of thermal radiation information, which should be widely concerned. The main purpose of infrared image enhancement is to enhance the edge sharpness of the image and improve the fuzzy state of the infrared image. It is more flexible and realistic to stretch the gray level in the adaptive histogram division. The brightness correction uses the moderate gray level of the infrared image, the gray level is more rich, and the expression is enhanced.

In this paper, an algorithm combining AHP and BC [33] is used to enhance infrared images. At first, we use Gaussian Filtering(GF) and locally weighted scatter plot smoothing [34] to perform on the gray histogram of the original image:

Where p(k) refers to the smoothed Probability Distribution Functon(PDF) after GF. 1 * (2w1 + 1) is the size of local window, in a 8-bit image L = 256. One-dimensional Gaussian kernel k(⋅) is defined as:
Where σ is a constant parameter deciding the weight of each neighboring PDF affecting the output, and current grayscale Ik is represents x0 = k.

Then partition the input image into two parts as: foreground Iirf and background Iirb. Thus, the final enchanced image Iire can be obtained as:

Where Iirf and Iirb can calculate by:

Rx is metric to decide the re-mappde range, ℵ(Ik) is the local contrast weighted distribution, Cj is the cumulative local contranst weighted distribution of [mj, mj+1].

Finally, the formulation of the objective function can be deduced by formula Eq (14) as:

Where Mo(m1) and MR stand for the mean of Iir and reference image IR. IR just a image which mean brightness is suitable for hunman vision system. It is utilized to generate a standard mean intensity value MR to optimize the objective fuction value F.

3.3 Enhanced image fusion method based on K-L transform

For simplicity, let us take Ivis(x, y) and Ivise(x, y) as the input. Arrange these image as column vectors of a matrix X¯. Make each row as an observation and each column as a variable to find the covariance matrix CXY of X¯. Calculate eigen values σ1, σ2 and eigen vectors ξ1=[ξ1(1)ξ1(2)] and ξ1=[ξ2(1)ξ2(2)] of CXY. The values of uncorrelated components KL1 and KL2 can giving by σmax:

The fused result Ivisef of K-L transform can be calculated by:
The extension for input images can be calculated by:
In the same way, the fused result Iiref is given by:
Where KL1 and KL2 represent the uncorrelated components in Iir and Iire.

3.4 Fusion of infrared and visible images

LP fusion is performed on each spatial frequency layer separately, so that different fusion operators can be used to highlight the features and details of specific frequency bands according to the characteristics and details of different frequency bands of different decomposition layers.

An important property of the LP is that it is a complete image representation: the steps used to construct the pyramid may be reversed to accurately merge the resulting new image, LP fusion results are as follows [35]:

Where LPl is the l-th level image decomposed from LP and Expand operator is the inverse of Reduce operator.

In the final fusion stage, in order to better retain detailed information, this paper superimpose the result of LP fusion with the detailed image obtained in the section 2. The final fusion results are defined as follows:

For the overall fusion process, it can be described in Algorithm 1.

Algorithm 1

Input: visible image Ivis, Infrared image Iir.

step 1. The detail layers Iird and Ivisd of infrared and visible image can be obtain by Eqs 17.

step 2. The visible enhanced image Ivise is obtained by Eqs 810. The infrared enhanced image Iire is obtained by Eqs 1116.

step 3. The K-L transform is performed on the source image and the enhanced image to obtain the enhanced fusion image Ivisef and Iiref by Eqs 1721.

step 4.Ivisef and Iiref were fused by LP to obtain Ilp by Eq 22.

step 5. Calculate the final fused image Iresult by Eq 23.

Output: Fused image Iresult

4 Experimental results and analysis

All experiments in this paper are established on Matlab R2018a platform under Windows 10 operating system. To ensure the reliability of the experimental effect, the infrared image and visible image data selected are all from TNO_Image_Fusion_Dataset [36] and Li [37]. The purpose of the experiment is to verify the proposed method with objective and subjective standards and compare it with existing methods.

In this paper, 4 groups of infrared and visible images that have been commonly used are selected as the test images, named “Natocamp”, “Nightstreet”, “Kaptein”, and “Gun”, respectively. The original registration test image is shown in Fig 3.

Original registration test image.
Fig 3

Original registration test image.

4.1 Comparison method and experimental parameter setting

Experimental parameter settings are shown in Table 1.

Table 1
Experimental parameter setting.
parameterMeaningDefault value
C1Enhancement coefficient of Power function16.176
C2Exponential percentage of Power function0.5339
w1Window radius of gaussian filter4
σ1Variance of gaussian filter0.7
w2Local minimum checks the length of the sliding window9
w3Length of the local entropy window7
ε1The lower bound of the weight function ℵ(x)0.0001
ε2The upper bound of the weight function ℵ(x)0.9999
tmaxThe maximum iteration of particle swarm optimization10
NThe maximum iteration of particle swarm optimization10
c1Learning factor of particle swarm optimization algorithm0.5
c2Learning factor of particle swarm optimization algorithm0.5
ωmaxThe maximum weight of inertia for particle swarm optimization0.9
ωminThe minimum inertia weight of particle swarm optimization0.1
NlThe number of Laplacian pyramid levels4

Where C1 and C2 are calculated for this paper, and the rest parameters are derived from Wan [32].

4.2 Filtering algorithm

It is very important to pre-process the source image and separate the detail layer. The quality of the detail layer directly determines the final fusion image’s performance in detail texture. In the experiment, the mean value [38], minimum value [39], Gaussian value [40] and median value [41] and the anisotropic diffusion filtering method are used to separate the infrared image and visible image at the detail layer. The detail separation layer results of Nato camp are shown in Fig 4.

Detail separation layer results of Nato camp.
Fig 4

Detail separation layer results of Nato camp.

In Fig 4, (b) and (d) can only get some detailed features, such as the outline of the building. However, the rest of the details are not reflected, such as roads and plants. the result of (c) is good, but contains a lot information do not belong to the background. The building of (e) can get a small amount of road information. The filtering algorithm (f) selected in this paper has rich details and clear texture, which is superior to the filtering decomposition results of other algorithms.

4.3 Fitting method to determine the visible image enhancement algorithm

In this study, the experiment is combined with the real life. By implementing white light illumination of different brightness, the image of color level under the illumination condition is taken and collected by the camera, and the relationship between the variation of color level in low light and different illumination is explored, so as to propose a simple and practical simulated lighting enhancement algorithm in dark scenes. The histogram of the chromatic scale image acquired under the brightness is shown in Fig 5.

Histogram of the gradation image collected at the lowest brightness.
Fig 5

Histogram of the gradation image collected at the lowest brightness.

It can be found that the color block with low gray value has two peaks in its histogram, because there is a white area around the color block, that is, the gray value under the current brightness of the color block with another peak of gray value of gray 255 similarly, the grayscale values of the remaining five chrominance level images are collected. The changes of gray values corresponding to l0 brightness to l1 brightness, l0 brightness to l2 brightness, l0 brightness to l3 brightness, l0 brightness to l4 brightness and l0 brightness to l5 brightness are shown in Tables 2 to 6:

Table 2
Correspondence of grayscale values between brightness l0 and brightness l1.
Brightness020406080100120140160180200220240255
l014202226363436444658708291100
l146606876819393105121129139143165185
Table 3
Correspondence of grayscale values between brightness l0 and brightness l2.
Brightness020406080100120140160180200220240255
l014202226363436444658708291100
l262808595105113117133137147155153176204
Table 4
Correspondence of grayscale values between brightness l0 and brightness l3.
Brightness020406080100120140160180200220240255
l014202226363436444658708291100
l268858997109117119133139152159153176208
Table 5
Correspondence of grayscale values between brightness l0 and brightness l4.
Brightness020406080100120140160180200220240255
l014202226363436444658708291100
l270859199109117121133139145155151174208
Table 6
Correspondence of grayscale values between brightness l0 and brightness l5.
Brightness020406080100120140160180200220240255
l014202226363436444658708291100
l270859197109117121133137145155155175208

By observing the objective data, it can be known that when the illumination reaches a certain level, the gray value of the acquired chroma level will tend to be a stable one. In order to ensure the robustness of the algorithm, the average illumination laverage is calculated, and the grayscale values changes corresponding to l0 brightness to laverage brightness is shown in Table 7.

Table 7
Correspondence of grayscale values between brightness l0 and brightness laverage.
Brightness020406080100120140160180200220240255
l014202226363436444658708291100
l263.27984.892.8102.6111.4114.2127.4134.6143.6152.6151173.2202.6

The data in Table 7 are fitted. The fitting function includes linear function, exponential function, logarithmic function and power function. The fitting results of each function are shown in Fig 6.

Fitting results of each function.
Fig 6

Fitting results of each function.

Goodness of Fit refers to the degree of fitting of regression lines to observed values. The statistical measure of Goodness of Fit is determined on coefficient R2. R2 has a maximum value of 1 and a minimum value of 0. The closer the value of R2 is to 1, the better the fitting degree of the regression line to the observed value is. On the contrary, when the value of R2 is close to 0, it indicates that the fitting degree of the regression line to the observed value is worse. From the perspective of objective index R2, the index AR2 of power function is the largest among the four functions, which proves that its goodness of fit is higher. The enhancement results of Nightstreet are shown in Fig 7.

Enhancement results of Nightstreet.
Fig 7

Enhancement results of Nightstreet.

In Fig 7, (a) is excessively enhanced, resulting in overexposure and poor practicability. In (b), there is a problem similar to (a), which is excessively enhanced and in the loss of much information in the image. In (c), the effect is better, but there is more noise after enhancement. In (d), it achieves the best enhancement effect.

The visible source image and the enhancement image are K-L transformed, and the fusion results of Nightstreet are shown in Fig 8.

Fusion results of Nightstreet.
Fig 8

Fusion results of Nightstreet.

In Fig 8, (a) is too bright that areas of high brightness overexposed, especially those marked by the red box. In (b) the brightness is low, and the area marked by red box is not conducive to observation. In (c), the fusion effect is good, but the marked corner area is not bright enough. In (d), it achieves the best fusion effect, complete image details, appropriate brightness, conducive to human eye observation.

4.4 Enhanced algorithm of infrared image

The infrared image processing in this study is based on an infrared image adaptive histogram Partition and brightness correction and enhancement algorithm proposed by Wan [30]. Wan proposed an infrared image adaptive histogram Partition and brightness correction and enhancement algorithm [33], and K-L transform of the results with the infrared source image to ensure that the original information of the source image is maintained while the infrared image is enhanced.

The enhancement and fusion results of Kaptein are shown in Fig 9.

Enhancement and fusion results of Kaptein.
Fig 9

Enhancement and fusion results of Kaptein.

4.5 Subjective evaluation

To better reflect the advantages of the proposed algorithm, enhanced infrared and visible images are used as new fusion source images, the experiment will test the proposed method and a variety of classical fusion algorithms in multiple fields: Gradient-Transform(GTF) algorithm [24], Pulse Coupled Neural Network(PCNN) algorithm [42], Dual Tree Complex Wavelet Transform(DTCWT) algorithm [43], CVT [17, 44] algorithm, Multi-resolution Singular Value Decomposition(MSVD) algorithm [45], Guided Filtering(GF) [46] algorithm, latent Low-Rank Representation(LRR) [47] algorithm, and a comparison between subjective and objective is conducted.

The fusion results of Natocamp are shown in Fig 10. In (a), the grass, the figure and the plants around the building are very blurred, with serious overall blurriness. In (b), the infrared target “human” is very clear, but the “plant” has been blurred, and the fusion distortion appears near the “building”. In (c) and (d), the area between the plant and the target is blurred. In (e) and (g), details are missing in the enclosure and plants are blurred. In (f), although there are prominent objectives, the background information is too vague to facilitate observation. In (h), roads, fences, plants, etc. are clearly defined and rich in detail.

Fusion results of Natocamp.
Fig 10

Fusion results of Natocamp.

The fusion results of Nightstreet are shown in Fig 11. In (a), the infrared target is relatively obvious, such as the tire of people and cars, but the effect is poor on the letters on signs, signs at the lower right corner, and the display of pedestrians at the upper right corner. In (b), there are many inconsistent fusion and artifact. In(c) and (d), the letters on the signs are better displayed, but the pedestrian and pavement outside the shops are more blurred. In (e), the overall picture is dark, with details such as letters and signs in the lower right corner are blurred. In (f), infrared targets are obvious, but many black spots are generated after fusion. In (g), the picture is generally good but lacking in details such as road and seats outside shops. In (h), the infrared target is clear, the lighting effect is good, the road and other details are complete.

Fusion results of Nightstreet.
Fig 11

Fusion results of Nightstreet.

The fusion results of Kaptein are shown in Fig 12. In (a), the trees above the building produce artifacts, the two sides of the building are seriously blurred, and the details of human feet are lost. In (b), the infrared target is obvious, but the fusion results in a large number of artifacts and detail misalignment. For example, the ground presents obvious shadows. In (c), the outline of the plants above the building is obvious, but there are a few artifacts on both sides of the plants and people. In (d), more artifacts are produced, such as ground and sky artifacts of varying degrees. In (e), the image is generally fuzzy and details of plants on both sides of the image are lost. In (g), the fused images are generally poor and lose their value for observation. In (f), the details of plants and people above the building are more complete, but the details of plants and ground on both sides of the road are missing. In (h), the details of people, rooms, roads and plants on either side are obvious and conducive to observation.

Fusion results of Kaptein.
Fig 12

Fusion results of Kaptein.

The fusion results of Gun are shown in Fig 13. In (a), the infrared target is obvious, but the information of face and background is lost and much artifact are produced. In (b), the fusion results in a “black block” on the face of the right character, and the hand and foot of the character cannot be recognized. In(c) to (e) infrared targets are relatively fuzzy, and the shadows in the background lose the gradual change characteristics and produce different degrees of artifact. In (f), the infrared target is obvious, but there are different degrees of “black” blocks on the left face and behind the other two people. In (g), only the infrared target of the gun is obvious, and the overall brightness is low. In (h), the infrared target is clear, and the texture after fusion is natura. For example, the shadow in the background can better reflect the real situation of the original scene, which is conducive to human observation.

Fusion results of Gun.
Fig 13

Fusion results of Gun.

4.6 Objective evaluation

Spatial Frequency(SF) [48], Mean Gradient(MG) [49], Energy Gradient(EG) [50], Edge Intensity(EI) [51], QAB/F [52], LAB/F [53], Mutual Information(MI) [54], Structural Similarity (SSIM) [55] are selected as objective evaluation indexes to evaluate the fusion results. SF reflects the overall activity of the image in the spatial domain. MG reflects the contrast of image expressiveness from large to small details, which can be used to evaluate the degree of image blurring. EG is used to evaluate the clarity of fused images. EI examines the grayscale change of each pixel in a certain field to reflect the intensity of the edge contours of the fused image. QAB/F is the important of evaluating the success of gradient information transfer from the inputs to the fused image. LAB/F is a measure of the information lost during the fusion process. MI represents the amount of information the fused image obtains from the input images. SSIM reflects the structural similarity between the input images and the fused image. The quality assessment results are shown from Tables 8 to 15. We rank the 8 algorithms in order from high to low according to the results of performance indicators.

Table 8
SF quality assessment results.
MethodNato campNightstreetKapteinGunRank
GTF9.413410.13757.439913.40922
PCNN10.937112.47619.893113.73524
DTCWT12.070413.48999.375820.65026
CVT12.154313.46179.527919.71555
MSVD9.313310.91518.392416.44833
GF12.796615.2739.324920.18557
LRR9.54949.9466.936613.46821
Ours22.417121.043517.527924.49528
Table 9
MG quality assessment results.
MethodNato campNightstreetKapteinGunRank
GTF3.41042.44832.46744.32722
PCNN3.92763.84413.43994.96954
DTCWT4.17143.24423.50858.18315
CVT4.33413.26813.68448.16226
MSVD2.82782.17622.54225.34373
GF4.74134.5093.67727.91757
LRR3.45642.41352.5293.86391
Ours8.22125.67797.061510.8138
Table 10
EG quality assessment results.
MethodNato campNightstreetKapteinGunRank
GTF4.11712.87513.46714.72241
PCNN4.72344.76684.18945.38624
DTCWT5.16854.21484.54758.93155
CVT5.27494.22724.71428.93816
MSVD4.30533.25974.16446.34863
GF5.62185.48334.42048.68227
LRR4.45713.21783.27086.25932
Ours10.66927.8099.2511.84098
Table 11
EI quality assessment results.
MethodNato campNightstreetKapteinGunRank
GTF3.57E+012.59E+012.36E+014.68E+012
PCNN4.10E+014.05E+013.51E+015.27E+014
DTCWT4.36E+013.42E+013.47E+018.67E+015
CVT4.54E+013.45E+013.63E+018.68E+016
MSVD2.81E+012.24E+012.26E+015.60E+011
GF5.02E+014.79E+013.76E+018.33E+017
LRR3.85E+012.71E+012.48E+016.13E+013
Ours8.41E+015.76E+016.78E+011.1496 e+028
Table 12
QAB/F quality assessment results.
MethodNato campNightstreetKapteinGunRank
GTF0.62120.67090.58350.51313
PCNN0.59450.57320.63860.56992
DTCWT0.73600.80180.79510.85588
CVT0.71780.79690.78590.84587
MSVD0.51670.55340.61280.69211
GF0.73720.83800.72660.83986
LRR0.69800.74300.67380.73554
Ours0.71900.84320.70000.84775
Table 13
LAB/F quality assessment results.
MethodNato campNightstreetKapteinGunRank
GTF0.36870.32190.40800.48192
PCNN0.37140.590.30120.41904
DTCWT0.24240.18080.17960.11957
CVT0.25120.18350.17930.12425
MSVD0.48280.44250.38610.30431
GF0.21840.10560.22110.14946
LRR0.29650.25460.32390.26243
Ours0.07830.05840.05840.07328
Table 14
MI quality assessment results.
MethodNato campNightstreetKapteinGunRank
GTF1.40361.93372.00821.07136
PCNN3.08913.04462.26952.41448
DTCWT1.02481.18481.2421.00473
CVT0.96991.08671.1610.98852
MSVD1.06171.5611.411.3775
GF1.36011.29181.97841.89047
LRR1.16111.22431.35961.21534
Ours0.90681.03870.97190.94851
Table 15
SSIM quality assessment results.
MethodNato campNightstreetKapteinGunRank
GTF0.6870.74380.73270.38784
PCNN0.65520.61040.64740.39641
DTCWT0.69030.6540.76460.48956
CVT0.68940.65350.75780.48745
MSVD0.71550.67860.77870.49997
GF0.63990.63320.67450.50653
LRR0.75210.76570.79740.52228
Ours0.60740.67850.61280.47952

The higher the objective index SF, MG, EG, EI, QAB/F, MI, and SSIM the better the fusion effect will be, while the LAB/F is the opposite. Combined with the data in Tables 8 to 11 and 13, it can be seen that in the performance of its own image, the algorithm in this paper performs well, has good edge contour characteristics, and the loss of fusion is the least. However, the connection with the source image is slightly weaker. The reason is that after Laplacian fusion, the features and details from different images may be merged, resulting in a change in the feature structure.

The run time is also an important standard to evaluate the quality of this algorithm. The run time of test image are shown from Table 16.

Table 16
The run time of test image.
MethodNato campNightstreetKapteinGunAverage time
GTF4.387534.919724.738221.036421.27045
PCNN69.4682236.8848229.5092185.1180.2405
DTCWT3.836529.223220.296316.826617.5456
CVT4.223230.154321.200517.666118.3110
MSVD3.845829.211920.686416.818817.6407
GF3.807929.346520.817816.901217.7183
LRR31.1413154.1903133.5535112.0761107.7403
Ours3.67628.785819.823416.435517.1801

According to Table 16, it can be seen that the running time of the algorithm in this paper is obviously less than that of other algorithms, so the time cost of this algorithm is the lowest. PCNN algorithm takes the longest time, the time cost of this algorithm is the highest.

5 Conclusion

On the one hand, the effect of image fusion depends on the quality of fusion algorithm, on the other hand, it also depends on the quality of source image. In order to better combine the characteristics of visible and infrared images and obtain more texture information, this paper proposes a new image enhancement fusion method combining K-L transform and LP fusion. Firstly, the anisotropic diffusion is used to extract the detail layer of the source image. According to the characteristics of the visible and infrared images, we select different enhancement algorithm. The power function enhancement algorithm is used to simulate the illumination of visible image to improve the brightness of the image and mine the details of the dark image. The infrared image is enhanced by adaptive histogram partition and brightness correction to highlight the characteristics of the target. Secondly, K-L transformation is performed between the enhanced images and the source image to form a new visible and infrared images to ensure that the image is enhanced without distortion and reduce artifacts. Finally, LP fusion is performed on the new visible and infrared images, and then the detail layer image is superimposed to obtain the fused image. The experimental results show that the method is subjectively clear texture, high visibility and good observability. In terms of objective indicators, the indicators of the image itself perform well, but the connection with the source image becomes weak. We will conduct further research on this issue in the later period to solve this problem.

References

Zeng Y, Huang W, Liu M, Zhang H, Zou B. Fusion of satellite images in urban area: Assessing the quality of resulting images. In: 2010 18th International Conference on Geoinformatics. IEEE; 2010. p. 1–4.

RShen, ICheng, JShi, ABasu. Generalized random walks for fusion of multi-exposure images. IEEE Transactions on Image Processing. 2011;20(12):36343646. 10.1109/TIP.2011.2150235

WbDing, DyBi, LyHe, et al Fusion of infrared and visible images based on shearlet transform and neighborhood structure features. Acta Optica Sinica. 2017;37(10):1010002 10.3788/AOS201737.1010002

Li HX, Guo XF. Research on Multi-Source Information Fusion Technology. In: International Academic Conference on Frontiers in Social Sciences and Management Innovation (IAFSM 2019). Atlantis Press; 2020. p. 24–28.

KTAhmed, SUmmesafi, AIqbal. Content based image retrieval using image features information fusion. Information Fusion. 2019;51:7699. 10.1016/j.inffus.2018.11.004

SFeng, QKai-yang, SWei, GHong. Image saliency detection based on region merging. J Comput Aided Des Comput Graph. 2016;28:16791687.

MLi, LKuang, SXu, ZSha. Brain Tumor Detection Based on Multimodal Information Fusion and Convolutional Neural Network. IEEE Access. 2019;7:180134180146. 10.1109/ACCESS.2019.2958370

XFeng, KHu, XLou. Infrared and visible image fusion based on the total variational model and adaptive wolf pack algorithm. IEEE Access. 2019;8:23482361. 10.1109/ACCESS.2019.2962560

JChen, XLi, LLuo, XMei, JMa. Infrared and visible image fusion based on target-enhanced multiscale transform decomposition. Information Sciences. 2020;508:6478.

10 

JMa, YMa, CLi. Infrared and visible image fusion methods and applications: A survey. Information Fusion. 2019;45:153178. 10.1016/j.inffus.2018.02.004

11 

RHou, DZhou, RNie, DLiu, LXiong, YGuo, et al VIF-Net: an unsupervised framework for infrared and visible image fusion. IEEE Transactions on Computational Imaging. 2020;6:640651. 10.1109/TCI.2020.2965304

12 

MNoushad, SPreetha. Image Pair Fusion using Weighted Average Method. Int J Sci Technol Eng. 2017;3:397402.

13 

ADogra, BGoyal, SAgrawal. From multi-scale decomposition to non-multi-scale decomposition methods: a comprehensive survey of image fusion techniques and its applications. IEEE Access. 2017;5:1604016067. 10.1109/ACCESS.2017.2735865

14 

PChai, XLuo, ZZhang. Image fusion using quaternion wavelet transform and multiple features. IEEE access. 2017;5:67246734. 10.1109/ACCESS.2017.2685178

15 

HLi, BManjunath, SKMitra. Multisensor image fusion using the wavelet transform. Graphical models and image processing. 1995;57(3):235245. 10.1006/gmip.1995.1022

16 

HJin, YWang. A fusion method for visible and infrared images based on contrast pyramid with teaching learning based optimization. Infrared Physics & Technology. 2014;64:134142. 10.1016/j.infrared.2014.02.013

17 

HZhang, XMa, YTian. An image fusion method based on curvelet transform and guided filter enhancement. Mathematical Problems in Engineering. 2020;2020.

18 

JPaul, BUShankar, BBhattacharyya. Change Detection in Multispectral Remote Sensing Images with Leader Intelligence PSO and NSCT Feature Fusion. ISPRS International Journal of Geo-Information. 2020;9(7):462 10.3390/ijgi9070462

19 

MYin, PDuan, WLiu, XLiang. A novel infrared and visible image fusion algorithm based on shift-invariant dual-tree complex shearlet transform and sparse representation. Neurocomputing. 2017;226:182191. 10.1016/j.neucom.2016.11.051

20 

RChao, KZhang, YjLi, et al An image fusion algorithm using wavelet transform. ACTA ELECTRONICA SINICA. 2004;32(5):750753.

21 

MNDo, MVetterli. The contourlet transform: an efficient directional multiresolution image representation. IEEE Transactions on image processing. 2005;14(12):20912106. 10.1109/TIP.2005.859376

22 

JAdu, MWang, ZWu, JHu. Infrared image and visible light image fusion based on nonsubsampled contourlet transform and the gradient of uniformity. International Journal of Advancements in Computing Technology. 2012;4(5):114121.

23 

JAdu, JGan, YWang, JHuang. Image fusion based on nonsubsampled contourlet transform for infrared and visible light image. Infrared Physics & Technology. 2013;61:94100. 10.1016/j.infrared.2013.07.010

24 

JMa, CChen, CLi, Huang J Infrared and visible image fusion via gradient transfer and total variation minimization. Information Fusion. 2016;31:100109. 10.1016/j.inffus.2016.02.001

25 

DPBavirisetti, RDhuli. Fusion of infrared and visible sensor images based on anisotropic diffusion and Karhunen-Loeve transform. IEEE Sensors Journal. 2015;16(1):203209. 10.1109/JSEN.2015.2478655

26 

ZFu, XWang, JXu, NZhou, YZhao. Infrared and visible images fusion based on RPCA and NSCT. Infrared Physics & Technology. 2016;77:114123. 10.1016/j.infrared.2016.05.012

27 

YHuang, DBi, DWu. Infrared and visible image fusion based on different constraints in the non-subsampled shearlet transform domain. Sensors. 2018;18(4):1169.

28 

JMa, WYu, PLiang, CLi, JJiang. FusionGAN: A generative adversarial network for infrared and visible image fusion. Information Fusion. 2019;48:1126. 10.1016/j.inffus.2018.09.004

29 

JMa, PLiang, WYu, CChen, XGuo, JWu, et al Infrared and visible image fusion via detail preserving adversarial learning. Information Fusion. 2020;54:8598. 10.1016/j.inffus.2019.07.005

30 

DLiu, DZhou, RNie, RHou. Infrared and visible image fusion based on convolutional neural network model and saliency detection via hybrid l0-l1 layer decomposition. Journal of Electronic Imaging. 2018;27(6):063036 10.1117/1.JEI.27.6.063036

31 

RHou, RNie, DZhou, JCao, DLiu. Infrared and visible images fusion using visual saliency and optimized spiking cortical model in non-subsampled shearlet transform domain. Multimedia Tools and Applications. 2019;78(20):2860928632. 10.1007/s11042-018-6099-x

32 

PPerona, JMalik. Scale-space and edge detection using anisotropic diffusion. IEEE Transactions on pattern analysis and machine intelligence. 1990;12(7):629639. 10.1109/34.56205

33 

MWan, GGu, WQian, KRen, QChen, XMaldague. Infrared image enhancement using adaptive histogram partition and brightness correction. Remote Sensing. 2018;10(5):682 10.3390/rs10050682

34 

WSCleveland. LOWESS: A program for smoothing scatterplots by robust locally weighted regression. American Statistician. 1981;35(1):54 10.2307/2683591

35 

WWang, FChang. A Multi-focus Image Fusion Method Based on Laplacian Pyramid. JCP. 2011;6(12):25592566.

36 

Toet A, et al. TNO Image fusion dataset. Figshare data. 2014.

37 

Li H, Wu XJ, Kittler J. Infrared and visible image fusion using a deep learning framework. In: 2018 24th International Conference on Pattern Recognition (ICPR). IEEE; 2018. p. 2705–2710.

38 

Fu B, Xiong X, Sun G. An efficient mean filter algorithm. In: The 2011 IEEE/ICME International Conference on Complex Medical Engineering. IEEE; 2011. p. 466–470.

39 

Liang H, Liu S, Yuan H. Optimal algorithms for running max and min filters on random inputs. In: International Computing and Combinatorics Conference. Springer; 2015. p. 507–520.

40 

Deng G, Cahill L. An adaptive Gaussian filter for noise reduction and edge detection. In: 1993 IEEE conference record nuclear science symposium and medical imaging conference. IEEE; 1993. p. 1615–1619.

41 

GArce, MMcLoughlin. Theoretical analysis of the max/median filter. IEEE transactions on acoustics, speech, and signal processing. 1987;35(1):6069. 10.1109/TASSP.1987.1165036

42 

QXiao-Bo, YJing-Wen, XHong-Zhi, ZZi-Qian. Image fusion algorithm based on spatial frequency-motivated pulse coupled neural networks in nonsubsampled contourlet transform domain. Acta Automatica Sinica. 2008;34(12):15081514. 10.1016/S1874-1029(08)60174-3

43 

Chabi N, Yazdi M, Entezarmahdi M. An efficient image fusion method based on dual tree complex wavelet transform. In: 2013 8th Iranian Conference on Machine Vision and Image Processing (MVIP). IEEE; 2013. p. 403–407.

44 

QZhang, BlGuo. Fusion of multisensor images based on the curvelet transform. Journal of Optoelectronics Laser. 2006;17(9):1123.

45 

VNaidu. Image fusion technique using multi-resolution singular value decomposition. Defence Science Journal. 2011;61(5):479.

46 

SLi, XKang, JHu. Image fusion with guided filtering. IEEE Transactions on Image processing. 2013;22(7):28642875. 10.1109/TIP.2013.2244222

47 

Li H, Wu XJ. Infrared and visible image fusion using latent low-rank representation. arXiv preprint arXiv:180408992. 2018.

48 

RShapley, PLennie. Spatial frequency analysis in the visual system. Annual review of neuroscience. 1985;8(1):547581. 10.1146/annurev.ne.08.030185.002555

49 

BPan, ZLu, HXie. Mean intensity gradient: an effective global parameter for quality assessment of the speckle patterns used in digital image correlation. Optics and Lasers in Engineering. 2010;48(4):469477. 10.1016/j.optlaseng.2009.08.010

50 

JGauss, JFStanton, RJBartlett. Coupled-cluster open-shell analytic gradients: Implementation of the direct product decomposition approach in energy gradient calculations. The Journal of chemical physics. 1991;95(4):26232638. 10.1063/1.460915

51 

XLuo, ZZhang, CZhang, XWu. Multi-focus image fusion using HOSVD and edge intensity. Journal of Visual Communication and Image Representation. 2017;45:4661. 10.1016/j.jvcir.2017.02.006

52 

CXydeas, VPetrovic. Objective image fusion performance measure. Electronics letters. 2000;36(4):308309. 10.1049/el:20000267

53 

Petrovic V, Xydeas C. Objective image fusion performance characterisation. In: Tenth IEEE International Conference on Computer Vision (ICCV’05) Volume 1. vol. 2. IEEE; 2005. p. 1866–1871.

54 

Seetha M, MuraliKrishna IV, Deekshatulu B. Data fusion performance analysis based on conventional and wavelet transform techniques. In: Proceedings. 2005 IEEE International Geoscience and Remote Sensing Symposium, 2005. IGARSS’05. vol. 4. IEEE; 2005. p. 2842–2845.

55 

ZWang, ACBovik, HRSheikh, EPSimoncelli. Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing. 2004;13(4):600612. 10.1109/TIP.2003.819861