Next Article in Journal
Changes in the Kinematic and Kinetic Characteristics of Lunge Footwork during the Fatiguing Process
Previous Article in Journal
Towards Automatic Depression Detection: A BiLSTM/1D CNN-Based Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Visible and Near-Infrared Image Synthesis Using PCA Fusion of Multiscale Layers

School of Electronic and Electrical Engineering, Kyungpook National University, 80 Daehakro, Buk-Gu, Daegu 702-701, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(23), 8702; https://doi.org/10.3390/app10238702
Submission received: 20 October 2020 / Revised: 3 December 2020 / Accepted: 3 December 2020 / Published: 4 December 2020

Abstract

:
This study proposes a method of blending visible and near-infrared (NIR) images to enhance their edge details and local contrast based on the Laplacian pyramid and principal component analysis (PCA). In the proposed method, both the Laplacian pyramid and PCA are implemented to generate a radiance map. Using the PCA algorithm, the soft-mixing method and the mask-skipping filter were applied when the images were fused. The color compensation method uses the ratio between the radiance map fused by the Laplacian pyramid and the PCA algorithm and the luminance channel of the visible image to preserve the chrominance of the visible image. The results show that the proposed method improves edge details and local contrast effectively.

1. Introduction

Recent improvements in image sensors have enabled cameras to realize better quality pictures. Many types of image-quality enhancement algorithms, such as high dynamic range (HDR), are also available today. In the field of ultrasmall cameras embedded in smartphones, cameras can be built with various functions by miniaturizing and improving performance. Cameras can now acquire both infrared and visible (VIS) images, and this capability is utilized in various ways, for example, for unlocking the smartphone function or capturing good-quality photos using the infrared camera. Infrared cameras are often used to capture a significant amount of information that lies beyond the VIS spectrum. The near-infrared (NIR) spectrum covers the wavelength range of 700–1100 nm. Therefore, infrared light is less scattered than VIS light, and hence, infrared images can capture more information, such as fog and shade. Capturing infrared images is also advantageous for obtaining details and edge information that are not generally acquired in the VIS spectrum. Many types of algorithms for fusing VIS and NIR images have been developed [1,2,3,4,5].
Most of these algorithms are intended for using NIR scenes to reveal hidden areas in VIS scenes. The conventional VIS and infrared images are used for dehazing. To synthesize VIS and NIR images, most existing methods use multiresolution decomposition and multiresolution-based convergence algorithms. Multiresolution is essential for obtaining details of NIR images. Infrared scenes can preserve image dehazing, high contrast, and edge information in dark areas better.
Laplacian–Gaussian pyramids and the local entropy-fusion algorithm are conventional VIS–NIR fusion algorithms [1]. Local entropy is utilized by synthetic algorithms to transmit the maximum information from VIS and NIR images. Conventional methods also use the multiresolution decomposition of the Laplacian–Gaussian pyramid, wherein the decomposed multiple resolution images are gently blended in each layer using a local-entropy synthesis method. In short, this conventional algorithm uses a Laplacian–Gaussian pyramid for multiresolution fusion and generates a weighting map using the local entropy to control local contrast and visibility to generate the fusion result. However, the details of the NIR image are not well preserved, and the color expression is not natural.
VIS and infrared image-fusion algorithm is a latent low-rank representation fusion algorithm (i.e., low-rank fusion) [5]. Low-rank fusion can be decomposed into low-rank parts (i.e., a global structure) and saliency parts (i.e., a local structure). The weighted-average strategy is used to fuse the low-rank parts to preserve edge information. Then, the saliency parts are fused using the sum strategy, and the final fusion image is obtained from the sum of the low-rank and saliency parts. The latent low-rank fusion algorithm decomposes an input VIS and an infrared image into a low-rank part and a saliency part of each other. After decomposition, each low-rank and saliency part of the VIS and infrared images are fused together by the sum strategy. Finally, the resulting image is obtained using the fused low-rank and saliency parts. However, this algorithm fails to capture the details of the NIR image and does not properly indicate the edge information during color compensation.
In this study, image fusion of VIS and NIR images using the Laplacian pyramid and principal component analysis (PCA) is proposed. Fusion with NIR images is adopted for enhancing image details. Hence, an image is separated by a Laplacian pyramid for details and a base layer in VIS and NIR image is used to enhance the sharpness of the image. The proposed method is presented to create a weighting map for a particular detail area and to skip the PCA calculation for unnecessary areas. The values of the weighting map are obtained properly from the PCA results of the VIS and NIR scenes to improve image details. After generating a radiance map, the proposed method is used to achieve the Stevens effect in a color appearance model [6] to enhance local contrast and details. Tone scaling is adopted for color compensation in the proposed method. Thus, with regard to color compensation, the proposed method calculates the brightness gap between the input VIS image and the generated radiance map. Consequently, compared with conventional methods, the proposed method results in less damage to the edge information in the NIR images improved local contrast and details, and better preservation of the color chrominance of the input VIS image. The proposed method is to preserve the detailed information by a higher value of PCA and enhance local contrast and details according to the visual properties of the Stevens effect. Thus, the proposed method preserves more detailed information and enhances the details effectively.

2. Laplacian Pyramid

The Laplacian pyramid is a multiscale processing algorithm that decomposes input images into multiscale images [7]. Laplacian-pyramid images can be separately processed at each pyramid level. Multiscale images of the Laplacian pyramid are obtained using Gaussian blur, up/downsampling, and image subtraction.
The first step in the Laplacian pyramid is based on the Gaussian pyramid, which produces multiple images with Gaussian blurring and downsampling. A filter with two dimensions of Gaussian blurring is shown in Equation (1):
w ( x , y ) = 1 2 π σ 2 · e x 2 + y 2 2 σ 2 ,
where w ( x , y ) is a function of the Gaussian blur whose size depends on the σ value. Generally, we use a 5 × 5 Gaussian blur prior to downsampling the images. The 5 × 5 Gaussian blur is to provide adequate filtering at a low computational cost [7].
Image downsampling is performed after blurring. The Gaussian pyramid aims to create a repetitive image that has been smoothened and downsampled. Gaussian blurring and downsampling are represented in Equation (2):
g l + 1 ( i , j ) = m = 2 2 n = 2 2 w ( m , n ) g l ( 2 i + m , 2 j + n ) ,
where w ( m , n ) is the Gaussian blur filter; g l is the input image; and g l + 1 ( i , j ) ( i , j ) is the Gaussian pyramid image of the results of blurring and downsampling of the input image. The downsampling method reduces the image size to one-half the row and column of the original image. Thus, the Gaussian pyramid is computed by blurring an input image of a different size. Therefore, the Gaussian pyramid is a sequence of low-pass and downsampled images.
Image blurring is calculated as the average pixel value of the surrounding pixels. Specifically, it is the removal of high-frequency components from an image-processing perspective. Therefore, the Gaussian blur is performed before image downsampling because it removes the high-frequency components that may cause aliasing. Aliasing is a result of a sampling-rate signal that is insufficient for changing other signals. Therefore, image blurring and smoothing are performed before downsampling to eliminate aliasing effects.
The second step in creating the Laplacian pyramid is to upsample the Gaussian pyramid and subtract the previous level image of the Gaussian pyramid to obtain a differential image. The upsampling of the Gaussian pyramid is as expressed in Equation (3):
E x p a n d ( g l + 1 ( i , j ) ) = 4 m = 2 2 n = 2 2 w ( m , n ) g l + 1 ( i m 2 , j n 2 ) ,
where E x p a n d ( g l + 1 ( i , j ) ) is an upsampled image of the Gaussian pyramid, and g l + 1 is an image of the Gaussian pyramid obtained using Equation (2).
A method for obtaining the differential images is given in Equation (4); the upsampling image obtained using Equation (3) is subtracted from the previous step image:
L 0 = g l E x p a n d ( g l + 1 ) ,
where L 0 is the image of the Laplacian pyramid; g l is the Gaussian pyramid obtained in the previous step; and E x p a n d ( g l + 1 ( i , j ) ) is the upsampling image obtained using Equation (3).
The reconstruction of the Laplacian pyramid is an iterative process of upsampling from the lowest-level image in the Laplacian pyramid and adding each level of differential images. To reconstruct the image without any error, the Gaussian-blurred images should be upsampled, and this blurred image is subtracted from the previous level of input images; this is shown in Equations (3) and (4). In summary, the Laplacian pyramid stores the differential images acquired by the Gaussian pyramid at each level. Figure 1 shows a block diagram of the Laplacian pyramid. The Laplacian pyramid image is shown in Figure 1 in normalized grayscale because detailed information in the original Laplacian pyramid images is difficult to see visually.

3. Principal Component Analysis

The PCA is a method of linearly combining the optimal weights for a data distribution [8] and employs the covariance matrix and eigenvector of the data. The eigenvector of the covariance matrix shows the direction of data distribution. Thus, the eigenvectors determine the influence of the principal component (PC). The direction of the eigenvector cannot be changed; however, it is possible to change its magnitude. Therefore, the eigenvector, as well as the PC, can be regarded as a reference axis with increasing or decreasing size. This allows one to determine the reference axis, which comprises the PC that is calculated using the covariance matrix and the eigenvector. The number of PC axes depends on the data dimension. The first PC axis is the maximum variance, which is the longest and widest extent of data distribution. The second PC axis is perpendicular to the first one, and it is the second-widest and longest extent of data distribution. In summary, PCA deals with the decomposition of the eigenvalue calculated using the input data covariance. When the PCA algorithm is adapted to image fusion, the column vectors of the input images must be produced as the first step.
E ( X ) = 1 / M k = 1 M X k ,
c o v ( X ,   Y ) = i = 1 n ( X i E ( X ) ) ( Y i E ( Y ) ) ( n 1 ) ,
where E(∙) is the average of the input data. X is an input image, and cov(∙) is the covariance matrix of the data pairs X and Y .
A x =   λ x ,   d e t ( A λ I ) = 0 ,   A = c o v ( x , y ) ,
P 1 = E v 1 E v ,   P 2 = E v 2 E v ,
where E v is a unique vector and the sum of the PCA weights P 1 and P 2 is 1.
PCA synthesis is accomplished using the mean weights of P 1 and P 2 , which are calculated by multiplying the weights of each input image in Equation (9).
I f u s e d = P 1 × I V I S +   P 2 × I N I R ,
where I f u s e d is the result of image synthesis calculated using the PCA algorithm [9], and I V I S and I N I R are the input VIS and NIR images, respectively. Therefore, the proposed method assigns the weight obtained from the PCA algorithm to fuse the VIS and NIR input images, as shown in Figure 2:

4. Radiance Map Generation

The proposed algorithm is used to generate a radiance map and to achieve color compensation. The flowchart of the proposed algorithm is shown in Figure 3, where L V I S is the luminance channel of the VIS image, and L N I R is the luminance channel of the NIR image. The input VIS and NIR images can be taken by a multi-image sensor device, which is a prism-based multi-band imaging device. It uses the same light path for VIS and NIR image capturing and can be synchronized between spatially separated VIS and NIR images, so it can be filmed without a time delay.
The proposed radiance map algorithm is used in the Laplacian pyramid, and PCA is used in the CIELAB luminance channel. The Laplacian pyramid of input images is decomposed into edge images and base layers. Each layer is fused by the PCA fusion algorithm using the soft-mixing method. The proposed PCA fusion algorithm has a PCA-skipping mask filter. It allows skipping PCA calculations for unnecessary areas by transferring the pixel value of such areas to the next pixel area without calculating the PCA, as shown in Figure 4.
The criteria to determine whether to implement the PCA is based on the pixel value in the PCA-skipping mask-filter area. If the pixel value in the area is less than 10% of the total pixel value in the histogram distribution of the input image, the PCA algorithm is not implemented. The pixel value that is not calculated using the PCA algorithm is set to 0.5 PC weight, as shown in Figure 4. The lower 10% of the histogram distribution of the entire pixel value is skipped because most edge-layer images have pixel values close to zero, except for the edge components. The proposed filter mask determines whether the pixels in the mask-filter area should be calculated using the PCA fusion algorithm. The size of the PCA filter mask depends on the size of the Laplacian pyramid.
The soft-mixing method fuses two images in the transition areas, as shown in Figure 5 [10]. The transition area range is a linear fusion of input images between t h 1   and t h 2 areas and those between t h 3 and t h 4 areas. In Figure 5, t h 1   represents the value of the lower 10%, and t h 2 represents the value of the top 10% of the total in the histogram distribution of PCA weights, as shown on the x-axis in Figure 5a. Further, t h 3   denotes the value of the lower 10%, and t h 4   denotes the value of the top 10% of the total in the histogram distribution of the normalized VIS luminance histogram, as shown on the x-axis of Figure 5b.
Equation (10) is the PCA fusion formula for the VIS and NIR edge layers adapted to Equation (9):
r a d d e t a i l = P d e t a i l ×   V I S d e t a i l +   ( 1   P d e t a i l )   × N I R d e t a i l ,
where P d e t a i l is the detail weighting map of the PCA weights applied by the soft-mixing method, and V I S d e t a i l and N I R d e t a i l are the edge-layer images calculated from the VIS and NIR input images, respectively, with the Laplacian pyramid.
Equation (11) is the fused expression of PCA in the base layer:
b a s e f u s e d = P b a s e × V I S b a s e + ( 1 P b a s e ) × N I R b a s e ,
where P b a s e the base weighting map of the PCA weights applied using the soft-mixing method, and V I S b a s e and N I R b a s e are the base-layer images obtained by calculating the VIS and NIR input images, respectively, using the Laplacian pyramid.
Equation (12) applies the fused image obtained using Equation (11) and the NIR images to enhance the local contrast and detail information in the base-layer images.
r a d b a s e = P r a d b a s e × N I R b a s e + ( 1     P r a d b a s e ) × b a s e f u s e d ,
where P r a d b a s e is the base weighting map of the PCA weights applied using the soft-mixing method in Figure 5b for improving the local contrast, and N I R b a s e is the NIR image obtained from the Laplacian pyramid, and b a s e f u s e d is the fusion image obtained using Equation (11).
Equation (13) shows a radiance map that applies the PCA image-fusion algorithm and the reconstruction process of the Laplacian pyramid:
Radiance   map =   n = 0 l { r a d n +   E x p a n d ( r a d n + 1 ) } ( base   layer   for   n = l ,   detail   layer   for   n = 0   ~ l 1 ) ,
where r a d n denotes the result of the fusion algorithm of the PCs calculated using Equations (10) and (12). E x p a n d ( r a d n + 1 ) represents the reconstruction process of the Laplacian pyramid (Equation (3)), and the radiance map denotes the radiance map obtained through the PCA and the Laplacian pyramid reconstruction process.
Figure 5a shows the results of applying the soft-mixing method for the edge layer. If the PCA weight is less than t h 1 , the weight shall be used only for the NIR image pixel value, and then the pixel value should be 1. Conversely, if the PCA weight exceeds t h 2 , the weight should only be used as the pixel value of the VIS image, and then the pixel value should be 1. The PCA weighting map is multiplied by two base-layer images and is adapted to the PCA fusion algorithm given in Equation (12).
Figure 5b shows the results of applying the soft-mixing method for the base layer. The PCA fusion algorithm is run differently from the detail-layer fusion algorithm because the base layer has low contrast. The PCA fusion is lower than the contrast of the VIS image only if the PCA weight of the base layer is used. Therefore, the base layer is considered in two stages of the PCA fusion algorithm.
In the first step, the PCA fusion algorithm is used in a manner similar to edge-layer PCA fusion. The VIS base layer and NIR base-layer images are implemented by the PCA fusion algorithm, as shown in Equation (11).
In the second step, the PCA fusion algorithm used in Equation (12) employs the PCA fusion base obtained from Equation (11) and the NIR base layer. Here, the NIR base layer was used because it contained edge information and higher contrast than the PCA fusion base and the VIS image. Moreover, if it were used with the VIS images, the fused image would have contained shade or reverses light. This would have reduced the tone compression effect.
To generate a radiance map, the Laplacian pyramid algorithm was reconstructed following Equation (13): the lower layer was upsampled and then combined with the previous layer. The lowest layer was an image comprising the base fusion of Equations (11) and (12), which upsampled this composed image and added an edge layer image. The images comprising the base and edge images were also upsampled, and then, the next edge-layer images were added.
However, the Laplacian pyramid-based multiscale image fusion algorithms are difficult for preserving details. It is necessary to use the detailed enhancement component on the Laplacian pyramid [11,12]. Thus, Based on the Stevens effect [6], which is the human visual characteristic of local contrast, a gamma value that changes locally according to regional adaptation luminance was applied to the detail layer to improve local contrast. After generating the radiance map, the proposed algorithm was used to realize the Stevens effect to increase the contrast. The Stevens effect increases local perceptual contrast in the detail layer, and the Stevens effect power function determines the luminance-dependent local contrast enhancement.
k = 1 / ( 5 L A + 1 ) ,
where L A is a function of 20% of the adaptation luminance
F = 0.2 k 4 ( 5 L A ) + 0.1 ( 1 k 4 ) 2 ( 5 L A ) 1 / 3 ,
where F is a prediction of a variety of luminance-dependent appearance effects in CIECAM02.
d e t a i l s s t e v e n s =   ( d e t a i l R a d i a n c e   m a p ) ( F + 0.8 ) 1 / 4 ,
where d e t a i l s s t e v e n s is a function that is adapted for the Stevens effect. Further, d e t a i l R a d i a n c e   m a p is a detail layer of the radiance map expressed in Equation (13).
R a d i a n c e   m a p f i n = b a s e R a d i a n c e   m a p d e t a i l s s t e v e n s ,
where R a d i a n c e   m a p f i n is the resultant radiance map, which is a combination of the base layer of the radiance map and d e t a i l s s t e v e n s . Further, b a s e R a d i a n c e   m a p is a base layer of the radiance map expressed in Equation (13).
In conclusion, the edge and base layers synthesized using PCA were implemented by reconstructing the Laplacian pyramid, and the final image was fused using the Stevens effect to increase local contrast. Thus, the proposed radiance map was obtained.

5. Color Compensation

Tone changing through the image fusion causes color distortions originating from the balance changes of RGB channels in the tone changing. The ratio of RGB signals changes, and color saturation would be reduced [13]. Tone changes should be applied to only the luminance channel, and the process sequences preserve the color information of the visible image. Converting from RGB signals to opponent color signals is necessary for constructing a uniform perceptual color space and correlated of various appearance attributes [14]. To correct this chromatic defect of tone mapping, we adopt a color compensation method that restores a ratio of color to luminance after tone mapping.
After VIS and NIR input images are blended to generate the proposed radiance map, the proposed color compensation uses a ratio between the radiance map and the VIS input image. The color compensation must consider the fact that tone scaling has a brightness gap. Therefore, the color compensation ratio is calculated using the radiance map and the VIS input image in the luminance channel in Equation (18). The ratio represents the variation in the luminance channel between the radiance map and the VIS input image:
r = R a d i a n c e   m a p L V I S ,
where r is the surrounding brightness variation between the VIS input image and the radiance map.
L s = ( g ( x , y ) L V I S ( x , y ) ) γ ,
where L s is a surrounding image with gamma correction, and g ( x , y ) is a Gaussian blur function. We used a 15 × 15 size; the δ (sigma) value of the Gaussian filter size is 15. L V I S is a VIS luminance channel image. Further, γ is the gamma value. We used the γ value as 0.1.
r a d p t = r × L s + ( 1 L s ) ,
where r a d p t is an adapted local brightness variation. Further, r is the surrounding brightness variation between the VIS input image and radiance map given in Equation (18), and L s represents a surrounding image with gamma correction, as shown in Equation (19).
a = ( a 128.0 ) × r a d p t + 128.0 ,
b = ( b 128.0 ) × r a d p t + 128.0 ,
where a and b are the compensated chrominance channels. The term r a d p t is introduced to eliminate color noise in dark areas. This term also increases the color gain in the bright areas and preserves the color, which is obtained from VIS input image, in the dark area. The chrominance channel is compensated by using r a d p t as proposed in Equations (21) and (22), where a and b are the chrominance channels in the CIELAB color space. This color compensation preserves the VIS input-image color to allow its natural application to the proposed radiance map.

6. Simulations

The proposed resulting images were evaluated against images obtained using conventional methods. As conventional methods, we considered the Laplacian fusion algorithm [1] and the low-rank fusion algorithm [5]. We used input images (i.e., reference images of the conventional method [1]). Both VIS and NIR input images were captured in the daytime. The conventional method (i.e., the low-rank fusion algorithm [5]) was adapted to the proposed color compensation algorithm to compare the Laplacian fusion algorithm [1] with the proposed method for color images.
The resulting images of the low-rank conventional method (d) contained noise, as shown in Figure 6, Figure 7, Figure 8, Figure 9, Figure 10, Figure 11 and Figure 12. Furthermore, we found that compared to the aforementioned conventional methods, the proposed method better represents the edges of objects such as clouds, trees, leaves, and mountains. If an object is clear in the input VIS image, the output images of the proposed method will preserve the hue of the color information represented in the input VIS image.
The resulting values were calculated from Table 1 to evaluate the image quality and image sharpness for comparing the results obtained using the conventional and proposed methods. Feature mutual information (FMI) [15] is a method of calculating image-fusion quality based on the calculation of information contained in the two input images. S3 [16] is the measurement criteria for image clarity, and it is measured according to the locally recognized sharpness of the image.
In Table 1 and Figure 13, a high two-metric score means that the image is of good quality. The average score of the image evaluation indicator for the proposed method is higher than that of the conventional method. The FMI method of evaluating fusion quality has higher scores than conventional methods. The S3 method of fusion quality evaluation, i.e., sharpness evaluation, has higher scores than the conventional methods. Thus, compared with the results of the conventional methods, the results of the proposed method were evaluated to be much better in terms of picture quality

7. Conclusions

We proposed the synthesis of VIS and NIR images. A radiance map was created using the Laplacian pyramid, and the soft-mixing method was employed based on PCA. Each pyramid level was implemented on the luminance channel in the CIELAB color space. The Laplacian pyramid was decomposed into its edge and base layers. These decomposed layers were calculated by combining the PCA fusion algorithm with the soft-mixing method. In particular, the base-layer soft-mixing method had a low contrast in the base layer. Therefore, the results were calculated differently from the applied PCA fusion algorithm. The base layer was fused with the NIR base layer to enhance the local contrast. Color compensation was used to preserve the VIS input image. To preserve the VIS image color, the proposed color compensation was adjusted by using the ratio of the luminance channels of the radiance map and the VIS input image. When the proposed color compensation process was completed, the fused image was converted from the CIELAB color space to the RGB color space. The advantages of the resulting image obtained by the proposed method are better local contrast and excellent detail and color representation. The possible real-life applications of the proposed method are to increase the visibility for a number of the camera embedded devices, such as a self-driving car, CCTV, drone and car black–box. In future work, we will attempt to fuse VIS and NIR images under nighttime conditions.

Author Contributions

Conceptualization, S.-H.L.; Methodology, S.-H.L.; Software, D.-M.S.; Supervision, S.-H.L.; Writing—original draft, D.-M.S.; Writing—review & editing, H.-J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) and the BK21 four project funded by the Ministry of Education, Korea (NRF-2019R1D1A3A03020225).

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this paper.

References

  1. Vanmali, A.V.; Gadre, V.M. Visible and NIR image fusion using weight-map-guided Laplacian–Gaussian pyramid for improving scene visibility. Sadhana 2017, 42, 1063–1082. [Google Scholar] [CrossRef] [Green Version]
  2. Kil, T.; Cho, N.I. Image Fusion using RGB and Near Infrared Image. J. Broadcast Eng. 2016, 21, 515–524. [Google Scholar] [CrossRef]
  3. Sadeghipoor, Z.; Lu, Y.M.; Süsstrunk, S. Correlation-based joint acquisition and demosaicing of visible and near-infrared images. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011; pp. 3165–3168. [Google Scholar] [CrossRef] [Green Version]
  4. Zhang, X.; Sim, T.; Miao, X. Enhancing photographs with Near Infra-Red images. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  5. Li, H.; Wu, X.-J. Infrared and visible image fusion using Latent Low-Rank Representation. arXiv 2018, arXiv:1804.08992. [Google Scholar]
  6. Kuang, J.; Johnson, G.M.; Fairchild, M.D. iCAM06: A refined image appearance model for HDR image rendering. J. Vis. Commun. Image Represent. 2007, 18, 406–414. [Google Scholar] [CrossRef]
  7. Burt, P.J.; Adelson, E.H. The Laplacian Pyramid as a Compact Image Code BT—Fundamental Papers in Wavelet Theory. Fundam. Pap. Wavelet Theory 1983, 31, 532–540. [Google Scholar]
  8. Desale, R.P.; Verma, S.V. Study and analysis of PCA, DCT & DWT based image fusion techniques. In Proceedings of the 2013 International Conference on Signal Processing, Image Processing & Pattern Recognition, Coimbatore, India, 7–8 February 2013; pp. 66–69. [Google Scholar] [CrossRef]
  9. Kaur, R.; Kaur, S. An Approach for Image Fusion using PCA and Genetic Algorithm. Int. J. Comput. Appl. 2016, 145, 54–59. [Google Scholar] [CrossRef]
  10. Kwon, H.J.; Lee, S.H.; Lee, G.Y.; Sohng, K.I. Radiance map construction based on spatial and intensity correlations between le and SE images for HDR imaging. J. Vis. Commun. Image Represent. 2016, 38, 695–703. [Google Scholar] [CrossRef]
  11. Li, Z.G.; Zheng, J.H.; Rahardja, S. Detail-Enhanced Exposure Fusion. IEEE Trans. Image Process. 2012, 21, 4672–4676. [Google Scholar] [CrossRef] [PubMed]
  12. Li, Z.; Wei, Z.; Wen, C.; Zheng, J. Detail-Enhanced Multi-Scale Exposure Fusion. IEEE Trans. Image Process. 2017, 26, 1243–1252. [Google Scholar] [CrossRef] [PubMed]
  13. Mantiuk, R.; Mantiuk, R.; Tomaszewska, A.; Heidrich, W. Color correction for tone mapping. Comput. Graph. Forum 2009, 28, 193–202. [Google Scholar] [CrossRef]
  14. Fairchild, M.D. Color Appearance Models, 3rd ed.; Wiley-IS&T: Chichester, UK, 2013. [Google Scholar]
  15. Haghighat, M.B.A.; Aghagolzadeh, A.; Seyedarabi, H. A non-reference image fusion metric based on mutual information of image features. Comput. Electr. Eng. 2011, 37, 744–756. [Google Scholar] [CrossRef]
  16. Vu, C.T.; Phan, T.D.; Chandler, D.M. S3: A spectral and spatial measure of local perceived sharpness in natural images. IEEE Trans. Image Process. 2012, 21, 934–945. [Google Scholar] [CrossRef]
Figure 1. Laplacian pyramid block diagram.
Figure 1. Laplacian pyramid block diagram.
Applsci 10 08702 g001
Figure 2. Principal component analysis (PCA) fusion block diagram.
Figure 2. Principal component analysis (PCA) fusion block diagram.
Applsci 10 08702 g002
Figure 3. Flowchart of the proposed algorithm.
Figure 3. Flowchart of the proposed algorithm.
Applsci 10 08702 g003
Figure 4. PCA-skipping mask-fusion method.
Figure 4. PCA-skipping mask-fusion method.
Applsci 10 08702 g004
Figure 5. Soft-mixing method for each layer: (a) Edge-layer fusion using a PCA level and (b) base-layer fusion using a visible (VIS) luminance level.
Figure 5. Soft-mixing method for each layer: (a) Edge-layer fusion using a PCA level and (b) base-layer fusion using a visible (VIS) luminance level.
Applsci 10 08702 g005
Figure 6. Input image pair and resulting images: (a) input VIS and NIR images, (b) Laplacian pyramid fusion [1], (c) low-rank fusion [5], and (d) the proposed method.
Figure 6. Input image pair and resulting images: (a) input VIS and NIR images, (b) Laplacian pyramid fusion [1], (c) low-rank fusion [5], and (d) the proposed method.
Applsci 10 08702 g006
Figure 7. Input image pair and resulting images: (a) input VIS and near-infrared (NIR) images, (b) Laplacian pyramid fusion [1], (c) low-rank fusion [5], and (d) the proposed method.
Figure 7. Input image pair and resulting images: (a) input VIS and near-infrared (NIR) images, (b) Laplacian pyramid fusion [1], (c) low-rank fusion [5], and (d) the proposed method.
Applsci 10 08702 g007
Figure 8. Input image pair and resulting images: (a) input VIS and NIR images, (b) Laplacian pyramid fusion [1], (c) low-rank fusion [5], and (d) the proposed method.
Figure 8. Input image pair and resulting images: (a) input VIS and NIR images, (b) Laplacian pyramid fusion [1], (c) low-rank fusion [5], and (d) the proposed method.
Applsci 10 08702 g008
Figure 9. Input image pair and resulting images: (a) input VIS and NIR images, (b) Laplacian pyramid fusion [1], (c) low-rank fusion [5], and (d) the proposed method.
Figure 9. Input image pair and resulting images: (a) input VIS and NIR images, (b) Laplacian pyramid fusion [1], (c) low-rank fusion [5], and (d) the proposed method.
Applsci 10 08702 g009
Figure 10. Input image pair and resulting images: (a) input VIS and NIR images, (b) Laplacian pyramid fusion [1], (c) low-rank fusion [5], and (d) the proposed method.
Figure 10. Input image pair and resulting images: (a) input VIS and NIR images, (b) Laplacian pyramid fusion [1], (c) low-rank fusion [5], and (d) the proposed method.
Applsci 10 08702 g010
Figure 11. Input image pair and resulting images: (a) input VIS and NIR images, (b) Laplacian pyramid fusion [1], (c) low-rank fusion [5], and (d) the proposed method.
Figure 11. Input image pair and resulting images: (a) input VIS and NIR images, (b) Laplacian pyramid fusion [1], (c) low-rank fusion [5], and (d) the proposed method.
Applsci 10 08702 g011
Figure 12. Input image pair and resulting images: (a) input VIS and NIR images, (b) Laplacian pyramid fusion [1], (c) low-rank fusion [5], and (d) the proposed method.
Figure 12. Input image pair and resulting images: (a) input VIS and NIR images, (b) Laplacian pyramid fusion [1], (c) low-rank fusion [5], and (d) the proposed method.
Applsci 10 08702 g012
Figure 13. Metric scores: (a) feature mutual information (FMI) metric score and (b) S3 metric score.
Figure 13. Metric scores: (a) feature mutual information (FMI) metric score and (b) S3 metric score.
Applsci 10 08702 g013
Table 1. Comparison of image quality metrics.
Table 1. Comparison of image quality metrics.
Laplacian Pyramid FusionLow-Rank FusionProposed Method
FMIS3FMIS3FMIS3
Figure 60.21890.15410.23650.15110.28210.3465
Figure 70.24920.17310.25710.12570.26680.3064
Figure 80.25720.25240.3160.17760.32420.382
Figure 90.29140.18330.2770.1640.32910.3574
Figure 100.2470.20270.26220.15720.31240.3582
Figure 110.41030.29020.34650.18990.34480.276
Figure 120.26460.50730.38810.31270.39930.4354
average0.27690.25190.29760.18260.32270.3517
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Son, D.-M.; Kwon, H.-J.; Lee, S.-H. Visible and Near-Infrared Image Synthesis Using PCA Fusion of Multiscale Layers. Appl. Sci. 2020, 10, 8702. https://doi.org/10.3390/app10238702

AMA Style

Son D-M, Kwon H-J, Lee S-H. Visible and Near-Infrared Image Synthesis Using PCA Fusion of Multiscale Layers. Applied Sciences. 2020; 10(23):8702. https://doi.org/10.3390/app10238702

Chicago/Turabian Style

Son, Dong-Min, Hyuk-Ju Kwon, and Sung-Hak Lee. 2020. "Visible and Near-Infrared Image Synthesis Using PCA Fusion of Multiscale Layers" Applied Sciences 10, no. 23: 8702. https://doi.org/10.3390/app10238702

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop