Next Article in Journal
Practical Guide to Ultrasound Beam Forming: Beam Pattern and Image Reconstruction Analysis
Previous Article in Journal
The Application of Deep Learning and Image Processing Technology in Laser Positioning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Halo-Free Multi-Exposure Image Fusion Based on Sparse Representation of Gradient Features

1
Faculty of Information Science and Engineering, Ningbo University, Ningbo 315211, China
2
Faculty of Information and Intelligent Engineering, Ningbo City College of Vocational Technology, Ningbo 315100, China
3
National Key Lab of Software New Technology, Nanjing University, Nanjing 210093, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2018, 8(9), 1543; https://doi.org/10.3390/app8091543
Submission received: 6 July 2018 / Revised: 26 August 2018 / Accepted: 29 August 2018 / Published: 3 September 2018
(This article belongs to the Section Optics and Lasers)

Abstract

:
Due to sharp changes in local brightness in high dynamic range scenes, fused images obtained by the traditional multi-exposure fusion methods usually have an unnatural appearance resulting from halo artifacts. In this paper, we propose a halo-free multi-exposure fusion method based on sparse representation of gradient features for high dynamic range imaging. First, we analyze the cause of halo artifacts. Since the range of local brightness changes in high dynamic scenes may be far wider than the dynamic range of an ordinary camera, there are some invalid, large-amplitude gradients in the multi-exposure source images, so halo artifacts are produced in the fused image. Subsequently, by analyzing the significance of the local sparse coefficient in a luminance gradient map, we construct a local gradient sparse descriptor to extract local details of source images. Then, as an activity level measurement in the fusion method, the local gradient sparse descriptor is used to extract image features and remove halo artifacts when the source images have sharp local changes in brightness. Experimental results show that the proposed method obtains state-of-the-art performance in subjective and objective evaluation, particularly in terms of effectively eliminating halo artifacts.

1. Introduction

There is abundant information about brightness and color in real natural scenes, and the dynamic range is very wide. Therefore, high dynamic range (HDR) images, which can truly represent natural scenes, are widely used in digital photography, aerospace, satellite meteorology, medical treatment, and military industries [1,2,3].
The HDR imaging technology can be divided into two categories. The first is the hardware-based method [4,5], which utilizes HDR devices to capture real natural scenes. However, these devices are too expensive to be popular with the public. The second uses software-based technology to obtain HDR images [6]. The notion of this category is that real natural scenes can be captured with a stack of low dynamic range (LDR) images with different exposures, and an HDR image can be obtained by using the camera response function [7,8]. Then, tone mapping technology [9,10,11] is used to reduce the dynamic range of the HDR image for display on the common device. However, the quality of the reconstructed HDR image depends heavily on the related exposure settings (e.g., exposure time and exposure value), which are often unknown to the user. Therefore, software-based HDR imaging technology through multi-exposure fusion [12,13,14] has attracted increasing amounts of attention. This technology can directly obtain greater information and produce a more perceptually appealing composite LDR image, reflecting natural scenes more closely. Furthermore, the result of a multi-exposure fusion can be directly presented on the LDR displays.
In recent years, many scholars have performed in-depth studies on multi-exposure fusion. Mertens et al. [15] proposed a multi-exposure fusion method based on weight maps, the factors of which are comprised of three quality measures (i.e., contrast, saturation, and exposure quality at the pixel level). Moreover, the multi-resolution approach, based on Laplacian pyramids, was also applied to pursue better human visual perception. However, this method placed too much emphasis on contrast. Thus, the details of brighter and darker regions were lost in the fused image. Vonikakis et al. [16] evaluated the exposure quality of pixels in the source images via illumination estimation filtering and then realized multi-exposure image fusion. Zhang et al. [17] achieved the multi-exposure fused image with gradient information, which can effectively preserve texture details. Li et al. [18] decomposed a number of multi-exposure source images into a detailed layer, containing small-scale details, and a base layer, containing large-scale intensity variations. Then, they used spatial consistency to fuse the basic and detailed layers. Shen et al. [19] proposed a multi-exposure fusion method, based on an improved pyramid, to preserve detail information. Liu et al. [12] introduced a dense scale invariant feature transform (SIFT) descriptor into image fusion to extract local details from multi-exposure source images. Ma et al. [13] proposed a multi-exposure fusion method, based on an effective structural path decomposition approach, which took account of the local and global characteristics of the source images and obtained better fused image quality. Prabhakar et al. [20] proposed an unsupervised deep learning framework for multi-exposure fusion (MEF), utilizing a no-reference quality metric as the loss function, which obtained excellent fusion results. On the basis of Ma et al. [13], Ma et al. [14] then proposed a multi-exposure fusion method by optimizing the color multi-exposure fusion structural similarity (MEF-SSIMc) index. However, the above fused methods did not consider that there were halo artifacts in the multi-exposure fused image. These halo artifacts always occur in the region of sharp local brightness changes, severely reducing the perceptual appeal.
Additionally, the theory of sparse representation shows that the linear combination of a “few” atoms from an over-complete dictionary can effectively describe the salient features of objects [21,22,23]. Thus, it has been widely used for image processing. Yang et al. [24] introduced sparse representation into image fusion. With their method, source images were divided into small patches and transformed into vectors via lexicographic ordering. They then realized sparse representation using an over-complete dictionary. Liu et al. [25] combined multi-scale transformation and sparse representation to constitute a general image fusion framework, which presented more details in the fused images. However, these works mainly focused on multi-focus and multi-modal fusion, and because they did not consider the characteristics of multi-exposure source images, they are still not appropriate for multi-exposure image fusion.
In this paper, we propose a halo-free multi-exposure fusion method based on sparse representation of gradient features. In our method, the local gradient sparse descriptor, obtained from a gradient map through sparse representation, is used for both local detail extraction and halo artifact removal. The main contributions of this paper can be summarized in the following three points:
(1)
By analyzing characteristics of the luminance gradient in multi-exposure source images, the cause of halo artifacts in the fused image is discussed. Because the local brightness changes in HDR scenes are violent, they can far exceed the dynamic range of a single photograph. Thus, the halo artifacts in the fused image are generated by the invalid large-amplitude gradients in the source images.
(2)
By calculating the l1-norm of the sparse coefficients in the luminance gradient map, we obtain the local gradient sparse descriptor as the feature of the images. This descriptor can effectively extract local details of source images.
(3)
A new multi-exposure fusion method based on the local gradient sparse descriptor is proposed, which can achieve a halo-free fused image by restraining the invalid large-amplitude gradients. Meanwhile, the fused image of the proposed method is not recovered by the over-complete dictionary, which effectively improves the fusion efficiency.
The remainder of this paper is organized as follows: In Section 2, we analyze the cause of halo artifacts in the multi-exposure fused images; Section 3 presents the proposed halo-free fusion method; the experimental results and discussions are described in Section 4; and finally, Section 5 concludes the paper.

2. Cause of Halo Artifacts in Fused Images

In HDR natural scenes, there is not only abundant information about brightness and color, but also plenty of sharp local brightness changes. In the process of multi-exposure fusion, if those sharp brightness changes cannot be fully considered, the halo artifacts will appear in the fused images. As seen in Figure 1a, which shows the fused result of Li et al. [18], halo artifacts have occurred in the sky outside the windows. By contrast, the sharp local brightness changes are fully considered in the proposed method. Thus, the fused image is more consistent with human visual perception through the removal of halo artifacts, as shown in Figure 1b.
The cause of these halo artifacts are revealed in Figure 2. In the multi-exposure source images of Figure 2a–c, the exposure of the scenery outside the window is increasing until over-exposure, while the brightness contrast increases step-by-step between indoors and outdoors. In the local luminance gradient amplitude of Figure 2k–o, the maximal gradient is found in Figure 2m. However, the region around the maximal gradient in the corresponding source image in Figure 2c is over-exposed and unsuitable for the fused image. However, in the same position of Figure 2a,b, the well-exposed region can be found, which are more suitable for the fused image than that of Figure 2c. Therefore, the maximal gradient existing in Figure 2m is an invalid gradient and the primary cause of halo artifacts in the fused image. The reason for this phenomenon is that a single photograph has very limited dynamic range, compared to real HDR scenes. Thus, the invalid large-amplitude gradients occur in the multi-exposure source images. If these invalid gradients cannot be suppressed, halo artifacts are inevitable in the fused image.
In most multi-exposure image fusion methods, the image features, such as contrast and saliency, are used as weight coefficients for image fusion, except for the gradient. However, these image features have more or less a direct connection with the gradient, which leads to halo artifacts of varying degrees in the fusion results. In the method by Mertens et al. [15], a Laplacian filter was applied to the grayscale version of each multi-exposure image to yield a simple indicator, C, for contrast. This indicator C was then used as one of the weight coefficients for image fusion. Generally, pixels with larger contrast always have larger gradients, so the pixels with invalid large gradients inevitably take part in the image fusion process with larger weight, generating the halo artifacts. Like Mertens et al.’s method [15], the Laplacian filter was implemented to each source image in the Li method [18] to obtain a high-pass image. Then, the Gaussian filter was used in the high-pass image to obtain the saliency map of the source image, which was used as the winner-take-all weight coefficient for image fusion. It can be noted that pixels with larger gradients always have greater saliency, so the pixels with invalid large gradients are also involved in image fusion with larger weighting factors. In the Ma method [13,14], signal contrast was used as one of benchmarks for image fusion, and the desired contrast of the fused image patch was determined by the highest contrast of all the source image patches, which would lead to the appearance of halo artifacts in the fused images.

3. Multi-Exposure Fusion via Sparse Representation of Gradient Features

Through further analysis of the local luminance gradient in Figure 2k–o, it can be seen that there is a maximal gradient existing in Figure 2m, whereas the gradients around this maximal gradient are almost at a zero level and are not changed. However, in Figure 2k,l, there are discernible gradient changes in those regions and most of the values are not zero. Considering the salience extraction capability of sparse representation, especially for a subtle feature, the local gradient sparse descriptor is proposed in our method, which can not only extract the local details, but can also effectively constrain the invalid gradients of large amplitude.
The schematic diagram of the proposed multi-exposure image fusion method, based on a sparse representation of gradient features, is shown in Figure 3. There are three steps in the proposed method: Image feature extraction, weight map calculation and weight term-based fusion. These image features are composed of a local gradient sparse descriptor and exposure quality.

3.1. Local Gradient Sparse Descriptor Extraction

The local gradient sparse descriptor is obtained by the sparse representation model. Since the digital signal can be represented with a linear combination of a “few” atoms from an over-complete dictionary [26,27], the inherent feature of the signal can be extracted through the sparse coefficient combination of the corresponding atoms. Mathematically, the digital signal y can be presented by the following model:
y = i = 1 , k d i α i = D A ,
where D = ( d 1 , d 2 , d k ) R n × k ( n < k ) denotes an over-complete dictionary, d i R n ( i = 1 , 2 , k ) is an atom of the over-complete dictionary, and A = ( α 1 , α 2 , α k ) R k indicates the sparse coefficient vector, which has a value that can be obtained by solving the following optimization problem:
min a a 0   subject   to   y = D A ,
where · 0 denotes the l0-norm: The number of nonzero components.
Generally, there is noise in the original digital signal. Thus, the sparse representation model, including noise, can be expressed as
min a a 0   subject   to   y D A 2 2 < ε ,
where ε > 0 is the error tolerance.
In the proposed method, the process of sparse representation is realized in two steps, namely dictionary learning and feature extraction.

3.1.1. Dictionary Learning in the Gradient Domain

During the process of sparse representation, it is meaningful to obtain a suitable over-complete dictionary [28]. Therefore, such a dictionary in the gradient domain should be obtained.
The dictionary learns from a large number of example image patches, using a certain training algorithm, such as K-means or K-SVD [29], and has better representative ability. Firstly, the luminance gradient map of the training image, I, is calculated:
S L = ( L ( I ) ) ,
where L is the operation for obtaining image luminance, and is the operation for obtaining the image gradient.
Then, the K-SVD algorithm is used to obtain the dictionary of gradient domain from the clear images of indoors and outdoors.
min D , A S L D A F 2   subject   to   i α i 0 < t   and   j d j 2 = 1 .
The dictionary learning framework in the gradient domain is shown in Figure 4. First, lots of luminance gradient maps are obtained from a large number of clear images, including indoors and outdoors. To construct the training set for dictionary learning from those luminance gradient maps, numerous image patches of size n × n are randomly sampled in the same manner as Reference [30], but the patch size is odd, n = 2r + 1. Then, all of image patches are transformed into vectors via column by column extraction. Finally, the over-completed dictionary of the gradient domain can be obtained by the K-SVD algorithm.

3.1.2. Local Gradient Sparse Descriptor Extraction

During the process of local gradient sparse descriptor extraction, luminance gradient maps of the multi-exposure source images are obtained by Equation (4), as shown in Figure 5a,b. Then, the luminance gradient map, SL, is divided into overlapping image patches, as in Yang’s method [24]. However, the patch size is odd, n = 2r + 1, and as big as the patch size used in the dictionary learning. The next step is to shape each image patch into a vector through column-by-column extraction, as shown in Figure 5c,d. It is noteworthy that the small red piece in this figure represents the center of the image patch. Finally, each vector is now represented by a gradient domain from the over-complete dictionary obtained in the previous section, as shown in Figure 5d, wherein the black piece represents the nonzero sparse coefficient. Thus, the image patches can be expressed as:
V L = [ d 1 d 2 d k ] [ s 1 ( 1 ) s 2 ( 1 ) s J ( 1 ) s 1 ( 2 ) s 2 ( 2 ) s J ( 2 ) s 1 ( k ) s 2 ( k ) s J ( k ) ] + ε ,
where V L R n 2 × J denotes the matrix that is lexicographically ordered from the luminance gradient maps, [ d 1 , d 2 , d k ] R n 2 × k ( n 2 < k ) is the over-complete dictionary, s j = [ s j ( 1 ) s j ( 2 ) s j ( k ) ] T is the sparse coefficient of each image patch, J is the number of image patches, and ε > 0 is the error tolerance. If D = [ d 1 , d 2 , d k ] , and S = [ s 1 , s 2 , s J ] T , which is a sparse matrix, Equation (6) can be expressed as:
V L = D S .
Then, the gradient sparse descriptor of the central pixel in the image patch, Ai(x, y), can be calculated as
A i ( x , y ) = s i j 1 ,
where · 1 denotes the l1-norm, s i j is the jth sparse matrix of the ith source image.
Generally, the larger the gradient value of the corresponding patch in multi-exposure source images, the better the exposure result of these regions, and the gradient sparse descriptor Ai(x, y) is also larger. However, in the HDR scenes, the invalid larger amplitude gradient may be found in a particular exposure source image, which does not represent a good exposure result. Fortunately, the gradients in the adjacent region of these invalid gradients are almost zero. So, the local gradient sparse descriptor Ai(x, y) of these regions is much smaller than the corresponding well-exposed image regions in other source images. Therefore, Ai(x,y) can effectively suppress the invalid large-amplitude gradients and remove halo artifacts from the fused image.

3.2. Exposure Quality Extraction

In the multi-exposure source images, there are more local details in the well-exposed regions. In the over-exposed or under-exposed regions, the luminance values are too extreme to present valid information. Therefore, the pixel luminance value is used to measure the local exposure quality. First, the luminance value of the images is normalized. Then, considering that there are more noises in the region under low illumination conditions, the exposure quality of pixels is calculated to obtain the exposure quality map:
B i ( x , y ) = { 1 , η + α < I ˜ i ( x , y ) < 1 α , 0 , o t h e r w i s e ,
where η ( 0 η 0.45 ) and α ( 0 α 0.5 η ) are the thresholds, I ˜ i ( x , y ) is the luminance map of the source image, and Bi(x, y) = 1 denotes that the related pixel is in the well-exposed region. Thus, the influence of over-exposed or under-exposed regions to the fused image can be effectively removed.

3.3. Initial Weight Map Estimation

After obtaining the local gradient sparse descriptor and exposure quality feature, the initial weight maps can be estimated by multiplying these two terms one-by-one:
T i ( x , y ) = A i ( x , y ) × B ( x , y ) .
Then, the weight maps, T i ( x , y ) are normalized:
W i ( x , y ) = T i ( x , y ) + τ i = 1 N ( T i ( x , y ) + τ ) ,
where τ denotes a small positive number (e.g., 10−25), which is used to avoid singularities in the weight value T ^ i ( x , y ) .

3.4. Final Weight Map Estimation and Fusion

Since the noises in the initial weight maps have greater influence on the fused image, the maps must be refined further. Considering the availability and real-time capability, the recursive filter [31], which can perform high-quality edge-preserving filtering of images in real time, is used to refine the initial weight maps:
W ^ i ( x , y ) = R F ( W i ( x , y ) , I i ( x , y ) ) ,
where RF(•) presents the recursive filter operation, and Wi denotes the initial weight map, and Ii are the multi-exposure source images which can intensify the edge information during filtering.
After the final weight map has been normalized, the fused image, IF, can be obtained by:
W ˜ i ( x , y ) = W ^ i ( x , y ) + τ i = 1 N ( W ^ i ( x , y ) + τ ) ,
and
I F ( x , y ) = i = 1 N W ˜ i ( x , y ) × I i ( x , y ) .

4. Experiment and Analyses

To compare the performance of the proposed multi-exposure fused method, 12 multi-exposure source image sequences, shot in different scenes with rich details and high brightness contrast, were used. Figure 6 shows two images of different exposures in each image sequence, as used in Ma et al.’s method [14], Merten’s method [15], and other databases. Additionally, more details of each multi-exposure source image sequence are described in Figure 6, including name, spatial resolution and number of source images. The experiment was run on a PC platform with an Intel® Core™ i5 processor with 4-GB memory.
In all of the following experiments, unless otherwise indicated, the parameters of the proposed method were as follows: the patch size, n, was fixed to 9 × 9; the global error was ε = 0.1; the threshold of exposure quality, η , was set to 0.15; another threshold of exposure quality, α, was set to 0.05; and the parameter of the recursive filter was set to the optimal value, as in Liu’s method [12].
Five representative multi-exposure fused methods (i.e., Liu15 method [12], Ma18 method [14], Mertens10 method [15], Vonikakis11 method [16], Li13 method [18]) were employed to make comparisons via subjective and objective evaluation methods. The code of these fusion methods is available on the corresponding websites [32,33,34,35,36], and default parameters were used in the comparisons.

4.1. Subjective Evaluation

The subjective evaluation of the multi-exposure fusion methods is shown in Figure 7, Figure 8, Figure 9 and Figure 10. In Figure 7a, the results of Mertens10 showed nice global contrast and color information, but the details in the brighter regions are lost in the clouds around the sun. Halo artifacts are produced at the boundary between the balloon and the blue sky. Vonikakis11 can preserve the proper details in the brighter regions, but the details of the darker regions are lost at the lawn in the bottom of the balloon in Figure 7b, and halo artifacts are present. The best global contrast can be obtained with Li13 and Ma18, but the halo artifacts in those methods are more prominent. There are almost no halo artifacts in Liu15, but shadow can be observed around the sun. The fused image of the proposed method presented better visual perception and preserved the proper global contrast and color information (e.g., clouds around the sun and the lawn below the balloon), while simultaneously removing the halo artifacts.
Figure 8 gives the results of an indoor scene. The results of Mertens10 showed nice global contrast, but the details, such as colors in the brighter regions, are lost (e.g., the details of the loudspeaker box on the left side). There are halo artifacts on the right side of the flowerpot. Additionally, the color is distorted, as shown in Figure 8e. Vonikakis11 also preserved the rich details in the brighter regions, but lost details in the darker regions (e.g., details of the outdoor scene in Figure 8b). Halo artifacts are also observed on the right side of the flowerpot. Halo artifacts in Li13, Liu15 and Ma18 are evident too. The proposed method had better global contrast and color information (e.g., vivid color in the loudspeaker box in Figure 8f), and overcame the halo artifacts present in the images produced via other methods.
The dynamic range of the scene in Figure 9 is not as wide as those in Figure 7 and Figure 8. However, it can also be seen that, in the Vonikakis11, Li13 and Ma18 method images, halo artifacts are observed near the doorframes in the middle of the fused images. There is also nice global contrast in the Mertens10 image, but the details, such as the background in the brighter regions, are lost (e.g., the details around the tree in the middle of the fused images). The results of Liu15 preserve the proper details in the brighter regions, but its global contrast is very limited. By contrast, the halo artifacts are removed with the proposed method, and proper global contrast and better visual perception are achieved.
As can be clearly seen in Figure 10, there are halo artifacts on the right side of the palace and around the head of the performer, except with Liu15 and the proposed method. Although the halo artifacts are removed by Liu15, the color in the sky is distorted and became pale. However, the proposed method not only maintained better color information of the sky, but also removed the halo artifacts in the fused image.
Except for the four multi-exposure fused images in Figure 7, Figure 8, Figure 9 and Figure 10, the fused images with respect to the other eight multi-exposure source images in Figure 6, obtained with the proposed method, are shown in Figure 11 owing to space limitations.

4.2. Objective Evaluation

In this subsection, two objective metrics are applied to evaluate the performance of the different fused methods: QAB/F [37], MEF-SSIMc [38], and the distortion identification-based image verity and integrity evaluation (DIIVINE) [39], which is a no-reference image evaluation metric.

4.2.1. Evaluation Using QAB/F

The fusion metric, QAB/F, which is widely used in the fused image evaluation [12,24], can be extended to evaluate the image fused with three or more source images. This metric is mainly used to analyze the edge information of the fused images. The more edge information remaining, the larger the QAB/F value is. The extended version of QAB/F for multi-exposure fusion is defined as:
Q A B / F = i = 1 N x = 1 H y = 1 W w I i ( x , y ) Q I i I F ( x , y ) i = 1 N x = 1 H y = 1 W w I i ( x , y ) ,
where I i ( x , y ) denotes the fused image to be evaluated, Q I i I F ( x , y ) = Q g I i I F ( x , y ) Q a I i I F ( x , y ) , Q g I i I F ( x , y ) and Q a I i I F ( x , y ) are the edge intensity and orientation preservation values of the pixel at position ( x , y ) , respectively, and w I i ( x , y ) is the gradient magnitude of I i ( x , y ) , which is used as the weighting factor of Q I i I F ( x , y ) .
The performance of the different methods on the 12 image sequences using QAB/F is listed in Table 1, in which the top two values are shown in bold. It can be seen from Table 1 that, in the fusion metric, QAB/F, the performance of the proposed method and Liu15 was obviously superior to the others.

4.2.2. The Structural Similarity Index (MEF-SSIMc)

The structural similarity index, MEF-SSIMc, represents the structure correlation between fused images and source images. The larger the MEF-SSIMc value, the closer the structure correlation. The formula of MEF-SSIMc is defined as:
Q ( { X k } , Y ) = 1 M j = 1 M S ( { x k ( j ) } , y ( j ) ) S ( { x k } , y ) = 2 σ x ¯ y + C 2 δ x ¯ 2 + δ y 2 + C 2 ,
where { x k } = { x k | 1 k K } denotes the set of co-located image patches in the source image sequence, y is the fused image. δ x ¯ 2 ,   δ y 2 , and σ x ¯ y denote the local variances of the source image x ¯ and the fused image y , and the local covariance between x ¯ and y , respectively. C 2 is a small positive stabilizing constant, j is the spatial patch index and M is the total number of patches.
The performance of the different methods on the 12 image sequences, using MEF-SSIMc, is listed in Table 2, where the top two values are shown in bold. According to the fusion metric, MEF-SSIMc, the performance of the proposed method was better than the other three compared methods and near to the Ma18 method. However, it was not as good as Mertens10, which uses contrast, saturation, and exposure quality to evaluate the source image exposure quality. However, this method may also lead to loss of details in the brighter regions, as shown in Figure 8a.

4.2.3. Evaluation Using DIIVINE

The multi-exposure fused image should not only retain more image information from the multi-exposure source images, but it should also match the human visual perception as closely as possible. Therefore, the non-reference image metric, DIIVINE, was introduced, based on perceived quality, into the fusion metric. DIIVINE evaluates the image quality through 88-dimensional image features, which are based on natural scene statistics. Meanwhile, DIIVINE is also capable of assessing the distorted image quality across multiple distortion categories. The larger the DIIVINE value, the more comfortable it is to human visual perception.
The performance of different methods on the 12 image sequences using DIIVINE is listed in Table 3. It can be seen that, in the fusion metric DIIVINE, the performance of the proposed method was superior to the other three comparative methods and near to the Liu15 method. However, it was not as good as Vonikakis11, which uses illumination estimation to evaluate the source image exposure quality. However, this method may also lead to loss of details in the blacker regions, as shown in Figure 9b.

4.3. Analysis of the Free Parameter

In this subsection, the influences of different parameters to objective fusion performances are presented and analyzed. The fusion performance was evaluated by the values of three fusion quality metrics, i.e., QAB/F, MEF-SSIMc and DIIVINE. When analyzing the influence of the patch size n, other parameters were set to ε = 0.1, α = 0.05 and η = 0.15. Then, when analyzing the influence of ɛ, other parameters were set to n = 9, α = 0.05 and η = 0.15. Next, α and η were analyzed in the same way. As shown in Figure 12, with the increase of n, the value of QAB/F and MEF-SSIMc increased slowly, and the value of DIIVINE showed almost no fluctuation. Furthermore, with the decrease of ε, the value of QAB/F and MEF-SSIMc changed a little, while the value of DIIVINE has improved. When changing with η and α, the value of QAB/F and MEF-SSIMc was worse if η and α were too large or too small, and the value of DIIVINE showed only a slight concave change.
To better understand the proposed method, a set of comparative experiments were carried out to analyze the effect of the recursive filter in the fusion process. The experimental results show that the recursive filter can effectively remove the noise of the initial weight map and optimize the fusion result, as shown in Figure 13.

4.4. Computational Efficiency

In this subsection, the computational efficiency of the above six fusion methods, that is, the execution time of the corresponding methods in the computer, is compared, as listed in Table 4. Because of the high computational complexity of the sparse representation model, the proposed method spent most of its time on calculating the local gradient sparse descriptor of the source images. However, in the above six fusion methods, the proposed method had great efficiency improvement, compared to Ma18.

5. Conclusions

In this paper, we first analyzed the cause of halo artifacts in multi-exposure fused images. Then, we proposed a new multi-exposure fusion method based on sparse representation of gradient features. The sparse representation model was used to obtain the local gradient sparse descriptor, which can not only remove the halo artifacts, but can also extract image details effectively. The experiment was assessed through both a subjective and objective evaluation of 12 multi-exposure image sequences. Experimental results show that the proposed method was able to remove the halo artifacts and obtain a higher quality fused image, presenting a more vivid natural scene. In the future, we plan to assess whether the proposed method could be applied to multi-exposure image fusion in dynamic scenes.

Author Contributions

H.S. and G.J. designed the algorithm and wrote the source code. They together wrote the manuscript. M.Y. provided suggestions on the algorithm and revised the entire manuscript. Y.S., H.J., Z.P. and F.C. provided suggestions on the experiments.

Funding

The work was supported by the Natural Science Foundation of China (61671258, 61871247, 61771269, 61620106012), the Research Fund of Zhejiang Education Department (Y201635391), K.C. Wong Magna Fund of Ningbo University.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Choi, I.; Baek, S.H.; Kim, M.H. Reconstructing Interlaced High-dynamic-range Video using Joint Learning. IEEE Trans. Image Process. 2017, 26, 5353–5366. [Google Scholar] [CrossRef] [PubMed]
  2. Nam, G.; Lee, H.; Oh, S.; Kim, M.H. Measuring Color Defects in Flat Panel Displays Using HDR Imaging and Appearance Modeling. IEEE Trans. Instrum. Meas. 2016, 65, 297–304. [Google Scholar] [CrossRef]
  3. Cauwerts, C.; Piderit, M.B. Application of High-Dynamic Range Imaging Techniques in Architecture: A Step toward High-Quality Daylit Interiors. J. Imaging 2018, 4, 19. [Google Scholar] [CrossRef]
  4. Artusi, A.; Richter, T.; Ebrahimi, T.; Mantiuk, R.K. High Dynamic Range Imaging Technology. IEEE Signal Process. Mag. 2017, 34, 165–172. [Google Scholar] [CrossRef]
  5. Richard, S. High dynamic range imaging. Opt. Eng. 2016, 52, 913–916. [Google Scholar]
  6. Kalantari, N.K.; Ramamoorthi, R. Deep High Dynamic Range Imaging of Dynamic Scenes. ACM Trans. Gr. 2017, 36, 1–12. [Google Scholar] [CrossRef]
  7. Huo, Y.; Zhang, X. Single image-based HDR image generation with camera response function estimation. Image Process. IET 2017, 11, 1317–1324. [Google Scholar] [CrossRef]
  8. Přibyl, B.; Chalmers, A.; Zemčík, P.; Hooberman, L.; Čadík, M. Evaluation of Feature Point Detection in High Dynamic Range Imagery. J. Vis. Commun. Image Represent. 2016, 38, 141–160. [Google Scholar] [CrossRef]
  9. Ji, W.; Park, R.; Chang, S. Local tone mapping using the K-means algorithm and automatic gamma setting. IEEE Trans. Consum. Electron. 2011, 57, 209–217. [Google Scholar]
  10. Khan, I.R.; Rahardja, S.; Khan, M.M.; Movania, M.M.; Abed, F. A tone-mapping technique based on histogram using a sensitivity model of the human visual system. IEEE Trans. Ind. Electron. 2018, 65, 3469–3479. [Google Scholar] [CrossRef]
  11. Eilertsen, G.; Mantiuk, R.K.; Unger, J. A Comparative Review of Tone Mapping Algorithms for High Dynamic Range Video. Comput. Gr. Forum 2017, 36, 565–592. [Google Scholar] [CrossRef]
  12. Liu, Y.; Wang, Z. Dense SIFT for Ghost-free Multi-exposure Fusion. J. Vis. Commun. Image Represent. 2015, 31, 208–224. [Google Scholar] [CrossRef]
  13. Ma, K.; Li, H.; Yong, H.; Wang, Z.; Meng, D.; Zhang, L. Robust Multi-Exposure Image Fusion: A Structural Patch Decomposition Approach. IEEE Trans. Image Process. 2017, 26, 2519–2532. [Google Scholar] [CrossRef] [PubMed]
  14. Ma, K.; Duanmu, Z.; Yeganeh, H.; Wang, Z. Multi-Exposure Image Fusion by Optimizing A Structural Similarity Index. IEEE Trans. Comput. Imaging 2018, 4, 60–72. [Google Scholar] [CrossRef]
  15. Mertens, T.; Kautz, J.; van Reeth, F. Exposure Fusion: A Simple and Practical Alternative to High Dynamic Range Photography. Comput. Gr. Forum 2010, 28, 161–171. [Google Scholar] [CrossRef]
  16. Vonikakis, V.; Bouzos, O.; Andreadis, I. Multi-exposure Image Fusion Based on Illumination Estimation. In Proceedings of the SIPA 2011, Crete, Greece, 22–24 June 2010; pp. 135–142. [Google Scholar]
  17. Zhang, W.; Cham, W.K. Gradient-directed Multi-exposure Composition. IEEE Trans. Image Process. 2012, 21, 2318–2323. [Google Scholar] [CrossRef] [PubMed]
  18. Li, S.; Kang, X.; Hu, J. Image Fusion with Guided Filtering. IEEE Trans. Image Process. 2013, 22, 28–64. [Google Scholar]
  19. Shen, J.; Zhao, Y.; Yan, S.; Li, X. Exposure Fusion using Boosting Laplacian Pyramid. IEEE Trans. Cybern. 2014, 44, 1579–1590. [Google Scholar] [CrossRef] [PubMed]
  20. Prabhakar, K.; Srikar, V.; Babu, R. DeepFuse: A Deep Unsupervised Approach for Exposure Fusion with Extreme Exposure Image. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4724–4732. [Google Scholar]
  21. Aharon, M.; Elad, M.; Bruckstein, A. K-SVD: An Algorithm for Designing Over-complete Dictionaries for Sparse Representation. IEEE Trans. Signal Process. 2006, 54, 4311–4322. [Google Scholar] [CrossRef]
  22. He, Z.; Yi, S.; Cheung, Y.M.; You, X.; Tang, Y.Y. Robust object tracking via key patch sparse representation. IEEE Trans. Cybern. 2017, 47, 354–364. [Google Scholar] [CrossRef] [PubMed]
  23. Qin, H.; Han, J.; Yan, X.; Zeng, Q.; Zhou, H.; Li, J.; Chen, Z. Infrared small moving target detection using sparse representation-based image decomposition. Infrared Phys. Technol. 2016, 76, 148–156. [Google Scholar] [CrossRef]
  24. Yang, B.; Li, S. Multi-focus Image Fusion and Restoration with Sparse Representation. IEEE Trans. Instrum. Meas. 2010, 59, 884–892. [Google Scholar] [CrossRef]
  25. Liu, Y.; Liu, S.; Wang, Z. A General Framework for Image Fusion Based on Multi-scale Transform and Sparse Representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  26. Bruckstein, A.M.; Donoho, D.L.; Elad, M. From Sparse Solutions of Systems of Equations to Sparse Modeling of Signals and Images. SIAM Rev. 2009, 51, 34–81. [Google Scholar] [CrossRef] [Green Version]
  27. Elad, M.; Figueiredo, M.A.T.; Ma, Y. On the Role of Sparse and Redundant Representations in Image Processing. Proc. IEEE 2010, 98, 972–982. [Google Scholar] [CrossRef] [Green Version]
  28. Cheng, J.; Liu, H.; Liu, T.; Wang, F.; Li, H. Remote Sensing Image Fusion via Wavelet Transform and Sparse Representation. ISPRS J. Photogramm. Remote Sens. 2015, 104, 158–173. [Google Scholar] [CrossRef]
  29. Cai, J.; Cheng, Q.; Peng, M.; Song, Y. Fusion of infrared and visible images based on nonsubsampled contourlet transform and sparse K-SVD dictionary learning. Infrared Phys. Technol. 2017, 82, 85–95. [Google Scholar] [CrossRef]
  30. Liu, Y.; Wang, Z. Simultaneous image fusion and denoising with adaptive sparse representation. Image Process. Iet 2014, 9, 347–357. [Google Scholar] [CrossRef]
  31. Gastal, E.S.L.; Oliveira, M.M. Domain transform for edge-aware image and video processing. ACM Trans. Gr. 2011, 30, 69–80. [Google Scholar] [CrossRef]
  32. T. Mertens Software. Available online: http://jankautz.com/ (accessed on 7 January 2018).
  33. S. Li Software. Available online: http://xudongkang.weebly.com/ (accessed on 7 January 2018).
  34. V. Vonikakis Software. Available online: http://vonikakis.com/ (accessed on 12 March 2018).
  35. Y. Liu Software. Available online: http://www.escience.cn/people/liuyu1/Codes.html (accessed on 7 January 2018).
  36. K. Ma Software. Available online: https://ece.uwaterloo.ca/~k29ma/ (accessed on 5 April 2018).
  37. Xydeas, C.S.; Petrovic, V.S. Objective Image Fusion Performance Measure. Electron. Lett. 2000, 36, 308–309. [Google Scholar] [CrossRef]
  38. Ma, K.; Zeng, K.; Wang, Z. Perceptual Quality Assessment for Multi-Exposure Image Fusion. IEEE Trans. Image Process. 2015, 20, 3345–3356. [Google Scholar] [CrossRef] [PubMed]
  39. Moorthy, A.K.; Bovik, A.C. Blind Image Quality Assessment: From Natural Scene Statistics to Perceptual Quality. IEEE Trans. Image Process. 2011, 20, 3350–3364. [Google Scholar] [CrossRef] [PubMed]
Figure 1. An example of multi-exposure image fusion. (a) The fused image by Li et al. [18], and (b) the fused image by the proposed method.
Figure 1. An example of multi-exposure image fusion. (a) The fused image by Li et al. [18], and (b) the fused image by the proposed method.
Applsci 08 01543 g001
Figure 2. Local luminance gradient comparison of multi-exposure source images: (ae) multi-exposure source images; (fj) luminance gradient maps; and (ko) luminance gradient amplitude of the red lines in (fj).
Figure 2. Local luminance gradient comparison of multi-exposure source images: (ae) multi-exposure source images; (fj) luminance gradient maps; and (ko) luminance gradient amplitude of the red lines in (fj).
Applsci 08 01543 g002
Figure 3. Schematic diagram of the proposed multi-exposure.
Figure 3. Schematic diagram of the proposed multi-exposure.
Applsci 08 01543 g003
Figure 4. Dictionary learning in the gradient domain: (a) Clear images of indoors and outdoors; (b) Luminance gradient map; (c) Lexicographical vector; (d) Over-complete dictionary.
Figure 4. Dictionary learning in the gradient domain: (a) Clear images of indoors and outdoors; (b) Luminance gradient map; (c) Lexicographical vector; (d) Over-complete dictionary.
Applsci 08 01543 g004
Figure 5. Local gradient sparse descriptor acquisition: (a) Input image; (b) Input luminance gradient map; (c) Luminance gradient patch; (d) Local gradient sparse descriptor‖S1.
Figure 5. Local gradient sparse descriptor acquisition: (a) Input image; (b) Input luminance gradient map; (c) Luminance gradient patch; (d) Local gradient sparse descriptor‖S1.
Applsci 08 01543 g005
Figure 6. Multi-exposure source image sequences used in the experiments: (a) “balloons”, 512 × 339 × 9 (the sequence contains a total of nine images, and the spatial resolution of each is 512 × 339); (b) “recorder”, 1024 × 768 × 13; (c) “Belgium House”, 512 × 384 × 9; (d) “mask”, 1200 × 800 × 3; (e) “StudioEV”, 1200 × 800 × 5; (f) “landscape”, 512 × 341 × 3; (g) “lighthouse”, 512 × 340 × 3; (h) “Kluki”, 512 × 341 × 3; (i) “desk”, 640 × 427 × 7; (j) “farm house”, 512 × 341 × 3; (k) “office”, 512 × 340 × 6; and (l) “Madison Capitol”, 512 × 384 × 30.
Figure 6. Multi-exposure source image sequences used in the experiments: (a) “balloons”, 512 × 339 × 9 (the sequence contains a total of nine images, and the spatial resolution of each is 512 × 339); (b) “recorder”, 1024 × 768 × 13; (c) “Belgium House”, 512 × 384 × 9; (d) “mask”, 1200 × 800 × 3; (e) “StudioEV”, 1200 × 800 × 5; (f) “landscape”, 512 × 341 × 3; (g) “lighthouse”, 512 × 340 × 3; (h) “Kluki”, 512 × 341 × 3; (i) “desk”, 640 × 427 × 7; (j) “farm house”, 512 × 341 × 3; (k) “office”, 512 × 340 × 6; and (l) “Madison Capitol”, 512 × 384 × 30.
Applsci 08 01543 g006
Figure 7. Performance comparison of different methods on the “balloons” image sequence. (a) Mertens10 [15]; (b) Vonikakis11 [16]; (c) Li13 [18]; (d) Liu15 [12]; (e) Ma18 [14]; and (f) the proposed method.
Figure 7. Performance comparison of different methods on the “balloons” image sequence. (a) Mertens10 [15]; (b) Vonikakis11 [16]; (c) Li13 [18]; (d) Liu15 [12]; (e) Ma18 [14]; and (f) the proposed method.
Applsci 08 01543 g007
Figure 8. Performance comparison of different methods on the “recorder” image sequence. (a) Mertens10 [15]; (b) Vonikakis11 [16]; (c) Li13 [18]; (d) Liu15 [12]; (e) Ma18 [14]; and (f) the proposed method.
Figure 8. Performance comparison of different methods on the “recorder” image sequence. (a) Mertens10 [15]; (b) Vonikakis11 [16]; (c) Li13 [18]; (d) Liu15 [12]; (e) Ma18 [14]; and (f) the proposed method.
Applsci 08 01543 g008
Figure 9. Performance comparison of different methods on the “Belgium House” image sequence. (a) Mertens10 [15]; (b) Vonikakis11 [16]; (c) Li13 [18]; (d) Liu15 [12]; (e) Ma18 [14]; and (f) the proposed method.
Figure 9. Performance comparison of different methods on the “Belgium House” image sequence. (a) Mertens10 [15]; (b) Vonikakis11 [16]; (c) Li13 [18]; (d) Liu15 [12]; (e) Ma18 [14]; and (f) the proposed method.
Applsci 08 01543 g009
Figure 10. Performance comparison of different methods on the “Mask” image sequence. (a) Mertens10 [15]; (b) Vonikakis11 [16]; (c) Li13 [18]; (d) Liu15 [12]; (e) Ma18 [14]; and (f) the proposed method.
Figure 10. Performance comparison of different methods on the “Mask” image sequence. (a) Mertens10 [15]; (b) Vonikakis11 [16]; (c) Li13 [18]; (d) Liu15 [12]; (e) Ma18 [14]; and (f) the proposed method.
Applsci 08 01543 g010
Figure 11. Other fused images of the proposed method. (a) Studio EV; (b) Landscape; (c) Lighthouse; (d) Kluki; (e)desk; (f) Farmhouse; (g) Office; and (h) Madison Capitol.
Figure 11. Other fused images of the proposed method. (a) Studio EV; (b) Landscape; (c) Lighthouse; (d) Kluki; (e)desk; (f) Farmhouse; (g) Office; and (h) Madison Capitol.
Applsci 08 01543 g011
Figure 12. Performance of the proposed method with different parameters, i.e., (a) QAB/F ~ n, (b) MEF-SSIMc ~ n, (c) DIIVINE ~ n, (d) QAB/F ~ ε, (e) MEF-SSIMc ~ ε, (f) DIIVINE ~ ε, (g) QAB/F ~ η, (h) MEF-SSIMc ~ η, (i) DIIVINE ~ η, (j) QAB/F ~ α, (k) MEF-SSIMc ~ α, (l) DIIVINE ~ α.
Figure 12. Performance of the proposed method with different parameters, i.e., (a) QAB/F ~ n, (b) MEF-SSIMc ~ n, (c) DIIVINE ~ n, (d) QAB/F ~ ε, (e) MEF-SSIMc ~ ε, (f) DIIVINE ~ ε, (g) QAB/F ~ η, (h) MEF-SSIMc ~ η, (i) DIIVINE ~ η, (j) QAB/F ~ α, (k) MEF-SSIMc ~ α, (l) DIIVINE ~ α.
Applsci 08 01543 g012
Figure 13. Performance comparison of fusion results between without recursive filter and with recursive filter. (ac) multi-exposure source images; (df) weight map without recursive filter; (gi) weight map with recursive filter; (j) fused image without recursive filter; and (k) fused image with recursive filter.
Figure 13. Performance comparison of fusion results between without recursive filter and with recursive filter. (ac) multi-exposure source images; (df) weight map without recursive filter; (gi) weight map with recursive filter; (j) fused image without recursive filter; and (k) fused image with recursive filter.
Applsci 08 01543 g013
Table 1. Performance of different methods on the 12 image sequences, using Q A B / F [37]. Top two values are shown in bold.
Table 1. Performance of different methods on the 12 image sequences, using Q A B / F [37]. Top two values are shown in bold.
ImageMethod
Mertens10 [15]Vonikakis11 [16]Li13 [18]Liu15 [12]Ma18 [14]The Proposed Method
balloons0.62100.54930.59400.63330.61370.6340
recorder0.67840.44830.67200.69700.64790.6923
Belgium0.65000.55630.66500.66080.64470.6724
mask0.74840.72010.74510.76460.73960.7636
Studio EV0.67480.62970.68120.68700.67120.6802
landscape0.75300.68050.73070.78740.74960.7749
lighthouse0.70370.68600.68230.73260.70120.7245
Kluki0.76040.75580.75390.78150.74350.7883
desk0.61880.53240.61790.62010.60880.6314
farmhouse0.73340.74090.74860.78550.74110.7758
office0.75000.74490.73310.76650.72680.7727
Madison0.72890.48840.71540.73160.69340.7389
Average0.70170.62770.69490.72070.69010.7208
Table 2. Performance of different methods on the 12 image sequences, using the structural similarity index (MEF-SSIMc) [38]. Top two values are shown in bold.
Table 2. Performance of different methods on the 12 image sequences, using the structural similarity index (MEF-SSIMc) [38]. Top two values are shown in bold.
ImageMethod
Mertens10 [15]Vonikakis11 [16]Li13 [18]Liu15 [12]Ma18 [14]The Proposed Method
balloons0.96940.88870.94760.92390.96190.9527
recorder0.97380.81280.93930.94410.94690.9567
Belgium House0.97090.83400.96390.90530.96600.9513
mask0.98540.96090.96700.96630.98620.9863
StudioEV0.97520.92560.95440.93970.95280.9530
landscape0.97620.95200.94210.97990.97750.9777
lighthouse0.98020.95470.95010.96810.95470.9690
Kluki0.98010.95600.94820.95500.96360.9638
desk0.94970.84210.92190.89700.92450.9227
farmhouse0.98100.93980.96470.96160.98040.9633
office0.98450.96090.96740.96630.98540.9786
Madison capitol0.97730.94330.94790.91750.96190.9472
Average0.97800.91420.95120.94370.96350.9602
Table 3. Performance of different methods on the 12 image sequences, using the distortion identification-based image verity and integrity evaluation (DIIVINE) [39]. Top two values are shown in bold.
Table 3. Performance of different methods on the 12 image sequences, using the distortion identification-based image verity and integrity evaluation (DIIVINE) [39]. Top two values are shown in bold.
ImageMethods
Mertens10 [15]Vonikakis11 [16]Li13 [17]Liu15 [12]Ma18 [14]The Proposed Method
balloons21.69029.39820.77325.04520.12625.593
recorder17.58136.83616.89725.91713.66625.491
Belgium24.09933.06523.30830.98721.11330.946
mask17.13623.30915.98519.98215.70219.219
Studio EV33.39137.52631.92636.21129.87837.892
landscape13.8439.85913.52320.89313.83716.623
lighthouse17.75224.92717.06821.06815.04520.761
Kluki6.36326.41536.21712.4435.4809.122
desk34.46839.93633.86339.39731.58039.907
farmhouse17.33421.60416.95421.10415.97622.142
office25.07731.47924.85030.80923.50729.931
Madison10.49022.56112.07614.70914.76115.102
Average19.93526.41019.45324.88018.38924.395
Table 4. Computational efficiency of different fused methods on the 12 image sequences (s).
Table 4. Computational efficiency of different fused methods on the 12 image sequences (s).
ImagesMethods
Mertens10 [15]Vonikakis11 [16]Li13 [17]Liu15 [12]Ma18 [14]The Proposed Method
balloons2.7294.8703.59515.720542.9170.95
recorder8.97023.74620.795109.3933648.04682.64
Belgium2.8755.1233.97116.913618.89184.52
mask4.4449.5177.85421.1722891.22227.85
Studio EV4.20712.59910.08829.6073076.38214.51
landscape0.8801.5281.0903.349466.0726.63
lighthouse0.8701.5581.1143.140587.4238.36
Kluki0.8681.5071.1593.371467.6961.65
desk2.3795.1344.29015.4561178.81116.54
farmhouse0.8851.4641.4063.392413.8778.03
office1.6232.6222.2928.320405.25112.92
Madison7.94712.93113.656125.060918.81789.35

Share and Cite

MDPI and ACS Style

Shao, H.; Jiang, G.; Yu, M.; Song, Y.; Jiang, H.; Peng, Z.; Chen, F. Halo-Free Multi-Exposure Image Fusion Based on Sparse Representation of Gradient Features. Appl. Sci. 2018, 8, 1543. https://doi.org/10.3390/app8091543

AMA Style

Shao H, Jiang G, Yu M, Song Y, Jiang H, Peng Z, Chen F. Halo-Free Multi-Exposure Image Fusion Based on Sparse Representation of Gradient Features. Applied Sciences. 2018; 8(9):1543. https://doi.org/10.3390/app8091543

Chicago/Turabian Style

Shao, Hua, Gangyi Jiang, Mei Yu, Yang Song, Hao Jiang, Zongju Peng, and Feng Chen. 2018. "Halo-Free Multi-Exposure Image Fusion Based on Sparse Representation of Gradient Features" Applied Sciences 8, no. 9: 1543. https://doi.org/10.3390/app8091543

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop