Next Article in Journal
Logistic Model to Support Service Modularity for the Promotion of Reusability in a Web Objects-Enabled IoT Environment
Previous Article in Journal
Design and Analysis of an Efficient Energy Algorithm in Wireless Social Sensor Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Dehaze Algorithm for Degraded Image of CMOS Image Sensors

1
College of Aeronautics and Astronautics Engineering, Air Force Engineering University, Xi’an 710038, China
2
Foundation Department, Air Force Engineering University, Xi’an 710051, China
3
Information and Navigation College, Air Force Engineering University, Xi’an 710077, China
*
Author to whom correspondence should be addressed.
Sensors 2017, 17(10), 2175; https://doi.org/10.3390/s17102175
Submission received: 29 August 2017 / Accepted: 18 September 2017 / Published: 22 September 2017
(This article belongs to the Section Physical Sensors)

Abstract

:
The CMOS (Complementary Metal-Oxide-Semiconductor) is a new type of solid image sensor device widely used in object tracking, object recognition, intelligent navigation fields, and so on. However, images captured by outdoor CMOS sensor devices are usually affected by suspended atmospheric particles (such as haze), causing a reduction in image contrast, color distortion problems, and so on. In view of this, we propose a novel dehazing approach based on a local consistent Markov random field (MRF) framework. The neighboring clique in traditional MRF is extended to the non-neighboring clique, which is defined on local consistent blocks based on two clues, where both the atmospheric light and transmission map satisfy the character of local consistency. In this framework, our model can strengthen the restriction of the whole image while incorporating more sophisticated statistical priors, resulting in more expressive power of modeling, thus, solving inadequate detail recovery effectively and alleviating color distortion. Moreover, the local consistent MRF framework can obtain details while maintaining better results for dehazing, which effectively improves the image quality captured by the CMOS image sensor. Experimental results verified that the method proposed has the combined advantages of detail recovery and color preservation.

1. Introduction

Complementary Metal-Oxide-Semiconductor (CMOS) [1] image sensors, given their high integration, low power consumption, small size, low cost and other advantages, have been widely used in environmental monitoring, intelligent navigation [2], outdoor tracking, and so on. The outdoor collection of visible light images will inevitably encounter cloudy weather (especially in recent years, haze weather), resulting in the reduction of visibility, contrast and sharpness of the scene, therefore making monitors, intelligent navigation, and outdoor object recognition difficult. A haze image is generated as shown in Figure 1.
As one of the most commonly used probabilistic graphical models, the Markov random field (MRF) provides not only an effective framework for modeling the statistical prior of natural images [3,4], but also a means of making inferences about images [5]. These inferences, widely investigated and applied to solve problems such as image reconstruction [6], segmentation [7], denoising [8,9,10], inpainting [11], concern underlying image and scene structure. In a CMOS for a single outdoor image, MRF has been informally used to model the haze scene and improve the dehaze result in some works. Fattal [12] assumed that the surface Lambertian shading factor and the scene transmission were locally independent to separate the haze from the scene, and then used a Gaussian MRF to smooth the transmission values. This was physically valid, but the assumption was too strong for a variety of images, and thus tended to under-estimate the haze thickness in practice. Based on the two observations that (1) images with enhanced visibility have more contrast than hazy images, and (2) air-light tends to be smooth, an interesting single image haze removal algorithm was proposed by maximizing the local contrast of the restored image in the framework of MRF in Reference [13]. The results were visually compelling; however, the method might not be physically valid. Nishino [14] introduced a novel Bayesian probabilistic method that modeled the image with a factorial MRF where the scene albedo and depth were two statistically independent latent layers and jointly estimated. Most details could be recovered from the haze image by this approach, but the result often suffered from oversaturation. Caraffa [15] proposed a novel MRF model for the single image dehazing problem, which could easily be refined to obtain better results on road images using the planar constraint; however, this method only outperformed the state-of-the-art in single road scene image dehazing.
As above-mentioned, these MRF-based techniques focus on traditional first-order MRF to model a single hazy image in References [12,13,14,15]. However, in a CMOS, the traditional first-order MRF only models the statistical dependency between the neighboring pixels and constrains image on the local prior information, not the whole image. To better capture the image structures, higher-order MRF extends its models to larger neighborhoods. Compared with first-order MRF, higher-order MRF incorporates more sophisticated qualitative and statistical priors, resulting in stronger modeling power. However, it is harder to minimize their energy functions and it is much more complex to estimate their parameters due to the explosive growth of their number. Therefore, higher-order MRF applications in image processing are still seldom used for the limitation of energy function optimization algorithm.
To enhance the first-order MRF constraint ability of the whole image and avoid too much complexity by the higher-order MRF algorithm, we propose a novel MRF dehazing method based on local consistent patches. The proposed method is unlike the traditional first-order MRF, which requires second-order clique potentials to constrain the neighboring relationships, and nor like the higher-order MRF that needs higher-order potentials defined on the overlapping cliques. There are two basic observations: the depth change of a hazy image is usually gradual, and the correct depth values in neighbor tend to be the same, hence, regardless of their scattering coefficient, the medium transmission map can be considered as a constant in a small patch; atmospheric light, tending to be smooth, mainly depends on the distance of scene and can be roughly considered as a constant in local. Relying on these two observations, we developed a linear model at small patches between the hazy image and the restored image, and then proposed a cost function in local consistent MRF. Our major contribution in this paper is that a local consistent range MRF was constructed by extending the neighboring range clique around the pixels to a non-neighboring range clique, where several similar patches in an effective window exist around each pixel. In this framework, both of the neighboring pixel structures and the correlations among the similar patches can be captured. This local consistent MRF cost function can be efficiently optimized by gradient-based methods. Thus, as our processing is carried out within the effective windows around each pixel, it can be considered as a local filter for a CMOS image sensor.
Very recently, He [16] developed a dehazing method based on local filtering which used a neighboring window-based operation and a local linear model, namely, soft matting. However, this method has an obvious limitation, that is, it may exhibit halo artifacts near some edges. To address this problem, many edge-aware weights were incorporated into the local-based filtering. Interestingly, by explicitly simplifying our cost function with respect to model parameters, we found that the cost function constructed by local consistent MRF had a dramatic relationship with the soft matting. Furthermore, unlike the soft matting, our method had an additional edge-preserving term. So, compared with the soft-matting, the proposed local consistent MRF model avoided the halo artifacts and restored more detailed information. Experimental results also verified its effectiveness.
The rest of this paper is organized as follows. In Section 2, we present the image degradation model due to the presence of haze in the scene, and in Section 3 we construct the local consistent MRF to better restrain the whole image. Considering that both the atmospheric light and medium transmission map satisfied local consistency, we solve the local consistent MRF in Section 4 and report the experiment results in Section 5. In Section 6, we summarize the proposed approach and discuss its limitations.

2. Degradation Model

During inclement weather conditions such as fog, haze, and mist, in images captured by CMOS sensor devices light passing through the scattering medium is attenuated along its original course and distributed to other conditions. This process is commonly modeled mathematically by an atmospheric scattering model, which is widely used in computer vision and image processing [17,18,19,20], and can be expressed as follows:
I ( x ) = J ( x ) t ( x ) + A ( x ) ( 1 t ( x ) )
The first term is the direct attenuation and the second term is the air-light. I(x) is the image capture from the CMOS sensor device; x is the 2D spatial location; and J(x) is the real scene to be recovered. A(x) is the atmospheric light, which describes the ambient light in the scene. Existing methods commonly assume that A(x) is globally constant and it is independent from location x; however, in practice, the variation of the values of A(x) is dependent on the scene depth, and cannot be deemed as a constant in a nutshell. t(x) is the medium transmission map which describes the portion of the light that is not scattered and reaches the camera:
t ( x ) = e β ( x ) d ( x )
where β(x) is the scattering coefficient and is regarded as a constant in a homogeneous atmosphere condition; and d(x) is the distance from the scene point to the camera. The aim of haze removal is to restore the haze-free image J(x) from the hazy image I(x). It is a challenging problem as haze is dependent on the unknown depth information d(x) as seen in Equation (2). In addition, it is ill-posed as the input is only a single hazy image while the components t(x), A(x) and J(x) are all unknown. To restore the haze-free image J(x), both the atmospheric light A(x) and the medium transmission map t(x) need to be estimated. Once A(x) and t(x) are estimated, the dehazed image J(x) is obtained by:
J ( x ) = 1 t ( x ) ( I ( x ) A ( x ) ) + A ( x )
where 0 < t(x) ≤ 1, and we define ax as the 1/t(x) and cx as the A(x). Therefore, in this paper, the atmospheric scattering model can be simplified as:
J ( x ) = a x ( I ( x ) c x ) + c x

3. Local Consistent Markov Random Fields

The energy functions for many commonly used MRF models can be written as a sum of unary and pairwise cliques [21,22]:
E ( x ) = m v ϕ m ( x m ) + ( m , n ε ) ψ m , n ( x m , x n )
where ν corresponds to the location set of all image pixels, and ε is the set of all neighboring pixels of m. The neighboring set is commonly chosen to be 4-neighborhood. The random variable xm denotes the configuration of pixel m of the image. Every possible assignment of the random variable x defines a restored image. The unary potential ϕ m is defined as the cost of a label being assigned to pixel m and can be computed using sophisticated potential functions such as color, texture, location, shape prior, and so on. The pairwise terms ψm,n, mostly considered as smooth terms, are typically defined as an edge feature based on the difference between neighboring pixels. The use of the pairwise terms in the MRF model makes it favor smooth object boundaries. Bearing the risk of undesirable side effects, it improves results in most cases, and restored images obtained by pairwise terms tend to be over-smooth and often do not extract the fine details of the scene. Additionally, pairwise terms modeling the image prior statistics in small image patches do not easily generalize to priors for entire images [23]. This limits their impact in machine vision applications.
To address the over-smooth problem and use this formulation to model the statistics of entire images [24,25], a higher-order method has been proposed. Unlike the conventional MRF model explained in the previous section, the clique potentials of a higher-order Markov random field model are extended by incorporating higher-order terms defined on sets or regions of pixels [3,26,27]. The cost function of this higher-order MRF model can be written as:
E ( x ) = m v ϕ m ( x m ) + ( m , n ε ) ψ m , n ( x m , x n ) + c S ϕ c ( x c )
where S refers to a set of image regions defined on super-pixels, and φc are higher-order terms defined on them. The framework described above is quite flexible and can be used to extract more detail from the restored images. However, the complexity of the algorithm for optimizing the cost function increases linearly with the size of the clique. This hinders wide use of the higher-order MRF model in image processing.

3.1. Basic Definition

Images degraded by haze are normally taken from outdoor natural scenes, therefore, the change in scene depth is usually gradual and the correct depth values generally satisfy the local smooth feature, except for pixels at depth discontinuities, whose number is relatively small. At the same time, we found that within the local pixel block, the same colors had the same scene depth changes. This discovery enabled us to use the image color feature to find the local consistent blocks about the scene depth information. As shown in Figure 2, we defined a local consistent MRF model based on these local consistent blocks.
Unlike the conventional cliques of MRF that are generally defined by the relationship between the neighboring pixels or the super-pixels, we designed the local consistent range of cliques which captured both the neighboring pixels’ structures and the correlations among the local similar patches. To take advantage of local consistent MRF in image dehazing, we first defined the local consistent blocks; and color moment is a kind of simple and effective representation in color feature. Thus, we used the second-order color moment to measure the similarity of the different image blocks in this paper.
In local consistent MRF, the clique of each pixel is composed of the neighboring pixels around the pixel and its top similar blocks, searched by block matching with squared error in all color channels as block similarity in an effective neighborhood around the pixel. As shown in Figure 2, the clique of pixel m (shown in the central point) combined its connected pixels (shown in the blue block) and the similar blocks (shown in the red blocks) in an effective neighborhood. Apparently, the local consistent blocks of pixel were adaptively located over the effective regions, and we selected the top six blocks by similarity as the local consistent blocks according to the squared errors.
Based on the above definitions, we denoted the clique of pixel m as F(xm), combined with two separations: 1-dimensional vector F0(xm) and Fk(xm). F0(xm) was defined over pixel and its neighboring pixels called the neighboring block. Fk(xm) was defined across all the top k similarity blocks of the pixel, as shown in Figure 3. The variable xm denoted the labeling of pixel m of the image. Then, F(xm) can be written as:
F ( x m ) = { F 0 ( x m ) , F 1 ( x m ) , , F k ( x m ) }
where the similarity between the neighboring block F0(xm) and the local consistent blocks Fk(xm) can be defined by squared errors, which can be denoted by:
σ k 2 ( x m ) = i = 1 3 ( F 0 ( x m i ) F k ( x m i ) ) ( F 0 ( x m i ) F k ( x m i ) ) T
where x m i is the color label of pixel m in channel i , i { r , g , b } .

3.2. Formulation of Local Consistent MRF Model

We used cliques composed of neighboring pixels around the pixel and its top similar blocks to develop the clique potentials instead of the pairwise and higher-order cliques as this method not only strengthened the constraint of conventional MRF models, but also avoided the complexity. Meanwhile, the cost function of local consistent MRF can be represented as:
E ( x ) = m n ϕ m ( x m ) + m n k ϕ ( F 0 ( x m ) , F k ( x m ) )
where the first term m ν ϕ m ( x m ) = | x m x ˜ m | 2 was used to constrain the relationship between the correct value x m ; and the estimated value x ˜ m in pixel m to ensure that the error was as small as possible. The second term k ϕ ( F 0 ( x m ) , F k ( x m ) ) = k λ F 0 ( x m ) F k ( x m ) 2 was used to describe the difference between the neighboring block and the local consistent blocks in pixel m. λ is a regularization parameter penalizing the difference.

4. The Solution of Local Consistent MRF

The problem described in Section 2 is a totally ill-posed problem; however, there are some clues or observations that can be considered as haze priors to resolve the ambiguity:
  • The input images degraded by haze are normally taken from outdoor natural scenes. Therefore, the scene depth change is usually gradual and the correct depth values of neighboring pixels tend to the same and, hence, the medium transmission map t(x) can be considered as a constant in a small patch, regardless of their scattering coefficient β(x).
  • The value variations of A(x) are dependent on the scene depth, that is, objects with the same depth have the same values of A(x). So, the values of A(x) tend to be the same in local except for the pixels at depth discontinuities, whose number is relatively small.
From the observations mentioned above, we can consider that atmospheric light A(x) and medium transmission map t(x) both satisfy the character of local consistency.

4.1. Construction of Local Consistent MRF Model

According to the clues above, the changes of A(x) and t(x) across the image tended to be smooth for the major pixels in local. Thus, atmospheric light and transmission map were assumed to be constants in one local patch. The image degradation model in Equation (4) was simplified as:
J ( x ) = a x ( I ( x ) c x ) + c x , b x = c x c x a x , x p x J ( x ) = a x I ( x ) + b x
which was assumed to have constant values of a x , b x , and c x in one small local patch p x centered at x . As such, Equation (10) builds up a linear relationship between the transmission map and the input image.
Motivated by the above descriptions, we used the haze-free image J(x) to develop a local consistent MRF model for the estimation of transmission map and atmospheric light. We calculated the potential function of MRF as:
E ( { a x , b x } | p x ) = E ( { J ( x ) } | p x ) = x ϕ x ( p x | { J ( x ) } ) + λ x k ϕ ( F 0 ( J x ) , F k ( J x ) )
where p x is a small patch centered at x′; λ is a regularization parameter penalizing the smoothness term; and k represents the number of consistent neighboring patches selected. We defined the first term, the data term, as:
ϕ x ( p x | { J ( x ) } ) = a x I ( x ) + b x J ˜ ( x ) 2
where J ˜ ( x ) is the initial value of the haze-free image J ( x ) ; and a x are two constants in local patch p x . The second term, the smoothness term, was defined as:
k ϕ ( F 0 ( J x ) , F k ( J x ) ) = k = 1 6 F 0 ( J x ) F k ( J x ) 2 = k = 1 6 i = 1 3 [ ( a x F 0 i ( I x ) + b x ) ( a x F k i ( I x ) + b x ) ] [ ( a x F 0 i ( I x ) + b x ) ( a x F k i ( I x ) + b x ) ] T
In one local patch p x , the cost function E ( a x , b x ) was equivalent to:
E ( a x , b x ) = x p x ( ( a x I ( x ) + b x J ˜ ( x ) ) 2 + λ k = 1 6 i = 1 3 [ ( a x F 0 i ( I x ) + b x ) ( a x F k i ( I x ) + b x ) ] [ ( a x F 0 i ( I x ) + b x ) ( a x F k i ( I x ) + b x ) ] T ) = x p x a x I ( x ) + b x J ˜ ( x ) 2 + N λ a x 2 k = 1 6 σ k 2 ( I x )
where N is the pixel number in local patch p x . We denoted Γ x ( x ) = 1 k = 1 6 ( σ k 2 ( I x ) + ε ) , ε as a small constant and its value is 0.001. Then, Equation (14) can be derived as:
E ( a x , b x ) = x p x a x I ( x ) + b x J ˜ ( x ) 2 + N λ Γ x ( x ) a x 2
Recent work in He [16] referred to a similar cost function as “soft matting”. It is important to note the difference in our approaches. Soft matting is used as a local filter to refine the transmission map. Interestingly, with Equation (15), our cost function based on local consistent MRF had a dramatic relationship with soft matting. In other words, with the local consistent MRF model, we established the dehazing relationship between the method based on random field and the method based on local filter, such as soft matting. However, unlike soft matting, our method has an edge-preserving term and larger weights are assigned to pixels at edges than to pixels in flat areas. Due to the edge-preserving term, the local consistent MRF model could avoid halo artifacts and recover most details from the haze image.
To obtain the variables a x , b x accurately, Equation (15) can be expressed as follows:
( a x , b x ) = a r g m i n E ( a x , b x )
To solve this, we first calculated the partial derivative of E ( a x , b x ) and made them equal to zero:
E ( a x , b x ) a x = 2 x p x ( ( a x I ( x ) + b x J ˜ ( x ) ) I ( x ) ) + 2 N λ Γ x ( x ) a x = 0
E ( a x , b x ) b x = 2 x p x ( a x I ( x ) + b x J ˜ ( x ) ) = 0
According to Equations (17) and (18), the estimation of variables a x , b x were:
a x = 1 N x p x I ( x ) J ˜ ( x ) b x N x p x I ( x ) λ Γ x ( x ) + 1 N x p x I ( x ) J ˜ ( x )
b x = x p x J ˜ ( x ) a x x p x I ( x ) N
Applying Equation (20) to Equation (21), we have:
a x = 1 N x p x I ( x ) J ˜ ( x ) 1 N x p x I ( x ) 1 N x p x J ˜ ( x ) λ Γ x ( x ) + 1 N x p x I ( x ) J ˜ ( x ) 1 N x p x I ( x ) 1 N x p x J ˜ ( x )
b x = 1 N ( x p x J ˜ ( x ) a x x p x I ( x ) )
As mentioned above, transmission map t(x) and atmospheric light A(x) both satisfied the feature of local consistent. Thus, once the values of a x , b x were available, it can be derived that
t ( x ) = 1 a ¯ x , A ( x ) = b ¯ x 1 a ¯ x
where a ¯ x = 1 N x p x a x ,   b ¯ x = 1 N x p x b x . N is the pixel number in patch px centered at x. a ¯ x and b ¯ x present the mean values of a x and b x in patch px.
At this point, the values of the atmospheric light A(x) and the medium transmission map t(x) became available, hence, the dehazed image J(x) could be recovered easily according to Equation (3). As the recovered scene radiance J(x) will suffer from noise when the transmission t(x) is close to zero, we restricted the value of t(x) by a lower bound t0, which was experimentally fixed to 0.1 in Reference [16]. The final function used for restoring dehazed image J(x) can be expressed by:
J ( x ) = 1 m a x ( t ( x ) , t 0 ) ( I ( x ) A ( x ) ) + A ( x )

4.2. Label Candidates and Initialization

As mentioned above, the energy minimization process presented in Equation (15) requires an initial estimate J ˜ ( x ) to start the operation. Theoretically, the choice of initial solution determines the optimization speed, thus we needed to select a close approximate to the solution as an initial estimate to speed up the process. The initial value J ˜ ( x ) cannot be accessed directly, therefore, to obtain the initial value of dehazed image J ˜ ( x ) , we approximately estimated it through the coarse atmospheric light and transmission map with Equation (3).
In this paper, we used the blurred Y ( x ) of the YIQ model [13] as the coarse value of atmospheric light A ˜ ( x ) . The value Y ( x ) and the blurred Y ( x ) , the atmospheric light, are shown in Figure 4.
With Equation (3), the transmission map can be derived from:
l n ( A ( x ) I ( x ) ) = l n ( A ( x ) J ( x ) ) + l n t ( x )
and Equation (25) can be simplified as:
I ˜ ( x ) = J ˜ ( x ) + T ˜ ( x )
We estimated the initial transmission map values as the largest possible transmission map at each pixel. The observed image contained three color channels, and a single transmission map can be obtained to estimate each channel c { r , g , b } . Thus, the largest possible transmission map value occurs when J ˜ ( x ) = 0 , and the corresponding transmission map estimate T ˜ c ( x ) is:
T ˜ c ( x ) = Ι ˜ c ( x )
where c is a specific color channel. Next, we noted that there was only one map estimate in the three transmission map estimates T ˜ c ( x ) , c { r , g , b } is the possible value for all three-color channels. Thus, we set the initial transmission map estimate t ˜ ( x ) to the closest one over all channels:
t ˜ ( x ) = e m a x c { r , g , b } T ˜ c ( x ) = e m a x c { r , g , b } I ˜ c ( x )
In other words, the largest transmission map t ˜ ( x ) was the valid map for all three-color channels. Note that when we maximized the I ˜ c ( x ) , we determined A ( x ) by picking image points corresponding to direct observations of the sky or the brightest scene points in Reference [12].
Figure 5 shows the initial transmission map estimate computed from the hazy image, and illustrates the initial estimate capturing the scene transmission map structure well.
Recently, He [16] proposed an estimation of the transmission map based on the dark channel prior (DCP). The DCP is used as the initial estimate for the soft matting that imposes conventional smoothing on the transmission map estimation. Furthermore, the estimate transmission map is computed in local leading to halo artifacts. The transmission map in this paper was also estimated in local areas with the edge-preserving terms in the proposed local consistent MRF cost function, and our method finally resulted in a finer estimation.
Above all, we obtained the estimated value of the dehazed image J ˜ ( x ) by applying the initial atmospheric light A ˜ ( x ) and transmission map t ˜ ( x ) to Equation (3). The redundant details in the atmospheric light and the transmission map were refined and correctly represented in Equation (16) during the joint minimization.

5. Experimental Results

To verify the effectiveness of the proposed dehazing method, we compared it with other MRF-based and soft-matting-based methods, including Tan [13], Nishino [14], Fattal [12] and He [16] on various hazy images. The experimental results in this paper consist of two parts. Part A discusses the dehazing results through a qualitative comparison of hazy images captured by CMOS sensor devices in the real-world. Part B presents the quantitative comparison of the respective dehazing results of the proposed method with other methods on real-world hazy images and synthetic hazy images. In addition, all the tests were implemented in the Matlab R2014a environment on an Intel Core (T) i7-3770 CPU @3.40 GHz processor with 8.00 GB of RAM running a Windows 7 operation system.

5.1. Qualitative Comparison among Real-World Hazy Images

We have stated that the local consistent MRF plays an important role for edge preservation. The edge-preserving term helps to recover better results. Most of the dehazing algorithms are able to obtain really good results in many hazy images. We used the dehazed images generated by different methods and their corresponding visible edges images to contrast the results markedly.
As shown in Figure 6, Figure 7, Figure 8 and Figure 9, a qualitative comparison of results with the four state-of-the-art dehazing algorithms mentioned in References [12,13,14,16] were measured on different real-world hazy images. Among them, Figure 6, Figure 7 and Figure 8 show the results of our method when compared with the MRF-based methods. Figure 9 shows the comparison results between our method and soft-matting.
As shown in Figure 6, most haze was removed in Tan’s [13] results, and the scene details were restored. However, the results suffered significantly from over-enhancement and did not fully recover the scene colors. For instance, the road region of the second image was dark and the swans in the first image became brown. This was because Tan’s algorithm was based on maximizing the local contrast of the restored image and had the inherent problem of overestimating the dehazed image. The results of Nishino [14] had a similar problem which tended to introduce color distortion in the dehazed image. As observed in Figure 6c, the restored images were over-saturated and distorted, especially in the first image where the sky color was turned to a darker blue and the wheat was not its real color. Furthermore, Nishino’s algorithm also over-enhanced the local contrast. In some results (the third image in Figure 7), Nishino’s results and its corresponding visible edges showed more details than our algorithm, while their visual pleasures were not as good as ours. In all cases, the results of Fattal [12] (shown in Figure 8) all showed color distortion, such as the darker color of the wheat bundles around the mountain in the first image, the brighter green in the middle forest of the second image, and in the third image, the grayish tint to the buildings in the middle and far distance. In contrast, our results were natural and remaining in colors across the entire image. Regarding detail contrast, the results by Fattal [12] lacked details in the far scene, while our algorithm recovered a dehazed image with finer granularity even in areas such as the mountain at the top region of our recovered image in Figure 8f.
In Figure 6, Figure 7 and Figure 8, all of the algorithms compared did not produce halo artifacts. It was clearly observed that our results were slightly sharper than the dehazed images by other haze removal methods, given the main contribution of the edge-preserving term in our local consistent MRF model.
Figure 9 shows the comparisons between the proposed method and soft-matting. Our results were more saturated and have no apparent haze degradation. Observing this comparison, the algorithm in Reference [16] may exhibit halo artifacts near some edges, for instance, the mountain and the plant leaves in the first image. Additionally, the blocking effect existed obviously in the sky of the first and second images. Moreover, He’s method also introduced color distortion in the regions with white objects such as the sky regions in the second and third images where the reason was that the scene brightness was similar to the atmospheric light as the method recovering the transmission map used in Reference [16] ( based on the DCP) was invalid. In addition, the atmospheric light is also an important factor for estimating the transmission in [16]. Therefore, to obtain the correct transmission map, we required an accurate estimation of the atmospheric light. However, the estimating method for atmospheric light in He’s algorithm was the approximate, not its accurate value. Hence, He’s method has its limitations. In contrast, with close initial estimations, our method estimated the atmospheric light and transmission map jointly in the framework of local consistent MRF, and also with the terms of edge-preserving, so our method could avoid the halo artifacts and recovered most of the scene details in the dehazed image.

5.2. Quantitative Comparison

The real-world hazy images dehazing results were compared by entropy and Peak Signal-to-Noise Ratio (PSNR). Entropy describes the image information. PSNR is used to characterize the full degree of the image’s information as well as structure. Note that the higher PSNR and entropy scores imply more excellent restored quality. In Table 1, the PSNR and entropy comparison of Figure 6, Figure 7, Figure 8 and Figure 9 are provided. As seen, our results mostly achieved a higher PSNR and entropy than the other methods. In Table 2, the execution time comparison of Figure 6, Figure 7, Figure 8 and Figure 9 are provided. The experimental results show that our algorithm is slower than that of Tan [13], Nishino [14] and Fattal [12] due to local consistent block, but our algorithm avoids the use of high order energy terms, so that is faster than He [16].
To evaluate the proposed algorithm comprehensively, we used synthetic hazy images with their corresponding known haze-free images to quantify the dehazed images on various dehazing algorithms. Figure 10 shows the comparison on synthetic hazy images, and Figure 11 and Figure 12 show the Mean Square Error (MSE) and Structural Similarity Index Measurement (SSIM) values produced by different algorithms in Figure 10, respectively.
Lower MSE implies a greater similarity between the dehazed image and the referenced ground-truth image. From Figure 11, our algorithm had the lowest value in almost all cases, which meant that our results were closer to the ground-truth images and the dehazing effects more natural.
SSIM evaluates the ability to preserve the structural information of the algorithms where a higher SSIM implies a greater similarity between the dehazed image and the ground-truth image. Figure 12 shows the SSIM comparison results in Figure 10 where it was obvious that the proposed algorithm could preserve the structures.
In Table 3, the execution time comparison of Figure 10 are provided. The experimental results show that our algorithm is slower than that of Caraffa [15] due to local consistent block, but our algorithm avoids the use of high order energy terms, so that is faster than He [16].

6. Conclusions

Due to its small volume and low cost, the CMOS image sensor has attracted increased interest in the last few years and is expected to play a major role in outdoor surveillance, intelligent navigation, object tracking, and so on. Unfortunately, in many cases, the images captured by CMOS are often of low clarity under the effect of fog and haze weather. To ensure the quality of the image captured is more suitable for analysis in various surveillance applications, we proposed a novel framework for a novel local consistent MRF model for robust dehazing based on two clues where both the atmospheric light and the transmission map satisfy the features of local consistency. Accordingly, we created a linear model for the dehazed image with a local consistent transmission map and atmospheric light, and constructed a cost function in local consistent MRF. Interestingly, we found that the cost function had a dramatic relationship with the soft-matting; however, unlike soft-matting, the cost function had an additional edge-preserving term which resulted in the avoidance of halo artifacts and restoring more details. Consequently, the local consistent MRF model proposed could effectively produce haze-free restoration results. Qualitative and quantitative results demonstrated that the proposed local consistent MRF model performed better in hazy image restoration on both real-world images and synthetic data. Nevertheless, the proposed model had some limitations, such as reduced execution time and local artifacts without considering the semantic information of the whole image. Both problems will be studied in our future work.

Acknowledgments

This work was financially supported in part by the National Natural Science Foundation of China (Project No. 61372167).

Author Contributions

Chen Qu conceived and designed the research; Ping Sui and Chen Qu performed the experiments; Du-Yan Bi revisited the paper and supervised the whole process; Ai-Nong Chao and Yun-Fei Wang provided valuable suggestions on CMOS Image sensors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kim, D.; Song, M.; Choe, B.; Kim, S. A Multi-Resolution Mode CMOS Image Sensor with a Novel Two-Step Single-Slope ADC for Intelligent Surveillance Systems. Sensors 2017, 17, 1497. [Google Scholar] [CrossRef] [PubMed]
  2. Sarkar, M.; Bello, D.S.S.; Chris, V.H.; Albert, J.P.T. Biologically inspired autonomous agent navigation using an integrated polarization analyzing CMOS image sensor. Procedia Eng. 2010, 5, 673–676. [Google Scholar] [CrossRef]
  3. Srivastava, A.; Lee, A.B.; Simoncelli, E.P.; Zhu, S.C. On advances in statistical modeling of natural images. J. Math. Imaging Vis. 2003, 18, 17–33. [Google Scholar] [CrossRef]
  4. Lyu, S.; Simoncelli, E.P. Modeling multiscale subbands of photographic images with fields of Gaussian scale mixtures. IEEE Trans. PAMI 2009, 31, 693–706. [Google Scholar] [CrossRef]
  5. Woodford, O.J.; Rother, C.L.; Kolmogorov, V. A Global Perspective on Map Inference for Low-level Vision. In Proceedings of the 2009 International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 2319–2326. [Google Scholar]
  6. Sun, J.; Zheng, N.N.; Shum, H.Y. Stereo matching using belief propagation. IEEE Trans. PAMI 2003, 25, 787–800. [Google Scholar] [CrossRef]
  7. Barbu, A.; Zhu, S.C. Generalizing Swendsen-Wang to sampling arbitrary posterior probabilities. IEEE Trans. PAMI 2005, 27, 1239–1253. [Google Scholar] [CrossRef] [PubMed]
  8. Samucl, K.G.G.; Tappen, M.F. Learning Optimized MAP Estimates in Continuously-valued MRF Models. In Proceedings of the 2009 Computer Vision and Pattern Recognition, Miami Beach, FL, USA, 20–26 June 2009; pp. 477–484. [Google Scholar]
  9. Barbu, A. Training an active random field for real-time image denoising. IEEE Trans. Image Process. 2009, 18, 2451–2461. [Google Scholar] [CrossRef] [PubMed]
  10. Tappen, M.F.; Liu, C.; Adelson, E.H.; Freeman, W.T. Learning Gaussian Conditional Random Fields for Low-level Vision. In Proceedings of the 2007 Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  11. Roth, S.; Black, M.J. Steerable Random Field. In Proceedings of the 2007 International Conference on Computer Vision, Rio de Janeiro, Brazil, 14–20 October 2007; pp. 1–8. [Google Scholar]
  12. Fattal, R. Single image dehazing. ACM Trans. Graph. 2008, 27, 1–9. [Google Scholar] [CrossRef]
  13. Tan, R.T. Visibility in Bad Weather from a Single Image. In Proceedings of the 2008 Computer Vision and Pattern Recognition, Anchorage, AK, USA, 24–26 June 2008; pp. 1–8. [Google Scholar]
  14. Nishino, K.; Kratz, L.; Lombardi, S. Bayesian defogging. Int. J. Comput. Vis. 2012, 98, 263–278. [Google Scholar] [CrossRef]
  15. Caraffa, L.; Tarel, J.P. Markov Random Field Model for Single Image Defogging. In Proceedings of the 2013 IEEE Conference Intelligent Vehicles Symposium, Gold Coast, Australia, 23–26 June 2013; pp. 994–999. [Google Scholar]
  16. He, K.M.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2341–2353. [Google Scholar] [CrossRef] [PubMed]
  17. Zhang, J.; Li, L.; Yang, G.; Zhang, Y.; Sun, J. Local albedo-insensitive single image dehazing. Vis. Comput. 2010, 26, 761–768. [Google Scholar] [CrossRef]
  18. Carr, P.; Hartley, R. Improved Single Image Dehazing Using Geometry. In Proceedings of the 2009 Digital Image Computing Techniques and Applications, Melbourne, Australia, 1–3 December 2009; pp. 103–110. [Google Scholar]
  19. Kopf, J.; Neubert, B.; Chen, B.; Cohen, M.F.; Cohenor, D.; Deussen, O.; Uyttendaele, M.; Lischinski, D. Deep photo: Model-based photograph enhancement and viewing. ACM Trans. Graph. 2008, 27, 1–116. [Google Scholar] [CrossRef]
  20. Narasimhan, S.G.; Nayar, S.K. Contrast restoration of weather degraded images. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 713–724. [Google Scholar] [CrossRef]
  21. Clifford, P. Markov random fields in statistics. In Disorder in Physical Systems: A Volume in Honour of John M. Hammersly; Grimmett, G., Welsh, D., Eds.; Oxford University Press: Oxford, UK, 1990; pp. 19–32. [Google Scholar]
  22. Roth, S.; Black, M.J. Fields of experts. Int. J. Comput. Vis. 2009, 82, 205–229. [Google Scholar] [CrossRef]
  23. The, Y.W.; Welling, M.; Osindero, S.; Hinton, G.E. Energy-based models for sparse overcomplete representations. J. Mach. Learn. Res. 2003, 41235–41260. [Google Scholar] [CrossRef]
  24. Besag, J. On the statistical analysis of dirty pictures. J. R. Stat. Soc. Ser. B 1986, 48, 259–302. [Google Scholar] [CrossRef]
  25. Geman, S.; Geman, D. Stochastic relaxation, Gibbs distributions, and the Bayesian restoration of images. IEEE Trans. Pattern Anal. Mach. Intell. 1984, 6, 721–741. [Google Scholar] [CrossRef] [PubMed]
  26. Huang, J.; Mumford, D. Statistics of Natural Images and Models. In Proceedings of the 1999 Computer Vision and Pattern Recognition, Ft. Collins, CO, USA, 23–25 June 1999; pp. 541–547. [Google Scholar]
  27. Portilla, J.; Strela, V.; Wainwright, M.J.; Simoncelli, E.P. Image denoising using scale mixtures of Gaussians in the wavelet domain. IEEE Trans. Image Process. 2003, 12, 1338–1351. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Image degradation model of the Complementary Metal-Oxide-Semiconductor (CMOS) image sensor.
Figure 1. Image degradation model of the Complementary Metal-Oxide-Semiconductor (CMOS) image sensor.
Sensors 17 02175 g001
Figure 2. Local consistent Markov random field (MRF). (a) The input image; and (b) The local consistent blocks.
Figure 2. Local consistent Markov random field (MRF). (a) The input image; and (b) The local consistent blocks.
Sensors 17 02175 g002
Figure 3. Vector representation of clique pixels.
Figure 3. Vector representation of clique pixels.
Sensors 17 02175 g003
Figure 4. Initial value of the atmospheric light. (a) Input hazy image; (b) The value of Y ( x ) ; and (c) the blurred Y ( x ) , namely the initial value of atmospheric light.
Figure 4. Initial value of the atmospheric light. (a) Input hazy image; (b) The value of Y ( x ) ; and (c) the blurred Y ( x ) , namely the initial value of atmospheric light.
Sensors 17 02175 g004
Figure 5. Initial value of the transmission map. (a) Input hazy image; and (b) the initial value of the transmission map.
Figure 5. Initial value of the transmission map. (a) Input hazy image; and (b) the initial value of the transmission map.
Sensors 17 02175 g005
Figure 6. Comparison of our method to the method discussed by Tan [13] using real-world images. (a,b) are input hazy images and their corresponding visible edges, respectively; (cf) are the dehazing results with corresponding visible edges generated by the method of Tan [13] and ours, respectively.
Figure 6. Comparison of our method to the method discussed by Tan [13] using real-world images. (a,b) are input hazy images and their corresponding visible edges, respectively; (cf) are the dehazing results with corresponding visible edges generated by the method of Tan [13] and ours, respectively.
Sensors 17 02175 g006
Figure 7. Comparison of our method to the method by Nishino [14] using real-world images. (a,b) input hazy images and their corresponding visible edges, respectively; (cf) are the dehazing results with corresponding visible edges generated by the method of Nishino [14] and ours, respectively.
Figure 7. Comparison of our method to the method by Nishino [14] using real-world images. (a,b) input hazy images and their corresponding visible edges, respectively; (cf) are the dehazing results with corresponding visible edges generated by the method of Nishino [14] and ours, respectively.
Sensors 17 02175 g007
Figure 8. Comparison of our method to the method by Fattal [12] using real-world images. (a,b) are input hazy images and their corresponding visible edges, respectively; (cf) are the dehazing results with corresponding visible edges generated by the method of Fattal [12] and ours, respectively.
Figure 8. Comparison of our method to the method by Fattal [12] using real-world images. (a,b) are input hazy images and their corresponding visible edges, respectively; (cf) are the dehazing results with corresponding visible edges generated by the method of Fattal [12] and ours, respectively.
Sensors 17 02175 g008
Figure 9. Comparison of our method to the He [16] soft matting based method using real-world images. (a,b) are input hazy images and their corresponding visible edges, respectively; (cf) are the dehazing results with corresponding visible edges generated by the method of He [16] and ours, respectively.
Figure 9. Comparison of our method to the He [16] soft matting based method using real-world images. (a,b) are input hazy images and their corresponding visible edges, respectively; (cf) are the dehazing results with corresponding visible edges generated by the method of He [16] and ours, respectively.
Sensors 17 02175 g009
Figure 10. A comparison of the results of synthetic hazy images. (a) Synthetic hazy images; (b) Caraffa’s [15] results; (c) He’s [16] results; (d) Our results; and (e) ground truth images.
Figure 10. A comparison of the results of synthetic hazy images. (a) Synthetic hazy images; (b) Caraffa’s [15] results; (c) He’s [16] results; (d) Our results; and (e) ground truth images.
Sensors 17 02175 g010
Figure 11. Mean Square Error (MSE).
Figure 11. Mean Square Error (MSE).
Sensors 17 02175 g011
Figure 12. Structural Similarity Index of different algorithms. Structural Similarity Index Measurement (SSIM) of different algorithms.
Figure 12. Structural Similarity Index of different algorithms. Structural Similarity Index Measurement (SSIM) of different algorithms.
Sensors 17 02175 g012
Table 1. Comparison of entropy and Peak Signal-to-Noise Ratio (PSNR) on the real-world image for each algorithm compared.
Table 1. Comparison of entropy and Peak Signal-to-Noise Ratio (PSNR) on the real-world image for each algorithm compared.
ImageEntropyPSNR
InputTanOursTanOurs
Figure 6Image (1)7.00797.22357.60439.986210.9585
Image (2)7.00797.34207.00799.35249.7479
Image (3)7.09557.61297.711410.606212.5667
ImageInputNishinoOursNishinoOurs
Figure 7Image (1)7.11437.67707.774210.183516.3444
Image (2)7.15786.93437.215715.857115.9809
Image (3)6.51147.09547.475412.185714.7134
ImageInputFattalOursFattalOurs
Figure 8Image (1)7.08787.32707.473912.679515.2639
Image (2)6.72726.91647.65959.320412.5983
Image (3)7.17737.27937.183216.145717.5572
ImageInputHeOursHeOurs
Figure 9Image (1)5.66106.84797.159714.461212.5766
Image (2)6.47886.96256.981910.648018.4398
Image (3)7.17737.27937.183216.145717.5572
Table 2. Comparisons of arithmetic effectivity on the real-world image for running times (second) of the haze removal algorithm.
Table 2. Comparisons of arithmetic effectivity on the real-world image for running times (second) of the haze removal algorithm.
ImageSizeTanOurs
Figure 6Image (1)624 × 41621.991330.5733
Image (2)596 × 39618.413125.8991
Image (3)1024 × 76853.918571.4230
ImageSizeNishinoOurs
Figure 7Image (1)600 × 40029.742133.8257
Image (2)465 × 38421.873926.1618
Image (3)440 × 44825.819329.8271
ImageSizeFattalOurs
Figure 8Image (1)512 × 34818.912325.2018
Image (2)351 × 2448.012512.9139
Image (3)576 × 76836.977743.5120
ImageSizeHeOurs
Figure 9Image (1)660 × 440107.999139.8194
Image (2)480 × 36084.337223.7916
Image (3)576 × 768181.039143.5120
Table 3. Comparisons of arithmetic effectivity in Figure 10/s.
Table 3. Comparisons of arithmetic effectivity in Figure 10/s.
InputSizeCaraffaHeOurs
Figure 10 Image (1)640 × 48036.8444102.340442.4456
Figure 10 Image (2)1376 × 1032113.3376387.5691125.9364
Figure 10 Image (3)512 × 38418.912459.418324.7468
Figure 10 Image (4)960 × 72057.1672197.856569.8188
Figure 10 Image (5)736 × 55242.3516173.001358.8756

Share and Cite

MDPI and ACS Style

Qu, C.; Bi, D.-Y.; Sui, P.; Chao, A.-N.; Wang, Y.-F. Robust Dehaze Algorithm for Degraded Image of CMOS Image Sensors. Sensors 2017, 17, 2175. https://doi.org/10.3390/s17102175

AMA Style

Qu C, Bi D-Y, Sui P, Chao A-N, Wang Y-F. Robust Dehaze Algorithm for Degraded Image of CMOS Image Sensors. Sensors. 2017; 17(10):2175. https://doi.org/10.3390/s17102175

Chicago/Turabian Style

Qu, Chen, Du-Yan Bi, Ping Sui, Ai-Nong Chao, and Yun-Fei Wang. 2017. "Robust Dehaze Algorithm for Degraded Image of CMOS Image Sensors" Sensors 17, no. 10: 2175. https://doi.org/10.3390/s17102175

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop