Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

1 Introduction

When taking a picture in presence of dust, smoke or water particles hanging in the air, the light emanating from the scene and reaching the camera’s sensor is scattered and attenuated. Similarly, the light coming from the light source is scattered by these particles, thus forming the so-called airlight [11] and resulting in an undesirable veil that reduces contrast and chroma in the picture [12]. In this context, the image formation can be modeled as the sum of the scene’s radiance and the airlight, weighted by a transmission factor t(x) as in Eq. 1:

$$\begin{aligned} I(x)=J(x)t(x) + A_{\infty }(1-t(x)) \end{aligned}$$
(1)

where x denotes the pixel location. I(x) is the image formed on the camera’s sensor. J(x) is the scene radiance. The transmission factor t(x) depends on scene depth d (distance to the sensor) and on the scattering coefficient \(\beta \) of the haze, such that \(t(x)=e^{-\beta .d(x)}\) [11]. Unlike more traditional image degradations, haze is a natural, depth-dependent noise that spans non-uniformly over the whole image. The degradation and the loss of information increases with depth, as the amount of fog between the imaged surface and the sensor increases. Hazy and foggy images have also different prevailing color, which depends on the scattering particles density and the ambient light [15]. The process to recover I(x) from J(x) by estimating first, the atmospheric light \(A_{\infty }\) and t(x), is known as dehazing or defogging.

Many dehazing methods have been proposed and compared to each other. Although, there exist a variety of dehazing algorithms for color images, not a single of them is usually accepted to work perfectly neither to work good enough. In addition, the resulting output of these algorithms is highly correlated to the concentration of scattering particles [7].

Image dehazing and its quality evaluation remain difficult processes. Image dehazing is a transdisciplinary challenge, as it requires knowledge from different fields: meteorology to model the fog, optical physics to understand how light is affected by this fog and computer vision, as well as image and signal processing to recover the parameters of the scene. Image quality assessment (IQA) of dehazed images may requires the hazy and the haze-free images of the same real scene to use full-reference image quality metrics, which are captured under the same conditions such as illuminant, viewing geometry and resolution. Existing databases do not meet this feature (cf. Sect. 2). For this reason, we introduce in Sect. 3 an original new color image database devoted to both haze model assessment and dehazing methods evaluation. We then discuss, in Sect. 4, the limitations to invert the haze model without considering the physical aspect of hazy image formation before to conclude.

2 Existing Hazy Image Databases

Several hazy image databases may be used for dehazing investigation. Hazy images databases including the haze-free reference image and other databases without reference. In the first category we find FRIDA (Foggy Road Image DAtabase) [10, 18] and FRIDA2 [10, 17] that represent evaluation databases for visibility and contrast restoration algorithms. These databases comprise a number of synthetic images with reference of urban road scenes and diverse road scenes, respectively. The view point is close to the one of the vehicle?s driver. The software \(SiVIC^{TM}\) was used to build physically-based road environments from a realistic complex urban model and to generate a moving vehicle with a physically-driven model of its dynamic behavior, and virtual embedded sensors. To each image without fog, four foggy images and a depth map are associated. The depth map is required to be able to add fog consistently in the images. Different kinds of fog are added to each of the four associated images: uniform fog, heterogeneous fog, cloudy fog, and cloudy heterogeneous fog. These four types of fog were inserted by applying the Koschmieder’s law [11] by weighting differently the attenuation coefficient and/or the atmospheric light with respect to the pixel position. Despite the different aspects that are addressed in this database, simulated images fail to represent accurately the natural phenomena effects [8]. The physical interaction of light with atmospheric particles modifies the perceived colors, while colors in the simulated image maintain their hue information and only their saturation component shifts between the original color (saturated), and the haze color (unsaturated).

WILD (Weather and Illumination Database) [6, 14] is an outdoor urban scene database, acquired every hour over seasons. These images are taken under different weather and illumination conditions. Atmospheric conditions, scene distances and temporal data are also associated to images. First, this database cannot be used to compare one dehazed image with a reference due to the variation of illuminant. Second, small changes that could occur would bias a pixel by pixel comparison. Therefore, it appears important to us to build a new database, which does not suffer from these problems and more adequate for such evaluation.

On the other hand, developers often use hazy images of natural scenes, with no reference, which have usually a small size to evaluate dehazing methods within a short time using the minimum amount of resources [2, 5]. However, the lack of the haze-free reference image makes the evaluation challenging and less reliable.

3 Proposed Database

In the CHIC database [1], we consider two indoor scenes, Scene A and Scene B. Scenes were set up in a closed rectangular room, which is large enough to simulate the effect of the distance and the fog density on the objects radiance (length = 6.35 m, width = 6.29 m, height = 3.20 m) with a large window (length = 5,54 m, width = 1.5 m) that allows a large amount of omnidirectional outdoor light to get in. It laterally covers the camera and the scenes. The photo session of each scene lasted 20 min. During this limited time, daylight is assumed to remain steady. This experiment was set up on February, from 1:00 to 4:00 p.m. when sunlight was not directly coming in through the window. Thus, the global light is close to the airlight in a cloudy day. Five Macbeth Color Checkers (MCCs) are placed in the scene at different distances to the camera. The farthest one serves to estimate the atmospheric light. This can be useful to follow up the color alteration when haze covers a scene. The scenes components present various colored surfaces types (reflective and glossy surfaces, rough surfaces, etc.).

A fog machine (FOGBURST 1500), which emits a dense vapor that appears similar to fog was used. A large quantity of fog is initially emitted until it is evenly distributed in the room and forms an opaque layer. Fog is then progressively evacuated through the window. This machine operates by evaporation. The vaporization of therein water-based liquid mixed with glycol, is done by heating. The particles of the ejected fog are water droplets, which have approximately the same radius size of the atmospheric fog (1 - 10 \(\mu \)m).

Thereby, in addition to the haze-free image, nine images of different fog levels images of different levels of fog are captured (Table 1), from level 1, the highest fog density, to level 9, the lowest density. This set of images are captured under two illuminants: outdoor daylight and a compound light (outdoor light + ceiling lamp light (fluorescent tube)). For each level, an RGB image was acquired using the color camera Nikon D7100 [4] providing NEF (RAW) and JPEG 6000\(\times \)4000 images. We used also the spectroradiometer Konica Minolta CS-2000 [3] to measure the transmission of the fog on a white patch of the MCC placed at the back of the scene for each fog density, perpendicularly to the optical path. It was calibrated focusing on the same patch without fog.

3.1 Scenes

  • Scene A: The shoot session of this scene was performed around 2:00 p.m. This scene shows a typical indoor view. We put on the table that is placed in the middle, a number of items with different characteristics such as shapes, colors, positions, surface types (glossy or rough surfaces) and textures. The wall behind the scene is half white and the top half with the white lines and the black holes represent distinctive elements to study algorithms handling near edges.

  • Scene B: The shoot session of scene B was done two hours later. During this time the temperature of the illuminant significantly changed. Unlike the first one, the distance from the camera to the farthest point is smaller. It contains bigger geometric shapes. The fog densities which are randomly determined are characterized by the transmittance spectrum of each fog level.

The camera stayed still over the shoot session of each scene. However, the illuminant is not the same for both scenes, nor the distance to the camera and the density of fog for the correspondent levels (Table 1).

Fig. 1.
figure 1

Scene A (a) and Scene B (b). The distances of different MCCs to the camera are shown. The resolution of the image is 6000\(\times \)4000.

In Table 1, the relative transmittance of fog is calculated with respect to airlight at a given distance over the black patch of original hazy images, as follows:

$$\begin{aligned} T = 1- \frac{S_{level x} - S_{airlight}}{S_{haze-free} - S_{airlight}} \end{aligned}$$
(2)

where \(S_{level x}\), \(S_{airlight}\) and \(S_{haze-free}\) are the spectral values of green in images of different fog levels, of the airlight image of our database where the scene is completely covered by fog and the haze-free image.

Table 1. Relative transmittance T of fog in original hazy images of scene A and scene B. Level 1: highest fog density. Level 9: lowest fog density.

Comparing to outdoor scenes, the light source is located at infinity and the particles forming the synthetic fog are water particles fall in the same range of particles’ size, therefore the same scattering law is applied. Since the room where the shoot session was running is not an open place, the airlight color seems to be close to dark gray. According to Fig. 2, the calculated values of A and t are almost the same over channels at all fog levels. All of these facts prove that, the outside conditions are almost fulfilled in our database.

4 Limitation of Koschmieder Model for Image Recovery

Single image dehazing methods, which consider the haze model given in Eq. 1, are usually based on strong assumptions to estimate \(A_{\infty }\) and t(x) from the single RGB hazy image [9, 16]. However, the only way to verify these assumptions and the model as well, is to have the haze-free image.

Fig. 2.
figure 2

Estimated A and t obtained for each level of haze for a given distance of the MCC placed at the back of the scene. First row: airlight, second row: transmission. First column: scene A, second column: scene B.

Fig. 3.
figure 3

The original hazy (first row) and the reconstructed images of Scene A (second row). A and t are estimated on the surrounded red color patch. The distance between it and the camera is 7 m. First row: original hazy images, second row: reconstructed images. Cropped image size is 1537\(\times \)2049.

Fig. 4.
figure 4

The original hazy (first row) and the reconstructed images of Scene B (second row). A and t are estimated on the surrounded red color patch. The distance between it and the camera is 4.25 m. First row: original hazy images, second row: reconstructed images. Cropped image size is 1537\(\times \)3073.

Consequently, \(A_{\infty }\) is the R, G and B values of the fog layer calculated from image level 1, which is uniformly covered by fog. For each level, the airlight \(A=A_{\infty }(1-t(x))\) is similarly calculated over the black patch of MCC placed at the scene’s back (color patch within the red surrounding in Figs. 3 and 4). We subtracted from it the offset values of R, G, and B of a patch of 20\(\times \)20 pixels of the same black patch from the original image without fog.

Once \(A_{\infty }\) and A are estimated, t(x) of a given scene depth is deduced from the second part of Eq. 1 over the same color chart. Since the distance of each chart is known, the scattering coefficient is deduced and the transmission matrix is then calculated over the entire image, using approximative depths of secondary objects. Based on the known distances of MCCs, images have been split into four zones. Particular focus should be given to the color patch within the red surrounding, which has an accurate known distance to the camera, and has also the lowest object’s transmission.

For the sake of simplicity and due to space constraints, one out of two adjacent levels of daylight images, is retained for image recovery study (Figs. 3 and 4).

Although Scene A and Scene B differ slightly from each other by depths and fog densities, corresponding images of both scenes of the same level provide close values of A and t. This gets reflected through reconstructed images (Figs. 3 and 4), where the area within the red surrounding is poorly recovered in level 1 and level 3 from \(t_{1} = 0.00 \) to \(t_{2}= 0.09\) in Scene A and from \(t_{1} = 0.00\) to \(t_{2}= 0.12\) in Scene B, respectively. From level 5 and on up, when t exceeds 0.17 in Scene A and 0.27 in Scene B, the inversion of the haze model (Eq. 1) succeeds to recover recognizable features comparing to the haze-free image. Since unknown parameters of all fog levels are similarly estimated, this means that the induced error is the same. When the fog density is relatively high (in our case, below level 5), even if the estimation is reasonably accurate, it seems difficult, if not impossible, to compensate thanks to it the lost transmitted light through scattering and absorption and getting a reconstructed image close to the haze-free image features. This is clearly shown on images of low levels of fog, where reconstructed images are noticeably noisy with perceptible saturation shift. This comes to confirm once more what has been pointed out by Narasimhan and Nayar [13], that this model is not valid for depths that are more than a few kilometers. Similarly, when the amount of fog greatly increases, even the radiance of near objects is no more well captured by the camera. Moreover, since the exponential attenuation comes quickly down to zero, the noise is greatly amplified for high fog densities.

5 Conclusion and Future Work

We proposed and described a new color hazy image database of two indoor static scenes. The haze-free image and the supplementary data help to evaluate the commonly accepted haze model, and to evaluate and compare dehazing methods. Such evaluation is done when we consider how much a dehazed image processed by a given method succeeds to meet haze-free image. No matter what circumstances lead up to visibility degradation. Although the parameters of the haze model are accurately estimated, from a level of data lost, when the transmission of the scene is very low, the haze-free image is no longer properly recovered. Koschmieder’s model is therefore considered to be not valid for high densities of fog.

Our future work will use this database for an objective assessment of several dehazing methods using IQA indices including full-reference metrics. We will study also the correlation of these metrics with the perceptual judgments as an efficient attempt to develop a proper haze-dedicated evaluating metric.