Next Article in Journal
A Novel Design of a Torsional Shape Memory Alloy Actuator for Active Rudder
Previous Article in Journal
Experimental and Density Functional Theory Simulation Research on PdO–SnO2 Nanosheet Ethanol Gas Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image Enhancement Thanks to Negative Grey Levels in the Logarithmic Image Processing Framework

Laboratoire Hubert Curien, UMR CNRS 5516, 18 Rue Professeur Benoît Lauras, 42000 Saint-Étienne, France
Sensors 2024, 24(15), 4969; https://doi.org/10.3390/s24154969
Submission received: 24 April 2024 / Revised: 21 June 2024 / Accepted: 11 July 2024 / Published: 31 July 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
The present study deals with image enhancement, which is a very common problem in image processing. This issue has been addressed in multiple works with different methods, most with the sole purpose of improving the perceived quality. Our goal is to propose an approach with a strong physical justification that can model the human visual system. This is why the Logarithmic Image Processing (LIP) framework was chosen. Within this model, initially dedicated to images acquired in transmission, it is possible to introduce the novel concept of negative grey levels, interpreted as light intensifiers. Such an approach permits the extension of the dynamic range of a low-light image to the full grey scale in “real-time”, which means at camera speed. In addition, this method is easily generalizable to colour images and is reversible, i.e., bijective in the mathematical sense, and can be applied to images acquired in reflection thanks to the consistency of the LIP framework with human vision. Various application examples are presented, as well as prospects for extending this work.

1. Introduction

Because of its multiple applications, the problem of image enhancement has been widely studied and has given rise to numerous papers, of which it is impossible to present an exhaustive survey. One of the oldest and best-known techniques is based on histogram equalization: [1], and, more recently, [2]. Among the multitude of methods concerning spatial or frequency domains, let us mention the fuzzy approach [3,4], an algorithm based on Fourier Transform with Highpass (sharpening) and Lowpass (smoothing) Filters [5], a solution using gamma correction [6].
Recently, deep learning methods have addressed the issue of image enhancement for various applications such as autonomous driving [7], iris recognition [8], and underwater images [9,10]. In this last paper, a novel method was proposed based on an embedded fusion mechanism. The authors observed that the quality of the results depends on the quality of the input image, so they performed a pre-processing step to enhance the initial image quality based on the white balance algorithm and contrast-limited adaptive histogram equalization.
Despite the multitude of references, very recent papers show that the subject is still relevant. Among these, we have selected two papers that will be used to compare our approach to state-of-the-art methods. In Ref. [11], a novel Retinex-based Network (R2RNet) is presented, which includes specialized subnets dedicated to denoising and contrast enhancement. Compared to various enhancement techniques, R2RNet appears efficient for all degraded images. In Ref. [12], the authors present a “Cyclic Generative Attention-Adversarial Network” (CGAAN). This network is unsupervised and performs the enhancement of low-light images without using paired datasets. To improve the algorithm’s efficiency in terms of visual quality, the authors propose a novel adaptative attention mechanism under the feature maps, allowing the algorithm to focus on the most significant regions. The network is then guided to recover the high-quality images. The effectiveness of CGAAN will be compared to many methods that will be described and used in Section 6.3, including DALE (Dark region-Aware Low-light image Enhancement), DRBN (Deep Recursive Band Network), and DSLR (Deep Stacked Laplacian Restorer).
Finally, in Ref. [13], the interested reader will find a comprehensive review of low-light image enhancement in terms of network structure, training data, and the presentation of evaluation metrics.
Most of these methods are empirical and lead to irreversible transformations (mathematically non-bijective), resulting in a significant loss of information.
The common objective of enhancement methods is to best compensate for the effects of poor acquisition conditions (low-light images) or variable acquisition conditions (unstable illumination). Given the multitude of algorithms available, there are two pitfalls to overcome when evaluating the effectiveness of a given method and comparing it to others.
The first and most important one is to give, if possible, a non-empirical justification (mathematical, physical, or optical) for the considered approach. The present study focuses on this point of view.
The second is to implement objective tools to evaluate the “quality” of the result. Let us remark that it is illusory to try to create absolute criteria that are relevant in all situations. Let us give an example. In the case of breast cancer detection on mammography images, the observation of small white spots representing microcalcifications constitutes information that should be preserved at all costs but that could be considered noise in other situations. Authors generally refer to quality parameters, which are extremely numerous in the literature, and they are most often presented as quality metrics without always satisfying the mathematical properties required for a metric. The most used are the Peak Signal-to-Noise Ratio (PSNR) parameter and the Structural Similarity Index Measure (SSIM) parameter. The PSNR parameter is known to be poorly correlated with human visual appreciation, while the SSIM and its many variants take the human visual system into account a little better but are not highly significant.
For all these reasons, we chose to work in the LIP (Logarithmic Image Processing) framework, which is presented in the following section. The LIP framework allows us to define an enhancement algorithm, “Full Dynamic Range Expansion” (FDRE), based on the new concept of negative grey levels, which are interpreted as light intensifiers. The advantages of this method are multiple. FDRE enhances a low-light image without a loss of information because it is based on a mathematically bijective transformation. Moreover, this algorithm inherits the properties intrinsic to the LIP model: it has a strong physical justification based on the Transmittance Law, is consistent with the human visual system, and runs at the camera rate. Finally, it is also adaptable to images acquired in reflection and easily extended to colour images.
To conclude this introduction, let us note that the present study extends a previous publication we authored in the journal Sensors [14], in which low-light images were enhanced thanks to the ability of the LIP addition law to simulate variable exposure times. In the same paper, the problem of denoising enhanced images was discussed (CNN approach) and the quality of the results was evaluated in terms of the PSNR and SSIM; thus, this question will not be studied here.

2. Recalls of the LIP Framework

The LIP model was introduced in 1987 [15]. Readers interested in more information on the subject can refer to [16].
Let I ( D ,   [ 0 ,   M [ ) represent the set of grey level images defined on the same spatial support D with values in the grey scale [ 0 ,   M [ . In a first step, images belonging to I ( D ,   [ 0 ,   M [ ) are considered acquired in transmission, so that we can associate to a pair ( f I ( D ,   [ 0 ,   M [ ) ,   x D ) the concept of transmittance T f x , defined as the ratio of the out-coming flux at x by the incoming flux (intensity of the source). Mathematically, T f ( x ) represents the probability, for a particle of the source incident at x , to go through the obstacle, i.e., to be seen by the sensor.
Remark 1:
To simplify notations, we will confuse in a same letter “ f ” the grey level image and the semi-transparent object having generated  f .
Two laws were defined on the space I ( D ,   [ 0 ,   M [ ) :
-
The addition law :
f g = f + g f . g M
which is a direct consequence of the Transmittance Law: T f g = T f   .   T g that represents the probability for a particle of the source to pass through the superposition of two obstacles f and g .
-
The scalar multiplication is associated with a real number λ and an image f :
λ f = M M ( 1 f M ) λ
Remark 2:
In the context of images acquired in transmission, the grey scale  [ 0 ,   M [  is inverted. In fact, stacking obstacles between the source and the sensor obviously darkens the resulting image so that  M  represents the black extremity of the scale, which corresponds to a limited situation of opacity. Conversely, the white extremity  0  is associated with a situation of transparency and represents the sensor’s response when it observes the source.
Let us recall [15} that the two previous laws possess all the properties required to give I ( D ,   [ 0 ,   M [ ) a Vector Space structure, except the existence, for an image f , of an opposite g = f .
Such an opposite must satisfy f g = 0 , which seems a priori unfeasible: indeed, adding g to a given object f would make a transparent object!
However, according to Formula (1), the equation f g = 0 leads to the formal expression f = f 1 f M , which takes its values in the interval ] ,   0 ] .
Under these conditions, we define an over-space of I ( D ,   [ 0 ,   M [ ) , noted F D , ,   M [ ) , representing the set of functions defined on D with values in ] ,   M [ . We clearly have the inclusion I ( D ,   [ 0 ,   M [ ) F D , ,   M [ ) . It is important to note that an element f of F D , ,   M [ ) can be considered as a virtual image with possible negative grey levels.

3. Major Properties of LIP Laws

Most image-enhancement methods are not based on sound science. We chose the LIP framework because it complies with physical, optical, mathematical and psychovisual properties:
-
Strong mathematical structure
It is easy to verify that the space F D , ,   M ,   ,   , satisfies all the conditions required (cf. [17], for example) to become a Real Vector Space. The consequences of such a remark are considerable: it gives access to countless concepts and properties introduced by mathematicians in the context of Vector Spaces, like interpolation, norms, and scalar products.
-
Consistency with the Human Visual System
We first considered the case of images acquired in transmission. However, all the results obtained within the LIP framework apply to images acquired in reflection, thanks to a paper published by Brailean [18] in which he established the consistency of the LIP model with Human Vision. Under these conditions, LIP tools can be used to interpret images in the same way as the human eye would.
-
Simulation of variable exposure time
This property was presented by Carré et al. [19], “LIP operators: Simulating exposure variations to perform algorithms independent of lighting conditions”, and was improved in 2021 in Sensors [14], “Extending Camera’s Capabilities in Low Light Conditions Based on LIP Enhancement Coupled with CNN Denoising”.
-
Simulation of variable thickness/opacity for images acquired in transmission
Such a property is useful in a variety of situations.
One example is the case of physical cuts performed with a microtome, where regularity in the thickness of successive slices is difficult to achieve. When acquiring images of such slices, variations in thickness lead to variations in opacity, which must be corrected to obtain homogeneous 3D reconstructions.
Another case is confocal microscopy. When acquiring images of a semi-transparent object (human skin and hair) at different depths of focus, the brightness of the images decreases sharply as the depth increases, requiring a step to compensate for this attenuation in brightness.
-
Changing the dynamic range of an image
This topic is developed in the following sections.

4. Image Enhancement with LIP Laws: A Quick Reminder

For images acquired in transmission, it is easy to explain the darkening/brightening effect of each law (cf. Figure 1 and Figure 2).
Law ⨹: If C [ 0 ,   M [ and f I ( D ,   [ 0 ,   M [ , calculating f C (resp. f C ) consists of subtracting (resp. adding) a uniform image C from (resp. to) f . Considering a very low-light image f , to obtain a satisfactory enhancement of f , the constant to be subtracted from f takes values close to M , at the risk of observing negative values of f C . Such a remark draws our attention to a physical interpretation of negative grey levels (cf. next section).
Law ⨻: If λ is a real number and f I ( D , [ 0 , M [ , calculating λ f consists of multiplying the thickness of f by λ , which obviously produces a brightening effect when λ ] 0,1 [ and a darkening effect when λ ] 1 , + [ , but not always an increase in the dynamic range of f .
Looking at Figure 1 and Figure 2, it is easy to imagine the potential effectiveness of LIP laws for image enhancement, which explains the numerous publications on the subject. For the record, we will just mention the initial work of two teams led by Professor Cahill (La Trobe University, Australia) and Professors Agaian and Panetta (respectively, the University of Texas and Tufts University, USA). In [20], Deng and Cahill describe a new implementation of Lee’s image enhancement algorithm. Based on the Logarithmic Image Processing (LIP) model, the proposed approach can simultaneously improve the overall contrast and sharpness of an image. In [21], Agaian and Panetta introduce a parameterized LIP (PLIP) model that covers both the linear arithmetic and LIP operations within a single unified model.
Image enhancement is usually performed by optimizing some parameter, such as the dynamic range or contrast concepts. It should be noted that the dynamic range must be treated with caution. Indeed, the presence of a single white pixel and a single black pixel in an image produces a maximum dynamic range without guaranteeing an interesting visual result.
For this reason, another approach was proposed in [16] for a low-light image f : it consists of calculating either the constant C 0 , maximizing the histogram standard deviation σ [ h f C of f C (cf. Figure 3), or t h e   r e a l   n u m b e r   λ 0 ,   m a x i m i z i n g   σ [ h λ f (cf. Figure 4), with the constraint, in both cases, of staying within the grey scale.

5. Optical Interpretation of Negative Grey Levels

In this section, we will take advantage of the Vector Space Structure of the space of functions F D ,   ,   M [ ) thanks to logarithmic laws ⨹ and ⨻. Moreover, this structure allows a well-founded optical interpretation of negative grey levels. Indeed, when we introduced the opposite f x = f ( x ) 1 f ( x ) M of a grey level f x , our goal was that equality f ( x ) f ( x ) = 0 be satisfied. This means that the union of the two “obstacles” f ( x ) and f x becomes transparent, as if the sensor was observing the source!
Thus, f ( x ) can be interpreted as a light generator that increases the intensity of emitted “photons” (or electrons, or X-rays depending on the source elements) to produce an outgoing flux at x equal to that of the initial source. In Figure 5, we present the grey level values along a straight line drawn on the considered image of a chart and their opposites in the interval ] ,   0 [ consisting of negative grey levels.
Similarly, the scalar multiplication λ f can be considered for negative values of λ : the previous interpretation in terms of thickness variation adapts to negative values thanks to the concept of “negative” thickness, which acts as a brightening operator.
To formalize this approach in optical terms, let us define the Generalized Grey Scale ] ,   M [ representing the set of all possible values of source intensity: the value corresponds to a theoretical infinite intensity, while M corresponds to a null one.
Choosing an initial source intensity I S   ] ,   M [ involves selecting an origin 0 S for the grey scale [ 0 S ,   M [ , producing the current space of images I ( D ,   [ 0 S , M [ ) . The over-space F D ,   , M [ ) then represents the set of all possible images when the reference source moves inside the interval ] , M [ , which leads to the following formula:
F D ,   , M [ ) =     0 S   ] , M [ I ( D ,   [ 0 S , M [ )
The following section presents the ability of negative grey levels to perform a maximum dynamization of a very-low-light image.

6. Full Dynamic Range Expansion (FDRE) Algorithm

For an image f , its dynamic range D R   ( f ) is calculated as D R ( f ) = f ( a ) f ( b ) , where
f a = S u p   f x ,   x D   a n d   f b = I n f   f x ,   x D
Case: subtraction of a constant
From the formula, f x C = f x C 1 C M , the dynamic range D R   ( f C ) obviously satisfies D R   ( f C ) = f ( a ) f ( b ) 1 C M   and increases with C when C [ 0 ,   M [ . Moreover, D R   f C tends toward + when C approaches M because 1   C M tends toward 0. Thus, a value
C 0 can be calculated maximizing D R   f C , i.e., satisfying
D R   ( f C 0 ) = M 1 , which corresponds to the maximum range for digitized images.
Remark 3:
Considering that f x C is expressed as an affine function α . f ( x ) + β , the proposed full expansion corresponds to a classical linear expansion, which consists of applying a bijection of the observed dynamic range   [ f ( b ) ,   f ( a ) ]   onto the whole grey scale   [ 0 ,   M 1 ] .
Case: scalar multiplication
When negative values of λ are allowed, the function λ f takes its values inside the interval ] ,   0 ] , and when λ varies from 0 to , D R ( λ f ) varies from 0 to + . Under these conditions,
D R ( f ) = f ( a ) f ( b ) and for λ < 0:
D R λ f = λ f b λ f a = M M 1 f b M λ M M 1 f a M λ D R λ f = M 1 f a M λ 1 f b M λ
It is important to note that there exists a unique value λ 0 < 0 such that D R λ 0 f = M 1 . Then,
M 1 f a M λ 0 1 f b M λ 0 = M 1   1 f a M λ 0 1 f b M λ 0   = 1 1 M
To achieve the FDRE of f , we must perform the two following steps:
Step 1: Computation of λ 0
In Equation (6), it is not possible to calculate λ 0 explicitly. Nevertheless, we can assert the uniqueness of λ 0 . Indeed, the dynamic range D R ( λ   f ) is expressed as the difference between two exponential functions and is then continuous with respect to the variable λ , meaning that it reaches each value lying in [ 0 , + [ only once—in particular, the value 1 1 M when λ   = λ 0   .
To obtain an approximate solution, we perform a dichotomy step to reduce an interval surrounding λ 0   . First, we choose the initial size of the interval, say 1, and determine the first value λ = 1 , 2 , 3 for which the following inequality applies:
1 f a M λ 1 f b M λ > 1 1 M
This means that the optimal value λ 0 belongs to the interval ] λ ,   λ + 1 ] . The principle of dichotomy consists then of dividing this interval by 2 depending on the location of λ 0 in ] λ ,   λ + 1 2 ] or ] λ + 1 2 ,   λ + 1 ] . The operation is iterated until the required precision for λ 0 is reached.
Step 2: Displaying the FDRE λ 0   f ( λ 0   ( f a ) of image f
Once λ 0 computed, the function λ 0   f must be displayed as an image. This is achieved by translating λ 0   f by the vector ( λ 0   ( f a ) , which results in a Full Dynamic Range Expansion of f . In Figure 6, two examples are presented:
The first one deals with the centrifugation of blood to separate red blood cells (erythrocytes) from plasma. The aim is to evaluate the position of the boundary between the two media on an image acquired in back-light condition, which is fairly easy, but also to be able to “read” the information on the label stuck to the test tube, in particular the batch reference and barcode. The second concerns the “Indoor scene” image in Figure 4.
Comments:
The FDRE technique is easily performed in real time (31 frames per second) for images of standard a size of 512 × 512 pixels.
The FDRE technique can be extended to colour images in a very simple way.

6.1. FDRE for Colour Images

We associate to a colour image f its components f R ,   f G ,   f B in the three channels Red, Green, and Blue, and we define the Global Dynamic Range G D R f of image f according to
G D R f = M a x x D f R ( x ) ,   f G ( x ) ,   f B ( x ) M i n x D f R ( x ) ,   f G ( x ) ,   f B ( x )
To apply the FDRE algorithm to a low-light colour image f , we compute the unique negative scalar λ 0 satisfying
G D R λ 0   f = M 1
Multiplying each of the three components f R ,   f G ,   f B by λ 0 results in three negative functions λ 0 f R ,   λ 0 f G ,   λ 0 f B . To bring these functions back into the grey scale [ 0 , M [ , we apply them a translation by the vector λ 0 M a x x D f R x ,   f G x ,   f B x . Examples are given in Figure 7.

6.2. Physical Interpretation of the Proposed FDRE Algorithm

As recalled in Section 3, the LIP framework is based on the Transmittance Law; the acquisition of an image inside this model is performed according to Figure 8.
The logarithmic addition (resp. subtraction) of a constant C to (resp. from) an image f can be represented in the same way (Figure 9).
Moreover, the LIP framework is consistent with the Human Visual System; its applicability is not limited to images acquired in transmission but naturally extends to images acquired in reflection that we wish to analyse as a human eye would.
To provide a well-founded optical interpretation of the FDRE algorithm, including the case of an image f acquired in reflection, the grey level f ( x ) of each pixel x is considered. In the LIP grey scale, where 0 represents the maximum intensity and M is the total opacity, f ( x ) obviously appears as an intermediate value between “white” and “black”, which means as an attenuation of the source, interpreted as a transmittance. It is then possible to design a virtual semi-transparent obstacle producing for each pixel x the value f ( x ) . Under such conditions, the computation of λ 0 corresponds to a theoretical negative thickness of this semi-transparent object to reach the maximum available dynamics, that is, the dynamics of the full grey scale.

6.3. FDRE Algorithm’s Efficiency Compared with Other Methods

Any author of a novel image enhancement algorithm has the legitimate ambition to evaluate its performance and compare it to existing methods in terms of visual quality. For this purpose, we could refer to classical parameters (PSNR, SSIM), but we know that such parameters are not strongly consistent with the human visual system. Moreover, it is illusory to hope to create a “universal” parameter suitable for all situations, like very-low-light images or dark images with bright regions, all being potentially acquired under varying light conditions. This is why the authors interested in image enhancement are generally satisfied with subjective assessments. In the following Section 6.3.1 and Section 6.3.2, we compare FDRE with various algorithms described in [11,12] cited in the introduction.

6.3.1. Comparison with SRIE, RetinexNet, and R2RNet

In Section 1, we presented the R2RNeT method proposed by Hai et al. [11]. In this study, R2RNeT is compared (see Figure 10) to two classical enhancement algorithms: Simultaneous Reflectance and Illumination Estimation (SRIE) (cf. [22]) and Retinex-Net (cf. [23]):
SRIE: To estimate reflectance and illumination from a given image, the authors proposed a novel weighted variational model based on Simultaneous Reflectance and Illumination Estimation (SRIE). Compared to conventional variational models, SRIE together reduces noise and preserves the estimated reflectance with more details.
Retinex-Net: The authors collected a low-light dataset containing low/normal-light image pairs and developed a deep Retinex-Net learned on this dataset. This Retinex-Net includes a Decom-Net for decomposition and an Enhance-Net for illumination correction.
Remark: 
Retinex theory, namely, the theory of the retinal cortex, established by Land and McCann, is based on the perception of colour by the human eye and the modeling of colour invariance [24].
In Figure 11, we display the result of our FDRE method.
Comment: Let us recall that the FDRE algorithm corresponds to a logarithmic expansion within the negative part of the grey scale (see Figure 5). Such an expansion acts less strongly on the darkest pixels than on the brightest ones, producing a visual appearance near saturation for the brightest regions (“Indoor scene” in Figure 7). It is important to note that no post-processing was applied to the image in Figure 11, unlike the image presented in Figure 10d after a denoising step.

6.3.2. Comparison with DALE, DRBN and CGAAN

In this section, we return to the algorithm “Cyclic Generative Attention-Adversarial Network” (CGAAN) presented in the introduction. Proposed by Zhen et al. [12], CGAAN is compared by the authors to two enhancement algorithms:
The novel enhancement method presented in [25] is called Dark region-Aware Low-light image Enhancement (DALE). It consists of applying a visual attention module to detect dark regions and apply a brightness enhancement. Such an approach preserves the colour of original images and normally avoids the saturation of illuminated regions.
A Deep Recursive Band Network (DRBN) is proposed in [26]. Based on paired low/normal-light images, it aims at recovering a linear band representation of an enhanced normal-light image. The band recomposition is learned with the perceptual guidance toward fitting perceptual regularization of high-quality images.
In Figure 12, we display the results of DALE, DRBN, and CGAAN (Figure 12b–d) applied to an initial low-light image (Figure 12a), and the result of our FDRE algorithm is proposed in Figure 12e.
To conclude this section, let us propose a comparison between a classical histogram equalization (HE) algorithm and the FDRE algorithm. In Figure 13, two low-light colour images are considered: “Indoor scene” and “Beach by night”. Classically, the HE algorithm is performed on the V-channel of their HSV representations.
We observe that on the very dark image, “Indoor scene”, HE produces false colours, while FDRE tends to saturate bright areas, as expected. Concerning “Beach by night”, which presents dark and bright regions, HE is ineffective, while FDRE highlights the information in the darkest areas.

7. Examples of Application

7.1. 3D Visualization of Scanner Dental Optical Cuts

The initial data consist of twelve slices of a jaw obtained using an X-ray scanner. To avoid over-irradiating the patient, dentists applied a low dose of X-rays. The resulting images therefore possess a limited contrast, requiring a FDRE step, based on scalar multiplication law , producing enhanced slices.
A major problem is raised by the scanner resolution anisotropy: the spatial resolution inside a slice is around 0.2 mm per pixel, while the thickness of each slice is around 1 mm. To overcome such a drawback, we applied a LIP interpolation computing four intermediate images between two successive slices to obtain cubic voxels. The formula for performing such an interpolation is very simple and deducted from a classical interpolation, written with LIP operators. We define the segment [ f , g ] associated with a pair of successive slices f and g according to
f , g =   λ f ( ( 1 λ ) g ) λ [ 0,1 ]
In our situation, we used λ = 1 5 ,   2 5 ,   3 5 ,   4 5 (cf Figure 14) to compute the four intermediate slices.
Since the initial slices are acquired in transmission, the LIP interpolation is perfectly adapted and produces intermediate grey levels in the sense of transparency. The grey level of a pixel belonging to slice f is assigned to the corresponding voxel.
Finally, we dispose of 56 images instead of the first 12.
In Figure 15a, the initial twelve under-lighted slices are displayed, as well as their anisotropic superposition. Applying the FDRE algorithm produces enhanced images (Figure 15b), highlighting the location of an included tooth considered the target.
The expectations of dentists were multiple and rather difficult to meet.
Their ultimate objective was to visualize in 3D all the optical sections of the scanner to improve their understanding and prepare for surgical procedures. For that, it is necessary to perform an automated segmentation of this 3D block whose segments correspond to the anatomical elements such as teeth, mandibular bone, and sinuses. In Figure 15c, a fairly simple example is proposed. In fact, it is not difficult to remove from the 3D block the regions corresponding to air and flesh, which are the darkest ones. This step is performed using a classic Region Growing algorithm.
Once an angle of view of the 3D block is chosen, the voxels located in the foreground are visualized with their grey level, while the others are not considered.
It is then easy to estimate the precise location of an included tooth not initially visible because it is entirely under the gum surface and thus to perform the minimum incision to achieve the extraction.

7.2. Other Fields of Application

Logarithmic image enhancement is applicable in many areas. We can cite the following:
- Thickness normalization of serialized slices made using a microtome or an ultra-microtome to standardize the dynamics of the corresponding images. This allows us to visualize homogeneous 3D blocks.
- Brightness correction of images acquired using a confocal microscope to observe semi-transparent objects (for example, skin and hair). Obviously, the dynamics of an image corresponding to a deep focal plane are strongly attenuated. In Figure 16, the acquisition of a human hair using a Tandem Scanning Confocal Microscope (TSM) is presented. We know that such a microscope can acquire the relief of an object. In fact, a pixel located at the object surface is assigned a maximum grey level when this pixel belongs to the focal plane of the microscope.
In Figure 16a, each pixel of the hair relief map is displayed with its maximum reflectance. When the focal plane of the TSM crosses the hair near its axis, we obtain a very low-light image of the medulla. After FDRE enhancement (Figure 16b), the segmentation of the medulla is performed thanks to a Region Growing algorithm that detects the hair boundaries and their grey levels together. Finally, the registration of such images, when the focal plane of the TSM moves through the entire thickness of the hair, allows a 3D visualization of the medulla (in false colour, Figure 16c) together with the hair surface seen from the inside.

8. Conclusions

This study focuses on low-light image enhancement, which has long been studied by many researchers with very different approaches that are often empirical. Our stated goals were to propose novel algorithms with a strong physical justification, consistent with the human visual system, with the shortest possible execution time. It has been demonstrated that the Logarithmic Image Processing (LIP) framework meets all these constraints and permits the introduction of negative grey levels to simulate an increase in the source intensity. Knowing that the logarithmic laws and , r e s p e c t i v e l y , model a variation of exposure time and a variation of opacity (thickness) of the observed object, we dispose of scientifically based and efficient tools to extend the dynamics of low-light images to the full dynamic range (FDRE algorithm). Due to the non-linearity of the logarithmic laws, FDRE acts less strongly on the darkest pixels than on the brightest ones, producing a visual aspect near saturation for the brightest regions. In conclusion, FDRE will achieve maximum efficiency for very-low-light images or regions.
In addition, the consistency of the model with the human visual system augurs for satisfactory results in terms of visual quality, while such an evaluation is not guaranteed by the usual “metrics” PSNR and SSIM. Moreover, the novel algorithm runs at a conventional camera speed (24 frames per second) using a standard PC.

9. Perspectives

Our goal now is to strengthen the relevance of the proposed tools by creating specialized quality evaluators, such as sharpness, contrast and resolution. The first criterion (sharpness) is particularly interesting, as it concerns different situations such as defocusing or the presence of a diffusing medium, as well as the application of Super-Resolution (SR) algorithms. By creating intermediate pixels, like any interpolation technique, SR algorithms require a de-blurring step, the effectiveness of which needs to be assessed. Since the blurring effect occurs near the contours (transition pixels), various papers have proposed quality criteria based on gradient operators. The problem is that most of these operators do not produce values naturally limited to the available grey scale but require adjustments that distort the information. We plan to use the notion of logarithmic contrast (LIP difference in terms of grey levels) between a pixel and its neighbors, with the advantage of remaining inside the grey scale. The initial results are promising and allow us to reclassify in the right order different defocused images of the same object. Moreover, such contrast concepts seem relevant to assess whether an image is well contrasted in the sense of human vision. It will be interesting to compare this approach with Artificial Intelligence methods concerning, for example, the learned similarity (cf. [27]).

Funding

This research received no external funding.

Data Availability Statement

Ethical review and approval were waived for this study. In fact, the only concerned section (Section 7.1) entitled “3D Visualization of Scanner Dental Optical Cuts” concerns private data (of family origin).

Acknowledgments

Many thanks are due to the Reviewers and especially to the Academic Editor for their demanding remarks and critiques, which have greatly contributed to improving the quality of the present manuscript.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; ter Haar Romeny, B.; Zimmerman, J.B.; Zuiderveld, K. Adaptive histogram equalization and its variations. Comput. Vis. Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  2. Garg, P.; Jain, T. A Comparative Study on Histogram Equalization and Cumulative Histogram Equalization. Int. J. New Technol. Res. 2017, 3, 41–43. [Google Scholar]
  3. Mahashwari, T.; Asthana, A. Image Enhancement Using Fuzzy Technique. Int. J. Res. Rev. Eng. Sci. Technol. 2013, 2, 1–4. [Google Scholar]
  4. Mittal, P.; Saini, R.K.; Jain, N.K. Image Enhancement Using Fuzzy Logic Techniques. In Soft Computing: Theories and Applications.Advances in Intelligent Systems and Computing; Ray, K., Sharma, T., Rawat, S., Saini, R., Bandyopadhyay, A., Eds.; Springer: Berlin/Heidelberg, Germany, 2019; Volume 742, pp. 978–981. [Google Scholar]
  5. Makandar, A.; Halalli, B. Image Enhancement Techniques using Highpass and Lowpass Filters. Int. J. Comput. Appl. 2015, 109, 12–15. [Google Scholar] [CrossRef]
  6. Rahman, S.; Rahman, M.; Abdullah-Al-Wadud, M.; Al-Quaderi, G.D.; Shoyaib, M. An adaptive gamma correction for image enhancement. EURASIP J. Image Video Process. 2016, 2016, 35. [Google Scholar] [CrossRef]
  7. Li, G.; Yang, Y.; Qu, X.; Cao, D.; Li, K. A deep learning based image enhancement approach for autonomous driving at night. Knowl.-Based Syst. 2021, 213, 106617. [Google Scholar] [CrossRef]
  8. Liu, M.; Zhou, Z.; Shang, P.; Xu, D. Fuzzified Image Enhancement for Deep Learning in Iris Recognition. IEEE Trans. Fuzzy Syst. 2020, 28, 92–99. [Google Scholar] [CrossRef]
  9. Ramkumar, G.; Ayyadurai, M.; Senthilkumar, C. An Effectual Underwater Image Enhancement using Deep Learning Algorithm. In Proceedings of the 5th International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 6–8 May 2021; pp. 1507–1511. [Google Scholar]
  10. Zhou, J.; Sun, J.; Zhang, W.; Lin, Z. Multi-view underwater image enhancement method via embedded fusion mechanism. Eng. Appl. Artif. Intell. 2023, 121, 105946. [Google Scholar] [CrossRef]
  11. Hai, J.; Xuan, Z.; Yang, R.; Hao, Y.; Zou, F.; Lin, F.; Han, S. R2RNet: Low-light image enhancement via Real-low to Real-normal Network. J. Vis. Commun. Image Represent. 2023, 90, 103712. [Google Scholar] [CrossRef]
  12. Zhen, T.; Peng, D.; Li, Z. Cyclic Generative Attention-Adversarial Network for Low-Light Image Enhancement. Sensors 2023, 23, 6990. [Google Scholar] [CrossRef]
  13. Tian, Z.; Qu, P.; Li, J.; Sun, Y.; Li, G.; Liang, Z.; Zhang, W. A Survey of Deep Learning-Based Low-Light Image Enhancement. Sensors 2023, 23, 7763. [Google Scholar] [CrossRef] [PubMed]
  14. Carré, M.; Jourlin, M. Extending Camera’s Capabilities in Low Light Conditions Based on LIP Enhancement Coupled with CNN Denoising. Sensors 2021, 21, 7906. [Google Scholar] [CrossRef] [PubMed]
  15. Jourlin, M.; Pinoli, J.-C. A model for Logarithmic Image Processing. J. Microsc. 1988, 149, 21–35. [Google Scholar] [CrossRef]
  16. Jourlin, M. Logarithmic Image Processing: Theory and Applications. Adv. Imaging Electron Phys. 2016, 195, 259. [Google Scholar]
  17. Schaefer, H. Topological Vector Spaces, 2nd ed.; Springer: New York, NY, USA, 1999. [Google Scholar]
  18. Brailean, J.; Sullivan, B.; Chen, C.; Giger, M. Evaluating the em algorithm using a human visual fidelity criterion. Int. Conf. Acoust. Speech Signal Process. 1991, 6, 2957–2960. [Google Scholar]
  19. Carré, M.; Jourlin, M. LIP operators: Simulating exposure variations to perform algorithms independent of lighting conditions. In Proceedings of the 2014 International Conference on Multimedia Computing and Systems (ICMCS), Marrakech, Morocco, 14–16 April 2014; pp. 122–126. [Google Scholar]
  20. Deng, G.; Cahill, L.W.; Tobin, G.R. The study of logarithmic image processing model and its application to image enhancement. IEEE Trans. Image Process. 1995, 4, 506–512. [Google Scholar] [CrossRef] [PubMed]
  21. Panetta, K.; Agaian, S.; Zhou, Y.; Wharton, E.J. Parameterized Logarithmic Framework for Image Enhancement. IEEE Trans. Syst. Man Cybern. 2011, 41, 460–473. [Google Scholar] [CrossRef]
  22. Fu, X.; Zeng, D.; Huang, Y.; Zhang, X.-P.; Ding, X. A weighted variational model for simultaneous reflectance and illumination estimation. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2782–2790. [Google Scholar]
  23. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep Retinex Decomposition for Low-Light Enhancement. arXiv 2018, arXiv:1808.04560. [Google Scholar]
  24. Land, E.H.; McCann, J.J. Lightness and Retinex theory. J. Opt. Soc. Amer. 1971, 61, 1–11. [Google Scholar] [CrossRef]
  25. Kwon, D.; Kim, G.; Kwon, J. Dale: Dark region-aware low-light image enhancement. arXiv 2020, arXiv:2008.12493. [Google Scholar]
  26. Yang, W.; Wang, S.; Fang, Y.; Wang, Y.; Liu, J. Band representation-based semi-supervised low-light image enhancement: Bridging the gap between signal fidelity and perceptual quality. IEEE Trans. Image Process. 2021, 30, 3461–3473. [Google Scholar] [CrossRef] [PubMed]
  27. Zhang, R.; Isola, P.; Efros, A.; Shechtman, E.; Wang, O. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 586–595. [Google Scholar]
Figure 1. LIP-Subtraction (resp. Addition) of a constant from (resp. to) an image f , “Epithelial cells”. From left to right: visualization of f C , f, and f C , case C = 64 grey levels.
Figure 1. LIP-Subtraction (resp. Addition) of a constant from (resp. to) an image f , “Epithelial cells”. From left to right: visualization of f C , f, and f C , case C = 64 grey levels.
Sensors 24 04969 g001
Figure 2. LIP-Multiplication by a real number. Image f , “Epithelial cells”. From left to right: visualization of 0.5 f , f , and 2 f .
Figure 2. LIP-Multiplication by a real number. Image f , “Epithelial cells”. From left to right: visualization of 0.5 f , f , and 2 f .
Sensors 24 04969 g002
Figure 3. Maximization of the standard deviation σ [ h f C . From left to right: initial low-light image f “Teenager” (exposure time of 8 ms). Enhanced image of f according to f   C 0 , with C 0   = 197 .
Figure 3. Maximization of the standard deviation σ [ h f C . From left to right: initial low-light image f “Teenager” (exposure time of 8 ms). Enhanced image of f according to f   C 0 , with C 0   = 197 .
Sensors 24 04969 g003
Figure 4. Maximization of the standard deviation σ [ h λ f   of the histogram of λ f . From left to right: Initial low light image f, “Indoor scene”, in the presence of bright points. Enhanced image of f according to λ 0 f , with λ 0   = 0.28 .
Figure 4. Maximization of the standard deviation σ [ h λ f   of the histogram of λ f . From left to right: Initial low light image f, “Indoor scene”, in the presence of bright points. Enhanced image of f according to λ 0 f , with λ 0   = 0.28 .
Sensors 24 04969 g004
Figure 5. Grey levels along the white straight line drawn on the chart and their opposites.
Figure 5. Grey levels along the white straight line drawn on the chart and their opposites.
Sensors 24 04969 g005
Figure 6. FDRE examples. From left to right: The test tube after centrifugation with the label location and the result of the LIP-multiplicative FDRE. Initial image “TIndoor scener” before and after multiplicative FDRE.
Figure 6. FDRE examples. From left to right: The test tube after centrifugation with the label location and the result of the LIP-multiplicative FDRE. Initial image “TIndoor scener” before and after multiplicative FDRE.
Sensors 24 04969 g006
Figure 7. Low-light colour images f ,   g and their enhanced images λ 0 f , λ 0 g . From left to right: indoor scene, human skin.
Figure 7. Low-light colour images f ,   g and their enhanced images λ 0 f , λ 0 g . From left to right: indoor scene, human skin.
Sensors 24 04969 g007
Figure 8. Image acquisition in transmission.
Figure 8. Image acquisition in transmission.
Sensors 24 04969 g008
Figure 9. Representation of f , C, f C , and f C in transmission. (a) Acquisition of initial image f. (b) Uniform image C. (c) Acquisition of f C . (d) Acquisition of f C .
Figure 9. Representation of f , C, f C , and f C in transmission. (a) Acquisition of initial image f. (b) Uniform image C. (c) Acquisition of f C . (d) Acquisition of f C .
Sensors 24 04969 g009
Figure 10. From a previous paper by Jiang Hai et al. [11]: (a) Input low-light image. (b) Simultaneous Reflectance and Illumination Estimation (SRIE) result (cf. [22]). (c) RetinexNet result (cf. [23]). (d) Result of the proposed algorithm.
Figure 10. From a previous paper by Jiang Hai et al. [11]: (a) Input low-light image. (b) Simultaneous Reflectance and Illumination Estimation (SRIE) result (cf. [22]). (c) RetinexNet result (cf. [23]). (d) Result of the proposed algorithm.
Sensors 24 04969 g010
Figure 11. Result of FDRE, without any denoising post-processing.
Figure 11. Result of FDRE, without any denoising post-processing.
Sensors 24 04969 g011
Figure 12. (a) Input, (b) DALE, (c) DRBN, (d) CGAAN, (e) FDRE.
Figure 12. (a) Input, (b) DALE, (c) DRBN, (d) CGAAN, (e) FDRE.
Sensors 24 04969 g012
Figure 13. Top: from left to right, initial image “Indoor scene”, histogram equalization, FDRE. Bottom: from left to right, initial image “Beach by night”, histogram equalization, FDRE.
Figure 13. Top: from left to right, initial image “Indoor scene”, histogram equalization, FDRE. Bottom: from left to right, initial image “Beach by night”, histogram equalization, FDRE.
Sensors 24 04969 g013
Figure 14. From left to right: initial voxel, interpolated voxels.
Figure 14. From left to right: initial voxel, interpolated voxels.
Sensors 24 04969 g014
Figure 15. Based on 12 X-rays slices of a jaw, the different steps to perform a 3D visualization.
Figure 15. Based on 12 X-rays slices of a jaw, the different steps to perform a 3D visualization.
Sensors 24 04969 g015aSensors 24 04969 g015b
Figure 16. Confocal microscopy: visualization of a human hair. (a) Surface visualization. (b) Hair medulla after FDRE. (c) Medulla 3D visualization.
Figure 16. Confocal microscopy: visualization of a human hair. (a) Surface visualization. (b) Hair medulla after FDRE. (c) Medulla 3D visualization.
Sensors 24 04969 g016
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jourlin, M. Image Enhancement Thanks to Negative Grey Levels in the Logarithmic Image Processing Framework. Sensors 2024, 24, 4969. https://doi.org/10.3390/s24154969

AMA Style

Jourlin M. Image Enhancement Thanks to Negative Grey Levels in the Logarithmic Image Processing Framework. Sensors. 2024; 24(15):4969. https://doi.org/10.3390/s24154969

Chicago/Turabian Style

Jourlin, Michel. 2024. "Image Enhancement Thanks to Negative Grey Levels in the Logarithmic Image Processing Framework" Sensors 24, no. 15: 4969. https://doi.org/10.3390/s24154969

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop