1. Introduction
Image enhancement is the process of applying specific techniques to boost an image’s visual quality. These techniques can imply diverse criteria, such as increasing the image contrast, noise reduction, highlighting relevant details, or adjusting brightness and color saturation. The main goal of image enhancements is to make the information within the image more easily interpretable and perceptible to human viewers, as well as to boost the automatic process of applications such as pattern recognition, medical analysis, and computer vision, among others [
1,
2].
Traditional image enhancement methods are divided into two well-defined categories [
3,
4]. The first is the spatial domain enhancement, which directly modifies pixel values to adjust aspects such as contrast and image detail. This category includes techniques such as gamma correction [
5], histogram equalization [
2], and sigmoid correction [
6], as well as more advanced techniques such as multi-dimensional adaptive local enhancements [
7] and recursive filters for color balance and shadow correction [
8]. The second category is frequency domain enhancement, which transforms the image into a mathematical domain to adjust the frequency components, allowing fine detail enhancement and reduction of undesirable patterns through techniques such as high-pass or low-pass filtering. In this family, methods such as wavelet transform [
9], discrete cosine transformation [
10], and Fourier transform [
11] are included. Both categories have specific applications and are chosen according to the type of improvement desired.
Remarkably, not all images require the same enhancement process, since the enhancement strategy relies upon the image’s specific characteristics. For example, a low-contrast image can significantly benefit from spatial domain enhancement techniques, like the contrast adjustment or the histogram equalization approach, which enhance the image’s sharpness [
12,
13]. On the other hand, in medical images such as magnetic resonances, the fine details highlighting can be critical; therefore, frequency domain enhancement techniques are more suitable for adjusting the high frequencies and spotlighting the image’s internal structures [
14]. In deep learning architectures that handle visual data such as point clouds [
15] or drone datasets [
16], image enhancement should aim to improve feature extraction by making relevant details more prominent, ensuring that critical features such as edges, shapes, or textures are effectively captured during the training process. Hence, there is no unique ideal operator for all images nor a unique quantitative metric that automatically evaluates the image quality. Automatic image enhancement is a process that produces enhanced images without human intervention and is an extremely complicated task in image processing [
17].
Image enhancement methods are typically parametric, which means their effectiveness relies significantly on the fine-tuning of various parameters, leading to increased complexity in achieving optimal results. Recently, multi-objective optimization has gained prominence in image processing as it addresses the challenge of simultaneously improving multiple, conflicting quality criteria. Therefore, the current work is focused on image enhancement, considering two essential properties of image processing: contrast and details. Improving these properties through transformation functions is generally compromised, meaning that increasing contrast can lead to a significant loss of details in the image. Hence, a multi-objective optimization problem with two objective functions is established: one related to the image contrast and the other to the image details. These functions are measured employing the entropy and the standard deviation for the image contrast, while the pixels’ quantity and intensity in high-frequency regions measure the image details. The contrast enhancement is performed by the sigmoid transform function, and the detail enhancement is performed by the unsharp masking and highboost filtering. To address the problem, NSGA-II [
18] is used with a posterior preference articulation, meaning that specific solutions are selected once the Pareto front is finally computed. The current work offers the following contributions:
The trade-off between the image contrast and details is set as a multi-objective problem. Unlike the traditional mono-objective approach, which only provides an optimal solution with a predefined priority, the current proposal offers the best solutions regarding the compromises between both criteria along the Pareto front.
A posterior preference operator is articulated, providing three key images from the Pareto front: the image with maximum contrast, the image with maximum detail, and the image at the knee of the front, representing the image closest to the utopia point. This operator allows end-users to select the most suitable solutions for their needs.
An experiment is conducted with images of two categories: medical and natural scene images. Both categories represent research fields where image processing is an essential endeavor. The results of this experiment demonstrate that the NSGA-II achieves images of superior quality compared to the original instances. Furthermore, a thorough analysis is conducted regarding the suitability of the obtained images according to the established preferences. For medical images, the evaluation focuses on how the selected solutions enhance the clarity and detail of relevant structures, which is crucial for diagnostics and analysis. For natural scene images, the analysis shows how the solutions improve contrast and detail, making the images more visually appealing and impactful.
The remainder of this paper is organized as follows:
Section 3 outlines the sigmoid correction and unsharp masking with highboost filtering methods used in this study. In
Section 4, we present the proposed multi-objective optimization model and posterior preference articulation method.
Section 5 provides the benchmark results, including experimental design, graphical analysis, and quantitative evaluations. Finally,
Section 6 offers conclusions and suggestions for future research directions.
2. Related Work
A significant challenge for image enhancement methods is tuning their parameters to achieve optimal results. This task can be complex, as each parameter may affect the image characteristics in different ways. Evolutionary algorithms (EAs) stand out as highly effective tools in this context. Pioneer work in this area is that of Bhandarkar et al. [
19], where a genetic algorithm (GA) was employed for an image segmentation problem. The outcomes showed that the GA outperforms traditional image segmentation methods regarding accuracy and robustness against noise. In subsequent years, numerous studies proposed diverse approaches to address image enhancement problems through EAs. An example of this is presented in [
20], where an optimization-based process employs the particle swarm optimization (PSO) algorithm to enhance the contrast and details of images by adjusting the parameters of the transformation function, taking into account the relationship between local and global information. Another instance of optimization-based image enhancements is presented in [
21], where an accelerated variant of PSO is used to optimize the above-mentioned transformation function, achieving a more efficient algorithm in terms of convergence.
Using EAs to solve image enhancement problems remains a common trend in recent years. This is observed in works such as [
22], where the differential evolution (DE) algorithm was employed to maximize image contrast through a modified sigmoid transformation function. This function adjusts parameters that control the contrast magnitude and the balance between bright and dark areas, with optimal values determined through the evolutionary process. Similarly, Bhandari and Maurya [
23] developed a novel optimized histogram equalization method, preserving average brightness while improving the contrast through a cuckoo search algorithm. The proposed method uses plateau boundaries to modify the image histogram, avoiding the extreme brightness changes often caused by traditional histogram equalization. Another interesting application integrating EAs and image histograms is presented in [
24], where a GA was applied to optimize a histogram equalization through optimal subdivisions considering different delimited light exposition regions. Particularly, these optimization-based strategies have been taken to fields beyond engineering, like [
25], where directed searching optimization was applied in a medical image enhancement process, improving the contrast while preserving the texture in specified zones through a threshold parameter. The popularity of EAs has led to more sophisticated approaches, as in [
26], where hybridization between whales optimization and chameleon swarm algorithms was proposed specifically to find the optimal parameters of the incomplete beta function and gamma dual correction. Several other EAs have been applied to image enhancement problems, such as monarch butterfly optimization [
27], chimp optimization algorithm [
28], sunflower optimization [
29], and slime mold algorithm [
30], among others. However, despite their contributions to notable improvements in image processing, these approaches primarily focus on maximizing or minimizing a single criterion. As a result, single-objective methods may inadequately address the complex relationships among various image characteristics, thereby limiting their effectiveness in real-world applications, where multiple criteria must be optimized simultaneously.
In the last decade, multi-objective optimization in image processing has also been the subject of several investigations. The relevance of the multi-objective approach lies in the need to balance multiple quality criteria simultaneously. In many cases, improving one image characteristic may worsen another. This inherent conflict between particular objectives requires an approach that obtains a set of optimal solutions, known as the Pareto front. For example, in [
31], a PSO variant was proposed to address a multi-objective problem aimed to simultaneously maximize the available information quantity (through the entropy evaluation) and minimize the resulting image distortion (measured by the structural similarity index). Similarly, in [
32], a multi-objective optimization using PSO is implemented to simultaneously optimize brightness, contrast, and colorfulness in a Retinex-based method. Another instance of multi-objective image enhancement problems is presented in [
33], where GA was employed to maximize the Poisson log-likelihood function (used to measure the quantitative accuracy) and the generalized scan-statistic model (measures the detection performance). In [
34], a multi-objective cuckoo search algorithm was used to enhance contrast by maximizing entropy and minimizing noise variance in adaptive histogram equalization. Another trade-off problem of image enhancement was solved in [
35], where through the Non-dominated Sorting Genetic Algorithm based on Reference Points (NSGA-III), the optimal parameters for anisotropic diffusion were found, aiming to produce effective filtering results while balancing the image noise and its contrast.
While the Pareto front approach provides a diverse range of solutions in image enhancement tasks, it often lacks mechanisms for preference articulation, which is essential for decision makers to select the most appropriate solution based on specific application needs [
36,
37,
38]. Despite its importance, the explicit application of preference articulation in the literature on multi-objective image enhancement remains limited. This work proposes a multi-objective approach to simultaneously enhance contrast and details, as these two parameters have primarily been addressed through single-objective methods. Consequently, there is a lack of models that consider their interdependence. Additionally, this approach incorporates preference articulation to facilitate the selection of a limited set of images that effectively capture the diverse characteristics within the dataset.
4. The Proposed Algorithm
This section presents the novel multi-objective optimization model and a posterior preference articulation method developed specifically for this study. This method is designed to find optimal trade-offs between the objective functions while providing preferential solutions capable of capturing different characteristics of the images. Furthermore, an analysis of the complexity of the proposed algorithm is conducted to evaluate its efficiency in handling the optimization process.
4.1. Multi-Objective Optimization Problem
This work focuses on enhancing images by considering two fundamental properties of image processing: contrast and details. The enhancement of these properties through transformation functions is often compromised, meaning that an increase in contrast can result in a significant loss of details in the image. Therefore, a multi-objective problem is defined with two objective functions:
Contrast Function: Defined as the product of entropy
and the normalized standard deviation of the pixel intensities
of the
I image. The contrast function is expressed as:
Details Function: Defined as the product of the number of pixels in high-frequency regions
and the intensity of these high-frequency pixels
of the
I image. The details function is denoted as:
Formally, the multi-objective optimization problem can be expressed as in (
8), where
represents the set of objective functions to be minimized. The decision vector
controls the behavior of the sigmoid correction function, with
as the lower and
as the upper boundary.
Before evaluating these objective functions, a series of image transformations are performed. Contrast enhancement is achieved using a sigmoid transformation function, where
and
are the decision variables that control the degree of contrast adjustment. Additionally, UMH is applied for detail enhancement. These transformations impact the objective functions by modifying the entropy, standard deviation, number of high-frequency pixels, and intensities.
Figure 2 illustrates the overall procedure for solving the multi-objective optimization problem using NSGA-II.
The selection of NSGA-II in this study is based on several algorithm strengths. Firstly, its elitist mechanism ensures the preservation of the best solutions across generations, preventing the loss of optimal solutions. Additionally, the crowding operator used by NSGA-II eliminates the need for sensitive parameters associated with niche techniques, simplifying the implementation process. Furthermore, NSGA-II exhibits low time complexity relative to other multi-objective evolutionary algorithms, which is advantageous for practical applications. It has also demonstrated efficient performance in multi-objective problems involving two objective functions, making it relevant for contrast and detail enhancement tasks. Although other more recent multi-objective evolutionary algorithms have succeeded in generic benchmarks, NSGA-II remains highly competitive in solving real-world problems [
41,
42], making it suitable for this work.
4.2. A Posterior Preference Articulation
The main objective of preference articulation is to select a limited set of images that effectively capture diverse characteristics within the data, which may be relevant across various contexts. This approach is particularly important given that the Pareto front can encompass as many solutions as there are members in the population or even more when using an external archive. This work focuses on choosing the extreme solutions of the Pareto front and its knee point. The knee point is identified as the solution on the Pareto front that is closest to the utopia point, which typically represents the ideal but unattainable objective values. Within the optimal trade-offs of the Pareto front, these three solutions represent opposing balances of contrast and detail, helping to capture essential image characteristics that may be relevant according to the specific requirements of each application.
To determine the knee point, the Pareto front is first normalized, ensuring that the values of each objective function fall within the range [0, 1]. The utopia point in this normalized space is a vector of zeros, representing the best possible values for each objective. The Euclidean distance of each solution on the normalized Pareto front to the utopia point is calculated as follows:
where
is the number of objective functions;
is the normalized value of the
j-th objective for the
i-th solution; and
is the value of the
j-th objective at the utopia point (typically 0 after normalization). The knee point is identified as the solution with the minimum distance to the utopia point.
4.3. Time Complexity
The most computationally expensive process in NSGA-II is the non-dominated sorting, which is used to categorize the individuals in the population [
18]. This step has a time complexity of
, where
M is the number of objectives and
N is the population size. Additionally, the population’s objective functions evaluation has a time complexity of
, where
F represents the time required to evaluate an individual. Since
F depends on the specific application and can vary, the total complexity of NSGA-II is expressed as
.
In this approach, the evaluation of each individual includes several image processes. The sigmoid function is applied to each pixel, resulting in an complexity, where P is the total number of pixels in the image. The smoothing filter has a complexity of , where is the filter size, implying that operations must be performed on neighbors. However, this term can be disregarded since k is a constant. The unsharp mask process, the application of detail thresholding, and the enhancement of details also have an complexity. Finally, calculating quality metrics, such as entropy and standard deviation, requires traversing each pixel, adding to the complexity. Thus, the total time complexity for evaluating the objective function is , meaning that the execution time grows linearly with the image size. Therefore, the overall complexity of the proposal is .
To empirically validate this, we conducted an experiment using different image resolutions, as shown in
Table 1. The results demonstrate the linear relationship between image size and execution time. This behavior is fundamental when handling high-resolution images, ensuring that processing remains feasible even as image size increases. The experiments are executed on a machine with an Apple M2 chip, an 8-core CPU, and 8 GB of unified memory.
Regarding implementation complexity, the proposed algorithm applies the correction methods before each objective function evaluation. At the end of the generations, preference articulation selects extremes and knee solutions from the Pareto front. These operations do not disrupt the conventional operators of the algorithm, enabling straightforward implementation in other evolutionary algorithms. To effectively support the preference articulation process, it is recommended that the algorithms be based on Pareto dominance.
5. Benchmark Results and Discussion
In this session, the fundamental aspects of the experiment are described. Subsequently, the results are presented and discussed in two parts: first, the visual analysis of the resulting images is conducted, followed by a discussion of the numerical results from the images concerning well-established indicators.
5.1. Experimental Design
A set of twenty images is selected to assess the effectiveness of the method developed in this work. This set is divided into two groups: the first includes 10 natural scene images extracted from the Kodak dataset [
43], specifically from kodim01 to kodim10 (hereinafter referred to as Natural1 to Natural10, respectively). The second group consists of 10 medical images (referred to as Medical1 to Medical10, respectively) selected from various libraries, including brain images [
44,
45], blood composition images (white blood cells of the basophil and eosinophil types) [
46,
47], X-rays [
48], ocular nodules [
49], dental infections [
50], microphotographs of pulmonary blood vessels [
51], and traumatic forearm positioning [
52].
The NSGA-II algorithm is executed for each image with a maximum of 30,000 function evaluations, aiming to produce a Pareto front containing the best solutions. From this front, solutions are extracted according to the defined articulation preference operator. The objective is to evaluate the quality of these solutions in terms of contrast and details, complemented by a visual analysis to determine the suitability of each image for specific purposes. Finally, the similarity of the enhanced images to the original ones is assessed using the structural similarity index (SSIM), which quantifies the degree of similarity between the processed and original images.
The parameter values used in this experiment are as follows: population size (), number of variables (), number of objective functions (), number of evaluations ( = 30,000), mutation probability (), crossover probability (), simulated binary crossover parameter (), polynomial mutation parameter (), details threshold (), lower bound (), and upper bound ().
5.2. Graphical Results
Table 2,
Table 3,
Table 4 and
Table 5 present the results obtained through the multi-objective optimization image enhancement approach. Specifically,
Table 2 and
Table 3 show the results for natural images, while
Table 4 and
Table 5 display medical images. The tables are organized as follows: the first and second columns list the image names and their corresponding original, unenhanced versions. The third to fifth columns showcase the selected points from the Pareto front, representing the maximum contrast, knee point, and maximum detail, in that order. The final column illustrates the obtained Pareto front through the optimization process, with red, green, and orange points indicating the images that achieved maximum contrast, knee point, and maximum detail, respectively.
As observed in the results, the images extracted from the Pareto front significantly maximize contrast and detail compared to the original images. In all study cases, the original image is dominated by the solutions extracted from the fronts, demonstrating the approach’s effectiveness in improving visual quality. However, the differences among the three enhanced images for each problem require a more detailed analysis.
In the natural images, the differences among the three preferred images are more subtle, given that these are high-quality images with inherently low contrast, specifically selected for contrast enhancement exercises. More pronounced differences are observed in Natural1, Natural6, Natural7, and Natural8 images. For the Natural7 image, there is a general improvement in overall details. However, specific regions, such as the highlighted flower in the yellow box, may lose details compared to the original image, which retains more information. This suggests that for future work, it may be advisable to apply local and/or adaptive image enhancement techniques to preserve details in specific regions while maintaining overall image quality.
For medical images, there are instances where differences are more perceptible. For example, in the Medical3 image, the maximum contrast solution makes it difficult to visualize the internal details of the basophil (a white blood cell highlighted in the box), which could result in a less accurate interpretation. In contrast, the knee and maximum detail solutions provide a clearer view of the interior of the white blood cell. Similarly, in the Medical5 image, the maximum contrast solution highlights the hand and arm bone structures. However, the maximum detail image offers a more precise view of the internal structures within the bones (see the highlighted region), which is crucial for a more detailed evaluation. Another notable example is the Medical8 image, where the maximum detail solution offers a more detailed view of the internal structure of the eosinophil (another type of white blood cell). However, the maximum contrast image improves the visibility of red blood cells. As shown in the yellow box, this solution reveals a red blood cell that is nearly imperceptible in the other solutions. An interesting case is the Medical6 image, where only a few non-dominated solutions are present on the Pareto front. Despite the similarities among the preferred solutions, the nodules are much more perceptible in the enhanced images than in the original image, as observed in the highlighted region.
The solutions extracted from the Pareto front represent optimal trade-offs between contrast and detail. For natural images, these three alternatives can be considered useful based on aesthetic criteria or in subsequent automatic processes that require prioritizing one property over another. In the case of medical images, these alternatives allow for a more precise evaluation suited to different diagnostic needs, providing a flexible approach to enhance the visualization of critical details according to the clinical context.
5.3. Quantitative Results
Table 6 and
Table 7 present noteworthy information regarding several criteria following the next structure. For each image whose name is presented in the first column, a set of three rows displays the outcomes of the Pareto front’s maximum contrast, knee point, and maximum details solutions. The results per individual regarding their entropy, normalized standard deviation, number of pixels, and pixel intensity are displayed from the third to the sixth column. The seventh and eighth columns display the objective function values of the multi-objective optimization problem. The last column displays the SSIM with respect to the original images, where all those images archived values above 0.7, i.e., SSIM > 0.7, are in boldface, implying that these images accomplished the enhancement with an acceptable similarity to the original image.
As can be seen, the maximum contrast solutions generally yield higher entropy and normalized standard deviation, indicating a broader range of pixel intensities and greater variability in the enhanced images. In contrast, maximum detail solutions focus on enhancing the finer details within the images. This often results in lower entropy and normalized standard deviation compared to maximum contrast solutions but may increase the number and the intensity of high-frequency pixels (indicating more detailed textures). These results highlight differences in contrast and detail between the solutions extracted from the Pareto front that may not be perceptible in the previous visual discussion. If we examine some cases, we can find images such as Natural2, Medical1, and Medical4, where the extreme points on the Pareto front do not show a visual difference. However, their associated values for contrast and detail exhibit numerical differences.
The reported values of the objective functions show that all exposed solutions are non-dominated, indicating that they represent an optimal trade-off between the two evaluated criteria. Regarding the SSIM index, 65% of the solutions exhibit values above 0.7, indicating a generally high level of structural similarity. Furthermore, by only analyzing the medical images, 85% of the solutions reached SSIM outcomes beyond 0.7, indicating that the proposal is a trustworthy tool for dealing with this kind of information. Nonetheless, if only the natural scene images are evaluated, the number of solutions that archived this SSIM outcome decreases to 50%. This may influence the artificially imposed low contrast in this set of images. Consequently, future work should consider incorporating SSIM as an additional objective function, especially for problems where image fidelity is crucial.
5.4. Image Quality Metrics
Besides the analysis related to the optimization problem, an evaluation of well-recognized image enhancement metrics is conducted.
Table 8 displays the metrics evaluations related to the natural set (left side) and medical image set (right side). The particular points of interest (Max. contrast, Knee, and Max. Detail) are divided by rows on each set. The study encompasses the Contrast-to-Noise Ratio (CNR), Blind/Referenceless Image Spatial Quality Evaluator (BRISQUE), and Natural Image Quality Evaluator (NIQE), which are tagged in that order in the column headers. Next, a brief explanation of how these metrics are addressed is offered.
The Contrast-to-Noise Ratio (CNR): This metric quantifies the contrast between a signal and the background. The higher the CNR outcome, the better the image enhancement. Values from zero to one indicate a poor contrast, while ranges from one to three indicate a moderated contrast, and greater values indicate a good contrast, implying that it is easy to identify features within the image [
53]. This metric is reference-based, i.e., it compares the enhanced image w.r.t. the original one. This work uses Otsu’s method to compute the threshold to differentiate noise from signal [
54].
The Blind/Referenceless Image Spatial Quality Evaluator (BRISQUE): This is a no-reference-based metric computed directly from the enhanced image. It measures the degradation of the image quality after an image processing task. Based on statistical approaches, it compares the image against a Gaussian mixture model that scores the image, where the lower the score, the better the image quality. Values lower than twenty mean a high-quality image, values from twenty to forty reflect good image quality, values from forty to sixty indicate fair image quality, and greater values reflect poor-quality images [
55].
The Natural Image Quality Evaluator (NIQE): This is also a no-reference-based metric that directly assesses the quality of images by comparing them to a statistical model built from natural images instead of a Gaussian mixture model. The lower the NIQE score, the better the image quality. Scores below five indicate high-quality images, a range between five and ten reflects good image quality, and greater values indicate poor quality [
56].
According to this information, on each column (evaluated metric) of
Table 8, the results that reflect high-quality images are in boldface. The following highlights are offered to provide a straightforward understanding of the outcomes.
Regarding the CNR metric, 93.33% (28/30) of the enhanced natural images achieved good noise contrast, indicating clarity in the obtained images. Similarly, 90% of the enhanced medical images (27/30) reached good noise contrast.
Regarding the BRISQUE criterion, 23.33% (7/30) of the enhanced natural images can be considered high-quality images, while the rest accomplish good quality standards. Interestingly, the enhanced medical images achieved the same number of high-quality images.
In terms of the NIQUE metric, 80% (24/30) of the enhanced natural images reached high-quality standards. On the other hand, only 73.33% (22/30) of the enhanced medical images reach this standard. Nonetheless, the rest of the images of both sets possess good-quality features.
These results demonstrate that the current proposal not only improves image contrast and details while providing a range of trade-off solutions but also preserves image quality and naturalness without introducing noise. This is particularly significant, as these criteria were not explicitly chosen for the optimization problem.
5.5. Comparison with Baseline Methods
In this section, a comparison is made between the results obtained by the proposed multi-objective approach and two well-known contrast enhancement methods: contrast stretching (CS) and adaptive histogram equalization (CLAHE), as illustrated in
Table 9. The first advantage of the multi-objective approach is that it produces a set of optimal solutions, each offering a different balance between contrast and detail. On the other hand, traditional methods only provide a single solution. For comparison, we selected the knee points from the Pareto front generated by the proposed method.
The comparison is performed by assessing the dominance of the obtained solutions in terms of contrast () and details (). If the solution from the multi-objective approach outperforms the traditional methods in both criteria, it is considered dominant (denoted by status “+”). If it improves only one of the two criteria, it is considered non-dominated (denoted by status “”). If the traditional method outperforms both criteria, the multi-objective solution is dominated (denoted by status “-”).
The results show that the multi-objective approach dominated CS in 19 images, being dominated only in the case of Medical5. Although CS is a classical method that adjusts the histogram to improve the overall contrast of an image, the multi-objective approach demonstrates superior performance as it focuses on enhancing the contrast–detail balance. Compared with CLAHE, 13 non-dominated solutions were obtained, 6 where the proposal dominates, and 1 where CLAHE outperforms our approach. Adaptive equalization performs better in detail enhancement (), which explains the lack of clear dominance in several cases compared to the multi-objective approach. Adaptive correction methods generally have advantages in highlighting contrast and details over general approaches. However, the optimization-based proposal can outperform the adaptive approach in several images while maintaining a competitive performance in most images, demonstrating the proposal’s robustness and flexibility in various scenarios.
To complement the analysis, a Wilcoxon rank-sum test was conducted to assess the statistical significance of the differences between the methods. The p-values indicate the following:
CS vs. Multi-objective approach (Contrast): , showing a significant difference in contrast performance, favoring the multi-objective approach.
CS vs. Multi-objective approach (Details): , suggesting no significant difference in detail enhancement.
CLAHE vs. Multi-objective approach (Contrast): , demonstrating a statistically significant improvement in contrast with the multi-objective approach.
CLAHE vs. Multi-objective approach (Details): , indicating no significant difference in detail enhancement between CLAHE and the multi-objective approach.
These statistical results reinforce the superiority of the multi-objective approach in improving contrast, particularly when compared with CS and CLAHE, while its performance in detail enhancement remains competitive.
6. Conclusions
The conflict between contrast and detail in image processing is presented as a multi-objective problem. This approach obtains a set of optimal solutions, forming a Pareto front in all cases, highlighting the trade-off between these two properties. Therefore, it is demonstrated that a single-objective approach to this problem will only lead to a particular solution among all the optimal solutions obtained through the multi-objective approach.
The proposed model integrates the sigmoid transformation function and UMH into the NSGA-II. Additionally, a posterior preference articulation is added, which selects three key solutions from the Pareto front: the maximum contrast solution, the maximum detail solution, and the knee point solution. These three solutions showed significant superiority in terms of contrast and detail compared to the original images. Furthermore, the outcomes visually and numerically demonstrated how these three image solutions, though all optimal solutions, differ in terms of entropy, standard deviation, number of detail pixels, and detail intensity. This variability allows fundamental characteristics to emerge across the images, underscoring the relevance of the proposed preferences across various contexts. Moreover, the proposed method demonstrated its effectiveness against traditional contrast enhancement techniques, such as contrast stretching and adaptive histogram equalization, by achieving a superior balance between contrast and detail, as evidenced by the dominance of its solutions in several cases.
A post hoc analysis regarding popular image quality metrics, such as CNR, BRISQUE, and NIQE, showed that part of the images created through the current proposal achieved high-quality image standards, while none of the generated enhanced images decreased below good-quality standards. These results demonstrate that the method is a trustworthy tool for image enhancement, offering a range of solutions that not only meet diverse user needs but also consistently maintain high-quality outcomes
Despite the overall improvement in image details, specific regions may lose details compared to the original image. This suggests that for future work, it may be advisable to apply local and/or adaptive image enhancement techniques to preserve details in specific regions while maintaining overall image quality. Future reserach should develop adaptive preference articulations that can identify all solutions from the Pareto front that reveal singular information within the image. Moreover, a broader investigation into the method’s versatility is planned, including experiments with alternative evolutionary algorithms. Furthermore, multi-objective micro-evolutionary algorithms are a promising and challenging research path in image enhancement tasks. Their potential lies in reducing computational overhead, though they also face challenges related to limited solution diversity and the risk of premature convergence.