1. Introduction
In the last few decades, we can observe significant development of the image fusion theory. The main task of image fusion is to combine relevant data from different source images to generate a single image that contains richer information [
1]. The most important part of this process is the effective extraction of image features and the use of appropriate fusion principles, which allow extracting useful information from source images and integrating it into the fused image without introducing any artifacts [
2]. For example, fusing a panchromatic (grayscale) photo presented in
Figure 1 and a multispectral (color) photo (
Figure 2) is used when images show the same view. If the panchromatic image shows more details than the multispectral one, it is worth combining them, getting one color photo of good quality (more details will be visible).
Figure 3 shows the effect of this kind of operation.
The presented example is only one of the various types of problems. Among them, we can highlight methods for sparse representation, multi-scale transformation, subspace, variation, neural network, saliency detection, and mixed models [
4]. A combination of useful information from source images is very beneficial for subsequent applications and is widely used in such fields as photography visualization [
5,
6,
7,
8], object tracking [
9,
10], medical diagnosis [
11,
12], and remote sensing monitoring [
13,
14].
Recently, among image fusion methods, the emergence of algorithms based on fractional differential calculus can be observed [
15,
16,
17,
18,
19]. They are applied in such problems as fingerprint detection [
20], exposing of essential elements in medical images [
21], brain photo analysis [
22], elimination of noise in images (improved image quality) [
23], or contrast enhancement [
24]. As the fractional derivative started to be used relatively recently, new applications are still being found.
The adaptation of the fractional calculus to image processing problems forced researchers to develop discrete 2D approximations of fractional derivatives operators. In the literature, we can find many different definitions, both for continuous functions and their discrete approximations. However, some discrete approximations were not correctly derived. In effect, the operators that only imitate, but are not actually, the fractional operators were proposed. The literature also shows deficiencies in the analysis of individual approximations, thus answers to the questions of which of them should be used, for which problem, and which order of the derivative should be applied are essential.
In this paper, we present an analysis of the method that uses a fractional order differential to multispectral images fusion [
15]. We analyze the correct basic definitions of the fractional order derivative. Original and corrected versions of the masks used in [
21] have been proposed. On the experimental side, various methods of determining derivatives of digital images are tested for different datasets than the one used in [
15], and the influence of fractional order change on the quality of fusion is presented. Additionally, achieved results are compared with the results obtained for methods where the integer order derivatives have been used. This novel comparison shows pros and cons of the application of fractional calculus in the context of image fusion as stated in
Section 5, and motivates further studies.
2. Fractional Order Derivative
The non-integer (fractional) order derivative is a generalization of the traditional integer derivative of the n-th order (, where is a set of natural numbers) to the real or even complex order. This is a topic of mathematical analysis that can be solved in different ways. Fractional order calculus is quite a rapidly developing field nowadays, finding applications in more and more new areas.
Although this calculus was discovered over three centuries ago, it remained only a purely mathematical notion that had few or no applications for a long time. However, in the second half of the 20th century and at the beginning of the 21st century, it began to be recognized that it could be used to solve real problems. New applications for these tools were found, and more and more different models of fractional order derivatives began to emerge.
Today, fractional calculus plays an essential role in control theory, viscoelasticity, heating processes, heat conduction, biotechnology, and particle physics. The application of this calculus in image processing is also an issue that has been intensively researched in recent years. The use of some models has shown that satisfactory results have been obtained, for example, in medical image processing [
21] or photo-based fingerprint recognition [
20]. This calculus is used in problems where there is a memory effect since having memory is a feature of fractional order operators [
21,
25].
One of the challenges with the non-integer derivatives remains their physical interpretation. Contrary to integer order, the physical meaning of fractional or even real order is not entirely clear.
Another problem is that fractional order derivatives lack local character. The result of the derivative of a function depends on the entire function. This causes a significant increase in the number of necessary calculations compared to the integer order derivatives. This results in the increase of time needed for determining this type of operator.
Numerical algorithms designed to derive non-integer order derivatives contain critical, nested loops, and their complexity increases with the increasing iteration number. Programming languages such as Matlab and Python are very often used in image processing. However, these are interpreted languages and are not efficient at handling nested loops.
There are methods to speed up the calculations. One of them is the “short memory principle” [
26]. It allows for getting an approximate result. Its use makes it possible not to consider the distant values of the function when calculating the derivative at a given point. At the expense of speeding up the operation, the accuracy of the final result is reduced.
Another method that can speed up the computation of non-integer derivatives is Oustaloup’s [
27] method. This method allows for obtaining a correct result in a predefined frequency band. The precision of the approximation is also limited to a narrow frequency spectrum. If this band is extended, the uncertainty of the obtained result increases.
Contrary to integer order derivatives, there is no single definition for non-integer order. Many definitions have been developed, but the most frequently used are three of them: Riemann–Liouville, Caputo, and Grünwald–Letnikov. Each of these definitions has advantages and disadvantages that will be discussed. For a broad class of functions, the definitions of Riemann–Liouville and Grünwald–Letnikov are equivalent. This makes it possible to use the Riemann–Liouville derivative definition at the beginning when defining the problem and then apply the Grünwald–Letnikov definition to obtain a solution.
2.1. Grünwald–Letnikov Definition
This definition was proposed in 1867 [
28]. For order
, the fractional derivative looks as follows [
29]:
where:
, and
However, definition (
1) based on the limit is practically useful only for a finite-difference implementation, hence the definition below is used:
where:
,
This definition is often used to determine the exact fractional derivative of a function. The achieved result is later used to determine the uncertainty obtained in calculating the derivative of the discrete approximation. If is sufficiently smooth, , then the Grünwald–Letnikov derivative is equivalent to the Riemann–Liouville definition.
In the literature, we can find GL definition in the following form:
where:
is a gamma function.
2.2. Riemann–Liouville Definition
The definition of the Riemann–Liouville derivative for a real order greater than 0 of the function
takes the form [
30]:
where:
is a gamma function.
The advantage of this definition is that the function under study does not have to be continuous at the origin. It does not have to be differentiable either [
31].
2.3. Caputo Definition
This definition was introduced by Michele Caputo in 1967 [
32]. Unlike the derivative calculated from the Riemann–Liouville definition, in this case, we do not need to define the fractional initial conditions. For a real order
, Caputo’s definition of the non-integer order derivative is as follows [
33]:
where:
is a gamma function.
The value of the derivative of the non-integer order based on Caputo’s definition satisfies the important relationship:
where
A is a constant value. The great advantage of this method is that it includes integer initial and final conditions.
When modeling non-integer order derivatives systems, it is possible to use more than one definition. A combination of the definitions of Riemann–Liouville and Grünwald–Letnikov is often used. In addition to the mentioned definitions, many others have also been created. However, they are not used that often. De Oliveira et al. collected and described many different definitions for non-integer order derivatives [
34].
2.4. Derivatives in Image Processing
The derivative determined on the images allows for studying the changes in brightness or color in the image. The natural application of this operator is edge detection. In this case, we need to find the derivative in two orthogonal (perpendicular) directions. They do not necessarily have to be vertical and horizontal. For example, it is possible to calculate diagonal derivatives of an image. The calculation of the first derivative of an image belongs to the group of gradient methods. They are a collection of the simplest edge detection operations to extract edges and remove the rest of the image.
The magnitude of the gradient is proportional to the rate of increase of the image function value for a given point. It also indicates how expressive the edge is. The higher the value, the more visible it is. If we want to define all the edges in the image, we should assume some limit value. For pixels whose magnitude is greater than or equal to this value, we will consider them as an edge.
Images are not continuous functions, and it is impossible to change the argument
t by an infinitely small amount. Thus, they must be considered as discrete functions, and it is necessary to use an approximate version of an operator that will be able to act on a discrete function. The minimum value we can move in the image is one pixel. Thus, the formula for the derivative of an image takes the form:
Based on this relationship, it is possible to create many different methods of determining the gradient.
2.4.1. Integer Order Derivative Mask
Many methods for determining the integer order derivatives in the images have been proposed. One such operator is Sobel’s mask [
35]. It is characterized by the fact that, during averaging, weights are used, giving the analyzed point the highest value. Sobel’s mask has the following form:
The second order derivative can be approximated by using Laplace mask:
Based on the selected definitions of calculating fractional order derivatives, various methods are proposed that can be used in image processing. Some of them make it possible to build an appropriate mask approximating the integer order of the derivative. Other methods transform images directly. Unfortunately, many of the proposed models do not designate non-integer derivatives, despite being described in this way.
2.4.2. Mask Based on Riemann–Liouville Definition
Amoako-Yirenkyi et al. [
21] proposed a mask based on the generalization of Riemann–Liouville definition and for any order
. However, apparently their derivation contains a mistake. In this paper, we present a correct derivation based on a standard Riemann–Liouville definition presented in
Section 2.2, and for order
.
For an analytical function
, such that
and
, a derivative operator is defined as:
By focusing on the integral expression in (
10), we can write:
and
This equation is for a one-dimensional function, but because an image has two dimensions, we have to transform this formula into a two-dimensional form by putting
. Finally, determining the directional derivatives with respect to
x and
y, we get the formulae for the gradient mask elements by finding the derivative in the horizontal and vertical directions, as follows:
where
and
with
is the size of the mask for every
, and
is a constant parameter.
The determined mask of the size
, proposed in [
21], for the horizontal direction takes the form:
while for the vertical direction:
Analyzing the masks Equations (
15) and (
16), it can be seen that the proposed masks do not follow Equations (
13) and (
14). Such inconsistency causes the final mask formulae to be contradictory to the reasoning coming from the definition of Riemann–Liouville fractional derivatives described in the formula (
4). The proper masks of the size
should have the following forms:
while for the vertical direction:
We compare the results obtained for correctly determining masks in the conducted experiments. Still, we also analyze the results obtained for the incorrect masks to assess the impact of this error on the results achieved.
2.4.3. Eight-Directions Non-Integer Order Mask
In [
15], a method was described that allows for building an eight-direction mask approximating non-integer order derivatives. This model is based on the Grünwald–Letnikov definition presented in (
3).
For images, the minimum value of
h change is equal to one pixel. After differentiating the finite function
with respect to
x, based on the expression (
3), one can get the Formulae (
19) and (
20):
Based on the above expressions, it is possible to build masks determining approximations of non-integer derivatives in eight different directions. Following [
15], the final mask composed of all eight directional masks should be more robust to image rotation due to its symmetry. The mask components for eight directions appear as follows:
Derivative mask in direction :
Derivative mask in direction :
Derivative mask in direction :
Derivative mask in direction
Derivative mask in direction left upper diagonal:
Derivative mask in direction left lower diagonal:
Derivative mask in direction right upper diagonal:
Derivative mask in direction right lower diagonal:
Based on the presented mask components, the resultant eight-direction mask can be obtained: where is a derivative order.
This mask appears to approximate fractional derivatives and will be further applied to the image fusion study. Its use is presented in [
15] and may be a good reference point for other methods.
2.4.4. FFT Approximation of a Fractional Order Derivative
In [
36], a method for determining the fractional order derivative was shown. It is based on the Riemann–Liouville definition of the following form:
where
is a fractional order of differ-integral of the function
and, for
, we have:
n for
,
for
The Fourier transform of the Riemann–Liouville fractional derivatives with the lower bound
is equal to:
For any two-dimensional function
absolutely integrable in
, the corresponding 2D Fourier transform is as follows [
30]:
Therefore, we can write the formulae for fractional order derivatives as:
where
is the inverse 2D continuous Fourier transform operator.
Finally, the result of image fractional derivative will be determined by the real part of the sum of the inverse transforms:
where
ℜ is a real part of a complex function.
3. Methods and Metrics Used
For our experiments, we used the component substitution method proposed in [
37]. The basis of this image fusion method is to extract the details of images by determining the difference between a panchromatic image and a linear combination of low-quality multispectral image channels. This method is effective when two combined images contain almost the same information.
The image fusion algorithm without a fractional order derivative was proposed in [
38] and can be written as follows:
where
and
is the
k-th band of the fused image,
is
k-th band of the low-resolution multispectral image,
represents the band weight, and
is a constant gain determined according to the relationship from [
39]:
for
, P is a panchromatic image, I is a linear combination of a low-resolution multispectral image,
N indicates the number of bands covering the spectral signature of the panchromatic image,
denotes covariance between the
X and
Y images, and
is a variance of image X. The band coefficients are determined in accordance with the AIHS approach (Adaptive Intensity-Hue-Saturation) [
40], which achieves good results in this issue.
Azarang et al. in [
15] proposed a modification of this method. Using the non-integer order derivative of the difference (P–I) should strengthen the edges and improve the fusion results. The modified form of the algorithm (
26) can be written as follows:
where
m is a proposed mask and ∗ is a convolution operator.
Metrics
A problem with comparing the quality of image fusion results is the lack of an appropriate metric. Our experiments described in
Section 4 support the claim that commonly used metrics such as the ones defined below are not always able to clearly determine which approach results in better visual quality of an image. Because of this, in the experiments, a selection of metrics is used.
To evaluate results, the following metrics were applied:
Root Mean Square Error (RMSE)—measures the changes in pixel values of the input band of the multispectral image
R and the sharpened image
F. This metric is fundamental if the images contain large, uniform areas. This error is calculated using the following formula [
41]:
Ideal value of this error is 0.
Relative dimensionless global error in synthesis (ERGAS) is a global quality factor. This error is sensitive to the change in the average pixel value of the image and the dynamically changing range. Its value tells about the amount of spectral distortion in the image. It is expressed as [
41]:
where:
is the ratio of the number of pixels of a panchromatic image to the number of pixels of a multispectral image,
is the mean of
-th band, while
N is the total number of bands. In the case of this metric, we aim for the error rate to be close to zero.
Spectral angle mapper (SAM) calculates spectral similarities by finding a spectral angle between two spectral vectors with a common origin. The length of a spectrum vector
is calculated by the following equation:
and the spectral angle
is calculated as [
42]:
where:
is the length of the endmember vector,
is the length of the modeled spectrum vector,
—reflection of the endmember. In the case of this metric, we aim for the error rate to be close to zero.
Correlation coefficient (CC)—shows the spectral correlation between two images. The value of this coefficient for the sharpened image
F and the input multispectral image
R is determined as follows [
41]:
and mean the average values of the F and R images, while m and n denote the shape of the images. The optimal value for this coefficient is 1.
Universal Image Quality Index (UIQI) is defined as [
41]:
The first factor presents the correlation coefficient between the images R and F, while the second factor is the luminance distance, and the last represents the contrast distortion. denotes the covariance between the images R and F. and are the mean values, and denote the standard deviation values of the R and F images, respectively. The best value for this index is 1. It can be reached if R equals F for all pixels.
Relative Average Spectral Error (RASE) is computed using the RMSE value using the following relation [
41]:
where
is the mean radiance of the
N spectral bands and
represents the
-th band of the input multispectral image. An optimal value of this error is equal to 0.
4. Experiments
For the experiments, a set of photos in PNG format, taken by the satellite from the Bing maps [
3], has been used. The creator of this collection divided the data into categories such as land and water. Our experiments were focused only on a part with land-based images containing 1078 images. This collection includes good-quality panchromatic photos. Based on this collection, two image sets were prepared. The first was the result of saving color photos in grayscale, representing high-quality panchromatic photos. The second contained multispectral color photos with reduced quality due to the use of a blurring mask.
As the first experiment, the mean error values for the image fusion algorithm for the basic version, without additional derivatives [
43], were checked. Its description and an indication of where the examined modification can be applied can be found in
Section 3. The result of the experiment is shown in
Table 1. Then, for the prepared data set, the method proposed in [
15], based on a Grünwald–Letnikov definition that allows for building an eight-direction mask defined in
Section 2.4.2, approximating the non-integer order derivatives, was examined. The results achieved for this method are presented in
Table 2. The experiments with masks presented in
Section 2.4.2 were shown in
Table 3 and the improved version of this mask in
Table 4. An experiment was also carried out in which the fractional derivative was determined using the Fast Fourier transform described in
Section 2.4.4. The results of this experiment are presented in
Table 5. For comparison, the experiments with the first derivative using Sobel’s filter, and the second derivative using Laplace operator, the description of which can be found in the
Section 2.4.1, were added. The results of Sobel’s and Laplace’s filters for image fusion are presented in
Table 6.
The results presented in the tables show that the best result of fusion for 4 out of 6 metrics was obtained by using an eight-direction mask proposed in [
15] for derivative order
(see
Table 2). This result is slightly better than results obtained without fractional order derivatives (see
Table 1). It can be observed that increasing the fractional order of derivative from 0.1 to 0.9 worsens the results. However, for the order in the range <0.6; 0.9>, this mask achieved the worst results from all examined methods.
The best results of all the proposed modifications were obtained for the FFT method and the order of
, which can be seen in
Table 5. Using the FFT-based derivative, the smallest fusion errors were obtained for the order
. Up to the order of
, the error increases slightly, while, from value 0.5, the errors increase with the order, but much slower than with the original method for which errors above
are already significant.
Although the lowest error results obtained by the FFT-based algorithm is twice as high as the best results obtained with the eight-directional mask, the visual assessment of the final images calls into question the determined error values. The FFT fusion images look optically similar, just as precise, and sometimes appear sharper, which can be observed in
Figure 4. It confirms that, in image processing, better metric values do not always guarantee higher visual quality.
An incorrect mask proposed by [
21] achieved the worst results from tested fractional order operators (see
Table 3). The corrected version of this mask improved results a little bit (see
Table 4). For these two masks, we can observe that maximum errors were obtained around the order of 0.5.
The Sobel and Laplace filters to approximate the first and the second order derivative gave abysmal results for all metrics (see
Table 6).
5. Conclusions
In the paper, the improved, corrected versions of the edge detecting masks based on fractional derivative have been proposed. The masks were implemented with different approximations of this derivative. Analyzed modification of different approximations did not improve the quality of image fusion much. In some cases, the introduced changes while improving the quality metrics of the fused images worsened their visual quality for the tested set.
Analysis of achieved metrics’ values shows that the original fractional order solution achieved the best metrics for order ; however, the fusion error for orders higher than 0.6 increased dramatically. More stable results were achieved by using the FFT algorithm for computing fractional order derivatives. Despite the error of the FFT algorithm for being two times higher, visual verification indicates that this algorithm received sharper fused images. This confirms the thesis that there are currently no metrics that would unambiguously assess the quality of image fusion.
The integer order derivative method produces a result that completely disrupts the algorithm. Comparing the results with various methods for determining non-integer order derivatives shows that these methods often have a much greater chance of success in applications where standard methods for integer orders do not bring positive results. Edge reinforcement is essential in this application. Fractional derivatives accomplish this to varying degrees, creating a greater chance that a particular combination will be better suited for a given application.
The experiments presented show the potential of applying fractional derivatives in the image fusion context. However, there is still a lot of future research required. It would be worth testing the operation of the best-performing filter for different data sets. In addition, new masks based on different definitions of a fractional derivative may lead to better results. A well-known problem of image quality metrics not clearly reflecting the image’s visual quality should be addressed in future work.