Next Article in Journal
Overview of Radar Alignment Methods and Analysis of Radar Misalignment’s Impact on Active Safety and Autonomous Systems
Previous Article in Journal
Effortless Totem-Pole Converter Control Using a Power Factor Correction Peak Current-Mode Controller
Previous Article in Special Issue
Distortion Correction and Denoising of Light Sheet Fluorescence Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Focus Images Fusion for Fluorescence Imaging Based on Local Maximum Luminosity and Intensity Variance

Department of Automation, Shanghai Jiao Tong University, Dongchuan Road 800, Shanghai 200240, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(15), 4909; https://doi.org/10.3390/s24154909
Submission received: 25 June 2024 / Revised: 15 July 2024 / Accepted: 22 July 2024 / Published: 29 July 2024
(This article belongs to the Special Issue Fluorescence Imaging and Sensing)

Abstract

:
Due to the limitations on the depth of field of high-resolution fluorescence microscope, it is difficult to obtain an image with all objects in focus. The existing image fusion methods suffer from blocking effects or out-of-focus fluorescence. The proposed multi-focus image fusion method based on local maximum luminosity, intensity variance and the information filling method can reconstruct the all-in-focus image. Moreover, the depth of tissue’s surface can be estimated to reconstruct the 3D surface model.

1. Introduction

Obtaining an all-in-focus image is very important for the automation of fluorescence image collection. However, due to the limited depth of field of high-resolution fluorescence microscope, it is hard to capture an image with all objects in focus, especially for thick tissue slices [1,2].
The confocal microscope was proposed and built to adjust the stage or illumination spot to find a focus position for every pixel. These pixels can then be combined to obtain an image with all parts in focus [3,4,5]. Depth information can also be used to reconstruct the 3D model of the visible area [6,7,8]. However, it is too time-consuming to use this technique for obtaining high-resolution images.
On the other side, a simpler type of fluorescence microscope called epifluorescence microscope can obtain one image at a time, but many parts of tissue may be out of focus. So, it is reasonable to obtain images at different depths and combine those images together to obtain an all-in-focus image [9].
Normal image fusion methods have been proposed by many people. Wang et al. proposed the complex shearlet features-motivated generative adversarial network. With of help of the generative adversarial network, the whole procedure of multi-focus fusion is modeled to be the process of adversarial learning [10]. Dong et al. proposed a new image fusion framework by utilizing area-based standard deviation in the dual tree contourlet transform domain [11]. De et al. proposed a method based on wavelet transformation and maximum sharpness [12]. Li et al. proposed a method based on guided filter [13]. Guo et al. proposed a method based on self-similarity and defocus information provided by a method given by Zhuo and Sim [14,15]. The latter two methods both assign a weighted average of pixels from different source images to the corresponding pixel in the fused image. The final images fused by these methods are not clear in some cases and it is hard to obtain the depth of tissue for every pixel as in a confocal microscope.
Actually, it is reasonable to make use of the property of fluorescence images that the luminosity intensity and clarity have a strong correlation due to the architecture of fluorescence microscopes [16]. This property has also been utilized to determine the depth of a pixel in confocal fluorescent microscopes. So, we introduce the local maximum luminosity into the sharpness evaluation function. In our method, we combine it with the variance of intensity that is usually used to measure the sharpness of common images to form a new sharpness evaluation function suitable for fluorescence images [17,18].
However, unlike a confocal microscope, which has a pinhole at the detector to block out-of-focus fluorescence, epifluorescence microscopes will receive out-of-focus fluorescence which will contaminate other portions of collected images. This kind of contamination also interferes with the methods using weighted averages in the works of Li et al. and Guo et al. [13,14].
As a result of out-of-focus fluorescence, the method used for confocal microscopes does not work well for images collected using epifluorescence microscopes [19]. In order to prevent out-of-focus fluorescence, we divide the whole visible area into different rectangular blocks and use the new sharpness evaluation function we propose as a criterion of clarity. Additionally, the depth of tissue’s surface in each block can be estimated according to the criterion. Meanwhile, in order to alleviate the blocking effects introduced during block segmentation, we propose an information filling method.
According to the aforementioned ideas, a block-based image fusion method for multi-focus fluorescence microscopic images is proposed. The depth information of the tissue’s surface for each pixel can be used to reconstruct the 3D surface of observed objects when we know the stride of the image collection and parameters of the fluorescence microscope. Our contributions are presented as follows:
  • A block-based image fusion method for multi-focus fluorescent imaging is proposed, and it is based on the local maximum luminosity, variance of intensity, and an information filling method. This method benefits from the architecture of fluorescence microscopy.
  • A method of information filling for neighboring blocks is proposed to deal with the blocking effect introduced by the common block-based method.
  • The depth information of each pixel can be obtained, and it can be used to reconstruct the 3D surface of these objects within source images.
The paper is organized as follows: Our methods are introduced in detail in Section 2. The comparison of images obtained by the proposed method and other methods is presented in Section 3. The discussion and conclusion are separately shown in Section 4 and Section 5.

2. Our Method

The proposed method is a block-based image fusion method which is mainly based on local maximum luminosity, variance of intensity within blocks, and the information filling method. The proposed method gives an explicit index per pixel which contains the depth information, allowing the feasible reconstruction of a 3D surface of the observed objects. The framework of our method can be viewed in Figure 1.
Figure 1 shows that in the first step, we collect images of the same visible area at different focal planes. Then, the whole visible area is divided into nonoverlapping rectangular blocks for block segmentation. After that, the depth can be estimated for every block, and a confidence map can be established. Confidence maps are used to adjust the depth map to further prevent the contamination of out-of-focus fluorescence. In the information filling step, an information filling method is used to alleviate blocking effects. Next, smoothing is applied to the block edges in depth space and the fused image. If the objects in the visible area are not separable, this fused image will be the final result. Otherwise, we will detect the objects and determine the segmentation block sizes according to the scale of objects. For every segmentation block size, we re-divide the whole visible area into blocks and fuse the source images again to form a new fused image. On the other side, we use the results of object detection to form masks to extract the objects from the corresponding fused images. Finally, we combine the results of different objects to obtain the final fused image.
Next, we will give more detailed information about the five key steps in our proposed method including image collection and block segmentation, the rough construction of the depth map according to the proposed sharpness estimation function and the construction of confidence map, depth adjustment according to the confidence map, pixel refinement using the information filling algorithm, and an optional step to decide the block size according to an object’s scale.

2.1. Image Collection and Block Segmentation

The first step is to collect source images at different focal planes and divide the whole visible area into nonoverlapping blocks. We use an automatic fluorescence scanner developed by Powerscin when collecting the data.When the stage does not move horizontally, the area we can see in the microscope is called the visible area for the sake of simplicity. In order to make sure every part of the visible area is clear in at least one source image, the range within which we collect images is between the focal planes of the deepest and shallowest part of a tissue slice in the visible area. Images are collected at different focal planes with a step size less than the microscope’s depth of field. Here, we assume N source fluorescence images are obtained, and these images are denoted by I i . The whole visible area is divided into nonoverlapping rectangular blocks with size m × n (denoted by B j , k ). The set of pixels in an image I i within a block B j , k is called a patch in this paper (denoted by p i , j , k ).
Here, we decide to first estimate the depth of every block and then obtain the depth and value of every pixel within this block rather than trying to obtain the value of every pixel directly. The reason is that we assume the depth of the tissue’s surface does not change seriously within a block in most cases if the block size is not large, and actually estimating sharpness for a block is much faster than estimating the sharpness for every pixel in the block with a receptive field of the same size. If we want to estimate sharpness for every pixel faster and take a smaller receptive field size, out-of-focus fluorescence from other parts of the tissue will probably interfere with the sharpness estimation of parts that themselves are not luminous.

2.2. Rough Construction of Depth Map

The second step gives a rough estimation of depth for every block B j , k and builds up a confidence map C j , k . In this subsection, we first introduce how to establish the depth map and show why we choose the multiplication of local maximum luminosity and variance of intensity as our sharpness evaluation function in this step. Then, we will explain how the confidence map is established.
In this step, we use the sharpness evaluation function to estimate the sharpness of every patch p i , j , k , and the sharpness value is s i , j , k respectively. Rather than estimating the depth for every block B j , k directly, we can first determine in which image block B j , k is the clearest. Because we can know the object distances at which these images are taken, it is easy to convert the index of image to the depth. The index of image in which the block B j , k is the clearest can be represented by
index j , k = arg max i s i , j , k
Meanwhile, the depth information of every block B j , k can be inferred from i n d e x j , k .
Because of the architecture of fluorescence microscopes, the luminosity intensity and clarity have a strong correlation, and in order to make use of this property, we introduce the local maximum luminosity into the sharpness estimation function. On the other side, the variance of intensity in a block itself can be a criterion to estimate sharpness and has good performance [17,18]. So, to take both luminosity and unevenness into consideration, we choose the multiplication of local maximum luminosity and variance of intensity as the sharpness estimation function, and it can be represented by
s i , j , k = m i , j , k × var i , j , k
where m i , j , k is the maximum pixel value in patch p i , j , k and
var i , j , k = 1 m n v r , s p i , j , k v r , s μ i , j , k 2
Here, μ i , j , k is the average flux for every patch p i , j , k , and v r , s is the pixel value in the corresponding patch p i , j , k . This criterion is better than other current sharpness evaluation functions for fluorescent images. This is better than the method which selects the maximum pixel value for every pixel which is used for confocal microscopes.
In order to show the performance, we compare our proposed sharpness evaluation function with other sharpness evaluation functions. The visible area is divided into blocks with size 20 × 20. For every sharpness evaluation function, we select a patch p a t c h j , k for every block B j , k and
patch j , k = p index j , k , j , k
Then, we stitch patches together to obtain a fused image for each sharpness evaluation function. Additionally, we show the result of the method used for confocal microscopes which selects the maximum pixel value for every pixel. We show the comparison in Figure 2. The images in the left column of Figure 2 show the whole fused images, and the images in the center and right columns of Figure 2 show two typical regions extracted from the whole visible area.
Figure 2a–c give the performance of maximum total luminosity method to extract the brightest values for every pixel, and the method is used in confocal microscopes [3,4]. We can see the texture contaminated by the out-of-focus fluorescence. The following images are reconstructed by selecting the patches p a t c h j , k for the corresponding sharpness evaluation functions.
Figure 2d–f use the Laplacian measure which is used by Zhuo and Sim as the criterion of clarity. The textures are not clear due to the contamination [15].
Figure 2g–i use the variance measure as a typical criterion of clarity [17,18]. This method has good performance for the most part, but it is not robust and the reconstructed image lacks continuity.
Figure 2j–l use the normalized-variance measure as the criterion of clarity [17,18]. It is even less robust than the variance measure.
Figure 2m–o use local maximum luminosity as the criterion of clarity. It obtains good performance for the most part and is robust. It also shows maximum continuity and good performance, but the textures in some parts are less clear than those of the variance measure.
Figure 2p–r use the proposed multiplication of the local maximum luminosity and intensity variance in the block as the criterion of clarity. It shows good performance and is very robust.
It is obvious that the performance of the multiplication of local maximum luminosity and variance of intensity is the best sharpness evaluation function for fluorescent images. The method used for the confocal microscope fails because of contamination of out-of-focus fluorescence. The Laplacian kernel is applied in the method proposed by Li et al., and it does not give good performance [13]. Compared with variance and normalized variance, it improves the robustness and continuity. Compared with local maximum luminosity, it is more likely to find the patch with a clearer texture. So, the multiplication of the local maximum luminosity and variance of intensity is selected as the sharpness evaluation for the second step of our method.
After that, we can establish the confidence map. The fluorescent substance will emit more light when it is in focus than when it is not. The depth estimation of the tissue parts that are luminous enough when in focus is less vulnerable to out-of-focus fluorescence from other parts. Based on this assumption, the confidence map has the following form:
C j , k = 0 m j , k threshold 0 1 m j , k > threshold 0
where t h r e s h o l d 0 here is a parameter that can be adjusted, and m j , k is the maximum pixel value among p a t c h j , k .

2.3. Depth Adjustment

In the third step, in order to prevent the contamination of out-of-focus fluorescence in the void regions, information in the confidence map C j , k is used to adjust the depth for each block. Because the surface of the tissue is continuous, the estimated depths of adjacent blocks should not differ a lot in most cases. During this step, if the depth of a block has a large difference from those of the surrounding blocks, the depth will be adjusted according to Algorithm 1. In this manner, lots of contaminated parts will be recovered.
Algorithm 1 Adjust i n d e x j , k
1:
v o t e _ n u m 0
2:
p o s _ v o t e 0
3:
for  s j = j k e r n e l : j + k e r n e l  do
4:
    for  s k = k k e r n e l : k + k e r n e l  do
5:
       if  c s j , s k = = 1  then
6:
           v o t e _ n u m v o t e _ n u m + 1
7:
          if  | i n d e x j , k i n d e x s j , s k | > t h r e s h o l d 1  then
8:
              p o s _ v o t e p o s _ v o t e + 1
9:
          end if
10:
      end if
11:
   end for
12:
end for
13:
if  p o s _ v o t e > r a t i o × v o t e _ n u m     AND     v o t e _ n u m > t h r e s h o l d 2  then
14:
    i n d e x j , k Average s j , s k i n d e x s j , s k
15:
end if
For every block B j , k , we adjust the i n d e x j , k with the following Algorithm 1.
Here, the kernel, t h r e s h o l d 1 , ratio, and t h r e s h o l d 2 are independent parameters which can be adjusted. After that, for every block where C j , k = 0 and there is a neighboring block that is confident, we set it to the average of i n d e x s j , s k where ( s j , s k ) near ( j , k ) and C s j , s k = 1 .

2.4. Information Filling

The blocking effects introduced by block segmentation can be obvious with the assumption that the depths of pixels within a block that are similar are invalid.The boundaries of objects are typical examples, and in this situation, the depths of some pixels within the block may be far away from the estimated depth of the block but close to the estimated depths of adjacent blocks. Therefore, in the fourth step, a method of information filling is used to alleviate the blocking effects and assign a depth to every pixel. After this step, every pixel has its own depth, and this information is useful for the 3D modeling of objects in the visible area. First, we initialize the index for every pixel p i x i n d e x p j , p k = i n d e x j , k where pixel p i x p j , p k belongs to B j , k . The part of this algorithm is given in Algorithm 2 for any two adjacent block B j 1 , k 1 , B j 2 , k 2 where at least one of their confidence values is 1.
Algorithm 2 Information filling.
1:
if  | i n d e x j 1 , k 1 i n d e x j 2 , k 2 | > t h r e s h o l d 3  then
2:
   for every pixel p i x p j , p k belongs to B j 1 , k 1 or B j 2 , k 2  do
3:
     The corresponding pixel value in every source image is v i , p j , p k
4:
     if  p p j , p k belongs to B j 1 , k 1  then
5:
         d = p i x i n d e x p j , p k ; x = i n d e x j 2 , k 2
6:
        if  v d , p j , p k < t h r e s h o l d 4 AND v d , p j , p k < v x , p j , p k  then
7:
           p i x i n d e x p j , p k = x
8:
        end if
9:
     else
10:
         d = p i x i n d e x p j , p k ; x = i n d e x j 1 , k 1
11:
        if  v d , p j , p k < t h r e s h o l d 4 AND v d , p j , p k < v x , p j , p k  then
12:
           p i x i n d e x p j , p k = x
13:
        end if
14:
     end if
15:
   end for
16:
end if
Here, t h r e s h o l d 3 and t h r e s h o l d 4 are two independent parameters which can be adjusted. After that, the smoothing is applied to the edge of each block in depth space and fused image. After this step, we can obtain a useful depth map.

2.5. Optional Step

Optional steps can be applied for objects that can be separated (an example can be seen in Figure 3); after, segmentation methods like DBSCAN or group finding methods like FoF can be applied to separate different objects [20,21].
To take suitable segmentation block sizes for different objects, the block sizes used to re-divide the visible area will be adjusted according to the scale of these objects. Objects of the same block size can be merged into a class to be processed together, and this will take less time. Then, pixels that do not belong to any class are assigned to the nearest classes, but one pixel that has been assigned to one class will not be assigned to other classes. The domains occupied by one class can be treated as a mask (the pixel of source images will be treated as 0 if the pixel is outside of this mask) when processing this class. The steps aforementioned will be used to fuse the images for every class.
After all the classes of objects are fused in the manner of algorithms mentioned before, they can be combined together to obtain the final fused image. The depth maps obtained for different classes of objects will be combined with the help of these masks. The flow chart can also be seen in Figure 1. This step has no effect on tissue that is not separable, so for this kind of tissue, we do not need to take this step.

3. Experiments

The source images are obtained using the fluorescence microscope at different object distances. All experiments are run on a laptop with 4 Cores, 2.6 GHz CPU, and 16 GB RAM. Here, the proposed method is compared with a method proposed by De et al., GFF (guided filtering-based fusion method), SSS (shared self-similarity) and SSSDI (shared self-similarity and depth information) [12,13,14].
The method proposed by De et al. makes use of the wavelet transformation [12]. In their method, they first use a nonlinear wavelet to decompose the source images into multi-resolution signals. Then, fusion happens at different resolutions to pursue maximum sharpness. Finally, they reconstruct the fused image through composition. GFF decomposes images into base layers and detail layers and uses a weighted average method for fusion separately for base layers and detail layers [13]. The weight maps are constructed with the help of guided filtering. The final fused image is obtained by combining the fused base layer and detail layer. SSS and SSSDI were both proposed by Guo et al., and they are very similar [14]. SSS is also a weighted average method. It makes use of shared self-similarity to generate adaptive regions and choose SML (some-of-modified-Laplacian) as a clarity metric. The fusion weights depend on this clarity metric. Different from SSS, SSSDI divides the clarity metric used by SSS by the square of the defocus scale to form a new clarity metric [14]. The defocus scale can be estimated by
σ ( x , y ) = 1 R ( x , y ) 2 1 σ 0
where σ 0 is a given fixed value, and R ( x , y ) can be represented by
R ( x , y ) = | i ( x , y ) | i 1 ( x , y )
Here, i ( x , y ) is the gradient of the original image at pixel ( x , y ) and i 1 ( x , y ) is the gradient of the image at pixel ( x , y ) after reblurring with a Gaussian kernel [13]. However, for the void region where pixel values are 0 ,   i ( x , y ) and i 1 ( x , y ) are both 0, and it is hard to define R ( x , y ) and σ ( x , y ) . So, SSSDI is invalid for some regions with no fluorescence in the source images.
The method proposed by De can fuse any number of source images. However, this method cannot make sure the lower bound of pixel value is greater than or equal to 0. The two methods are tested: (1) rescaling values for every pixel according to the lower bound and upper bound of the fused image, and (2) putting all the values below 0 to 0. The second method produces better fused images, and the images processed by the second method are selected for comparison. Meanwhile, the code provided by Li et al. for GFF and Guo et al. for SSS and SSSDI can fuse two source images at a time. In order to fuse N source images, images of an odd index are fused with the following images with an even index. If the last image has an odd index, it remains to the next round of fusion, and we iterate this process until the final fused image is obtained. For the proposed method, the multiplication of local maximum luminosity and variance of intensity is chosen as the s i , j , k .

3.1. Fusion Performance

The performance of different methods on one set of source images is shown in Figure 3. This set of source images includes 81 fluorescence images of one kind of pollen. The parameters for GFF, SSS and SSSDI are the default parameters in their code. The parameters for our method are set as t h r e s h o l d 0 = 60 , k e r n e l = 2 , t h r e s h o l d 1 = 40 , t h r e s h o l d 2 = 4 , r a t i o = 0.5 , t h r e s h o l d 3 = 10 , t h r e s h o l d 4 = 80 , and the tentative block size is 15 × 15 .
It can be seen that the method proposed by De and Chanda is not vulnerable to contamination of out-of-focus fluorescence, but it suffers from artifact-like blocking effects [14]. Additionally, it cannot make sure the lower bound is zero, and this is another problem of this method. The result of GFF suffers from out-of-focus fluorescence because of the weighted average method [13]. The texture of the pollen grains in the image fused by GFF is not as clear as that in ours, especially the top left and the bottom. SSS and SSSDI methods seem more vulnerable to this interference, and SSSDI performs no better than SSS in this situation [14]. We can see that the texture of the top left pollen grain is dramatically contaminated by the out-of-focus fluorescence.
The proposed method is less vulnerable to out-of-focus fluorescence and obtains better performance. But the stability of our method in the object’s boundary is not as good as that within the boundary because the third step cannot work very well in this situation, as it is hard to obtain enough positive votes. As we can see, the proposed method has obtained state-of-the-art performance and has better performance on some parts of the pollen grains.
In order to show the generality of the proposed method, the results of different methods on another set of source images are shown in Figure 4. This set of images contains 60 fluorescence images of a spinach stem. The parameters for our method are set as t h r e s h o l d 0 = 150 , k e r n e l = 2 , t h r e s h o l d 1 = 40 , t h r e s h o l d 2 = 4 , r a t i o = 0.5 , t h r e s h o l d 3 = 10 , and t h r e s h o l d 4 = 80 , and the tentative block size is 20 × 20 .
Even though most textures in the image fused by the method proposed by De are clear, it still suffers from the blocking effect [12]. SSS and SSSDI methods are still affected by the contamination of out-of-focus fluorescence if we focus on the bottom right part [14]. The fused image using the GFF method is much better than SSS and SSSDI but not as clear as that of our method [13]. What is more, our method is even faster than the other three methods in this situation [12,13,14]. We do not show the image fused by our proposed method with this optional step because this tissue should be treated as a whole, and the result will be almost the same.

3.2. Computation Time

The computational times of all the methods spent on the different sets of source images are shown in Table 1.
As can be seen, the proposed method without the optional step is the fastest, and the proposed with the optional step will be much slower. However, the optional step is not necessary in many situations. The method proposed by De and Chanda and the method proposed by Li et al. are fast but not as fast as the proposed one without the optional step [12,13]. The SSSDI method proposed by Guo et al. is the slowest [14]. SSSDI is claimed to take about five times as long as SSS in their paper [14]. Actually, SSSDI takes more than eight times as long as SSS in our experiments, and this is because the laptop runs out of memory RAM, and it needs to make use of the disk.

3.3. Depth Map

As aforementioned, the proposed method can obtain a determined depth map which can be used to reconstruct the 3D surface of observed objects if the parameters of the fluorescent are known. Here, an example of a depth map and point cloud of one pollen grain’s surface are shown in Figure 5. Figure 5a is the fused image of the first set of fluorescence images. Figure 5b is the depth map constructed in our algorithms, and the depth is set to 0 for the void region. Figure 5c is the point cloud of one pollen, which is marked by a red rectangle in Figure 5a. Better reconstruction of the 3D surface of the observed area may be achievable using the depth information obtained by our method, but it is out of the range of this paper.

4. Discussion

From the result of the experiment aforementioned, it can be seen that, first, our infusion method for fluorescence imaging can achieve start-of-art performance. Second, the computation time for the proposed method is on an average level, and for inseparable tissue, it is faster than the other three methods because the optional part is not necessary and the computation time will decrease. Third, the blocking effects are obviously alleviated by our information filling method. Moreover, the depth for every pixel can be obtained from the proposed method.
The proposed method can make use of the properties of epifluorescence microscopes to finish the work of a confocal microscope such as obtaining clear images and obtaining the depth information of every pixel. This information can be used to reconstruct the 3D model of the visible area similar to a confocal microscope. Compared to using other devices like a light field microscope for high-quality microscopic imaging, our method has lower equipment costs and higher algorithm efficiency. Additionally, the speed of obtaining a clear image will increase prominently. However, due to equipment limitations, we did not collect ground-truth images to conduct quantitative analysis of our algorithm. Moreover, whether our multi-depth fusion algorithm based on the sharpness evaluation function can be applied to more scenarios still requires further experiments and verification.

5. Conclusions

In this paper, a fluorescence image fusion method based on local maximum luminosity, variance of intensity within blocks, and the information filling method has been proposed and compared with a state-of-the-art image fusion method. In order to estimate clarity, the relationship of luminosity and clarity has been used, and a new sharpness measure has been proposed. Additionally, the performance of it has been compared with other sharpness measures. The information filling method has been proposed to alleviate blocking effects. In the experiment, it shows that we have obtained state-of-the-art performance in fusing fluorescent images, and our method is faster than the other three methods when the optional step is not necessary. What is more, the estimation of the depth can also be obtained to reconstruct the 3D model of the visible area.

Author Contributions

Conceptualization, H.C.; methodology, H.C.; validation, H.C. and K.W.; formal analysis and data curation, H.C. and K.W. The manuscript was written by H.C. with the assistance of D.M. and C.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

This work is supported by National Major Scientific Re-search Instrument Development Project (62227811), Second Institute of Oceanography, MNR “Deep Blue Plan” research program (No. SL2021ZD205). We express our sincerest gratitude to our collaborator Min Hou from the Department of Reproductive Medicine Center, Shanghai East Hospital, Tongji University School of Medicine.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lichtman, J.W.; Conchello, J.A. Fluorescence microscopy. Nat. Methods 2005, 2, 910–919. [Google Scholar] [CrossRef] [PubMed]
  2. Renz, M. Fluorescence microscopy—A historical and technical perspective. Cytom. Part A 2013, 83, 767–779. [Google Scholar] [CrossRef] [PubMed]
  3. White, J.; Amos, W.; Fordham, M. An evaluation of confocal versus conventional imaging of biological structures by fluorescence light microscopy. J. Cell Biol. 1987, 105, 41–48. [Google Scholar] [CrossRef] [PubMed]
  4. Inoué, S. Foundations of confocal scanned imaging in light microscopy. In Handbook of Biological Confocal Microscopy; Springer: Berlin/Heidelberg, Germany, 2006; pp. 1–19. [Google Scholar]
  5. Pacheco, S.; Wang, C.; Chawla, M.K.; Nguyen, M.; Baggett, B.K.; Utzinger, U.; Barnes, C.A.; Liang, R. High resolution, high speed, long working distance, large field of view confocal fluorescence microscope. Sci. Rep. 2017, 7, 13349. [Google Scholar] [CrossRef]
  6. Zhang, Q.; Tian, L.; Huang, L.; Wang, X.; Zhu, H. Rendering distortion estimation model for 3D high efficiency depth coding. Math. Probl. Eng. 2014, 2014, 940737. [Google Scholar] [CrossRef]
  7. Olsovsky, C.; Shelton, R.; Carrasco-Zevallos, O.; Applegate, B.E.; Maitland, K.C. Chromatic confocal microscopy for multi-depth imaging of epithelial tissue. Biomed. Opt. Express 2013, 4, 732–740. [Google Scholar] [CrossRef] [PubMed]
  8. Jin, L.; Tang, Y.; Wu, Y.; Coole, J.B.; Tan, M.T.; Zhao, X.; Badaoui, H.; Robinson, J.T.; Williams, M.D.; Gillenwater, A.M. Deep learning extended depth-of-field microscope for fast and slide-free histology. Proc. Natl. Acad. Sci. USA 2020, 117, 33051–33060. [Google Scholar] [CrossRef] [PubMed]
  9. Chen, Y.; Deng, N.; Xin, B.; Xing, W.; Zhang, Z. A novel multi-focus image fusion method of nonwovens based on GHM multiwavelet transform technology. Text. Res. J. 2019, 89, 2870–2879. [Google Scholar] [CrossRef]
  10. Wang, L.; Liu, Z.; Huang, J.; Liu, C.; Zhang, L.; Liu, C. The Fusion of Multi-Focus Images Based on the Complex Shearlet Features-Motivated Generative Adversarial Network. J. Adv. Transp. 2021, 2021, 5439935. [Google Scholar] [CrossRef]
  11. Dong, M.; Dong, C.; Guo, M.; Wang, Z.; Mu, X. Multi-focus image fusion based on area-based standard deviation in dual tree contourlet transform domain. In Proceedings of the 9th International Conference on Graphic and Image Processing (ICGIP 2017), Qingdao, China, 14–16 October 2017; SPIE: Bellingham, WA, USA, 2018; pp. 878–885. [Google Scholar]
  12. De, I.; Chanda, B. A simple and efficient algorithm for multifocus image fusion using morphological wavelets. Signal Process. 2006, 86, 924–936. [Google Scholar] [CrossRef]
  13. Li, S.; Kang, X.; Hu, J. Image fusion with guided filtering. IEEE Trans. Image Process. 2013, 22, 2864–2875. [Google Scholar]
  14. Guo, D.; Yan, J.; Qu, X. High quality multi-focus image fusion using self-similarity and depth information. Opt. Commun. 2015, 338, 138–144. [Google Scholar] [CrossRef]
  15. Zhuo, S.; Sim, T. Defocus map estimation from a single image. Pattern Recognit. 2011, 44, 1852–1858. [Google Scholar] [CrossRef]
  16. Kim, J.; Yoon, K.; Kim, K. Design of a long-pass filter with effects on fluorescence image observation for surgical fluorescence microscope applications. Symmetry 2021, 13, 1571. [Google Scholar] [CrossRef]
  17. Groen, F.C.; Young, I.T.; Ligthart, G. A comparison of different focus functions for use in autofocus algorithms. Cytom. J. Int. Soc. Anal. Cytol. 1985, 6, 81–91. [Google Scholar] [CrossRef] [PubMed]
  18. Schlangen, S.; Ihme, M.; Rahlves, M.; Roth, B. Autofocusing system for spatial light modulator-based maskless lithography. Appl. Opt. 2016, 55, 1863–1870. [Google Scholar] [CrossRef] [PubMed]
  19. Leray, A.; Odin, C.; Le Grand, Y. Out-of-focus fluorescence collection in two-photon microscopy of scattering media. Opt. Commun. 2008, 281, 6139–6144. [Google Scholar] [CrossRef]
  20. Ester, M.; Kriegel, H.-P.; Sander, J.; Xu, X. A density-based algorithm for discovering clusters in large spatial databases with noise. In Proceedings of the 2nd International Conference on Knowledge Discovery and Data Mining (KDD), Portland, OR, USA, 2–4 August 1996; pp. 226–231. [Google Scholar]
  21. Davis, M.; Efstathiou, G.; Frenk, C.S.; White, S.D. The evolution of large-scale structure in a universe dominated by cold dark matter. Astrophys. J. 1985, 292, 371–394. [Google Scholar] [CrossRef]
Figure 1. Framework of our proposed method.
Figure 1. Framework of our proposed method.
Sensors 24 04909 g001
Figure 2. Performance of maximum total luminosity method and the results of different sharpness estimation functions: (ac) maximum total luminosity method [3,4]. (df) the Laplacian measure which is used by Zhuo and Sim [15]. (gi) a method using the variance measure as a typical criterion of clarity [17,18]. (jl) a method using the normalized-variance measure as the criterion of clarity [17,18]. (mo) a method using local maximum luminosity as the criterion of clarity. (pr) our method.
Figure 2. Performance of maximum total luminosity method and the results of different sharpness estimation functions: (ac) maximum total luminosity method [3,4]. (df) the Laplacian measure which is used by Zhuo and Sim [15]. (gi) a method using the variance measure as a typical criterion of clarity [17,18]. (jl) a method using the normalized-variance measure as the criterion of clarity [17,18]. (mo) a method using local maximum luminosity as the criterion of clarity. (pr) our method.
Sensors 24 04909 g002
Figure 3. Images fused by different methods: (a) a method proposed by De and Chanda [12]. (b) GFF proposed by Li et al. [13]. (c) SSS proposed by Guo et al. [14] (d) SSSDI proposed by Guo et al. [14]. (e) Our proposed method without an optional step. (f) Our proposed method with an optional step.
Figure 3. Images fused by different methods: (a) a method proposed by De and Chanda [12]. (b) GFF proposed by Li et al. [13]. (c) SSS proposed by Guo et al. [14] (d) SSSDI proposed by Guo et al. [14]. (e) Our proposed method without an optional step. (f) Our proposed method with an optional step.
Sensors 24 04909 g003
Figure 4. Images fused by different method: (a) a method proposed by De and Chanda [12]. (b) GFF proposed by Li et al. [13]. (c) SSS proposed by Guo et al. [14]. (d) SSSDI method proposed by Guo et al. [14]. (e) Our proposed method without an optional step.
Figure 4. Images fused by different method: (a) a method proposed by De and Chanda [12]. (b) GFF proposed by Li et al. [13]. (c) SSS proposed by Guo et al. [14]. (d) SSSDI method proposed by Guo et al. [14]. (e) Our proposed method without an optional step.
Sensors 24 04909 g004
Figure 5. Example of depth map obtained from our method and the point cloud of one pollen: (a) the fused image of the first set of fluorescence images. (b) the depth map constructed in our algorithms. (c) the point cloud of one pollen.
Figure 5. Example of depth map obtained from our method and the point cloud of one pollen: (a) the fused image of the first set of fluorescence images. (b) the depth map constructed in our algorithms. (c) the point cloud of one pollen.
Sensors 24 04909 g005
Table 1. Computational time for different methods (in minutes).
Table 1. Computational time for different methods (in minutes).
DeGFFSSSSSSDIOursOurs (with Optional Step)
Set 1 ( 1920 × 1200 )3.484.38254.672212.071.5117.17
Set 2 ( 1920 × 1200 )2.973.12111.671555.050.89-
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cheng, H.; Wu, K.; Gu, C.; Ma, D. Multi-Focus Images Fusion for Fluorescence Imaging Based on Local Maximum Luminosity and Intensity Variance. Sensors 2024, 24, 4909. https://doi.org/10.3390/s24154909

AMA Style

Cheng H, Wu K, Gu C, Ma D. Multi-Focus Images Fusion for Fluorescence Imaging Based on Local Maximum Luminosity and Intensity Variance. Sensors. 2024; 24(15):4909. https://doi.org/10.3390/s24154909

Chicago/Turabian Style

Cheng, Hao, Kaijie Wu, Chaochen Gu, and Dingrui Ma. 2024. "Multi-Focus Images Fusion for Fluorescence Imaging Based on Local Maximum Luminosity and Intensity Variance" Sensors 24, no. 15: 4909. https://doi.org/10.3390/s24154909

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop