Next Article in Journal
An Effective and Robust Parameter Estimation Method in a Self-Developed, Ultra-Low Frequency Impedance Spectroscopy Technique for Large Impedances
Previous Article in Journal
Heart Pulse Transmission Parameters of Multi-Channel PPG Signals for Cuffless Estimation of Arterial Blood Pressure: Preliminary Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimizing Edge Detection Efficiency with a Grünwald–Letnikov Fractional Network

by
Diaa Eldin Elgezouli
1,* and
Abdulrahman B. M. Alzahrani
2
1
Basic Science Department, Common First Year, King Saud University, P.O. Box 145111, Riyadh 11362, Saudi Arabia
2
Department of Mathematics, College of Science, King Saud University, P.O. Box 2455, Riyadh 11451, Saudi Arabia
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(16), 3298; https://doi.org/10.3390/electronics13163298
Submission received: 24 July 2024 / Revised: 17 August 2024 / Accepted: 17 August 2024 / Published: 20 August 2024
(This article belongs to the Section Artificial Intelligence)

Abstract

:
Edge detection is an essential image processing act that is crucial for many computer vision applications such as object detection, image segmentation, face recognition, text recognition, medical imaging, and autonomous vehicles. Deep learning is the most advanced and widely used tool of them all. In this paper, we present a novel deep learning model and use image datasets to test it. Our model uses a fractional calculus tool, which could enhance gradient approaches’ performances. Specifically, we approximate the fractional-order derivative-order neural network (GLFNet) using a Grünwald–Letnikov fractional definition. First, the original dataset is subjected to a Grünwald–Letnikov fractional order. After that, the CNN model is updated with the new dataset, concluding the standard CNN procedure. The training rate, the improvement in the F-measure for identifying the effective edge while maintaining the CNN model’s memory consumption, and the values of the loss errors between the prediction and training processes were all tested using the MNIST dataset. Our experiments show that GLFNet considerably enhances edge detection. GLFNet outperformed CNN with an average loss error ratio of 15.40, suggesting fewer loss mistakes. The F1-measure ratio of 0.81 indicates that GLFNet can compete with CNN in terms of precision and recall. The training time for GLFNet was lowered by an average ratio of 1.14 when compared to CNN, while inference time was faster with a ratio of 1.14, indicating increased efficiency. These findings demonstrate the efficacy of introducing Grünwald–Letnikov fractional convolution into deep learning models, resulting in more precise and reliable edge detection while preserving comparable memory utilization.

1. Introduction

Edge detection is a fundamental aspect of image processing and is the foundation for numerous computer vision applications. Accurate edge detection is crucial for object recognition, picture segmentation, and scene understanding. However, traditional edge detection methods, such as the Sobel, Canny, and Laplacian filters, have long been instrumental in delineating boundaries within images. However, traditional edge detection methods often face challenges such as unequal performance across different image formats and noise sensitivity [1,2].
Modern deep learning models achieve state-of-the-art accuracy in many different tasks, including image classification [3,4,5], semantic segmentation [6], object detection [7,8], posture detection [9,10], and more. The fundamental goal of research in this field is to create increasingly complex architectures with millions, if not billions, of trainable parameters to obtain performance improvements. Many of these astounding findings have yet to be made due to memory and computational limits in large-scale models. To bridge the gap between cloud and edge computing, a slew of deep learning techniques and optimization methods have been proposed in recent years, including lightweight deep learning models, network compression, and efficient neural architecture [11,12,13].
As artificial intelligence evolves and advances, researchers and engineers are investigating new and inventive methods for neural network topologies that can overcome the limits of old models. One such emerging concept is the fractional neural network. Fractional calculus, a development of classical calculus that applies the concepts of differentiation and integration to non-integer orders, has emerged as a promising field in this context. This mathematical field gives new techniques for analyzing and processing signals and functions, providing more precise control over the systems’ memory and smoothness. Researchers want to use the increased flexibility and complexity of fractional-order operations in CNNs, potentially leading to considerable gains in performance and interpretability. This mathematical framework has been discovered to be a powerful tool for modeling complex systems with memory effects and non-ocal phenomena [14]. Fractional CNN seeks to leverage fractional calculus’s unique properties to improve traditional neural networks’ performance and capabilities. These models seek to capture the long-term dependencies and memory effects frequently found in real-world data, which can be difficult for ordinary neural networks to handle well [15].
The Riemann–Liouville, Caputo, and Grünwald–Letnikov fractional derivatives are the three most often used fractional derivatives in CNNs. The Riemann–Liouville fractional derivative is appropriate for jobs requiring long-range dependencies in data. An integral transform defines it, and it is especially helpful for modeling processes with memory effects. = Unlike conventional integer-order derivatives, this derivative enables the inclusion of fractional-order terms into CNNs, giving rise to a means of capturing more intricate dynamics and patterns [16]. However, the definition of the Caputo fractional derivative is slightly different, and it fits in better with initial value problems. It also makes sense more intuitively and integrates more easily into neural network frameworks. Because these derivatives introduce fractional-order operations that increase the network’s sensitivity to different scales and features in the data, they help CNNs reach better feature extraction and representation capabilities. These fractional derivatives enable the modeling of complex behaviors and correlations in CNN architectures, which may result in more resilient and flexible models for tasks like segmentation, time-series prediction, and image classification [17].
Discrete approximation, which encapsulates the idea of fractional calculus through a finite difference limit process, is used to derive the Grünwald–Letnikov fractional derivative. This derivative is especially valued since it is easy to use and understand, which makes it a valuable option for CNNs that need to incorporate fractional calculus. The GL derivative is relatively easy to integrate into neural network topologies since it approximates fractional differentiation through a series expansion of the difference operator. This method improves the network’s capacity to learn intricate features and patterns by efficiently simulating memory effects and long-range relationships in data [18]. Due to the computational and practical properties of GL attributes, the Grünwald–Letnikov fractional-order derivative is applied in this work for conventional CNN. The proposed method is new in the context of neural networks. First, the original dataset is subjected to a Grünwald–Letnikov fractional order. After that, the standard CNN method is completed by adding the new dataset to the CNN model. The main contribution of this work is summarized as follows:
  • We provide a novel method for optimizing convolutional neural networks (CNNs) based on the Grünwald–Letnikov fractional-order derivative. This derivative allows the network to capture complicated temporal and spatial correlations in data, boosting its capacity to detect subtle patterns that may be critical for accurate classification or segmentation tasks.
  • The proposed approach significantly enhances CNN performance across a variety of criteria. It successfully reduces loss mistakes by using the fractional-order derivative to improve feature extraction and model training operations. Second, it optimizes the F1 measure, an important statistic in binary classification tasks that balances precision and recall, increasing the network’s total predictive power. Third, it minimizes training and inference times by enhancing the network’s computational efficiency, resulting in speedier processing without sacrificing accuracy.
  • Unlike existing methods, which are frequently confined by specific CNN designs, our methodology operates on the dataset before CNN processing, making it adaptable and suitable to various network structures and datasets. This adaptability enables practitioners to seamlessly include fractional calculus into current CNN systems without requiring significant architectural changes.
  • To validate the success of our approach, we analyze GLFNet (Grünwald–Letnikov fractional network) on benchmark datasets and compare its performance to that of a baseline CNN without fractional derivatives. The experimental results show that GLFNet regularly outperforms its integer-order counterpart regarding accuracy, noise robustness, and generalization capabilities over various datasets and tasks.
The remainder of the paper is organized as follows: Section 2 looks at relevant work, while Section 3 provides essential fractional derivative definitions and tools for future usage. Section 4 describes the proposed method. Section 5 contains the experimental results and a discussion. Finally, Section 6 provides the work’s conclusion.

2. Related Work

The field of convolutional neural networks (CNNs) has seen significant advancements, particularly with fractional convolution operations that enhance the network’s capability to capture intricate patterns and features. Fractional convolutions, utilizing non-integer stride and dilation factors, have demonstrated improvements across various computer vision tasks, such as optical metrology and image categorization.
Jia et al. [19] proposed a pioneering approach using fractional-order differential equations to enhance deep CNNs for image denoising. Their FOCNet architecture, based on discretized fractional-order differential equations, improves memory management and feature integration in both forward and backward passes. Additionally, integrating multi-scale feature interactions enhances FOCNet’s dynamic system control, significantly boosting performance in image-denoising tasks. Chen et al. [20] introduced the fractional-order variational residual CNN (FVCNN) for low-dose CT (LDCT) denoising, incorporating a fractional-order total variational (FTV) term in the loss function to enhance texture preservation. FVCNN effectively removes noise while retaining critical texture features, validated through extensive experiments.
Crack detection in civil engineering and transportation remains challenging due to hazy boundaries and irregular shapes of crack images. Cao et al. [21] proposed a method based on fractional differential and fractal dimension analysis for fracture detection, offering promising results in crack pathology assessment. Authors in [22] developed fractional convolutional neural networks (FRCNNs), enhancing feature extraction in image classification tasks and outperforming conventional CNNs. Zheng et al. [23] extended DeepLabv3+ with a fractional-order image enhancement convolution kernel for semantic segmentation, improving segmentation accuracy, particularly in edge detection tasks.
Sahlol et al. [24] focused on reducing computational complexity in COVID-19 image classification using an FO-MPA (fractional-order marine predators algorithm) hybrid approach. This method combines CNNs with fractional calculus to optimize feature extraction and selection, demonstrating robust performance in medical imaging tasks. Chandra et al. [25] proposed a linear fractional mesh-free partial differential equation (FPDE) for image enhancement, which preserves details in smooth areas while enhancing high-frequency information, beneficial for tumor identification in CAD models. Badashah et al. [26] introduced the F-HHO method for malignant bone sarcoma detection, combining fractional calculus with harmony search optimization to improve early diagnosis capabilities.
Chen et al. [27] developed the fractional-order residual CNN (FRCNN) for LDCT denoising, balancing radiation dose reduction with image quality improvement. Zhao et al. [28] proposed the fractional Gabor convolutional network (FGCN) for efficient feature fusion and comprehensive feature extraction from multisensor remote sensing data. Zamora et al. [29] presented a fractional-order CNN technique with chaos error mapping to improve the accuracy of chatter detection in machining processes. This methodology highlights ongoing attempts to optimize manufacturing processes using innovative technologies and marks a significant advancement in addressing non-inear machining dynamics.
The study “Convolutional filter approximation using fractional calculus” by Raubitzek et al. [30] proposed using fractional calculus to more effectively approximate convolutional filters. This investigation aimed to improve CNN performance or maintain it while potentially decreasing computational costs. Liu et al. [31] introduced the fractional-order fusion model (FFM) for lightweight detection of low-light lane curvature, enhancing real-time performance in automated driving scenarios. Kuo et al. [32] proposed the fractional-order CNN (FOCNN) for chatter detection in machining processes, addressing production costs and tool wear issues.
Recent work has extensively explored integrating machine learning techniques with fractional derivatives. Kuo et al. [32] summarized approaches that use fractional calculus to improve modeling performance and prediction accuracy across various fields. This highlights the potential synergy between machine learning algorithms and sophisticated mathematical techniques, continuing efforts to develop methodologies for better predictive modeling and handling complex system dynamics. Cheng et al. [33] developed the DSD-Net, integrating fast Fourier transform and fractal dimension estimation into a dual encoder structure for crack detection in low-light environments. Ding et al. [34] focused on edge detection using fractional masks derived from the Atangana–Baleanu–Caputo fractional integral, demonstrating improved accuracy in segmenting water-repellent images of insulators through numerical simulations. Kumar et al. [35] presented novel activation functions derived from the Riemann–Liouville conformable fractional derivative (CFD), investigating their impact on classification task performance in multilayer perceptron (MLP) models.
This compilation highlights the diverse applications and advancements facilitated by fractional calculus in enhancing deep learning models across various domains of computer vision and image processing and Table 1, Table 2, Table 3, Table 4 and Table 5 give set parameters to compare the differences between related works and the proposed method [19,20].
Comparison of GLFNet, FOCNet, and FVCNN.

3. Mathematical Definitions and Tools

Fractional derivatives, a generalization of conventional integer-order derivatives, have received much interest in recent years because they can represent the inherent memory and non-local features of complex systems, while the subject of fractional calculus has advanced quickly, the proliferation of numerous fractional derivative definitions has caused confusion and difficulty for researchers in selecting the best model for their applications [36].
One of the most applicable fractional derivatives is the Grünwald–Letnikov derivative, a valuable fractional calculus technique that allows one to extend standard differentiation to non-integer orders. This derivative, which bears the names of the mathematicians Grünwald and Letnikov, provides a unique method for examining functions with fractional differentiation orders. This concept, which permits the differentiation of functions with non-integer orders, captures the core of fractional calculus. A flexible framework for studying signals and systems with fractal-like characteristics and simulating phenomena with intricate dynamics is provided by the Grünwald–Letnikov derivative. Accordingly, the Grünwald–Letnikov fractional derivative is defined as follows [37]:
  • Grünwald–Letnikov Left-Sided Derivative ( GLD α , a + ):
    G L a + D α f ( t ) = lim h 0 1 h α k = 0 ( 1 ) k ( α + 1 ) k k ! ( α k + 1 ) f ( t k h ) ,
  • Grünwald–Letnikov Right-Sided Derivative ( GLD α , b ):
    G L b D α f ( t ) = lim h 0 1 h α k = 0 ( 1 ) k ( α + 1 ) k k ! ( α k + 1 ) f ( t + k h ) .
The Riemann–Liouville derivative is one of the most often utilized fractional derivatives. Nevertheless, there are many drawbacks to the Riemann–Liouville concept, including its incapacity to meet the classical beginning requirements and its unclear physical meaning. The Caputo fractional derivative alters the Riemann–Liouville definition, which flips the order of integration and differentiation to solve these problems. Consequently, the fractional derivative of Riemann–Liouville is defined as follows [38]:
  • Riemann–Liouville Left-Sided Derivative ( RLD α , a + ):
    R L a + D α f ( t ) = 1 Γ ( n α ) d n d t n a t ( t ξ ) n α 1 f ( ξ ) d ξ , t a ,
  • Riemann–Liouville Right-Sided Derivative ( RLD α , b ):
    R L b D α f ( t ) = ( 1 ) n 1 Γ ( n α ) d n d t n t b ( ξ t ) n α 1 f ( ξ ) d ξ , t b .
Caputo derivatives provide a fundamental idea that extends beyond conventional integer-order derivatives. These derivatives, which Michele Caputo introduced in 1967, immediately incorporate beginning conditions into their definition, so addressing the drawbacks of Riemann–Liouville derivatives. Caputo derivatives concentrate on the current state and initial circumstances, in contrast to Riemann–Liouville derivatives, which call for knowledge of a function’s history up to this point. This makes them especially useful for modeling dynamic systems having distinct beginning points; the fractional derivative of Caputo is defined as follows [38]:
  • Caputo Left-Sided Derivative ( CD α , a + ):
    C a + D α f ( t ) = 1 Γ ( n α ) a t ( t ξ ) n α 1 d n d ξ n f ( ξ ) d ξ , t a ,
  • Caputo Right-Sided Derivative ( CD α , b ):
    C b D α f ( t ) = ( 1 ) n 1 Γ ( n α ) t b ( ξ t ) n α 1 d n d ξ n f ( ξ ) d ξ , t b .

4. The Proposed Method

This section provides a complete description of the technique used in the proposed method. The Grünwald–Letnikov fractional derivative was chosen for the proposed method because it can be calculated using the discrete approximation method, which uses a finite difference limit process to encapsulate the concept of fractional calculus. Because this derivative is simple to use and grasp, CNNs that must integrate fractional calculus may find it highly beneficial. The GL derivative, which approximates fractional differentiation via a series extension of the difference operator, is simple to integrate into neural network topologies. This technique correctly simulates memory effects and long-range correlations in data, boosting the network’s ability to learn complex features and patterns. The Grünwald–Letnikov fractional neural network (GLFNet), which performs convolutions in the same way as convolution networks for image edge detection, substitutes fractional derivative for the original pixels in the local feature map patch covered by the convolution kernels during a convolutional operation.
To apply the proposed method, firstly we must first transform Grünwald–Letnikov derivatives from one to two variables ( s , t ) so Equations (1) and (2) take the following forms:
Grünwald–Letnikov left-sided derivative ( GLD α , a + ) with respect to s:
G L a + D s α f ( s , t ) = lim h 0 1 h α k = 0 ( 1 ) k ( α + 1 ) k k ! ( α k + 1 ) f ( s k h , t ) ,
and,
Grünwald–Letnikov right-sided derivative ( GLD α , b ) with respect to s:
G L b D s α f ( s , t ) = lim h 0 1 h α k = 0 ( 1 ) k ( α + 1 ) k k ! ( α k + 1 ) f ( s + k h , t ) .
By the same way can defined Grünwald–Letnikov derivative with respect to t as follows:
Grünwald–Letnikov left-sided derivative ( GLD α , a + ) with respect to t:
G L a + D t α f ( s , t ) = lim h 0 1 h α k = 0 ( 1 ) k ( α + 1 ) k k ! ( α k + 1 ) f ( s , t k h ) ,
and,
Grünwald–Letnikov right-sided derivative ( GLD α , b ) with respect to t:
G L b D t α f ( s , t ) = lim h 0 1 h α k = 0 ( 1 ) k ( α + 1 ) k k ! ( α k + 1 ) f ( s , t + k h ) .
where ( α + 1 ) k = ( α + 1 ) ( α ) ( α 1 ) ( α k + 1 ) denotes the Pochhammer symbol.
Secondly, we apply Grünwald–Letnikov derivatives in two variables to image pixels using the following steps:
  • Image Representation: Consider a grayscale image represented as a matrix I, where each element I ( s , t ) denotes the intensity value at pixel coordinates ( s , t ) .
  • Grünwald–Letnikov Derivative Configuration: For a pixel at coordinates ( s , t ) , the Grünwald–Letnikov derivative concerning ( s , t ) involves computing fractional differences with neighboring pixels within a defined local patch.
  • Iterative Calculation: Apply this procedure for every pixel ( s , t ) in the image, changing h and α by the needs of the application. A thorough examination of local intensity variations is made possible by the iteration’s ability to record fractional differences across the spatial dimensions s and t.
  • Aggregation and Output: The Grünwald–Letnikov derivative values can be obtained by aggregating these fractional differences over the entire image. These derivatives can then be used directly to process deeper layers of a convolutional neural network and enhance network performance.

5. Experimental Evaluation

This section examines GLFNet with the same model without the Grünwald–Letnikov fractional step and discusses the training and validation performance of our Grünwald–Letnikov fractional neural network (GLFNet) model intended for edge detection. To evaluate the model’s learning behavior and capacity to generalize to new data, we examine the important metrics—loss and accuracy—across the epochs. The training dataset was iterated over a predetermined number of epochs using fractional orders α = 1 , α = 0.95 , α = 0.97 , and α = 0.99 to examine their effects on convolutional neural networks. α = 1 represents the standard integer-order convolution, serving as a baseline. The selected fractional orders near 1 ( α = 0.95 , 0.97 , and 0.99 ) were chosen to evaluate how different degrees of fractional differentiation influence model performance. This range was selected to balance theoretical insights with empirical observations of performance improvements and practical considerations related to computational efficiency. The model was modified based on the optimization of the chosen loss functions for each fractional order.

5.1. Our CNN Architecture and Dataset MNIST

The convolutional neural network (CNN) architecture and dataset utilized in this study is structured as follows [39,40,41]:

5.1.1. Convolutional Layers

  • First Convolutional Layer:
    -
    Filters: 32 filters of size 3 × 3
    -
    Activation Function: ReLU (rectified linear unit)
    -
    Padding: ‘same’, to ensure that the spatial dimensions of the input are preserved
    -
    Input Shape: ( 28 , 28 , 1 ) , where 28 × 28 represents the spatial dimensions and 1 denotes the single grayscale color channel
    -
    Number of Parameters: 320
  • Second Convolutional Layer:
    -
    Filters: 1 filter of size 3 × 3
    -
    Activation Function: Sigmoid
    -
    Padding: ‘same’, to maintain the output feature map dimensions equal to the input
    -
    Number of Parameters: 289

5.1.2. Activation Functions

  • ReLU Activation: Applied in the first convolutional layer to introduce non-linearity, facilitating the network’s ability to learn complex patterns.
  • Sigmoid Activation: Utilized in the second convolutional layer to produce binary edge maps, which are essential for edge detection tasks.

5.1.3. Padding

Both convolutional layers use ’same’ padding, ensuring that the dimensions of the output feature maps match the spatial dimensions of the input images.

5.1.4. Optimizer

  • Optimizer: Adam
  • Learning Rate: 0.001
  • Description: Adam is chosen for its adaptive learning rate capabilities, which enhances its performance in training neural networks.

5.1.5. Loss Function

  • Loss Function: Binary Crossentropy
  • Application: Suitable for binary classification tasks, such as edge detection, where the objective is to differentiate between foreground and background.

5.1.6. Training and Evaluation

  • Dataset: MNIST
  • Training Procedure: The model is trained for 15 epochs with a batch size of 128.
  • Training Dataset Size: 1000 samples
  • Validation Dataset Size: 100 samples

5.1.7. Parameter Count and Efficiency

  • Total Number of Parameters: 609
  • Parameters per Layer:
    -
    First Convolutional Layer: 320
    -
    Second Convolutional Layer: 289
Table 6, Table 7 and Table 8 provide a comprehensive comparison between the CNN and GLFNet models across various values of the parameter α ( α = 1, α = 0.95, α = 0.97, and α = 0.99). Each table highlights key performance metrics, including loss, F1 score, training time, and inference time, for both models.
Figure 1, Figure 2 and Figure 3 offer a detailed visual comparison between the CNN and GLFNet models at different values of the parameter α .

5.2. Comparative Analysis

This comparison focused on convolutional neural networks (CNNs) because GLFNet is intended to be a preprocessing step rather than a complete network architecture. Unlike current approaches, which use fractional operations within CNN layers, our method leverages fractional calculus to improve feature extraction before to CNN application. Using a normal CNN as a baseline helps us to illustrate how our preprocessing technique affects CNN performance. This approach demonstrates how GLFNet increases the quality of inputs to CNNs, increasing their effectiveness. We did not compare GLFNet to other CNN architectures because its key contribution is in preprocessing capacity. When compared to the traditional CNN strategy, the proposed method improved training and inference speeds while significantly lowering loss error. Furthermore, the model’s performance is as indicated by the F1-measure demonstrated significant enhancements using GLFNet.

5.2.1. Loss Error

The loss error values for the first test image are shown in Table 6. The GLFNet demonstrated lower loss error values compared to the CNN:
GLFNet Loss Error: −0.7447, −2.1166, −2.4064, −2.228 CNN Loss Error: 0.122, 0.1186, 0.11147, 0.148 These results show that GLFNet consistently generated fewer loss errors across a variety of test scenarios.

5.2.2. F1-Measure

The F1-measure, which assesses the balance of precision and recall, was greater for the CNN in most circumstances, while GLFNet performed competitively:
GLFNet F1-Measure: 0.1395, 0.6629, 0.6203, 0.6591 CNN F1-Measure: 0.8000, 0.6552, 0.6129, 0.6514 Although the F1-measure values of the CNN were often closer, GLFNet performed better, especially in several test instances.

5.2.3. Training Time

The training times (in seconds) for GLFNet and the CNN are summarized below. GLFNet showed a reduction in training time compared to the CNN: GLFNet Training Time: 16.6261, 15.4130, 15.7058, 20.5462 CNN Training Time: 20.5560, 20.6240, 20.5701, 17.3720.

5.2.4. Inference Time

Inference times (in seconds) were also measured. GLFNet performed faster in most cases: GLFNet Inference Time: 0.0986, 0.0849, 0.0750, 0.0810 CNN Inference Time: 0.1028, 0.1050, 0.0944, 0.0838.

5.2.5. Parameters

The CNN has 609 parameters, while GLFNet maintains a similar parameter count but improves performance metrics.
Also, in Figure 1, Figure 2 and Figure 3, we observe that at the different values of α in (GLF net), both the training and validation losses initially decrease steadily as the epochs progress and the curves. This indicates that the model is more effectively learning from the training data than the CNN. As the same epochs (15 epochs) proceed, we can see from Figure 1, Figure 2 and Figure 3 that both the training and validation losses in the GLFNet drop steadily at different values of α compared to the CNN model. This suggests that the model is learning from the training set of data more successfully than the CNN. Additional comparisons of performance metrics between various test situations are given in Table 7 and Table 8, which further demonstrate the efficacy of GLFNet.

5.3. Discussion

GLFNet consistently produced reduced loss errors throughout test scenarios, demonstrating its usefulness in minimizing disparities between projected and actual values when compared to the CNN. It also demonstrated increased computational efficiency, with training and inference times decreased by about 1.14 times, which is beneficial for real-time processing applications. Although the CNN produced better F1-measure values in most cases, GLFNet remained competitive and, in some circumstances, surpassed the CNN, demonstrating its ability to balance precision and recall well. However, GLFNet’s F1-measure performance was often lower than that of the CNN, indicating that there is still space for improvement in the balance of precision and recall. The use of fractional calculus complicates parameter tuning, especially in determining the ideal value for α . This complexity needs more extensive testing. Furthermore, while GLFNet demonstrated promise in the test settings, its performance on bigger or more diverse datasets warrants further investigation. Despite having a similar parameter count to the CNN, GLFNet’s fractional calculus technique may not always result in significant improvements, necessitating greater investigation into its parameter efficiency and overall computing benefits. Overall, GLFNet showed increased efficiency in terms of reduced loss errors and quicker inference times. Although the F1-measure of the CNN was higher in the majority of test instances, GLFNet’s competitive performance and shorter training time demonstrate its potential for effective implementation.

6. Conclusions

In conclusion, this study offers three significant contributions. First, we introduce the Grünwald–Letnikov fractional neural network (GLFNet), which innovatively merges convolutional neural networks (CNNs) with fractional derivatives. This novel integration enhances edge recognition by leveraging Grünwald–Letnikov fractional convolution, leading to notable improvements in precision and reliability for identifying image edges. Second, our GLFNet design surpasses traditional CNNs in several key metrics. It achieves a substantial reduction in loss error, with values ranging from −0.7447 to −2.4064 for GLFNet, compared to 0.11147 to 0.148 for CNN. Furthermore, GLFNet improves computational efficiency, with training durations decreased by around 1.14 times and inference times lowered by a comparable ratio. Although GLFNet’s F1-measure spans from 0.1395 to 0.6629, whereas CNNs vary from 0.6129 to 0.8000, GLFNet still provides a competitive combination of precision and recall in edge detection, all while using comparable memory. Third, we carried out comprehensive edge detection tests on the MNIST dataset. Despite these advances, GLFNet has several limitations, with the F1-measure being lower than CNNs in some circumstances. Also, the parameter α requires tuning, indicating room for improvement. Future research will focus on developing GLFNet and applying it to more difficult and diverse vision tasks, to increase its utility and efficacy in broader vision applications.

Author Contributions

Conceptualization, D.E.E. and A.B.M.A.; methodology D.E.E. and A.B.M.A.; software. D.E.E.; formal analysis, D.E.E. and A.B.M.A.; writing—original draft preparation, D.E.E. and A.B.M.A.; writing—review and editing, D.E.E. and A.B.M.A.; funding acquisition, A.B.M.A.; All authors have read and agreed to the published version of the manuscript.

Funding

The Researchers Supporting Project number (RSPD2024R920), King Saud University, Saudi Arabia.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare that there are no conflicts of interest regarding the publication of this article.

References

  1. Ziou, D.; Tabbone, S. Edge detection techniques-an overview. Pattern Recognit. Image Anal. Adv. Math. Theory Appl. 1998, 8, 537–559. [Google Scholar]
  2. Spontón, H.; Cardelino, J. A review of classic edge detectors. Image Process. Line 2015, 5, 90–123. [Google Scholar] [CrossRef]
  3. Bharadiya, J. Convolutional neural networks for image classification. Int. J. Innov. Sci. Res. Technol. 2023, 8, 673–677. [Google Scholar]
  4. Kaur, R.; Kumar, R.; Gupta, M. Deep neural network for food image classification and nutrient identification: A systematic review. Rev. Endocr. Metab. Disord. 2023, 24, 633–653. [Google Scholar] [CrossRef] [PubMed]
  5. Maurício, J.; Domingues, I.; Bernardino, J. Comparing vision transformers and convolutional neural networks for image classification: A literature review. Appl. Sci. 2023, 13, 5521. [Google Scholar] [CrossRef]
  6. Emek Soylu, B.; Guzel, M.S.; Bostanci, G.E.; Ekinci, F.; Asuroglu, T.; Acici, K. Deep-learning-based approaches for semantic segmentation of natural scene images: A review. Electronics 2023, 12, 2730. [Google Scholar] [CrossRef]
  7. Naranjo, M.; Fuentes, D.; Muelas, E.; Diez, E.; Ciruelo, L.; Alonso, C.; Abenza, E.; Gómez-Espinosa, R.; Luengo, I. Object detection-based system for traffic signs on drone-captured images. Drones 2023, 7, 112. [Google Scholar] [CrossRef]
  8. Wu, T.; Dong, Y. YOLO-SE: Improved YOLOv8 for remote sensing object detection and recognition. Appl. Sci. 2023, 13, 12977. [Google Scholar] [CrossRef]
  9. Chen, Z.; Lu, J.; Wang, H. A review of posture detection methods for pigs using deep learning. Appl. Sci. 2023, 13, 6997. [Google Scholar] [CrossRef]
  10. Stern, L.; Roshan Fekr, A. In-bed posture classification using deep neural network. Sensors 2023, 23, 2430. [Google Scholar] [CrossRef] [PubMed]
  11. Ruan, X.; Liu, Y.; Yuan, C.; Li, B.; Hu, W.; Li, Y.; Maybank, S. Edp: An efficient decomposition and pruning scheme for convolutional neural network compression. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4499–4513. [Google Scholar] [CrossRef]
  12. Chen, D.; Li, Y.; Qiu, M.; Wang, Z.; Li, B.; Ding, B.; Deng, H.; Huang, J.; Lin, W.; Zhou, J. Adabert: Task-adaptive bert compression with differentiable neural architecture search. arXiv 2020, arXiv:2001.04246. [Google Scholar]
  13. Joshi, S.; Ojo, S.; Yadav, S.; Gulia, P.; Gill, N.S.; Alsberi, H.; Rizwan, A.; Hassan, M.M. Object detection and classification from compressed video streams. Expert Syst. 2023, e13382. [Google Scholar] [CrossRef]
  14. Arora, S.; Mathur, T.; Agarwal, S.; Tiwari, K.; Gupta, P. Applications of fractional calculus in computer vision: A survey. Neurocomputing 2022, 489, 407–428. [Google Scholar] [CrossRef]
  15. Zhao, C.; Ye, J.; Zhu, Z.; Huang, Y. FLRNN-FGA: Fractional-Order Lipschitz Recurrent Neural Network with Frequency-Domain Gated Attention Mechanism for Time Series Forecasting. Fractal Fract. 2024, 8, 433. [Google Scholar] [CrossRef]
  16. Coelho, C.; Costa, M.F.P.; Ferrás, L.L. Fractional Calculus Meets Neural Networks For Computer Vision: A Survey. AI 2024, 5, 1391–1426. [Google Scholar] [CrossRef]
  17. Joshi, M.; Bhosale, S.; Vyawahare, V.A. A survey of fractional calculus applications in artificial neural networks. Artif. Intell. Rev. 2023, 56, 13897–13950. [Google Scholar] [CrossRef]
  18. Zhao, H.; Chen, J.; Jiang, P.; Zeng, Z. Optimizing Neural Network Image Classification with Fractional Order Gradient Methods. Available online: https://ssrn.com/abstract=4673972 (accessed on 17 August 2024).
  19. Jia, X.; Liu, S.; Feng, X.; Zhang, L. Focnet: A fractional optimal control network for image denoising. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6054–6063. [Google Scholar]
  20. Chen, M.; Pu, Y.F.; Bai, Y.C. A fractional-order variational residual CNN for low dose CT image denoising. In Proceedings of the Intelligent Computing Theories and Application: 15th International Conference, ICIC 2019, Nanchang, China, 3–6 August 2019; Proceedings, Part I 15. Springer: Berlin/Heidelberg, Germany, 2019; pp. 238–249. [Google Scholar]
  21. Cao, T.; Wang, W.; Tighe, S.; Wang, S. Crack image detection based on fractional differential and fractal dimension. IET Comput. Vis. 2019, 13, 79–85. [Google Scholar] [CrossRef]
  22. Zamora Esquivel, J.; Cruz Vargas, A.; Camacho Perez, R.; Lopez Meyer, P.; Cordourier, H.; Tickoo, O. Adaptive activation functions using fractional calculus. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Seoul, Republic of Korea, 27–28 October 2019. [Google Scholar]
  23. Zheng, T. Semantic Segmentation Based on Deeplabv3+ With Fractional-Order Image Enhancement Convolution Kernel. In Proceedings of the 3rd International Conference on Control and Computer Vision, Macau, China, 23–25 August 2020; pp. 34–39. [Google Scholar]
  24. Sahlol, A.T.; Yousri, D.; Ewees, A.A.; Al-Qaness, M.A.; Damasevicius, R.; Elaziz, M.A. COVID-19 image classification using deep features and fractional-order marine predators algorithm. Sci. Rep. 2020, 10, 15364. [Google Scholar] [CrossRef]
  25. Chandra, S.K.; Bajpai, M.K. Fractional mesh-free linear diffusion method for image enhancement and segmentation for automatic tumor classification. Biomed. Signal Process. Control 2020, 58, 101841. [Google Scholar] [CrossRef]
  26. Badashah, S.J.; Basha, S.S.; Ahamed, S.R.; Subba Rao, S. Fractional-Harris hawks optimization-based generative adversarial network for osteosarcoma detection using Renyi entropy-hybrid fusion. Int. J. Intell. Syst. 2021, 36, 6007–6031. [Google Scholar] [CrossRef]
  27. Chen, M.; Pu, Y.F.; Bai, Y.C. Low-dose CT image denoising using residual convolutional network with fractional TV loss. Neurocomputing 2021, 452, 510–520. [Google Scholar] [CrossRef]
  28. Zhao, X.; Tao, R.; Li, W.; Philips, W.; Liao, W. Fractional Gabor convolutional network for multisource remote sensing data classification. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–18. [Google Scholar] [CrossRef]
  29. Zamora, J.; Vargas, J.A.C.; Rhodes, A.; Nachman, L.; Sundararajan, N. Convolutional filter approximation using fractional calculus. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 383–392. [Google Scholar]
  30. Raubitzek, S.; Mallinger, K.; Neubauer, T. Combining fractional derivatives and machine learning: A review. Entropy 2022, 25, 35. [Google Scholar] [CrossRef]
  31. Liu, Y.; Wang, Y.; Li, Y.; Li, Q.; Wang, J. SETR-YOLOv5n: A lightweight low-light lane curvature detection method based on fractional-order fusion model. IEEE Access 2022, 10, 93003–93016. [Google Scholar] [CrossRef]
  32. Kuo, P.H.; Tseng, Y.R.; Luan, P.C.; Yau, H.T. Novel fractional-order convolutional neural network based chatter diagnosis approach in turning process with chaos error mapping. Nonlinear Dyn. 2023, 111, 7547–7564. [Google Scholar] [CrossRef]
  33. Cheng, J.; Chen, Q.; Huang, X. An Algorithm for Crack Detection, Segmentation, and Fractal Dimension Estimation in Low-Light Environments by Fusing FFT and Convolutional Neural Network. Fractal Fract. 2023, 7, 820. [Google Scholar] [CrossRef]
  34. Ding, Y.; Nan, X. On Edge Detection Algorithms for Water-Repellent Images of Insulators Taking into Account Efficient Approaches. Symmetry 2023, 15, 1418. [Google Scholar] [CrossRef]
  35. Kumar, M.; Mehta, U.; Cirrincione, G. Enhancing neural network classification using fractional-order activation functions. AI Open 2024, 5, 10–22. [Google Scholar] [CrossRef]
  36. Teodoro, G.S.; Machado, J.T.; De Oliveira, E.C. A review of definitions of fractional derivatives and other operators. J. Comput. Phys. 2019, 388, 195–208. [Google Scholar] [CrossRef]
  37. Atici, F.M.; Chang, S.; Jonnalagadda, J. Grünwald-Letnikov fractional operators: From past to present. Fract. Differ. Calc. 2021, 11, 147–159. [Google Scholar]
  38. Caputo, M. Linear models of dissipation whose Q is almost frequency independent—II. Geophys. J. Int. 1967, 13, 529–539. [Google Scholar] [CrossRef]
  39. MNIST Dataset. Available online: http://yann.lecun.com/exdb/mnist/ (accessed on 12 August 2024).
  40. Keras. Keras Documentation. Keras. 2024. Available online: https://keras.io/getting_started/ (accessed on 12 August 2024).
  41. Chollet, F. Deep Learning with Python; Manning Publications Co.: Shelter Island, NY, USA, 2018; Available online: https://tanthiamhuat.wordpress.com/wp-content/uploads/2018/03/deeplearningwithpython.pdf (accessed on 12 August 2024).
Figure 1. (a) CNN and GLFNet at α = 1 ; (b) CNN and GLFNet at α = 0.95 ; (c) CNN and GLFNet at α = 0.97 ; (d) CNN and GLFNet at α = 0.99 .
Figure 1. (a) CNN and GLFNet at α = 1 ; (b) CNN and GLFNet at α = 0.95 ; (c) CNN and GLFNet at α = 0.97 ; (d) CNN and GLFNet at α = 0.99 .
Electronics 13 03298 g001
Figure 2. (a) CNN and GLFNet at α = 1 ; (b) CNN and GLFNet at α = 0.95 ; (c) CNN and GLFNet at α = 0.97 ; (d) CNN and GLFNet at α = 0.99 .
Figure 2. (a) CNN and GLFNet at α = 1 ; (b) CNN and GLFNet at α = 0.95 ; (c) CNN and GLFNet at α = 0.97 ; (d) CNN and GLFNet at α = 0.99 .
Electronics 13 03298 g002
Figure 3. (a) CNN and GLFNet at α = 1 ; (b) CNN and GLFNet at α = 0.95 ; (c) CNN and GLFNet at α = 0.97 ; (d) CNN and GLFNet at α = 0.99 .
Figure 3. (a) CNN and GLFNet at α = 1 ; (b) CNN and GLFNet at α = 0.95 ; (c) CNN and GLFNet at α = 0.97 ; (d) CNN and GLFNet at α = 0.99 .
Electronics 13 03298 g003aElectronics 13 03298 g003b
Table 1. Overview of Methodologies.
Table 1. Overview of Methodologies.
AspectGLFNetFOCNetFVCNN
Processing StagePre-Convolutional (Uses Fractional Calculus)Within Convolutional Layers (Focus Mechanism)Within Convolutional Layers (Focused View)
Mathematical BasisFractional Derivatives (Grünwald–Letnikov)Standard Convolutional Layers with FocusStandard Convolutional Layers with Focused View
Feature Extraction MethodFractional Calculations Applied Before ConvolutionsConvolutional Layers with Focused ExtractionConvolutional Layers with Enhanced Focus
Table 2. Comparison of Feature Processing.
Table 2. Comparison of Feature Processing.
AspectGLFNetFOCNetFVCNN
Initial ProcessingFractional Calculus Applied Directly to Raw DataDirect Convolutional Layer ProcessingDirect Convolutional Layer Processing
Focus MechanismNot Used (Fractional Processing Precedes Convolution)Focus Mechanism Applied to Highlight RegionsEnhanced Focus on Specific Image Regions
Feature Extraction ApproachFractional Derivatives Enhance Feature RepresentationStandard Convolutional Filters UsedStandard Convolutional Filters with Focused Views
Table 3. Methods for Classification and Localization.
Table 3. Methods for Classification and Localization.
AspectGLFNetFOCNetFVCNN
Classification StagePerformed After Fractional Feature ExtractionClassification Within Focused AreasClassification Within Focused Areas
Bounding Box PredictionNot Specifically AddressedBounding Boxes Estimated Around Detected ObjectsBounding Boxes Estimated Around Detected Objects
Table 4. Post-Processing and Additional Characteristics.
Table 4. Post-Processing and Additional Characteristics.
AspectGLFNetFOCNetFVCNN
Post-ProcessingNot SpecifiedUses Non-Maximum Suppression for RefinementEmploys Similar Post-Processing to Standard CNNs
Handling Long-Range CorrelationsEffectively Captures Using Fractional CalculusLess Focus on Long-Range CorrelationsLess Focus on Long-Range Correlations
Memory EffectsSimulates Memory Effects Through Fractional CalculusStandard CNN Methods with Minimal Memory FocusStandard CNN Methods with Minimal Memory Focus
Table 5. Complexity and Suitability Analysis.
Table 5. Complexity and Suitability Analysis.
AspectGLFNetFOCNetFVCNN
Complexity of IntegrationSimple Integration with Fractional CalculusModerate Complexity Due to Focus MechanismModerate Complexity with Focused View Mechanism
Suitability for Complex FeaturesHigh Suitability Due to Enhanced Feature LearningModerate Suitability with Focused Region AccuracyModerate Suitability with Focused Region Accuracy
Table 6. Loss and F1 Score for CNN and GLFNet models for the first image test.
Table 6. Loss and F1 Score for CNN and GLFNet models for the first image test.
CNN and GLF net at α = 1
ModelLossF1 ScoreTraining Time (s)Inference Time (s)
CNN0.1220.820.55660.1028
GLF net−0.74470.139516.62610.0986
CNN and GLFNet at α = 0.95
ModelLossF1 ScoreTraining Time (s)Inference Time (s)
CNN0.11860.655220.6240.105
GLF net−2.11660.662915.4130.0849
CNN and GLFNet at α = 0.97
ModelLossF1 ScoreTraining Time (s)Inference Time (s)
CNN0.11470.612920.57010.0944
GLF net−2.40640.620315.70580.075
CNN and GLFNet at α = 0.99
ModelLossF1 ScoreTraining Time (s)Inference Time (s)
CNN0.1480.651417.3720.0838
GLF net−2.2280.659120.54620.081
Table 7. Loss and F1 Score for CNN and GLFNet models for the second image test.
Table 7. Loss and F1 Score for CNN and GLFNet models for the second image test.
CNN and GLFNet at α = 1
ModelLossF1 ScoreTraining Time (s)Inference Time (s)
CNN0.12880.884615.7520.0843
GLF net−0.64280.38320.55540.1056
CNN and GLF net at α = 0.95
ModelLossF1 ScoreTraining Time (s)Inference Time (s)
CNN0.12050.64320.56680.0794
GLF net−2.28620.675315.350.0969
CNN and GLFNet at α = 0.97
ModelLossF1 ScoreTraining Time (s)Inference Time (s)
CNN0.13240.674420.5770.0817
GLF net−2.56810.707520.55120.086
CNN and GLFNet at α = 0.99
ModelLossF1 ScoreTraining Time (s)Inference Time (s)
CNN0.12790.674420.55540.0808
GLF net−2.07720.707520.56380.0842
Table 8. Loss and F1 Score for CNN and GLFNet models for the Third image test.
Table 8. Loss and F1 Score for CNN and GLFNet models for the Third image test.
CNN and GLFNet at α = 1
ModelLossF1 ScoreTraining Time (s)Inference Time (s)
CNN0.1440.813620.55660.087
GLF net−0.53640.317120.56230.0831
CNN and GLFNet at α = 0.95
ModelLossF1 ScoreTraining Time (s)Inference Time (s)
CNN0.13680.688520.58510.0744
GLF net−2.07020.717920.56590.0776
CNN and GLFNet at α = 0.97
ModelLossF1 ScoreTraining Time (s)Inference Time (s)
CNN0.11890.675220.5590.077
GLF net 2.0227 0.704415.35860.0762
CNN and GLFNet at α = 0.99
ModelLossF1 ScoreTraining Time (s)Inference Time (s)
CNN0.12520.719215.39530.0895
GLF net−2.08830.749215.96550.079
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Elgezouli, D.E.; Alzahrani, A.B.M. Optimizing Edge Detection Efficiency with a Grünwald–Letnikov Fractional Network. Electronics 2024, 13, 3298. https://doi.org/10.3390/electronics13163298

AMA Style

Elgezouli DE, Alzahrani ABM. Optimizing Edge Detection Efficiency with a Grünwald–Letnikov Fractional Network. Electronics. 2024; 13(16):3298. https://doi.org/10.3390/electronics13163298

Chicago/Turabian Style

Elgezouli, Diaa Eldin, and Abdulrahman B. M. Alzahrani. 2024. "Optimizing Edge Detection Efficiency with a Grünwald–Letnikov Fractional Network" Electronics 13, no. 16: 3298. https://doi.org/10.3390/electronics13163298

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop