Next Article in Journal
SaMDE: A Self Adaptive Choice of DNDE and SPIDE Algorithms with MRLDE
Next Article in Special Issue
Enhanced Pelican Optimization Algorithm with Deep Learning-Driven Mitotic Nuclei Classification on Breast Histopathology Images
Previous Article in Journal / Special Issue
Multi-Strategy Improved Sand Cat Swarm Optimization: Global Optimization and Feature Selection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bio-Inspired Spotted Hyena Optimizer with Deep Convolutional Neural Network-Based Automated Food Image Classification

1
Department of Computer Science, College of Science & Art at Mahayil, King Khalid University, Muhayil 61321, Saudi Arabia
2
Department of Information Systems, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
3
Department of Special Education, College of Education, King Saud University, Riyadh 12372, Saudi Arabia
4
Unit of Scientific Research, Applied College, Qassim University, Buraydah 51425, Saudi Arabia
5
Research Center, Future University in Egypt, New Cairo 11835, Egypt
6
Department of Computer and Self Development, Preparatory Year Deanship, Prince Sattam bin Abdulaziz University, Al-Kharj 11942, Saudi Arabia
*
Author to whom correspondence should be addressed.
Biomimetics 2023, 8(6), 493; https://doi.org/10.3390/biomimetics8060493
Submission received: 13 September 2023 / Revised: 23 September 2023 / Accepted: 28 September 2023 / Published: 18 October 2023
(This article belongs to the Special Issue Bionic Artificial Neural Networks and Artificial Intelligence)

Abstract

:
Food image classification, an interesting subdomain of Computer Vision (CV) technology, focuses on the automatic classification of food items represented through images. This technology has gained immense attention in recent years thanks to its widespread applications spanning dietary monitoring and nutrition studies to restaurant recommendation systems. By leveraging the developments in Deep-Learning (DL) techniques, especially the Convolutional Neural Network (CNN), food image classification has been developed as an effective process for interacting with and understanding the nuances of the culinary world. The deep CNN-based automated food image classification method is a technology that utilizes DL approaches, particularly CNNs, for the automatic categorization and classification of the images of distinct kinds of foods. The current research article develops a Bio-Inspired Spotted Hyena Optimizer with a Deep Convolutional Neural Network-based Automated Food Image Classification (SHODCNN-FIC) approach. The main objective of the SHODCNN-FIC method is to recognize and classify food images into distinct types. The presented SHODCNN-FIC technique exploits the DL model with a hyperparameter tuning approach for the classification of food images. To accomplish this objective, the SHODCNN-FIC method exploits the DCNN-based Xception model to derive the feature vectors. Furthermore, the SHODCNN-FIC technique uses the SHO algorithm for optimal hyperparameter selection of the Xception model. The SHODCNN-FIC technique uses the Extreme Learning Machine (ELM) model for the detection and classification of food images. A detailed set of experiments was conducted to demonstrate the better food image classification performance of the proposed SHODCNN-FIC technique. The wide range of simulation outcomes confirmed the superior performance of the SHODCNN-FIC method over other DL models.

1. Introduction

Food image detection and identification are the existing research subjects in the domain of Computer Vision (CV). “Food” is one of the developing areas of interest for the CV community as well as multimedia [1], whereas image detection and identification remain a highly significant problem in the medical field as well. In the literature, a new food recording tool called “FoodLog” has been developed that supports users to record their daily meals with the aid of an image recovery technique [2]. However, it is extremely challenging to perform food image analyses. For instance, the identification of food products in images is still a challenging process due to low inter-class variance and high intra-class variance [3]. Furthermore, many food classes have not yet been effectively classified. Therefore, automated food detection is a developing area of research not only in the image recognition domain but also in social media research [4]. A significant number of researchers paid attention to this domain due to its improving advantages from a medical viewpoint [5]. Automated food identification tools can support and facilitate decision-making methods in terms of calorie calculation, food quality detection, diet monitoring systems to overcome obesity, etc. [6]. In general, food is naturally distorted and has a broad difference in appearance [7]. Food images may have high intra-class and low inter-class variances, owing to which standard techniques may not be able to detect complex features in the images. This drawback in the food identification process makes it a challenging task since complex features cannot be detected using conventional methods [8].
Recently, several developments have occurred in the domain of dietary valuation depending on multimedia approaches, e.g., food image analysis [9]. In the literature, an automated image-based nutritional assessment technique was proposed in which the technique had the following key stages: food image identification, recognition of food products, weight or quantity valuation, and lastly, nutritional and caloric value assessment [10]. In recent years, developments in Machine Learning (ML), image processing, and specifically Convolutional Neural Networks (CNN), and Deep-Learning (DL) techniques have heavily benefited the image classification and detection processes, comprising the issue of food image identification [11]. Researchers have developed diverse phases of food detection systems, despite which it remains challenging to find a satisfactory and efficient solution for food identification and classification with high accuracy. This is because there exist extensive types of food products and extremely complicated hybrid food products in food images [12]. Therefore, it is tremendously challenging to detect all food items accurately since a variety of food items can appear similar in terms of shape, color, or context, and are not even differentiable to the human eye [13].
Given this background, the current research article develops the Bio-Inspired Spotted Hyena Optimizer with a Deep Convolutional Neural Network-based Automated Food Image Classification (SHODCNN-FIC) approach. The presented SHODCNN-FIC method exploits the DL model with hyperparameter tuning approaches for the classification of food images. To achieve this, the SHODCNN-FIC method exploits the DCNN-based Xception model to derive the feature vectors. In addition to this, the SHODCNN-FIC technique uses the SHO algorithm for the optimal hyperparameter selection of the Xception model. The SHODCNN-FIC technique uses the Extreme Learning Machine (ELM) model for the detection and classification of food images. A detailed set of experiments was conducted to illustrate the better food image classification performance of the SHODCNN-FIC technique. The key contributions of the current study are summarized below.
(a)
The development of an automated SHODCNN-FIC algorithm, including Xception feature extraction, SHO-based parameter tuning, and ELM-based classification for food classification. To the best of the authors’ knowledge, the SHODCNN-FIC approach has never been reported in the literature.
(b)
The development of a new technique, i.e., SHODCNN-FIC, by combining bio-inspired optimization and DL for automatic food image classification. The proposed technique is highly useful in many real-time applications involving dietary analysis and restaurant menu management.
(c)
The SHODCNN-FIC leverages the power of deep learning using the DCNN-based Xception model for extracting the feature vectors from food images. Furthermore, the optimum fine-tuning of the hyperparameters for the Xception model, using the SHO technique, improves the performance of the DL model by fine-tuning its parameters.
(d)
The application of the ELM model for the actual detection and classification of food images. ELM is known for its high accuracy and fast training features in different machine-learning tasks.
The rest of the paper is organized as follows. Section 2 discusses the related works, and Section 3 details the proposed model. Then, Section 4 provides the analytical results, and Section 5 concludes the paper.

2. Related Works

Shah and Bhavsar [14] introduced the depth-restricted-CNN (DRCNN) method in which the Transfer Learning (TL) technique was applied to a few frameworks, such as the Alexnet, Resnet-50, Inceptionv3, VGG16, and the VGG19 framework. This method was presented as a Batch Normalization (BN) approach that heavily enhances performance with a lower number of parameters. Chopra and Purwar [15] introduced a food image detection system composed of CNN, GA, and PSO to improve outcomes. CNN, as an approach, was utilized in this study for the classification of food images. The reason for supplementing the CNN technique with GA and PSO is to ensure an efficient classification outcome. In the literature [16], an enhanced VGG16 framework was proposed through a food classification technique. This approach employed the Asymmetric Convolution Block (ACB) to change the convolution kernels and enhance the effectiveness of the standard technique. This technique also involved BN and pooling layers to enrich the normalization. The attention mechanism should be integrated with the CNN technique due to its complications, such as higher texture similarity, complex context, and contextual intervention.
Chopra and Purwar [17] developed the Squirrel Search Algorithm (SSA) to provide optimum solutions for multiple thresholds. This technique implemented the CNN method to identify food images. Then, the study suggested that the Enhanced SSA (ESSA) increases food detection accuracy. Yadav and Chand [18] recommended automatic food classification techniques with the help of the DL algorithm. In this study, both VGG-16 and SqueezeNet CNNs were exploited for the classification of food images. These networks demonstrated significantly high effectiveness due to two tasks, namely fine-tuning the hyperparameters and data augmentation. The developed VGG16 framework then enhanced the performance of the automated food image classification process. In the study conducted earlier [19], the CNN approach was introduced and employed to recognize and classify food images. A pre-trained Inceptionv3 CNN algorithm was implemented using TL to stimulate the original customized CNN model. By utilizing the pre-trained method, the learning approach increased, and therefore, more proficient results were achieved. Therefore, data augmentation must be executed on the training set, since it enhances the performance.
Pan et al. [20] recommended a novel classification technique based on the DL approach for the automatic identification of food items. A combinational CNN (CBNet) was created with a subnet integrating method in this study. Primarily, two different NNs were employed to learn important features. Secondarily, a highly developed feature fusion element combined the features from sub-networks. Shermila et al. [21] introduced a new DL-based Food Item Classification (DEEPFIC) method in which the image was processed using the sigmoid stretching algorithm to improve the quality of the images and eliminate the noise. Afterward, the preprocessed image was segmented by employing the Improved Watershed Segmentation (IWS2) technique. In this study, the RNN approach was utilized for the extraction of the features, which were then normalized through the dragonfly algorithm. The Bi-LSTM was employed in this study for the classification of food items.
Though existing automatic food image classification algorithms are valuable, these methods have critical shortcomings that need to be resolved. One important limitation is that these methods often have a narrow scope in identifying food items from certain cultural contexts or cuisines, therefore resulting in poor generalization whenever it encounters unconventional or diverse dishes. Furthermore, these models struggle when handling variations in food presentation techniques, including changes in angles, lighting, or plating styles, which are common in real-time scenarios. Therefore, a research need exists to develop a highly effective and efficient hyperparameter optimization method, particularly a customized one for food image classification tasks. This is because the hyperparameter tuning process is a crucial aspect in enhancing model performance. This involves the exploration of novel techniques or the adaptation of existing ones to overcome the unique challenges posed by food image datasets. The hyperparameter tuning process ultimately affects the generalization ability and performance of the models. In this scenario, the DL model is extremely complicated and has various hyperparameters, namely batch sizes, learning rates, regularization strengths, and layer depths, among others. This hyperparameter considerably affects the performance of the model in terms of learning from data and its capability to fit patterns while preventing over-fitting issues. Without accurate tuning, the DL model converges slowly, becomes trapped in a sub-optimal solution, or fails to adapt to certain features of the dataset. By systematically adjusting the hyperparameters through techniques such as random search, grid search, or metaheuristic optimization algorithms, DL algorithms can be fine-tuned to accomplish high accuracy, fast convergence rate, and best generalization. These outcomes make the model highly effective in different applications. Addressing these research gaps can advance the field of automatic food image classification using hyperparameter tuning and contribute to the development of highly efficient, accurate, and interpretable models with real-time applications in fields such as food waste reduction, dietary analysis, and restaurant menu management.

3. The Proposed Model

The current research article is focused on the design and development of an automated food image detection and classification algorithm named the SHODCNN-FIC approach. The main objective of the SHODCNN-FIC method is to recognize and classify food images into distinct types. The presented SHODCNN-FIC technique exploits the DL model with hyperparameter tuning strategies for the classification of food images. It involves different stages of operations, namely Xception, SHO-based hyperparameter tuning, and ELM classification. Figure 1 shows the entire procedure of the SHODCNN-FIC algorithm.

3.1. Feature Extraction Using Xception Model

The SHODCNN-FIC technique uses the DCNN-based Xception model to derive the feature vectors. The CNN model has proved to be an extraordinary implementer of different image-grouping problems in various fields [22]. The concept of sharing the load in CNN makes the image segmentation process a difficult one by finding the high-level components in the images and diminishing the dissipating tendency problem. The development of the CNN technique incorporates the related layer, convolution layer, and pooling layer. The convolution layer deals with channels, whereas the chief aim is to eliminate the features from the images. Both pooling and convolution layers yield low performances when looking at and holding the basic data in food images. The final layer is the related layer that uses ReLU and takes a certain level component from the food image to gather them into different classifications with marks.
In the XceptionNet model, the conventional convolutional layers are exchanged for depth-wise separable convolution layers. The CNN feature map enables cross-channel and decoupling spatial correlation, whereas the mapping of both correlations is added to the basic operations of the network. Finally, the XceptionNet replaces the main structure of the Inception model. XceptionNet, with 36 convolution layers, is divided into 14 modules. First, the actual image is transformed into defining the possibility contained over different input channels to arrive at the unified images. The following scheme exploits 11 depth-wise convolution layers.

3.2. Hyperparameter Tuning Using the SHO Algorithm

The SHO algorithm is applied for the hyperparameter tuning process. This technique is based on the hunting strategy of the hyena predation model [23]. It comprises four phases: searching, siege, hunting, and attacking. It continuously approaches and encircles the prey after recognizing its location. The individual search for a target is the optimal searching point, and the rest update their positions [24]. The distance equation between the prey and the spotted hyenas is given in Equation (1).
D h = B P P t P t
where   B shows the coefficient vector, D h signifies the distance between the captured prey and the searched individual, and t denotes the iteration count. P P and P show the position of the target and the individuals searched for during t iteration.
P t + 1 = P P t E D h
where E signifies the coefficient vector. The individual search location at the t + 1 iteration is related to the target point and the distance between them. Equations (1) and (2) contain a coefficient vector and the expression is as follows:
B = 2 r d 1 E = 2 h r d 2
where r d 1 and r d 2 are random numbers that lie in the range [0, 1]; h indicates the control factor that drops linearly from 5 to 0 as follows:
h = 5 I i n t e r 5 M i n t e r  
where M i n t e r denotes the maximal iteration count; I i n t e r shows the natural numbers except 0. The spotted hyenas frequently engage in groups to encircle the target. Assume that a better-searched individual is much closer to the target, whereas the rest defines the location of the better-searched individuals as the target location, which forms a cluster and cooperatively moves toward the optimum point location. The computation equation for this scenario is attained from the succeeding formula:
D h = B P P P k P k = P h E D h
where P h indicates the optimum location for the hyena group; P k shows the location of the residual hyenas.
C h = P k + P k + 1 + + P k + N N = C n o s P h , P h + 1 , P h + 2 , , P h + M
where   C h shows a set of N optimum solutions; N refers to the number of spotted hyenas; and C n o s indicates the number of solutions attained. The coefficient vector E changes continuously, whereas the control factor h gradually decreases. Once the absolute value of E becomes less than 1, then it is the attack moment. Otherwise, it continues to search for prey. The computation equation for this scenario is as follows:
P t + 1 = C h N
C h is the optimum search individual set, where the individuals disperse and pursue the target. The condition is that E is higher than 1, after which the distance between the target and spotted hyenas is forcefully limited. Extending the search phase might assist in finding a better hunting position and ensure the successful implementation of global search. Figure 2 depicts the steps involved in the SHO algorithm.
The SHO system progresses an FF to offer the highest classification solution. It expresses a positive integer to exemplify the optimal solution of the candidate performance. The reduction in classifier errors is assumed to be FF.
f i t n e s s x i = C l a s s i f i e r E r r o r R a t e x i = N o .   o f   m i s c l a s s i f i e d   i n s t a n c e s   T o t a l   n o .   o f   i n s t a n c e s   × 100

3.3. Image Classification Using the ELM Model

In this study, the ELM algorithm has been applied to the food image classification process. ELM is an FFNN model for ML that provides various benefits compared to other techniques, including RBFNN and BPNN [25]. It does not need adjustment of the structural parameters, which makes it an easier and highly effective one. In ELM, the weights connected between the hidden and input layers, along with the threshold of the HL neurons, are randomly generated and do not require adjustment during training. Consider that there exist N training instances X i ,   Y i , where 1 i N ,   X j = [ X i 1 ,   X i 2 ,   ,   X i n ] T R n refers to the input vector of the ith sample, and Y i = [ y i 1 , y i 2 ,   , y i m ] T R m indicates the output vector of the ith samples as follows
Σ l f β f g w f X i + b f = T i
In Equation (9), g x denotes the activation function;   T i = [ t i 1 ,   t i 2 ,   ,   t i m ] T implies the output vector of the ith sample. w f = [ w f 1 , w f 2 , , w f n ] T indicates the input weight; b f shows the threshold of f th HL neuron; and β f = β f 1 , β f 2 , , β f m represents the output weight. The objective of the ELM technique is to minimize the output error as follows:
Σ i = 1 N T i Y i = 0  
H β = Y
where H = g W 1 X 1 + b 1 g W l X 1 + b l   g W 1 X N + b 1 g W l X N + b l signifies the layer output matrix of the network; β = [ β 1 T , β 2 T ,   , β l T ] l × m T ; and Y = [ Y 1 T , Y 2 T ,   ,   T N T ] N × m T . The output weight β is attained by resolving the least-square solution as follows:
min   = H β Y
β ^ = H + Y  
Here, the generalized inverse matrix of the output matrix H is represented as   H + .

4. Results and Discussion

The proposed model was simulated using the Python 3.8.5 release. The proposed model was executed on a PC configured with specifications as follows: i5-8600k, GeForce 1050Ti 4 GB, 16 GB RAM, 250 GB SSD, and 1 TB HDD. The food classification outcomes of the SHODCNN-FIC algorithm were tested using the Indian food classification dataset [26]. The dataset includes a total of 1800 samples under six classes, as defined in Table 1. Figure 3 represents some of the sample images.
Figure 4 illustrates the classification outcomes of the SHODCNN-FIC method on 60:40 of the TR set/TS set. Figure 4a,b depict the confusion matrix generated by the SHODCNN-FIC approach. The outcome indicates that the SHODCNN-FIC method detected and categorized all six class labels. Likewise, Figure 4c demonstrates the PR examination results of the SHODCNN-FIC method. The figure infers that the SHODCNN-FIC technique attained the maximum PR outcome under all six classes. Lastly, Figure 4d shows the ROC examination outcomes of the SHODCNN-FIC system. The figure shows that the SHODCNN-FIC technique achieved promising outcomes with maximum ROC values under all six class labels.
The food classification results of the SHODCNN-FIC technique with 60:40 of TR set/TS set are reported in Table 2 and Figure 5. The outcomes infer the proficient performance of the SHODCNN-FIC technique on different food classes. On the 60% TR set, the SHODCNN-FIC technique attained the average a c c u y , p r e c n , r e c a l , F s c o r e , and MCC values of 85.90%, 60.55%, 57.95%, 57.54%, and 50.42%, respectively. Also, on the 40% TS set, the SHODCNN-FIC method accomplished the average a c c u y , p r e c n , r e c a l , F s c o r e , and MCC values of 85.69%, 60.39%, 57.10%, 57.49%, and 49.79%, respectively.
To evaluate the performance of the SHODCNN-FIC method on 60:40 TR set/TS set, the TR and TS a c c u y curves were plotted and are shown in Figure 6. The TR and TS a c c u y values illustrate the performance of the SHODCNN-FIC technique over various number of epochs. The figure shows meaningful insights into the learning task and the generalization abilities of the SHODCNN-FIC method. With an increase in the number of epochs, both TR and TS a c c u y curves improved. The SHODCNN-FIC technique attained improved testing accuracy, which can detect patterns in the TR and TS datasets.
Figure 7 displays the overall TR and TS loss values of the SHODCNN-FIC method on 60:40 of the TR set/TS set over a different number of epochs. The TR loss outcomes show that the model’s loss reduced over an increasing number of epochs. Primarily, the loss values were reduced as the model modified the weight to minimize the prediction error on TR and TS datasets. The loss curves illustrate the extent to which the model fits the training data. Both TR and TS loss values steadily decreased, and this shows that the SHODCNN-FIC technique effectually learned the patterns exhibited in the TR and TS datasets. The SHODCNN-FIC approach adjusted the parameters to minimize the discrepancy between the prediction and the original training label.
Figure 8 shows the classification outcomes of the SHODCNN-FIC method at 70:30 of the TR set/TS set. Figure 8a,b show the confusion matrix generated by the SHODCNN-FIC technique. The outcome indicates that the SHODCNN-FIC method detected and categorized all six class labels. Likewise, Figure 8c demonstrates the PR examination outcomes of the SHODCNN-FIC method. The figure infers that the SHODCNN-FIC technique attained the maximum PR performance under all six classes. Lastly, Figure 8d depicts the ROC examination outcomes of the SHODCNN-FIC approach. The figure portrays the promising performance of the SHODCNN-FIC approach with maximum ROC values under all six class labels.
The food classification results of the SHODCNN-FIC technique with 70:30 of the TR set/TS set are reported in Table 3 and Figure 9. The outcomes found the proficient performance of the SHODCNN-FIC method on different food classes. On the 70% TR set, the SHODCNN-FIC technique achieved average a c c u y , p r e c n , r e c a l , F s c o r e , and MCC values of 85.98%, 60.95%, 57.79%, 58.68%, and 50.76%, respectively. Also, on the 30% TS set, the SHODCNN-FIC technique yielded average a c c u y , p r e c n , r e c a l , F s c o r e , and MCC values of 84.81%, 58.08%, 54.51%, 55.32%, and 46.88%, respectively.
To assess the performance of the SHODCNN-FIC method on the 70:30 TR set/TS set, the TR and TS a c c u y curves were determined and are shown in Figure 10. The TR and TS a c c u y curves illustrate the performance of the SHODCNN-FIC technique over several epochs. The figure offers meaningful insights into the learning task and generalization capabilities of the SHODCNN-FIC model. With an increase in the number of epochs, the TR and TS a c c u y curves were enhanced. It can be observed that the SHODCNN-FIC model obtained enhanced testing accuracy, which can detect the patterns in both TR and TS datasets.
Figure 11 shows the overall TR and TS loss values of the SHODCNN-FIC method at 70:30 of the TR set/TS set over a varying number of epochs. The TR loss values illustrate that the model loss reduced over an increasing number of epochs. Primarily, the loss values were reduced as the technique modified the weight to minimize the prediction error on TR and TS data. The loss curves show the extent to which the model fits the training data. Both TR and TS loss values steadily reduced, which shows that the SHODCNN-FIC model effectually learned the patterns displayed in both TR and TS data. The SHODCNN-FIC method adjusted the parameters to minimize the discrepancy between the predicted and the original training label.
In Table 4 and Figure 12, the overall comparative analysis outcomes between the proposed SHODCNN-FIC system and other approaches are given. The outcomes show that the ResNet50 model achieved the worst results, whereas the NASNetLarge, MobileNet, ResNet101, and ResNet152 models obtained slightly closer performances. Meanwhile, the InceptionResNet model gained a considerably high performance. However, the SHODCNN-FIC technique demonstrated promising performance with the maximum a c c u y , p r e c n , r e c a l , F s c o r e , and MCC values of 85.98%, 60.95%, 57.79%, 58.68%, and 50.76% respectively.
The Computation Time (CT) analysis outcomes of the SHODCNN-FIC technique and other existing DL approaches are demonstrated in Table 5 and Figure 13. The outcomes show the enhanced classification results of the SHODCNN-FIC technique with a minimal CT of 2.03 s. At the same time, it can be observed that the SHODCNN-FIC technique exhibits an enhanced food image classification outcome.

5. Conclusions

This paper designs an automated food image detection and classification algorithm named SHODCNN-FIC. The main objective of the SHODCNN-FIC technique is to recognize and classify distinct types of food images. The presented SHODCNN-FIC technique exploits the DL model with hyperparameter tuning strategies for the classification of food images. It involves different stages of operations, namely the Xception, SHO-based hyperparameter tuning, and the ELM classification. To accomplish this, the SHODCNN-FIC technique employs a DCNN-based Xception model to derive feature vectors. In addition, the SHODCNN-FIC technique uses the SHO approach for the selection of the optimum hyperparameters for the Xception model. The SHODCNN-FIC technique uses the ELM model for both the detection and classification of food images. A detailed set of experiments was conducted to demonstrate the enhanced food image classification performance of SHODCNN-FIC. The extensive simulation values portray the improved performance of the SHODCNN-FIC method over other DL approaches. In the future, the SHODCNN-FIC approach could be used to handle multi-modal inputs, such as the integration of image data with textual descriptions or nutritional information. This could enable a highly comprehensive and accurate food recognition and classification system. Future work should focus on adapting SHODCNN-FIC to a real-time basis. Edge computing environments are increasingly relevant, especially for applications like dietary monitoring or mobile food recognition apps.

Author Contributions

Conceptualization, H.M. and G.A.; methodology, G.A.; software, A.A.A.; validation, N.S.A., H.M. and I.I.; formal analysis, N.S.A.; investigation, G.A.; resources, A.M.; data curation, G.A.; N.S.A.; writing—original draft preparation, H.M., G.A., N.S.A., I.I. and A.M.; writing—review and editing, H.M., A.M. and A.A.A.; visualization, N.S.A.; supervision, G.A.; project administration, H.M.; funding acquisition, H.M., G.A. and N.S.A. All authors have read and agreed to the published version of the manuscript.

Funding

The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work through a large group Research Project under grant number (RGP2/95/44). Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2023R387), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. Research Supporting Project number (RSPD2023R521), King Saud University, Riyadh, Saudi Arabia. This study is supported via funding from Prince Sattam bin Abdulaziz University project number (PSAU/2023/R/1444). This study is partially funded by the Future University in Egypt (FUE).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable to this article as no dataset was generated during the current study.

Conflicts of Interest

The authors declare that they have no conflict of interest. The manuscript was written through the contributions of all authors. All authors have approved the final version of the manuscript.

References

  1. VijayaKumari, G.; Vutkur, P.; Vishwanath, P. Food classification using transfer learning technique. Glob. Transit. Proc. 2022, 3, 225–229. [Google Scholar]
  2. Termritthikun, C.; Jamtsho, Y.; Muneesawang, P.; Zhao, J.; Lee, I. Evolutionary neural architecture search based on efficient CNN models population for image classification. Multimed. Tools Appl. 2023, 82, 23917–23943. [Google Scholar] [CrossRef]
  3. Chun, M.; Jeong, H.; Lee, H.; Yoo, T.; Jung, H. Development of Korean Food Image Classification Model Using Public Food Image Dataset and Deep Learning Methods. IEEE Access 2022, 10, 128732–128741. [Google Scholar] [CrossRef]
  4. Chen, W.; Song, R. A new deep learning-based food recognition system for the mobile terminal. In Proceedings of the 2023 IEEE 12th Data-Driven Control and Learning Systems Conference (DDCLS), Xiangtan, China, 12–14 May 2023; IEEE: Piscataway, NJ, USA, 2003; pp. 112–117. [Google Scholar]
  5. He, L.; Cai, Z.; Ouyang, D.; Bai, H. Food Recognition Model Based on Deep Learning and Attention Mechanism. In Proceedings of the 2022 8th International Conference on Big Data Computing and Communications (BigCom), Xiamen, China, 6–7 August 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 206–216. [Google Scholar]
  6. Li, J.; Chen, W.; Zhu, Y.; Xuan, K.; Li, H.; Zeng, N. Intelligent detection and behavior tracking under ammonia nitrogen stress. Neurocomputing 2023, 59, 126809. [Google Scholar] [CrossRef]
  7. Wu, P.; Wang, Z.; Li, H.; Zeng, N. KD-PAR: A knowledge distillation-based pedestrian attribute recognition model with multi-label mixed feature learning network. Expert Syst. Appl. 2023, 237, 121305. [Google Scholar] [CrossRef]
  8. Liu, M.; Wang, Z.; Li, H.; Wu, P.; Alsaadi, F.E.; Zeng, N. AA-WGAN: Attention augmented Wasserstein generative adversarial network with application to fundus retinal vessel segmentation. Comput. Biol. Med. 2023, 158, 106874. [Google Scholar] [CrossRef] [PubMed]
  9. Mezgec, S.; Seljak, B.K. Deep neural networks for image-based dietary assessment. JoVE J. Vis. Exp. 2021, 169, e61906. [Google Scholar]
  10. Nr, D.; Gk, D.S.; Kumar Pareek, D.P. A Framework for Food recognition and predicting its Nutritional value through Convolution neural network. In Proceedings of the International Conference on Innovative Computing & Communication (ICICC) 2022, Delhi, India, 19–20 February 2022. [Google Scholar]
  11. Alahmari, S.S.; Salem, T. Food State Recognition Using Deep Learning. IEEE Access 2022, 10, 130048–130057. [Google Scholar] [CrossRef]
  12. Tasci, E. Voting combinations-based ensemble of fine-tuned convolutional neural networks for food image recognition. Multimedia Tools Appl. 2020, 79, 30397–30418. [Google Scholar] [CrossRef]
  13. Du, J.; Coumba, B.Y.; Jin, X. Senegalese Food Recognition for Tourism Information Using Convolution Neural Network. In Proceedings of the 2021 5th International Conference on Electronic Information Technology and Computer Engineering, Xiamen, China, 22–24 October 2021; pp. 1177–1182. [Google Scholar]
  14. Shah, B.; Bhavsar, H. Depth-restricted convolutional neural network—A model for Gujarati food image classification. Vis. Comput. 2023, 1–16. [Google Scholar] [CrossRef]
  15. Chopra, M.; Purwar, A. Food Image Recognition by Optimizing CNN with PSO and GA. In Proceedings of the 2022 Fourteenth International Conference on Contemporary Computing 2022, Noida, India, 4–6 August 2022; pp. 37–42. [Google Scholar]
  16. Liu, H.; Gong, H.; Ding, X. Food image recognition algorithm based on improved VGG16. In Proceedings of the 2021 IEEE 2nd International Conference on Information Technology, Big Data and Artificial Intelligence (ICIBA), Chongqing, China, 17–19 December 2021; IEEE: Piscataway, NJ, USA, 2021; Volume 2, pp. 899–903. [Google Scholar]
  17. Chopra, M.; Purwar, A. Food recognition using enhanced squirrel search optimisation algorithm and convolutional neural network. Int. J. Data Anal. Tech. Strateg. 2023, 15, 238–254. [Google Scholar] [CrossRef]
  18. Yadav, S.; Chand, S. Automated food image classification using deep learning approach. In Proceedings of the 2021 7th International Conference on Advanced Computing and Communication Systems (ICACCS), Coimbatore, India, 19–20 March 2021; IEEE: Piscataway, NJ, USA, 2021; Volume 1, pp. 542–545. [Google Scholar]
  19. Chaitanya, A.; Shetty, J.; Chiplunkar, P. Food Image Classification and Data Extraction Using Convolutional Neural Network and Web Crawlers. Procedia Comput. Sci. 2023, 218, 143–152. [Google Scholar] [CrossRef]
  20. Pan, L.; Li, C.; Pouyanfar, S.; Chen, R.; Zhou, Y. A Novel Combinational Convolutional Neural Network for Automatic Food-Ingredient Classification. Comput. Mater. Contin. 2020, 62, 731–746. [Google Scholar] [CrossRef]
  21. Shermila, P.J.; Ahilan, A.; Shunmugathammal, M.; Marimuthu, J. DEEPFIC: Food item classification with calorie calculation using dragonfly deep learning network. Signal Image Video Process. 2023, 17, 3731–3739. [Google Scholar] [CrossRef]
  22. Ogundokun, R.O.; Li, A.; Babatunde, R.S.; Umezuruike, C.; Sadiku, P.O.; Abdulahi, A.T.; Babatunde, A.N. Enhancing Skin Cancer Detection and Classification in Dermoscopic Images through Concatenated MobileNetV2 and Xception Models. Bioengineering 2023, 10, 979. [Google Scholar] [CrossRef] [PubMed]
  23. Dhiman, G.; Kumar, V. Spotted hyena optimizer: A novel bio-inspired based metaheuristic technique for engineering applications. Adv. Eng. Softw. 2017, 114, 48–70. [Google Scholar] [CrossRef]
  24. Yang, X.; Cheng, L. Hyperspectral Image Pixel Classification based on Golden Sine and Chaotic Spotted Hyena Optimization Algorithm. IEEE Access. 2023, 11, 89757–89768. [Google Scholar] [CrossRef]
  25. Dong, X.; Xu, H.; Cao, H.; Cui, T.; Sun, Y. Temperature Compensation of Wind Tunnel Balance Signal Detection System Based on IGWO-ELM. Sensors 2023, 23, 7224. [Google Scholar] [CrossRef] [PubMed]
  26. Available online: https://www.kaggle.com/datasets/l33tc0d3r/indian-food-classification (accessed on 17 July 2023).
Figure 1. Overall flow of the SHODCNN-FIC algorithm.
Figure 1. Overall flow of the SHODCNN-FIC algorithm.
Biomimetics 08 00493 g001
Figure 2. Steps involved in the SHO algorithm.
Figure 2. Steps involved in the SHO algorithm.
Biomimetics 08 00493 g002
Figure 3. Sample images.
Figure 3. Sample images.
Biomimetics 08 00493 g003
Figure 4. 60:40 of TR set/TS set (a,b) Confusion matrices, (c) PR_curve, and (d) ROC.
Figure 4. 60:40 of TR set/TS set (a,b) Confusion matrices, (c) PR_curve, and (d) ROC.
Biomimetics 08 00493 g004
Figure 5. Average values of the SHODCNN-FIC algorithm at 60:40 TR set/TS set.
Figure 5. Average values of the SHODCNN-FIC algorithm at 60:40 TR set/TS set.
Biomimetics 08 00493 g005
Figure 6. A c c u y curve of the SHODCNN-FIC algorithm at 60:40 of the TR set/TS set.
Figure 6. A c c u y curve of the SHODCNN-FIC algorithm at 60:40 of the TR set/TS set.
Biomimetics 08 00493 g006
Figure 7. Loss curve of the SHODCNN-FIC algorithm at 60:40 of the TR set/TS set.
Figure 7. Loss curve of the SHODCNN-FIC algorithm at 60:40 of the TR set/TS set.
Biomimetics 08 00493 g007
Figure 8. 70:30 of TR set/TS set (a,b) Confusion matrices, (c) PR_curve, and (d) ROC.
Figure 8. 70:30 of TR set/TS set (a,b) Confusion matrices, (c) PR_curve, and (d) ROC.
Biomimetics 08 00493 g008
Figure 9. Average values of the SHODCNN-FIC algorithm at 70:30 TR set/TS set.
Figure 9. Average values of the SHODCNN-FIC algorithm at 70:30 TR set/TS set.
Biomimetics 08 00493 g009
Figure 10. A c c u y curve of the SHODCNN-FIC algorithm at 70:30 of the TR set/TS set.
Figure 10. A c c u y curve of the SHODCNN-FIC algorithm at 70:30 of the TR set/TS set.
Biomimetics 08 00493 g010
Figure 11. Loss curve of the SHODCNN-FIC algorithm at 70:30 of the TR set/TS set.
Figure 11. Loss curve of the SHODCNN-FIC algorithm at 70:30 of the TR set/TS set.
Biomimetics 08 00493 g011
Figure 12. Comparative analysis outcomes of the SHODCNN-FIC algorithm and other recent methods.
Figure 12. Comparative analysis outcomes of the SHODCNN-FIC algorithm and other recent methods.
Biomimetics 08 00493 g012
Figure 13. CT outcomes of the SHODCNN-FIC algorithm and other recent methods.
Figure 13. CT outcomes of the SHODCNN-FIC algorithm and other recent methods.
Biomimetics 08 00493 g013
Table 1. Details on the database.
Table 1. Details on the database.
ClassNo. of Samples
Butter_Naan300
Burger300
Chapati300
Dal_Makhani300
Fried_Rice300
Idli300
Total Samples1800
Table 2. Food classification outcomes of the SHODCNN-FIC algorithm at 60:40 TR set/TS set.
Table 2. Food classification outcomes of the SHODCNN-FIC algorithm at 60:40 TR set/TS set.
Class A c c u y P r e c n R e c a l F s c o r e MCC
Training Phase (60%)
Butter_Naan80.6543.0577.8455.4447.53
Burger88.4374.0250.5460.0654.91
Chapati86.2068.0343.0152.7046.73
Dal_Makhani86.6762.5667.0164.7156.55
Fried_Rice86.1156.7751.4653.9945.91
Idli87.3158.9057.8358.3650.88
Average85.9060.5557.9557.5450.42
Testing Phase (40%)
Butter_Naan77.2242.2963.9150.9038.19
Burger88.3369.7446.4955.7950.73
Chapati89.4470.1350.4758.7053.77
Dal_Makhani87.3654.5569.9061.2854.45
Fried_Rice85.9761.8656.5959.1150.74
Idli85.8363.7955.2259.2050.88
Average85.6960.3957.1057.4949.79
Table 3. Food classification outcomes of the SHODCNN-FIC algorithm at 70:30 TR set/TS set.
Table 3. Food classification outcomes of the SHODCNN-FIC algorithm at 70:30 TR set/TS set.
Class A c c u y P r e c n R e c a l F s c o r e MCC
Training Phase (70%)
Butter_Naan89.9271.7168.0669.8363.82
Burger90.9580.5962.8470.6266.07
Chapati90.4878.3458.8567.2162.64
Dal_Makhani80.4842.1458.4248.9638.04
Fried_Rice85.0055.5654.2554.8945.90
Idli79.0537.3444.3340.5428.09
Average85.9860.9557.7958.6850.76
Testing Phase (30%)
Butter_Naan88.8965.7959.5262.5056.09
Burger91.1176.5659.7667.1262.71
Chapati88.1572.1348.3557.8952.70
Dal_Makhani78.5243.9266.3352.8541.09
Fried_Rice82.2245.6547.7346.6736.02
Idli80.0044.4445.3644.9032.68
Average84.8158.0854.5155.3246.88
Table 4. Comparative analysis outcomes of the SHODCNN-FIC algorithm and other recent approaches.
Table 4. Comparative analysis outcomes of the SHODCNN-FIC algorithm and other recent approaches.
Methods A c c u y P r e c n R e c a l F s c o r e MCC
SHODCNN-FIC85.9860.9557.7958.6850.76
InceptionResNet81.9158.9755.7554.5448.34
NasNetLarge77.9151.9753.6753.8746.94
MobileNet75.3658.1751.4756.9744.97
ResNet10173.758.9754.1553.5147.38
ResNet15271.3145.9747.9741.9745.93
ResNet5068.2744.9556.9754.9746.53
Table 5. CT outcomes of the SHODCNN-FIC algorithm and other recent methods.
Table 5. CT outcomes of the SHODCNN-FIC algorithm and other recent methods.
ModelComputational Time (s)
SHODCNN-FIC2.03
InceptionResNet4.57
NasNetLarge4.45
MobileNet3.37
ResNet-1013.65
ResNet-1523.42
ResNet-505.83
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mahgoub, H.; Aldehim, G.; Almalki, N.S.; Issaoui, I.; Mahmud, A.; Alneil, A.A. Bio-Inspired Spotted Hyena Optimizer with Deep Convolutional Neural Network-Based Automated Food Image Classification. Biomimetics 2023, 8, 493. https://doi.org/10.3390/biomimetics8060493

AMA Style

Mahgoub H, Aldehim G, Almalki NS, Issaoui I, Mahmud A, Alneil AA. Bio-Inspired Spotted Hyena Optimizer with Deep Convolutional Neural Network-Based Automated Food Image Classification. Biomimetics. 2023; 8(6):493. https://doi.org/10.3390/biomimetics8060493

Chicago/Turabian Style

Mahgoub, Hany, Ghadah Aldehim, Nabil Sharaf Almalki, Imène Issaoui, Ahmed Mahmud, and Amani A. Alneil. 2023. "Bio-Inspired Spotted Hyena Optimizer with Deep Convolutional Neural Network-Based Automated Food Image Classification" Biomimetics 8, no. 6: 493. https://doi.org/10.3390/biomimetics8060493

APA Style

Mahgoub, H., Aldehim, G., Almalki, N. S., Issaoui, I., Mahmud, A., & Alneil, A. A. (2023). Bio-Inspired Spotted Hyena Optimizer with Deep Convolutional Neural Network-Based Automated Food Image Classification. Biomimetics, 8(6), 493. https://doi.org/10.3390/biomimetics8060493

Article Metrics

Back to TopTop