Next Article in Journal
Chronic Headache Attributed to Vertebrobasilar Insufficiency
Previous Article in Journal
Clinicopathological Relevance of PAX8 Expression Patterns in Acute Kidney Injury and Chronic Kidney Diseases
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Dynamic Weight Agnostic Neural Networks and Medical Microwave Radiometry (MWR) for Breast Cancer Diagnostics

1
School of Informatics, University of Edinburgh, Edinburgh EH8 9YL, UK
2
Department of Computing, Imperial College London, London SW7 2BX, UK
3
Faculty of Mathematics and Information Technology, Volgograd State University, 400062 Volgograd, Russia
4
Medical Microwave Radiometry (MMWR) Ltd., Edinburgh EH10 5LZ, UK
5
Pushchino, Library for Natural Sciences of the Russian Academy of Sciences, 119991 Moscow, Russia
6
Biological Systems Unit, Okinawa Institute of Science and Technology, Onna 904-0495, Japan
*
Author to whom correspondence should be addressed.
Diagnostics 2022, 12(9), 2037; https://doi.org/10.3390/diagnostics12092037
Submission received: 27 July 2022 / Revised: 14 August 2022 / Accepted: 21 August 2022 / Published: 23 August 2022
(This article belongs to the Section Point-of-Care Diagnostics and Devices)

Abstract

:
Background and Objective: Medical microwave radiometry (MWR) is used to capture the thermal properties of internal tissues and has usages in breast cancer detection. Our goal in this paper is to improve classification performance and investigate automated neural architecture search methods. Methods: We investigated extending the weight agnostic neural network by optimizing the weights using the bi-population covariance matrix adaptation evolution strategy (BIPOP-CMA-ES) once the topology was found. We evaluated and compared the model based on the F1 score, accuracy, precision, recall, and the number of connections. Results: The experiments were conducted on a dataset of 4912 patients, classified as low or high risk for breast cancer. The weight agnostic BIPOP-CMA-ES model achieved the best average performance. It obtained an F1-score of 0.933, accuracy of 0.932, precision of 0.929, recall of 0.942, and 163 connections. Conclusions: The results of the model are an indication of the promising potential of MWR utilizing a neural network-based diagnostic tool for cancer detection. By separating the tasks of topology search and weight training, we can improve the overall performance.

1. Introduction

Medical microwave radiometry (MWR) is used to obtain the internal tissue temperature of the body [1]. This is done by measuring the naturally emitted thermal radiation from the tissues. Additionally, it is a noninvasive, nonionizing, and cost-effective approach. Due to the device’s accuracy, there are already multiple clinical applications using the temperature readings and patterns to identify various conditions [1,2,3,4,5,6,7,8,9,10,11]. In this paper, we focus on using MWR to detect breast cancer. This is viable since the growth rate of tumors is correlated with the tissues’ temperature [12,13]. In addition to the thermal information of the tissue, we can derive from MWR the cancer cells’ reproduction rate and mutagenesis risk levels [14].
MWR is a relatively new clinical imaging technique. Thus, for it to be adopted successfully, an artificial intelligence (AI) diagnostic tool needs to be developed in parallel. The diagnostic tool alleviates the need for training clinicians to interpret the data and prevents workload increase, while also providing a more accurate prediction. Thus, our first objective is to further improve the diagnostic accuracy of the model. Furthermore, while this research is focused on breast cancer, MWR has clinical applications at different anatomical locations [1,14] and in various conditions. To reduce the development time of models for each of these cases, we explore adapting automatic machine learning (AutoML) techniques for MWR data.
There has been previous work with AutoML for MWR using a cascade correlation neural network (CCNN) [15]. Subsequent improvements were made by expanding the pool of layers and activation functions the model could explore [15,16]. Despite the improvements, it was not able to outperform predefined architectures [16]. However, it resulted in a small network that was desirable when considering edge computing and hardware limitations. Various classification models have been explored in the past, such as deep neural networks, convolution neural networks, support vector machines, and random forests [15,16]. Additionally, a rule-based classification model was introduced that improved the interpretability of the results [16].
In summary, our contributions in the field of MWR for breast cancer detection are two-fold. First, we evaluate the application of weight agnostic Neural Network (WANN) [17] on MWR data and compare it against the CCNN that was used in previous research [15]. Secondly, we improve the WANN model for MWR classification. Once the topology of the network is found using WANN, we use the bi-population covariance matrix adaptation evolution strategy (BIPOP-CMA-ES) [18] to find the optimal weight candidates. Combining the WANN and BIPOP-CMA-ES strategies, we obtain state-of-the-art classification performance on MWR breast cancer data. Furthermore, we conclude that a random search strategy to optimize the weights yields better results than those achieved by a gradient descent method for architectures generated by WANNs for MWR data.

2. Methods

2.1. Cascade Correlation Neural Network

Cascade correlation neural network (CCNN) is an early neural architecture search (NAS) technique for supervised tasks [19]. The idea of a CCNN is to start with a minimum-sized network, input and output layers, and add one additional node at a time until convergence. The steps of the algorithm are as follows [19]:
  • Initialize network topology with input and output nodes.
  • Create a pool of candidate hidden layer nodes initialized at different starting weights. The hidden layer node takes input from all previous layers. Its output is connected to the output layer nodes. Each candidate node is trained until convergence.
  • From the pool of candidates, select and add to the network the candidate node that maximizes the magnitude of the correlation between the output and target on the validation set. The input weights of the added hidden layer nodes are frozen.
  • If the correlation does not improve or improves by a small margin, then terminate the algorithm. Otherwise, proceed to step two.
An example of the connections created after adding the 3rd hidden layer can be seen in Figure 1.
The loss function we used is the cross-correlation loss and optimize using stochastic gradient descent. The optimizer’s learning rate was set to 5 × 10 6 . Each of the nodes used a sigmoid activation function. Furthermore, their weights were sampled from a Gaussian distribution with a mean of 0 and 0.5 standard deviation, and the bias was set to 0. A combination of batch normalization and dropout layers (rate of 0.5) was added after the node. The candidate node pool size was set to 30. Additionally, we reinitialized the weights of the output layers after each iteration to avoid being stuck in a bad local minimum.

2.2. Weight Agnostic Neural Network

Another NAS method is the weight agnostic neural network (WANN) approach [17]. The main difference between WANN and CCNN is that during the architecture search, the weights of the model are not trained. Instead, a set of fixed shared weights are used to evaluate the average performance. According to the authors, the idea behind this is to automatically find architectures that have inductive biases and can perform well in their given task without training.
Inspired by genetic evolution, WANN starts with a population of small initial networks. The initial networks consist of the input and output layers. However, the nodes between the layers are sparsely connected. Then, once the population of networks is established, a series of fixed shared-value weights, which we set to [−2, −1.5, −1, −0.5, +0.5, +1, +1.5, +2], are used to evaluate the performance of each topology. The evaluation metric we used is the geometric mean between the predicted and actual values.
The topologies were ranked on the basis of three criteria: mean performance across all fixed shared weights, best performance between any of the fixed shared weights, and the number of connections. Similar to the authors of WANN, we used the Non-dominated Sorting Genetic Algorithm II (NSGA-II) [20] to sort the network topologies based on the previous criteria. NSGA-II is a multi-objective sorting genetic algorithm that combines elitism and does not require a priori selection of shared parameters. The highest-ranking topologies were selected for the next step using the tournament algorithm [21].
Once the new population was selected, they were subsequently varied to generate the next generation of population. There were three mutation operations used to increase the complexity of the model. First, a node can be inserted between two connected nodes. Secondly, a new connection can be added between two existing nodes. Finally, the activation function of a node can be changed according to the list in Table 1. This process of evaluating, ranking, and generating a new population was repeated until there was no longer improvement.
The hyperparameters for the WANN model are summarized in Table 2. These hyperparameters were defined through extensive experimentation. Specifically, we searched 200 generations, each having a population size of 200. The existence of the initial connections between input and output layers was set to 0.2. For topology variation, we set a likelihood of changing the activation function to 0.5, adding a node to 0.25, and creating a new connection to 0.25. Finally, for the tournament algorithm, we set the size to 4.

2.3. Weight Agnostic Neural Network BIPOP-CMA-ES

For MWR breast cancer data, we can determine patterns and relationships between the points to identify high-risk patients. By using a WANN model, the resulting architecture relies on creating node connections to identify similar properties, in addition to new ones. Thus, the architecture acts as a prior. However, there are more subtle cases to distinguish between those that are low- and high-risk. Specifically, these cases will be when the tumors’ growth rate slows down. This can be achieved through weight optimization once the optimal architecture has been found.
Based on the research results of WANN, while it performs better than chance in most cases, it is not able to outperform fixed topologies that have had their parameters tuned [17]. A way to circumvent this is by taking the best topology found by the WANN model and proceeding to optimize the parameters via a gradient descent algorithm. However, a network with various activation functions results in a difficult gradient traversal [17].
Thus, a better way of optimizing the weights is through a black-box optimization method such as the CMA-ES algorithm [17,22]. With the randomized search of the CMA-ES, it is suitable for a rigged landscape in which there are many bad local minima, discontinuities, and noise. The steps of the CMA-ES algorithm are summarized in Figure 2.
We used a variant of the CMA-ES algorithm, the bi-population covariance matrix adaptation evolution strategy (BIPOP-CMA-ES) [18]. Through our experiments, we found it to perform better than CMA-ES. BIPOP-CMA-ES uses a variable population size. It initially starts with a small population size, which we set to 50, and doubles after each restart. Additionally, to speed up convergence, we fine-tuned the initial single shared weight of the model. We achieved this by linearly evaluating values between −2 and 2. Finally, the cross-entropy loss was used to find the best fit.

3. Results

3.1. Data

The MWR breast cancer data were captured using the MMWR-2020 (RTM-01-RES) (http://www.mmwr.co.uk accessed on 26 July 2022) device in clinics over the world. The data were classified as either low or high risk for breast cancer. Classification of the patients was done by clinicians using MWR, mammography, and/or biopsy data as necessary. In total, 4912 cases were recorded, with 4377 as low-risk and 535 as high-risk. Subsequently, we class-balanced and separated our data into train, validate, and test sets using 60%, 20%, and 20%, respectively. The data are publicly available from http://www.mmwr.co.uk/dataset_clean breast.csv accessed on 26 July 2022.
The MWR data consist of temperature readings of the skin surface and internally at a depth of 3–5 cm. There were in total 44 points recorded on the mammary glands and surrounding regions. On each gland, a point was recorded on the nipple and nine equidistant around the nipple. Additionally, two reference points were captured just below the chest. Finally, two more points were captured under each axillary area. Each of these points was captured both on the skin and at a depth. The positions described are shown in detail in Figure 3.

3.2. Experimental Results

Each model was trained three times using a different random number generator seed. For the analysis, we obtained from the test set results the mean and standard deviation. We explored additional NAS techniques that are available from the Neural Network Intelligence [24] package. Specifically, the exploration strategies we investigated are Differentiable Architecture Search (DARTS) [25,26], regularized evolution [27], and Tree-structured Parzen Estimator (TPE) [28]. The model search space is based on a previously developed fully connected (FC) network [15], with the extended search space specified in Table 3. To also consider the impact of our model, we compared the output results of our model against the other models using a statistical significance t-test. We consider a p-value of <0.05 to be statistically significant.
From the models presented, the best performing on the test set across all metrics is the proposed WANN BIPOP-CMA-ES. The model obtained an average F1-score of 0.933, accuracy of 0.932, precision of 0.929, recall of 0.942, and required 163 connections. It needed 10 6 function evaluations before converging on the validation set. When assessing the p-value, our model’s results are statistically significant compared to all other models. All but the WANN model result in a much lower p-value of at most 0.005. The WANN model is still statistically significant but with a marginal value of 0.045. The summary of the results is shown in Table 4. The accuracy and loss during training of the WANN BIPOP-CMA-ES can be seen in Figure 4.
To further analyze the models, we also compared the mean element-wise difference of the values between WANN BIPOP-CMA-ES against WANN and FC-Evolution (second-best-performing model). They obtained respectively a mean difference of 0.49 (±0.57) and 0.76 (±0.78). Furthermore, we observed from the networks’ output that most erroneous cases from the WANN were close to the threshold of value 0.5. On the other hand, the erroneous cases of the FC models differed by a large amount. In contrast to WANN’s p-value, it was the worst-performing model when no weight optimizations were done.
The architecture and optimized weights obtained from the CCNN had an average performance on the F1-score with a value of 0.809. It required a total of 17 iterations before terminating due to the validation loss not decreasing given patience of five iterations for this execution. The accuracy and loss during training can be viewed in Figure 5. However, it required a total of 672 connections before converging.
Additionally, we evaluated the robustness of the WANN’s generated topology to weight changes. Specifically, we evaluated the performance of the network when using fixed, randomly generated, and fine-tuned shared weights. For each of these cases, we ran them 10 times and took their average. In Table 5, we show a summary of the results of the test set. By tuning the weights, we improved the performance. In contrast, using random weights, we obtained the worst performance.

4. Discussion

Similar to the observations of the authors of WANN [17], the generated model we obtained can achieve better performance than that of chance based on the generated topology of the network. We can further improve the performance by optimizing the weights. However, despite the improvement when we used a gradient descent optimizer, the performance is still subpar to that of the other models. We gained a much larger improvement in performance when utilizing a random search evolution strategy.
Our proposed model obtains the best performance on all metrics evaluated and has the least number of connections. The WANN and CCNN models have a small number of connections as they start from a minimum-sized network and gradually expand. In contrast, FC-Evolution, FC-TPE, and FC-DARTS search from a predefined architecture space that allows them to start from a large or small network. The trend of these approaches was to opt for larger network sizes early in their architecture search, as they have a higher learning capacity. Additionally, we showed that the predicted results from the network are statistically significant when paired against all other evaluated models. However, while the WANN’s performance is the worst, it has the highest p-value. This is probably an indication of the importance of the architecture and that inductive biases are maintained to some degree, despite weight training.
A general summary of the advantages and disadvantages of all models is shown in Table 6. Furthermore, there are domain-specific advantages of WANN BIPOP-CMA-ES and an extension of NAS for healthcare applications. First, the generated topology of the network is optimized to have a small number of parameters and sparse connections due to the inclusion of the model size as a minimization objective [17]. This allows the model to be deployed on low-end devices and on already existing clinical hardware, which is particularly important for accessibility to low- and mid-income countries. Second, we decrease the development time, as architecture tunning through manual trial-and-error is reduced. Additionally, the model becomes more accessible to nontechnical experts, such as clinicians, as they do not require vast knowledge of machine learning to develop a model. Without this barrier, they can more effectively contribute to and improve the diagnostic tool. Finally, from the aforementioned benefits, it reduces the complexity and time of adapting MWR for different anatomical locations and pathologies.
The mutation operations of WANN only increase the complexity of the network topology. For future work, we will expand it to include operations such as deleting nodes and deleting connections so there is more flexibility in defining the architecture. In addition, there is no crossover mutation operation in the WANN model, which will reduce population diversity. Hence, we will explore different crossover mutation operations [29,30,31] and restart techniques to increase model performance [32,33]. Finally, we will investigate cross-trial information sharing, such as including in the mutation pool more complicated building blocks generated from previous trials.
Furthermore, we will look at improving our model by searching for the best loss function [34], utilizing a one-shot learning search [35,36], and conducting a hyperparameter search [37]. We will also compare against additional NAS methods such as reinforcement learning-based searching methods [38,39], ensemble methods [40], and transfer learning [41]. Our main aim is to determine if a performance improvement can be made without significantly increasing computational complexity.

Author Contributions

Conceptualization, C.G.; Data curation, L.O.; Funding acquisition, A.L.; Investigation, S.V.; Methodology, C.G.; Project administration, I.G.; Resources, L.O.; Software, J.L. and L.P.; Supervision, C.G. and A.L.; Validation, L.P.; Writing–original draft, J.L.; Writing–review & editing, C.G., T.K. and I.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are publicly available from http://www.mmwr.co.uk/dataset_clean breast.csv accessed on 26 July 2022.

Acknowledgments

Authors thank Medical Microwave Radiometry (MMWR) LTD for providing the dataset.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Goryanin, I.; Karbainov, S.; Shevelev, O.; Tarakanov, A.; Redpath, K.; Vesnin, S.; Ivanov, Y. Passive microwave radiometry in biomedical studies. Drug Discov. Today 2020, 25, 757–763. [Google Scholar] [CrossRef]
  2. Levshinskii, V.; Galazis, C.; Losev, A.; Zamechnik, T.; Kharybina, T.; Vesnin, S.; Goryanin, I. Using AI and Passive medical Radiometry for diagnostics (MWR) of venous diseases. Comput. Methods Programs Biomed. 2021, 215, 106611. [Google Scholar] [CrossRef] [PubMed]
  3. Shevelev, O.; Petrova, M.; Smolensky, A.; Osmonov, B.; Toimatov, S.; Kharybina, T.; Karbainov, S.; Ovchinnikov, L.; Vesnin, S.; Tarakanov, A. Using medical microwave radiometry for brain temperature measurements. Drug Discov. Today 2021, 27, 881–889. [Google Scholar] [CrossRef]
  4. Rodrigues, D.B.; Stauffer, P.R.; Pereira, P.J.; Maccarini, P.F. Microwave radiometry for noninvasive monitoring of brain temperature. In Emerging Electromagnetic Technologies for Brain Diseases Diagnostics, Monitoring and Therapy; Springer: Cham, Switzerland, 2018; pp. 87–127. [Google Scholar]
  5. Tarakanov, A.V.; Tarakanov, A.A.; Vesnin, S.; Efremov, V.V.; Goryanin, I.; Roberts, N. Microwave Radiometry (MWR) temperature measurement is related to symptom severity in patients with Low Back Pain (LBP). J. Bodyw. Mov. Ther. 2021, 26, 548–552. [Google Scholar] [CrossRef]
  6. Osmonov, B.; Ovchinnikov, L.; Galazis, C.; Emilov, B.; Karaibragimov, M.; Seitov, M.; Vesnin, S.; Losev, A.; Levshinskii, V.; Popov, I.; et al. Passive microwave radiometry for the diagnosis of coronavirus disease 2019 lung complications in Kyrgyzstan. Diagnostics 2021, 11, 259. [Google Scholar] [CrossRef] [PubMed]
  7. Kaprin, A.D.; Kostin, A.A.; Andryukhin, M.I.; Ivanenko, K.V.; Popov, S.V.; Shegai, P.V.; Kruglov, D.P.; Mangutov, F.S.; Leushin, V.Y.; Agasieva, S.V. Microwave radiometry in the diagnosis of various urological diseases. Biomed. Eng. 2019, 53, 87–91. [Google Scholar] [CrossRef]
  8. Zamechnik, T.V.; Losev, A.G.; Petrenko, A.Y. Guided classifier in the diagnosis of breast cancer according to microwave radiothermometry. Math. Phys. Comput. Modeling 2019, 22, 52–66. [Google Scholar] [CrossRef]
  9. Losev, A.G.; Medevedev, D.A.; Svetlov, A.V. Neural networks in diagnosis of breast cancer. In Institute of Scientific Communications Conference; Springer: Cham, Switzerland, 2020; pp. 220–227. [Google Scholar]
  10. Bardati, F.; Iudicello, S. Modeling the visibility of breast malignancy by a microwave radiometer. IEEE Trans. Biomed. Eng. 2007, 55, 214–221. [Google Scholar] [CrossRef]
  11. Levick, A.; Land, D.; Hand, J. Validation of microwave radiometry for measuring the internal temperature profile of human tissue. Meas. Sci. Technol. 2011, 22, 065801. [Google Scholar] [CrossRef]
  12. Gautherie, M. Temperature and blood flow patterns in breast cancer during natural evolution and following radiotherapy. Prog. Clin. Biol. Res. 1982, 107, 21–64. [Google Scholar] [PubMed]
  13. Gautherie, M.; Gros, C.M. Breast thermography and cancer risk prediction. Cancer 1980, 45, 51–56. [Google Scholar] [CrossRef]
  14. Vesnin, S.G.; Turnbull, A.K.; Dixon, J.M.; Goryanin, I. Modern microwave thermometry for breast cancer. J. Mol. Imaging Dyn. 2017, 7, 136. [Google Scholar] [CrossRef]
  15. Galazis, C.; Vesnin, S.; Goryanin, I. Application of artificial intelligence in microwave radiometry (MWR). In Proceedings of the 12th International Joint Conference on Biomedical Engineering systems and Technologies (BIOSTEC 2019), Prague, Czech Republic, 22–24 February 2019; De Maria, E., Gamboa, H., Fred, A., Eds.; Scitepress: Setúbal, Portugal, 2019; pp. 112–122. [Google Scholar]
  16. Levshinskii, V.; Galazis, C.; Ovchinnikov, L.; Vesnin, S.; Losev, A.; Goryanin, I. Application of data mining and machine learning in microwave radiometry (MWR). In International Joint Conference on Biomedical Engineering Systems and Technologies; Rogue, A., Tomczyk, A., de Maria, E., Putze, F., Moucek, R., Fred, A., Gamboa, H., Eds.; Springer Nature: Cham, Switzerland, 2020; pp. 265–288. [Google Scholar]
  17. Gaier, A.; Ha, D. Weight agnostic neural networks. In Proceedings of the 33rd Conference on Neural Information Processing Systems (NeurIPS 2019), Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  18. Loshchilov, I. CMA-ES with restarts for solving CEC 2013 benchmark problems. In Proceedings of the 2013 IEEE Congress on Evolutionary Computation, Cancun, Mexico, 20–23 June 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 369–376. [Google Scholar] [CrossRef] [Green Version]
  19. Fahlman, S.E.; Lebiere, C. The cascade-correlation learning architecture. In Advances in Neural Information Processing Systems 2 (NIPS 1989); Morgan Kaufmann: Burlington, MA, USA, 1990; pp. 524–532. [Google Scholar]
  20. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef] [Green Version]
  21. Fang, X.; Tettamanti, T.; Piazzi, A. Online calibration of microscopic road traffic simulator. In Proceedings of the SAMI 2020: IEEE 18th World Symposium on Applied Machine Intelligence and Informatics, Herl’any, Slovakia, 23–25 January 2020; pp. 275–280. [Google Scholar] [CrossRef]
  22. Hansen, N.; Müller, S.D.; Koumoutsakos, P. Reducing the time complexity of the derandomized evolution strategy with covariance matrix adaptation (CMA-ES). Evol. Comput. 2003, 11, 1–18. [Google Scholar] [CrossRef]
  23. Gagganapalli, S.R. Implementation and Evaluation of CMA-ES Algorithm. Master’s Thesis, North Dakota State University, Fargo, ND, USA, 2015. [Google Scholar]
  24. Microsoft, Neural Network Intelligence. v2.8. 2021. Available online: https://github.com/microsoft/nni (accessed on 26 July 2022).
  25. Liu, H.; Simonyan, K.; Yang, Y. DARTS: Differentiable Architecture Search. In Proceedings of the International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  26. Wan, A.; Dai, X.; Zhang, P.; He, Z.; Tian, Y.; Xie, S.; Wu, B.; Yu, M.; Xu, T.; Chen, K.; et al. Fbnetv2: Differentiable neural architecture search for spatial and channel dimensions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12965–12974. [Google Scholar]
  27. Real, E.; Aggarwal, A.; Huang, Y.; Le, Q.V. Regularized evolution for image classifier architecture search. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 4780–4789. [Google Scholar]
  28. Bergstra, J.; Bardenet, R.; Bengio, Y.; Kégl, B. Algorithms for hyper-parameter optimization. In Advances in Neural Information Processing Systems 24 (NIPS 2011); Curran Associates Inc.: Red Hook, NY, USA, 2011. [Google Scholar]
  29. Lu, Z.; Whalen, I.; Boddeti, V.; Dhebar, Y.; Deb, K.; Goodman, E.; Banzhaf, W. Nsga-net: Neural architecture search using multi-objective genetic algorithm. In Proceedings of the Genetic and Evolutionary Computation Conference, Prague, Czech Republic, 13–17 July 2019; pp. 419–427. [Google Scholar]
  30. Xue, Y.; Jiang, P.; Neri, F.; Liang, J. A multi-objective evolutionary approach based on graph-in-graph for neural architecture search of convolutional neural networks. Int. J. Neural Syst. 2021, 31, 2150035. [Google Scholar] [CrossRef] [PubMed]
  31. Hassanat, A.; Almohammadi, K.; Alkafaween, E.; Abunawas, E.; Hammouri, A.; Prasath, V.B.S. Choosing mutation and crossover ratios for genetic algorithms—A review with a new dynamic approach. Information 2019, 10, 390. [Google Scholar] [CrossRef] [Green Version]
  32. Molina, D.; LaTorre, A.; Herrera, F. An insight into bio-inspired and evolutionary algorithms for global optimization: Review, analysis, and lessons learnt over a decade of competitions. Cogn. Comput. 2018, 10, 517–544. [Google Scholar] [CrossRef]
  33. Yu, H.; Tan, Y.; Sun, C.; Zeng, J. A generation-based optimal restart strategy for surrogate-assisted social learning particle swarm optimization. Knowl.-Based Syst. 2019, 163, 14–25. [Google Scholar] [CrossRef]
  34. Li, C.; Yuan, X.; Lin, C.; Guo, M.; Wu, W.; Yan, J.; Ouyang, W. Am-lfs: Automl for loss function search. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 8410–8419. [Google Scholar]
  35. Li, X.; Lin, C.; Li, C.; Sun, M.; Wu, W.; Yan, J.; Ouyang, W. Improving one-shot nas by suppressing the posterior fading. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 13836–13845. [Google Scholar]
  36. Zela, A.; Elsken, T.; Saikia, T.; Marrakchi, Y.; Brox, T.; Hutter, F. Understanding and robustifying differentiable architecture search. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020; Volume 2. [Google Scholar]
  37. Zela, A.; Klein, A.; Falkner, S.; Hutter, K. Towards automated deep learning: Efficient joint neural architecture and hyperparameter search. In Proceedings of the International Conference in Machine Learning, Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
  38. Zoph, B.; Le, Q.V. Neural architecture search with reinforcement learning. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  39. Baker, B.; Gupta, O.; Naik, N.; Raskar, R. Designing neural network architectures using reinforcement learning. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  40. Zaidi, S.; Zela, A.; Elsken, T.; Holmes, C.; Hutter, F.; The, Y.W. Neural ensemble search for uncertainty estimation and dataset shift. In Advances in Neural Information Processing Systems 34 (NeurIPS 2021); 2021; pp. 7898–7911. Available online: https://proceedings.neurips.cc/paper/2021/hash/41a6fd31aa2e75c3c6d427db3d17ea80-Abstract.html (accessed on 26 July 2022).
  41. Zhao, Y.; Wang, L.; Tian, Y.; Fonseca, R.; Guo, T. Few-shot neural architecture search. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 12707–12718. [Google Scholar]
Figure 1. The cascade correlation neural network after adding three hidden layers.
Figure 1. The cascade correlation neural network after adding three hidden layers.
Diagnostics 12 02037 g001
Figure 2. The pseudo-code of the CMA-ES algorithm [23].
Figure 2. The pseudo-code of the CMA-ES algorithm [23].
Diagnostics 12 02037 g002
Figure 3. Capture positions for each mammary gland (0–8) and at the axillary point (9). There are the additional T1 and T2 reference points under the chest. Each position was captured on the skin and at a depth of 3–5 cm.
Figure 3. Capture positions for each mammary gland (0–8) and at the axillary point (9). There are the additional T1 and T2 reference points under the chest. Each position was captured on the skin and at a depth of 3–5 cm.
Diagnostics 12 02037 g003
Figure 4. Cross-entropy loss and accuracy curve of the BIPOP-CMA-ES optimization algorithm during training of the best WANN topology.
Figure 4. Cross-entropy loss and accuracy curve of the BIPOP-CMA-ES optimization algorithm during training of the best WANN topology.
Diagnostics 12 02037 g004
Figure 5. The geometric mean loss and accuracy of the training and validation set of the cascade correlation neural network during training.
Figure 5. The geometric mean loss and accuracy of the training and validation set of the cascade correlation neural network during training.
Diagnostics 12 02037 g005
Table 1. Summary of the pool of activation functions that the weight agnostic neural network samples from.
Table 1. Summary of the pool of activation functions that the weight agnostic neural network samples from.
NameEquationRange
Linear
f ( x ) = x
( , )
Binary step
f ( x ) = { 1     f o r     x 0   0     f o r     x < 0 }
{ 0 ,   1 }
Sin
f ( x ) = s i n ( x )
[ 1 ,   1 ]
Cosine
f ( x ) = c o s ( x )
[ 1 ,   1 ]
Sigmoid
f ( x ) = 1 1 + e x
( 0 ,   1 )
Gaussian
f ( x ) = e x 2
( 0 , 1 ]
TanH
f ( x ) = t a n h ( x )
( 1 , 1 )
Inverse
f ( x ) = x
( , )
Absolute Value
f ( x ) = a b s ( x )
[ 0 , )
ReLu
f ( x ) = { x     f o r     x 0   0     f o r     x < 0 }
[ 0 , )
Squared
f ( x ) = x 2
[ 0 , )
Table 2. The hyperparameter selection for the weight agnostic neural network.
Table 2. The hyperparameter selection for the weight agnostic neural network.
HyperparameterValue
Generations200
Population Size200
Change Activation Probability (%)50
Add Node Probability (%)25
Add Connection Probability (%)25
Initial Active Connections (%)20
Tournament Size4
Table 3. The model search space of a fully connected neural network. It was used by DARTS, regularized evolution, and TPE NAS algorithms.
Table 3. The model search space of a fully connected neural network. It was used by DARTS, regularized evolution, and TPE NAS algorithms.
Model Search SpaceRange
Number of Layers[2, 8]
Number of Units per Dense Layer{32, 64, 128, 256, 512}
Activation FunctionSee Table 1
Batch Normalization{Include, not Include}
Dropout Rate{0, 0.1, 0.25, 0.5}
Skip Connection{Include, not Include}
Table 4. The summary of the results on the test set for all models. Each model was trained three times, in which a different random number generator seed was used each time, and the metrics presented are the average and standard deviation of those runs.
Table 4. The summary of the results on the test set for all models. Each model was trained three times, in which a different random number generator seed was used each time, and the metrics presented are the average and standard deviation of those runs.
ModelF1-ScoreAccuracyPrecisionRecallConnectionsp-Value
FC-Evolution0.905 ± 0.0040.903 ± 0.0040.892 ± 0.0120.919 ± 0.0091342 k ± 373 k<<0.05
FC-TPE0.9 ± 0.0180.9 ± 0.0170.903 ± 0.0160.897 ± 0.021254 k ± 124 k<<0.05
FC-DARTS0.849 ± 0.0370.846 ± 0.040.834 ± 0.0470.866 ± 00271166 k ± 124 k<<0.05
CCNN0.809 ± 0.0110.816 ± 0.0110.825 ± 0.0120.795 ± 0.033672 ± 28<<0.05
WANN0.673 ± 0.0130.697 ± 00140.727 ± 0.0420.631 ± 0.047163 ± 9<0.05
WANN BIPOP-CMA-ES0.933 ± 0.0070.932 ± 0.008 0.929 ± 0.0050.942 ± 0.021163 ± 9-
Table 5. Performance of WANN on the test set using different weight schemes.
Table 5. Performance of WANN on the test set using different weight schemes.
WANNF1-ScoreAccuracyPrecisionRecall
Random weight0.52090.55910.55710.4892
Shared weight0.59790.66280.72120.5105
Tuned shared weight0.65460.69470.73630.5892
Table 6. Summary of the main advantages and disadvantages of the models explored.
Table 6. Summary of the main advantages and disadvantages of the models explored.
ModelAdvantagesDisadvantages
FC-Evolution
  • With an evolutionary algorithm, it is simpler to define population mutations.
  • Evolution strategies can be parallelized easier.
  • The diversity of the connections is limited to predefined layers.
  • Has a fixed outer structure, such as the number of layers or units.
  • Favors larger model sizes.
FC-TPE
  • It is a sequential model-based optimization method that approximates the architecture’s hyperparameters based on previous results.
  • Architecture search is computationally efficient.
  • The diversity of the connections is limited to predefined layers.
  • Has a fixed outer structure, such as the number of layers or units.
  • Favors larger model sizes.
FC-DARTS
  • Gradient descent is used to search for optimal architecture hyperparameters.
  • Since it is a one-shot NAS approach, it can converge to an optimal architecture and weights faster. To achieve this, it uses two-step training: one to search for model architecture and the other for parameter optimization.
  • Training data are divided into architecture and weight training. For a small (pathological) dataset like the one used, it limits the performance as the divided data do not share the same distribution.
  • The diversity of the connections is limited to predefined layers.
  • Has a fixed outer structure, such as the number of layers or units.
  • Favors larger model sizes.
CCNN
  • Small network size.
  • Freezes previously added nodes to improve training time.
  • Slow to train as each node in the pool has to be trained individually.
  • Depends on a larger node pool size to avoid bad local minima.
  • Training individual nodes at a time can make the model more prone to overfitting.
  • Node connections are predefined. Thus, there is limited diversity and complexity in the resulting architecture.
WANN
  • Small network size (model size is an optimization objective).
  • Evolution strategies can be parallelized easier.
  • Promotes connection sparsity between the nodes.
  • No weight optimization leads to faster architecture convergence.
  • Allows for diverse and complex connections between nodes that are not practical to define manually.
  • Finds a more optimal network architecture compared to other methods explored. The performance is exclusively obtained from the network’s architecture (which has inductive biases).
  • Subpar performance when compared to other models.
  • The model’s weights are not optimized for the given task and hence they are underutilized.
  • The WANN architecture search space is unconstrained. Thus, there are no clear stopping criteria. Too few trials might result in a poor-performing network, or too many can lead to unnecessary high compute time.
WANN BIPOP-CMA-ES
  • Small network size (model size is an optimization objective).
  • Evolution strategies can be parallelized easier.
  • Promotes connection sparsity between the nodes.
  • Allows for diverse and complex connections between nodes that are not practical to define manually.
  • By separating the architecture search and weight training, it benefits from the inductive biases from the WANN and weights fine-tuning from the BIPOP-CMA-ES to improve performance.
  • A small architecture during the first phase results in fast weight training during the second phase.
  • Low performance when using the more common gradient descent methods used in neural network training.
  • It is currently unclear which WANN model should be selected to be trained. The highest-performing WANN on a given set of metrics might not yield the best performance when weight has been optimized.
  • While it is computationally efficient to train a single WANN, it is computationally demanding when training multiple WANN models.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, J.; Galazis, C.; Popov, L.; Ovchinnikov, L.; Kharybina, T.; Vesnin, S.; Losev, A.; Goryanin, I. Dynamic Weight Agnostic Neural Networks and Medical Microwave Radiometry (MWR) for Breast Cancer Diagnostics. Diagnostics 2022, 12, 2037. https://doi.org/10.3390/diagnostics12092037

AMA Style

Li J, Galazis C, Popov L, Ovchinnikov L, Kharybina T, Vesnin S, Losev A, Goryanin I. Dynamic Weight Agnostic Neural Networks and Medical Microwave Radiometry (MWR) for Breast Cancer Diagnostics. Diagnostics. 2022; 12(9):2037. https://doi.org/10.3390/diagnostics12092037

Chicago/Turabian Style

Li, Jolen, Christoforos Galazis, Larion Popov, Lev Ovchinnikov, Tatyana Kharybina, Sergey Vesnin, Alexander Losev, and Igor Goryanin. 2022. "Dynamic Weight Agnostic Neural Networks and Medical Microwave Radiometry (MWR) for Breast Cancer Diagnostics" Diagnostics 12, no. 9: 2037. https://doi.org/10.3390/diagnostics12092037

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop