Evidential Interpretation Approach for Deep Neural Networks in High-Frequency Electromagnetic Wave Processing
Abstract
1. Introduction
- The proposed approach can indicate whether a subset of DNN representations contains an HF-MEW semantic supporting one or more classes. At the same time, it can also explore the subset by the HF-MEW semantic that a human can understand. The subset is trustworthy for high-dimensional representation explanations with small uncertainty. The proposed approach outperforms the other DNN interpretation approaches in HF-EMW processing.
- The explained subsets can be applied to evaluate the DNN learning process to avoid over-fitting. The application visually shows the two-phase learning processes of how a DNN captures HF-MEW semantics. Furthermore, the explained subsets can also be used for semantics-guided reinforcement learning, which can improve the performance of a DNN.
2. Background
2.1. HF-EMW Semantics
2.2. Dempster–Shafer Theory
2.2.1. Mass Functions
2.2.2. Dempster’s Rule
2.2.3. Weights of Evidence
3. Interpretation of High-Dimension Representations
3.1. Evidential Reasoning on Class Set
- Mass indicates that subset highly support that the truth is class . Thus, subset may contain one or more HF-EMW semantic contents related to class .
- Mass indicates that subset cannot determine that the true class is in subsets A or B. Thus, subset may have the common HF-EMW semantics related to all classes in the intersection between A and B.
- Mass indicates that subset cannot provide any useful supports to any class on . Thus, subset has very low HF-EMW semantics.
3.2. Evidential Reasoning on HF-EMW Semantics
3.2.1. Continuous DST-Based Model
- Step 1.
- The similarity between subset and a prototype vector in the continuous DST-based model is computed as
- Step 2.
- The similarity w.r.t is then converted into a GRFN as
- Step 3.
- The Z GRFNs from prototypes are then aggregated by a generalized Dempster’s rule operation ⊞ as such that
- The output represents the estimate of the conditional expectation of HF-EMW semantics F. A small distance indicates that subset has useful information to support the semantics.
- The variance output represents the conditional variability w.r.t F when the given input is , which can be regarded as aleatory uncertainty. A large value of indicates a large aleatory uncertainty. The large value might be caused by random noise or some elements in , which do not contain related information about F. In this case, there may be a strict subset in , which is the learned semantic knowledge about F.
- The precision output represents the conditional precision of F when the input is , which can be regarded as epistemic uncertainty. A small value of indicates a large epistemic uncertainty. The large value might originate from the fact that does not include enough information about F or contains some conflicting information.
3.2.2. Evidential Belief Prediction Interval
3.2.3. Interpretation of DNN Learning Processes
- A curve of vs. epoch is plotted to visualize the change about the evidence of supports to class based on subset , where is the subset of a training set , including samples belonging to class .
- A curve of vs. epoch is plotted to visualize the change about lack of evidence (no information in provides related classification information).
3.2.4. Learning Strategy
4. Sampling Method for Representation Subsets
Algorithm 1 Sampling algorithm for local representations |
Input: Output: Require: as a threshold for ,
|
5. Numerical Experiments
5.1. Experiment Setting
5.1.1. Datasets
5.1.2. Network Details
5.1.3. Training Details
5.1.4. Comparison Study and Trustworthiness Metrics
- Most Relevant First (MoRF) [49]: Partial elements in the three subsets are replaced by random values, and the output change is measured to evaluate whether the subsets are important for the related task. A large change indicates that the subsets are trustworthy and important for HF-EMW semantics.
- Remove and retrain (ROAR) [50]: In the MoRF method, the output change might result from the network not being trained well. Thus, the ROAR method retrains the network on the subset with the random values. The subsets are not highlighted as necessary if the accuracy does not drop.
5.2. Semantic Explanations and Trustworthiness Evaluations
5.2.1. Experiment on the AREM Dataset
5.2.2. Experiment on the UOEM Dataset
5.3. Applications of HF-EMW Semantics Explanations
5.3.1. Learning Evaluation
5.3.2. Semantics-Guided Reinforcement Learning
6. Conclusions
- The evidential discrete models can indicate whether a subset of representations contains an HF-MEW semantic supporting one or more classes, while an interpretable continuous DST-based model interprets the subset as the HF-MEW semantic that humans can understand.
- The trustworthiness evaluation indicates that the representation subsets from the proposed approach are trustworthy for high-dimensional representation explanations with small uncertainty. The proposed approach outperforms the other interpretation approaches in MoRF and ROAR testing views, achieving an absolute fractional output change of 39.84% with 10% removed elements in most important features.
- The explained subsets can be applied to evaluate the learning process to avoid under- and over-fitting. The application visually shows the two-phase learning processes of how the subset captures semantic intuition. Furthermore, the explained subsets can also be used for semantics-guided reinforcement learning, where the semantic-guided reinforcement learning make an improvement of 4.23% on classification accuracy.
- Regarding limitations, the proposed approach cannot interpret the non-formalized semantics of electromagnetic signals, such as the ones in Figure 1. To address this issue, we consider converting the interpretable continuous DST-based model into an evidential signal inversion model, which can directly invert the subsets of DNN representations into the distribution of an electromagnetic property in a 2D/3D space. One potential way is to use the important semantic subsets, such as , , and in Section 5, as the inputs of FWI to predict the the permittivity distribution in the propagation path of the HF-EMW. Then, the permittivity distribution of a 2D/3D space can be predicted using the HF-EMWs in the space.
Author Contributions
Funding
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Appendix A. Proof of the Predictive GRFN
References
- Wu, H.T.; Li, H.; Chi, H.L.; Kou, W.B.; Wu, Y.C.; Wang, S. A hierarchical federated learning framework for collaborative quality defect inspection in construction. Eng. Appl. Artif. Intell. 2024, 133, 108218. [Google Scholar] [CrossRef]
- Yang, H.; Ma, T.; Huyan, J.; Han, C.; Wang, H. Aggregation segregation generative adversarial network (AG-GAN) facilitated multi-scale segregation detection in asphalt pavement paving stage. Eng. Appl. Artif. Intell. 2024, 129, 107663. [Google Scholar] [CrossRef]
- Yu, C.M.; Lin, Y.H. The docking control system of an autonomous underwater vehicle combining intelligent object recognition and deep reinforcement learning. Eng. Appl. Artif. Intell. 2025, 139, 109565. [Google Scholar] [CrossRef]
- Alzubaidi, L.; Chlaib, H.K.; Fadhel, M.A.; Chen, Y.; Bai, J.; Albahri, A.S.; Gu, Y. Reliable deep learning framework for the ground penetrating radar data to locate the horizontal variation in levee soil compaction. Eng. Appl. Artif. Intell. 2024, 129, 107627. [Google Scholar] [CrossRef]
- Virieux, J.; Operto, S. An overview of full-waveform inversion in exploration geophysics. Geophysics 2009, 74, WCC1–WCC26. [Google Scholar] [CrossRef]
- Plessix, R.E. A review of the adjoint-state method for computing the gradient of a functional with geophysical applications. Geophys. J. Int. 2006, 167, 495–503. [Google Scholar] [CrossRef]
- Adler, A.; Araya-Polo, M.; Poggio, T. Deep learning for seismic inverse problems: Toward the acceleration of geophysical analysis workflows. IEEE Signal Process. Mag. 2021, 38, 89–119. [Google Scholar] [CrossRef]
- Kazei, V.; Ovcharenko, O.; Plotnitskii, P.; Peter, D.; Zhang, X.; Alkhalifah, T. Mapping full seismic waveforms to vertical velocity profiles by deep learning. Geophysics 2021, 86, R711–R721. [Google Scholar] [CrossRef]
- Gebraad, L.; Boehm, C.; Fichtner, A. Bayesian elastic full-waveform inversion using Hamiltonian Monte Carlo. J. Geophys. Res. Solid Earth 2020, 125, e2019JB018428. [Google Scholar] [CrossRef]
- Sun, H.; Demanet, L. Extrapolated full-waveform inversion with deep learning. Geophysics 2020, 85, R275–R288. [Google Scholar] [CrossRef]
- Mosser, L.; Dubrule, O.; Blunt, M.J. Stochastic seismic waveform inversion using generative adversarial networks as a geological prior. Math. Geosci. 2020, 52, 53–79. [Google Scholar] [CrossRef]
- Wu, Y.; McMechan, G.A. CNN-boosted full-waveform inversion. In Proceedings of the SEG International Exposition and Annual Meeting, Online, 11–16 October 2020; SEG: Houston, TX, USA, 2020; p. D031S057R003. [Google Scholar]
- Guo, C.; Pleiss, G.; Sun, Y.; Weinberger, K.Q. On calibration of modern neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 1321–1330. [Google Scholar]
- Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should I trust you?” Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
- Zhang, K.; Yan, J.; Zhang, F.; Ge, C.; Wan, W.; Sun, J.; Zhang, H. Spectral-Spatial Dual Graph Unfolding Network for Multispectral and Hyperspectral Image Fusion. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5508718. [Google Scholar] [CrossRef]
- Simonyan, K.; Vedaldi, A.; Zisserman, A. Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv 2013, arXiv:1312.6034. [Google Scholar]
- Keshk, M.; Koroniotis, N.; Pham, N.; Moustafa, N.; Turnbull, B.; Zomaya, A.Y. An explainable deep learning-enabled intrusion detection framework in IoT networks. Inf. Sci. 2023, 639, 119000. [Google Scholar] [CrossRef]
- Zhang, Q.; Yang, Y.; Ma, H.; Wu, Y.N. Interpreting cnns via decision trees. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6261–6270. [Google Scholar]
- Yan, J.; Zhang, K.; Zhang, F.; Ge, C.; Wan, W.; Sun, J. Multispectral and hyperspectral image fusion based on low-rank unfolding network. Signal Process 2023, 213, 109223. [Google Scholar] [CrossRef]
- Sun, L.; Zhang, K.; Zhang, F.; Wan, W.; Sun, J. Deep Rank-N Decomposition Network for Image Fusion. IEEE Trans. Multimed. 2024, 26, 7335–7348. [Google Scholar] [CrossRef]
- Zou, L.; Tosti, F.; Alani, A.M. Nondestructive inspection of tree trunks using a dual-polarized ground-penetrating radar system. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–8. [Google Scholar] [CrossRef]
- Liu, B.; Ren, Y.; Liu, H.; Xu, H.; Wang, Z.; Cohn, A.G.; Jiang, P. GPRInvNet: Deep learning-based ground-penetrating radar data inversion for tunnel linings. IEEE Trans. Geosci. Remote Sens. 2021, 59, 8305–8325. [Google Scholar] [CrossRef]
- Fang, Y.; Ma, T.; Tong, Z. Ground-Penetrating Radar Wave Response Simulation of Pipe Leakage in Subgrade Soil of Urban Road with Couple of Leakage and Radio Frequency. In Proceedings of the 22nd 103rd Transportation Research Board (TRB) Annual Meeting, Washington, DC, USA, 8–12 January 2023. [Google Scholar]
- Yager, R.R.; Liu, L. Classic Works of the Dempster-Shafer Theory of Belief Functions; Springer: Berlin/Heidelberg, Germany, 2008; Volume 219. [Google Scholar]
- Dempster, A.P. Upper and lower probabilities induced by a multivalued mapping. In Classic Works of the Dempster-Shafer Theory of Belief Functions; Springer: Berlin/Heidelberg, Germany, 2008; pp. 57–72. [Google Scholar]
- Shafer, G. A Mathematical Theory of Evidence; Princeton University Press: Princeton, NJ, USA, 1976; Volume 42. [Google Scholar]
- Denœux, T. Logistic regression, neural networks and Dempster–Shafer theory: A new perspective. Knowl.-Based Syst. 2019, 176, 54–67. [Google Scholar] [CrossRef]
- Denœux, T. An evidential neural network model for regression based on random fuzzy numbers. In Proceedings of the International Conference on Belief Functions, Paris, France, 26–28 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 57–66. [Google Scholar]
- Huang, L.; Ruan, S.; Decazes, P.; Denœux, T. Lymphoma segmentation from 3D PET-CT images using a deep evidential network. Int. J. Approx. Reason. 2022, 149, 39–60. [Google Scholar] [CrossRef]
- Tong, Z.; Xu, P.; Denoeux, T. An evidential classifier based on Dempster-Shafer theory and deep learning. Neurocomputing 2021, 450, 275–293. [Google Scholar] [CrossRef]
- Tong, Z.; Xu, P.; Denoeux, T. Evidential fully convolutional network for semantic segmentation. Appl. Intell. 2021, 51, 6376–6399. [Google Scholar] [CrossRef]
- Denoeux, T. Distributed combination of belief functions. Inf. Fusion 2021, 65, 179–191. [Google Scholar] [CrossRef]
- Denoeux, T.; Kanjanatarakul, O.; Sriboonchitta, S. A new evidential k-nearest neighbor rule based on contextual discounting with partially supervised learning. Int. J. Approx. Reason. 2019, 113, 287–302. [Google Scholar] [CrossRef]
- Tong, Z.; Ma, T.; Zhang, W.; Huyan, J. Evidential transformer for pavement distress segmentation. Comput.-Aided Civ. Infrastruct. Eng. 2023, 38, 2317–2338. [Google Scholar] [CrossRef]
- Shwartz-Ziv, R.; Tishby, N. Opening the black box of deep neural networks via information. arXiv 2017, arXiv:1703.00810. [Google Scholar] [CrossRef]
- Denœux, T. Quantifying prediction uncertainty in regression using random fuzzy sets: The ENNreg model. IEEE Trans. Fuzzy Syst. 2023, 31, 3690–3699. [Google Scholar] [CrossRef]
- Zadeh, L.A. Fuzzy sets as a basis for a theory of possibility. Fuzzy Sets Syst. 1978, 1, 3–28. [Google Scholar] [CrossRef]
- Denœux, T. Reasoning with fuzzy and uncertain evidence using epistemic random fuzzy sets: General framework and practical models. Fuzzy Sets Syst. 2023, 453, 1–36. [Google Scholar] [CrossRef]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 26, 84–90. [Google Scholar] [CrossRef]
- Zhang, Y.; Tong, Z.; She, X.; Wang, S.; Zhang, W.; Fan, J.; Cheng, H.; Yang, H.; Cao, J. SWC-Net and Multi-Phase Heterogeneous FDTD Model for Void Detection Underneath Airport Pavement Slab. IEEE Trans. Intell. Transp. Syst. 2024, 25, 20698–20714. [Google Scholar] [CrossRef]
- Tong, Z.; Zhang, Y.; Ma, T. Permittivity measurement with uncertainty quantification in cement-based composites using ENNreg-ANet and high-frequency electromagnetic waves. Measurement 2025, 244, 116537. [Google Scholar] [CrossRef]
- Zeng, J.; Huang, L.; Bai, X.; Wang, K. DBFFT: Adversarial-robust dual-branch frequency domain feature fusion in vision transformers. Inf. Fusion 2024, 108, 102387. [Google Scholar] [CrossRef]
- Van Der Linden, I.; Haned, H.; Kanoulas, E. Global aggregations of local explanations for black box models. arXiv 2019, arXiv:1907.03039. [Google Scholar] [CrossRef]
- Srinivas, S.; Fleuret, F. Full-gradient representation for neural network visualization. Adv. Neural Inf. Process. Syst. 2019, 32, 4124–4133. [Google Scholar]
- Chefer, H.; Gur, S.; Wolf, L. Transformer interpretability beyond attention visualization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 782–791. [Google Scholar]
- Wang, H.; Wang, Z.; Du, M.; Yang, F.; Zhang, Z.; Ding, S.; Mardziel, P.; Hu, X. Score-CAM: Score-weighted visual explanations for convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; pp. 24–25. [Google Scholar]
- Mothilal, R.K.; Sharma, A.; Tan, C. Explaining machine learning classifiers through diverse counterfactual explanations. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, Barcelona, Spain, 27–30 January 2020; pp. 607–617. [Google Scholar]
- Kim, B.; Wattenberg, M.; Gilmer, J.; Cai, C.; Wexler, J.; Viegas, F.; Sayres, R. Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav). In Proceedings of the International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 2668–2677. [Google Scholar]
- Samek, W.; Binder, A.; Montavon, G.; Lapuschkin, S.; Müller, K.R. Evaluating the visualization of what a deep neural network has learned. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 2660–2673. [Google Scholar] [CrossRef]
- Hooker, S.; Erhan, D.; Kindermans, P.J.; Kim, B. A benchmark for interpretability methods in deep neural networks. Adv. Neural Inf. Process. Syst. 2019, 32, 9737–9748. [Google Scholar]
Examples | Element Number | ||||
---|---|---|---|---|---|
1 | 2 | 3 | 4 | 5 | |
Sample 1 | |||||
Sample 2 | |||||
Sample 3 | |||||
Sample 4 |
Frequency | 200 MHz | 450 MHz | 800 MHz | 1.2 GHz | |
---|---|---|---|---|---|
No. | Training | 240 | 360 | 900 | 900 |
Validation | 80 | 120 | 300 | 300 | |
Testing | 80 | 120 | 300 | 300 | |
Total | 400 | 600 | 1500 | 1500 |
RNN | YOLO v8 | DetTransformer | StreamPETR | SWC-Net | Semantics-Guided SWC-Net | |
---|---|---|---|---|---|---|
Classification accuracy/% | 76.85 | 89.91 | 90.32 | 88.26 | 91.27 | 94.26 |
CIoU/% | 79.34 | 85.28 | 81.58 | 82.36 | 87.15 | 90.03 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Li, X.; Su, M.; Zhu, Y.; Ma, S.; Liu, S.; Tong, Z. Evidential Interpretation Approach for Deep Neural Networks in High-Frequency Electromagnetic Wave Processing. Electronics 2025, 14, 3277. https://doi.org/10.3390/electronics14163277
Li X, Su M, Zhu Y, Ma S, Liu S, Tong Z. Evidential Interpretation Approach for Deep Neural Networks in High-Frequency Electromagnetic Wave Processing. Electronics. 2025; 14(16):3277. https://doi.org/10.3390/electronics14163277
Chicago/Turabian StyleLi, Xueliang, Ming Su, Yu Zhu, Shansong Ma, Shifu Liu, and Zheng Tong. 2025. "Evidential Interpretation Approach for Deep Neural Networks in High-Frequency Electromagnetic Wave Processing" Electronics 14, no. 16: 3277. https://doi.org/10.3390/electronics14163277
APA StyleLi, X., Su, M., Zhu, Y., Ma, S., Liu, S., & Tong, Z. (2025). Evidential Interpretation Approach for Deep Neural Networks in High-Frequency Electromagnetic Wave Processing. Electronics, 14(16), 3277. https://doi.org/10.3390/electronics14163277