sensors-logo

Journal Browser

Journal Browser

Sensor-Fusion-Based Deep Interpretable Networks

A special issue of Sensors (ISSN 1424-8220). This special issue belongs to the section "Sensor Networks".

Deadline for manuscript submissions: 15 November 2024 | Viewed by 590

Special Issue Editors


E-Mail Website
Guest Editor
Department of Information Science, Xi’an University of Technology, Xi’an 710048, China
Interests: visual information processing; pattern recognition
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
School of Electrical and Computer Engineering, Oklahoma State University, Stillwater ,OK 74078,USA
Interests: image processing; machine learning; pattern recognition; computer vision; biomedical imaging and multimedia applications
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
College of Electrical and Information Engineering, Hunan University, Changsha 410082, China
Interests: computer vision; machine learning; pattern recognition
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
School of Automation and Electrical Engineering, University of Science and Technology Beijing, Beijing 100083, China
Interests: image and video semantic segmentation; deep learning; industrial process control; industrial intelligence; natural language processing; knowledge graph
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues,

Sensor fusion is a technology that combines data and information from multiple sensors to obtain more comprehensive, accurate, and reliable perception results. Sensor fusion based on deep interpretable networks is more advanced, utilizing the powerful modeling and abstraction capabilities of deep learning to process multi-source sensor data, while emphasizing the interpretability of the model, making the output results of the model more convincing and credible.

In this network, data fusion technology can make full use of redundant information and complementarity between different sensors to improve the overall perception accuracy and robustness. By accurately calibrating and synchronizing various sensors, deep learning models can learn deeper features and complex patterns in the fused data. In addition, the interpretability of the model aids in understanding its internal working principles and decision making, which is crucial for the safety and reliability of key application scenarios, such as autonomous driving and intelligent manufacturing.

This Special Issue aims to put together original research and review articles on recent advances, technologies, solutions, applications, and new challenges in the field of sensor fusion-based deep interpretable networks.

Potential topics include but are not limited to the following:

  • Sensor Fusion for Comprehensive Perception;
  • Multi-View Adaptive Fusion Network for Object Detection, Recognition and Understanding;
  • Deep Learning and Sensor Fusion for Enhanced Decision Making;
  • Open-Source Sensor Fusion for different application scenarios;
  • Kalman and Complementary Filters in Sensor Fusion;
  • Autonomous Driving with Scene Understanding;
  • Sensor Fusion for Motion Tracking Capabilities in Smartphones and Tablets.

Dr. Guangfeng Lin
Prof. Dr. Guoliang Fan
Dr. Zhigang Ling
Dr. Jiangyun Li
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Sensors is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2600 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • sensors fusion
  • reliability networks
  • causal reasoning
  • interpretable networks
  • graph neural networks
  • attention fusion mechanism

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • e-Book format: Special Issues with more than 10 articles can be published as dedicated e-books, ensuring wide and rapid dissemination.

Further information on MDPI's Special Issue polices can be found here.

Published Papers (1 paper)

Order results
Result details
Select all
Export citation of selected articles as:

Research

14 pages, 35441 KiB  
Article
Smart Ship Draft Reading by Dual-Flow Deep Learning Architecture and Multispectral Information
by Bo Zhang, Jiangyun Li, Haicheng Tang and Xi Liu
Sensors 2024, 24(17), 5580; https://doi.org/10.3390/s24175580 - 28 Aug 2024
Viewed by 351
Abstract
In maritime transportation, a ship’s draft survey serves as a primary method for weighing bulk cargo. The accuracy of the ship’s draft reading determines the fairness of bulk cargo transactions. Human visual-based draft reading methods face issues such as safety concerns, high labor [...] Read more.
In maritime transportation, a ship’s draft survey serves as a primary method for weighing bulk cargo. The accuracy of the ship’s draft reading determines the fairness of bulk cargo transactions. Human visual-based draft reading methods face issues such as safety concerns, high labor costs, and subjective interpretation. Therefore, some image processing methods are utilized to achieve automatic draft reading. However, due to the limitations in the spectral characteristics of RGB images, existing image processing methods are susceptible to water surface environmental interference, such as reflections. To solve this issue, we obtained and annotated 524 multispectral images of a ship’s draft as the research dataset, marking the first application of integrating NIR information and RGB images for automatic draft reading tasks. Additionally, a dual-branch backbone named BIF is proposed to extract and combine spectral information from RGB and NIR images. The backbone network can be combined with the existing segmentation head and detection head to perform waterline segmentation and draft detection. By replacing the original ResNet-50 backbone of YOLOv8, we reached a mAP of 99.2% in the draft detection task. Similarly, combining UPerNet with our dual-branch backbone, the mIoU of the waterline segmentation task was improved from 98.9% to 99.3%. The inaccuracy of the draft reading is less than ±0.01 m, confirming the efficacy of our method for automatic draft reading tasks. Full article
(This article belongs to the Special Issue Sensor-Fusion-Based Deep Interpretable Networks)
Show Figures

Graphical abstract

Back to TopTop