Neural Networks for Feature Extraction

A special issue of Electronics (ISSN 2079-9292). This special issue belongs to the section "Computer Science & Engineering".

Deadline for manuscript submissions: 15 January 2025 | Viewed by 653

Special Issue Editors


E-Mail Website
Guest Editor
School of Artificial Intelligence, Xidian University, Xi’an 710071, China
Interests: artificial intelligence
Academy of Advanced Disciplinary Research, Xidian University, Xi’an 710071, China
Interests: artificial intelligence; medical imaging; machine Learning

Special Issue Information

Dear Colleagues,

Neural networks are powerful machine learning algorithms that can automatically learn feature representations of input data. Compared with traditional methods, using neural networks for feature extraction has the following advantages: as an unsupervised feature learning method, neural networks can automatically discover features in the dataset without manual feature engineering. Deep neural networks can learn more abstract and advanced features of the data, containing information about the attributes and structure of the data, while the features learned by traditional methods tend to be more-superficial and harder-to-learn, abstract features. The features learned by neural networks can be well generalized to new data, while the features learned by traditional methods tend to be too dependent on training data. Neural networks can learn the intrinsic dependencies between features, and such feature representations are often more powerful than simply stacking features.

Neural networks have been widely used in feature extraction tasks, and current research focuses on the following aspects: improving the effectiveness of feature learning, e.g., through regularization and pre-training; and designing new neural network structures to learn more abstract and efficient features, e.g., convolutional neural networks, recurrent neural networks, and spike neural networks.
Combining neural networks with other methods can be carried out to form a more powerful feature learning framework.

In summary, neural networks have the advantages of automatically learning features, learning abstract features, good feature generalization, and convergence to stable features, which often make the features learned by neural networks more powerful than artificial features and widely applicable to downstream tasks. Neural networks provide a powerful tool for feature learning and representation.

Dr. Zhen Cao
Dr. Zhang Guo
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Electronics is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2400 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Published Papers (1 paper)

Order results
Result details
Select all
Export citation of selected articles as:

Research

19 pages, 10890 KiB  
Article
Exploring Neighbor Spatial Relationships for Enhanced Lumbar Vertebrae Detection in X-ray Images
by Yu Zeng, Kun Wang, Lai Dai, Changqing Wang, Chi Xiong, Peng Xiao, Bin Cai, Qiang Zhang, Zhiyong Sun, Erkang Cheng and Bo Song
Electronics 2024, 13(11), 2137; https://doi.org/10.3390/electronics13112137 - 30 May 2024
Viewed by 335
Abstract
Accurately detecting spine vertebrae plays a crucial role in successful orthopedic surgery. However, identifying and classifying lumbar vertebrae from arbitrary spine X-ray images remains challenging due to their similar appearance and varying sizes among individuals. In this paper, we propose a novel approach [...] Read more.
Accurately detecting spine vertebrae plays a crucial role in successful orthopedic surgery. However, identifying and classifying lumbar vertebrae from arbitrary spine X-ray images remains challenging due to their similar appearance and varying sizes among individuals. In this paper, we propose a novel approach to enhance vertebrae detection accuracy by leveraging both global and local spatial relationships between neighboring vertebrae. Our method incorporates a two-stage detector architecture that captures global contextual information using an intermediate heatmap from the first stage. Additionally, we introduce a detection head in the second stage to capture local spatial information, enabling each vertebra to learn neighboring spatial details, visibility, and relative offset. During inference, we employ a fusion strategy that combines spatial offsets of neighboring vertebrae and heatmap from a conventional detection head. This enables the model to better understand relationships and dependencies between neighboring vertebrae. Furthermore, we introduce a new representation of object centers that emphasizes critical regions and strengthens the spatial priors of human spine vertebrae, resulting in an improved detection accuracy. We evaluate our method using two lumbar spine image datasets and achieve promising detection performance. Compared to the baseline, our algorithm achieves a significant improvement of 13.6% AP in the CM dataset and surpasses 6.5% and 4.8% AP in the anterior and lateral views of the BUU dataset, respectively. Full article
(This article belongs to the Special Issue Neural Networks for Feature Extraction)
Show Figures

Figure 1

Back to TopTop