Lifelong Machine Learning-Based Efficient Robotic Object Perception

A special issue of Electronics (ISSN 2079-9292). This special issue belongs to the section "Artificial Intelligence".

Deadline for manuscript submissions: 15 September 2024 | Viewed by 2132

Special Issue Editors


E-Mail Website
Guest Editor
State Key Laboratory of Robotics, Shenyang Institute of Automation, Institutes for Robotics and Intelligent Manufacturing, Chinese Academy of Sciences, Shenyang 110016, China
Interests: lifelong learning; robot perception; incremental learning; domain adaptation

E-Mail Website
Guest Editor
Department of Robotics, Hunan University, Changsha 410082, China
Interests: aerial robots; swarm robots; visual servo; embedded system applications

E-Mail Website
Guest Editor
School of Telecommunications Engineering, Xidian University, Xi'an 710071, China
Interests: pattern recognition; machine learning; multi-view learning

Special Issue Information

Dear Colleagues,

This Special Issue (SI) is intended to present scholarly papers that address efficient robotic object perception problem from the perspective of lifelong machine learning. Lifelong machine learning aims to utilizes knowledge from past tasks to efficiently and effectively learn new tasks over a lifetime, which is more suitable for the robotic learning scenarios, i.e., perceive the objects or environments in a never-ending manner. Then, the question emerges: “how to lifelong perceive objects with a robot?” To answer this question, we invite scientists, researchers, and robotic specialists together with academics to share their insights of the lifelong robot perception learning. What will the robot learn in a lifelong manner? What kind of knowledge or experience is most suitable for robot perception? How does the robot learn when encountering a new task? Meanwhile, humans can learn from just one or a handful of examples (i.e., few- or zero-shot learning) with vision–audio–touch senses; can robot achieve very long-term learning in this manner as humans do? All of these are important discussions at the moment and this Special Issue will help all those interested in the topic to promote their vision and ideas.

Overall, this Special Issue focuses on learning with fewer labels for robot multi-modality perception tasks such object classification, object detection, semantic segmentation, robot navigation, SLAM, and many others, and the topics of interest include (but are not limited to) the following areas:

  • Life-long/continual/incremental learning methods
  • New methods for few-/zero-shot learning
  • Robot embodied intelligence
  • Federated robot perception learning
  • Meta-learning methods
  • Self-supervised learning methods
  • Novel domain adaptation methods
  • Lifelong semi-supervised learning methods
  • Lifelong unsupervised learning methods
  • Lifelong multi-modal learning methods

Therefore, we invite contributions from experimental researchers and theorists of high-quality manuscripts for publication in this SI.

Dr. Gan Sun
Dr. Hang Zhong
Dr. Qianqian Wang
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 100 words) can be sent to the Editorial Office for announcement on this website.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Electronics is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2400 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

 

Keywords

  • lifelong/continual/incremental learning methods
  • new methods for few-/zero-shot learning
  • robot embodied intelligence
  • federated robot perception learning
  • meta-learning methods
  • self-supervised learning methods
  • novel domain adaptation methods
  • lifelong semi-supervised learning methods
  • lifelong unsupervised learning methods
  • lifelong multi-modal learning methods

Published Papers (2 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

16 pages, 7134 KiB  
Article
Dual-Stage Attribute Embedding and Modality Consistency Learning-Based Visible–Infrared Person Re-Identification
by Zhuxuan Cheng, Huijie Fan, Qiang Wang, Shiben Liu and Yandong Tang
Electronics 2023, 12(24), 4892; https://doi.org/10.3390/electronics12244892 - 5 Dec 2023
Viewed by 785
Abstract
Visible–infrared person re-identification (VI-ReID) is an emerging technology for realizing all-weather smart surveillance systems. To address the problem of pedestrian discriminative information being difficult to obtain and easy to lose, as well as the wide modality difference in the VI-ReID task, in this [...] Read more.
Visible–infrared person re-identification (VI-ReID) is an emerging technology for realizing all-weather smart surveillance systems. To address the problem of pedestrian discriminative information being difficult to obtain and easy to lose, as well as the wide modality difference in the VI-ReID task, in this paper we propose a two-stage attribute embedding and modality consistency learning-based VI-ReID method. First, the attribute information embedding module introduces the fine-grained pedestrian information in the attribute label into the transformer backbone, enabling the backbone to extract identity-discriminative pedestrian features. After obtaining the pedestrian features, the attribute embedding enhancement module is utilized to realize the second-stage attribute information embedding, which reduces the adverse effect of losing the person discriminative information due to the deepening of network. Finally, the modality consistency learning loss is designed for constraining the network to mine the consistency information between two modalities in order to reduce the impact of modality difference on the recognition results. The results show that our method reaches 74.57% mAP on the SYSU-MM01 dataset in All Search mode and 87.02% mAP on the RegDB dataset in IR-to-VIS mode, with a performance improvement of 6.00% and 2.56%, respectively, proving that our proposed method is able to reach optimal performance compared to existing state-of-the-art methods. Full article
(This article belongs to the Special Issue Lifelong Machine Learning-Based Efficient Robotic Object Perception)
Show Figures

Figure 1

15 pages, 2597 KiB  
Article
Object Detection Network Based on Module Stack and Attention Mechanism
by Xinke Dou, Ting Wang, Shiliang Shao and Xianqing Cao
Electronics 2023, 12(17), 3542; https://doi.org/10.3390/electronics12173542 - 22 Aug 2023
Viewed by 793
Abstract
Currently, visual computer applications based on convolutional neural networks are rapidly developing. However, several problems remain: (1) high-quality graphics processing equipment is needed, and (2) the trained network model has several unnecessary convolution operations. These problems result in a single-stage target detection network [...] Read more.
Currently, visual computer applications based on convolutional neural networks are rapidly developing. However, several problems remain: (1) high-quality graphics processing equipment is needed, and (2) the trained network model has several unnecessary convolution operations. These problems result in a single-stage target detection network that often requires unnecessary computing power and is difficult to apply to equipment with insufficient computing resources. To solve these problems, based on YOLOv5, a YOLOv5-L (YOLOv5 Lightweight) network structure is proposed. This network is improved using YOLOv5. First, to enhance the inference speed of the detector on the CPU, the PP-LCNet (PaddlePaddle-Lightweight CPU Net) is employed as the backbone network. Second, the focus module is removed, and the end convolution module in the head network is replaced by a deep separable convolution module, which eliminates redundant operations and reduces the amount of computation. The experimental results show that YOLOv5-L enables a 48% reduction in model parameters and computation compared to YOLOv5, a 35% increase in operation speed, and a less than 2% reduction in accuracy, which is significant in the environment of low-performance computing equipment. Full article
(This article belongs to the Special Issue Lifelong Machine Learning-Based Efficient Robotic Object Perception)
Show Figures

Figure 1

Back to TopTop