Next Article in Journal
Optimization Design of CMUT Sensors with Broadband and High Sensitivity Characteristics Based on the Genetic Algorithm
Previous Article in Journal
DMAF-NET: Deep Multi-Scale Attention Fusion Network for Hyperspectral Image Classification with Limited Samples
Previous Article in Special Issue
Attention Mechanism and LSTM Network for Fingerprint-Based Indoor Location System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Compressed Communication and Application in Multi-Robot 2D-Lidar SLAM: An Intelligent Huffman Algorithm

School of Electrical Engineering and Automation, Anhui University, Hefei 230093, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(10), 3154; https://doi.org/10.3390/s24103154
Submission received: 7 April 2024 / Revised: 9 May 2024 / Accepted: 14 May 2024 / Published: 16 May 2024
(This article belongs to the Special Issue UAV and Sensors Applications for Navigation and Positioning)

Abstract

:
Multi-robot Simultaneous Localization and Mapping (SLAM) systems employing 2D lidar scans are effective for exploration and navigation within GNSS-limited environments. However, scalability concerns arise with larger environments and increased robot numbers, as 2D mapping necessitates substantial processor memory and inter-robot communication bandwidth. Thus, data compression prior to transmission becomes imperative. This study investigates the problem of communication-efficient multi-robot SLAM based on 2D maps and introduces an architecture that enables compressed communication, facilitating the transmission of full maps with significantly reduced bandwidth. We propose a framework employing a lightweight feature extraction Convolutional Neural Network (CNN) for a full map, followed by an encoder combining Huffman and Run-Length Encoding (RLE) algorithms to further compress a full map. Subsequently, a lightweight recovery CNN was designed to restore map features. Experimental validation involves applying our compressed communication framework to a two-robot SLAM system. The results demonstrate that our approach reduces communication overhead by 99% while maintaining map quality. This compressed communication strategy effectively addresses bandwidth constraints in multi-robot SLAM scenarios, offering a practical solution for collaborative SLAM applications.

1. Introduction

Recent years have witnessed significant advancements in multi-robot SLAM algorithms due to their potential applications in search and rescue operations, environmental monitoring, and the exploration of unknown or hazardous environments [1,2,3,4,5]. Single-robot SLAM, where a solitary robot independently navigates and maps its surroundings, is often constrained by time-consuming exploration and limited observation capabilities in large areas. In contrast, multi-robot SLAM has the advantage of efficiently exploring large-scale environments, enhancing mapping accuracy, and improving robustness through the collaborative efforts of multiple robots in various challenging scenarios.
Traditional multi-robot SLAM is developed directly from expanding the single SLAM into multiple robots under a centralized scheme. For example, the CCM-SLAM [6] stands out as a well-established centralized system for visual-inertial multi-robot SLAM. In this system, a central server takes charge of multi-robot map management, fusion, and optimization. By building upon [6], COVINS [7,8] is an extension that has been demonstrated to scale effectively to 12 robots. Another recent addition to centralized CSLAM systems is CVIDS [9], which has achieved the collaborative localization and dense reconstruction of multiple agents in a unified co-ordinate system by using a monocular visual-inertial sensor suite and a centralized loosely coupled framework. LAMP [10,11] is a computationally efficient and outlier-resilient centralized multi-robot SLAM system that consists of a scalable loop closure detection module and GNC (graduated nonconvexity)-based pose map optimization module. Though centralized systems are simply realized and with high precision, they require stable communications and are not robust to the failure of the central node.
Distributed systems can alleviate these problems by reducing the reliance on a central server. Ref. [12] proposed a metric-semantic 3D mesh model for a distributed multi-robot SLAM system, which includes an outlier-resistant fully distributed trajectory estimation method based on GNC and a novel outlier-resistant distributed PGO algorithm. In [13], Kimera-Multi [12,14] is improved to adapt to large-scale real-world deployments, with special emphasis on handling intermittent and unreliable communications. DOOR SLAM [15] uses a distributed frontend to detect inter-robot loop closures without exchanging raw sensor data, successfully rejecting outliers and obtaining accurate trajectory estimates while requiring low communication bandwidth. At the same time, lidar-based distributed multi-robot SLAM is also underway. In [16], the Scan-Context algorithm is proposed to project a 3D point cloud onto a low-resolution 2D plane, thus converting it into a compact feature vector for retrieving the created map and enabling localization, greatly reducing the amount of data and computational complexity. In [17], a distributed backend is used to remove inter-robot loop closure redundancy and perform position map optimization to optimize the global robot cluster position.
The selection of appropriate map representations of multi-robot SLAM directly impacts its performance on computation efficiency, memory usage, and communication burden. For instance, in environments where ground robots navigate indoor spaces, employing a 2D map is sufficient [18]. Studies have demonstrated that occupancy grid maps provide a compact and accurate solution compared to feature-based maps [19,20]. However, certain applications necessitate 3D representations despite the associated computational and storage complexities. This complexity presents challenges, particularly for resource-constrained robotic platforms. Given the communication limitations in multi-robot SLAM systems, there is a preference for compact or sparse map representations, such as the topological maps used in [21,22]. Additionally, efforts are underway to develop semantic-based representations, such as sparse maps annotated with labeled regions [23]. The choice of map representation is critical for long-term operations due to increasing memory requirements, posing a persistent challenge in multi-robot SLAM [24]. Communication bottlenecks in multi-robot SLAM systems typically arise from the exchange of sensor data or representations used for computing inter-robot loop closures [25]. Robots must exchange sufficient data to determine if other robots have explored the same areas and subsequently estimate map alignment using overlapping map sections. Therefore, advancements in the frontend of multi-robot SLAM systems often involve efficient methods for searching loop closure candidates across a team while considering communication constraints. The increasing challenge of balancing efficientinformation transmission with limited communication bandwidth is a pressing issue in multi-robot SLAM systems.
Distributed multi-robot SLAM highly relies on the exchange of data among the network to realize co-ordination processes, such as place recognition, relative pose computation, loop closure detection, and so on. There is increasing work attempting to relieve the communication burden such that the multi-robot system can be deployed to critical environments with less communication capacity. Instead of communicating the merged full map, only the key points and descriptors from parts of the keyframes (using the camera) or submaps (using lidar) are exchanged to realize these co-ordination processes. As depicted in Table 1, segments are extracted in [26] from the source point cloud to generate a source map that contains lists of low-dimensional segment descriptors. In [9], BRIEF descriptors are extracted from each frame and exchanged with another robot. In [27], the keyframes and map points of each robot are transmitted to the server and then distributed by the server to other robots. In [28], two communication modes are designed in different situations: compact and greedy. In compact mode, the communication overhead is minimized by sending compact descriptors. In greedy mode, each drone shares as much information as possible, which is suitable for good network conditions. In [17], NetVlAD descriptors are transmitted between the robots to perform place recognition. In [12], the data flow between different robots are key points and feature descriptors. Compact representations have been explored, incorporating semantic features [29] that rely on objects as landmarks. This approach requires communicating only object labels and poses to other robots, presenting a condensed object-based descriptor that depends on the configuration of objects in a scene for place recognition. In addition to compacting representations, ensuring the sharing of only pertinent information is valuable. However, the data types transmitted by these methods are partial or compact representations of the data, which will lead to the loss of detail. In a wide range of task scenarios, if there will be an accumulation of errors, it is meaningful to study the transmission of the merged full map.
We present a multi-robot SLAM system utilizing 2D occupancy grid maps as a representation of the environment, enabling merged, full-map transmission. Due to the specific format characteristics of raster maps, we have devised intelligent compression algorithms aimed at significantly reducing redundancy in the information contained within these maps while preserving their overall integrity. These intelligent compression algorithms consist of a lightweight network for both map feature extraction and recovery, along with a map feature coder and decoder.
The main innovations and contributions of this paper are as follows:
  • We present a compressed communication framework for multi-robot SLAM, enabling merged, full-map transmission among robots, which can reduce the duration of exploration and produce a map of the whole environment quickly. Such a whole map plays a key role in many robotic tasks, such as path planning, collision avoidance et al.
  • According to the characteristics of the occupancy grid map, we designed an intelligent compression algorithm by combining a convolutional neural network, Huffman coding, and RLE coding that compress the transmitted full map by both image downsampling and stream encoding.
  • We use Ultra-Wide Band (UWB) as a communication medium to validate the multi-robot SLAM system proposed in this paper. The results show that our method is able to reduce the communication burden by up to 99 percent and the localization error is less than 5 cm compared to when no compression is employed.
The remainder of the paper is organized as follows: In Section 2, we describe the proposed multi-robot SLAM method based on an intelligent compression algorithm. In Section 3, the experimental results are provided to validate the performance of the intelligent compression algorithm. Finally, we conclude this paper in Section 4.

2. The Compressed Communication Approach

The multi-robot SLAM framework based on compressed communication consists of the following three main components: full-map creation, full-map compression and transmission, and full-map fusion. In Figure 1, the robots first use Gmapping to build the full map. The map is then compressed by our proposed algorithm and transmitted to other robots using the UWB device. After receiving maps, a robot can decompress maps and merge them with its own created full maps. Eventually, the robot can navigate through the environment using the merged maps.

2.1. Full-Map Creation

Gmapping [30] is a SLAM algorithm that combines particle filtering and FastSLAM. The algorithm predicts the robot’s position using a motion model and then uses sensor data for position correction and map updates. As the robot moves through the environment, it gradually builds an accurate occupancy grid map and is able to determine the robot’s position on the map. We use Gmapping to create full maps for each robot.

2.2. Full-Map Compression and Decompression

Map compression consists of a map feature extraction CNN CNN f e a t u r e and an encoder. The map feature m f e a t u r e that retains the key information of the full map is generated by CNN f e a t u r e . Then, m f e a t u r e is fed into the encoder to form a string encoding e. Corresponding to the above, map decompression consists of a decoder and a map recovery CNN CNN r e c o v e r y . The decoder recovers e into m f e a t u r e , and then the m f e a t u r e is processed through CNN r e c o v e r y to obtain the new full map. Although the new full map is slightly different from the map before compression, it does not affect its ability to provide navigation for the robot, as demonstrated in the experiments in Section 3.
(1) CNN f e a t u r e : As shown in Figure 2, CNN f e a t u r e consists of three weight layers, which preserve the structural information of the input map while obtaining storage space-saving m f e a t u r e . The first layer consists of convolution (set the convolution kernel to 3 × 3 ) and ReLU (enhance the model’s expressive power) for the purpose of extracting the features of the map. Considering the computing power of the processing unit on the robot, we set 16 sets of 3 × 3 filters. The second layer, composed of convolution, Batch Normalization (BN), and ReLU, aims to downsample and enhance features. In order to change the map resolution to a quarter of the original, the parameter stride of the convolution was set to 2, and 16 filters of size 3 × 3 × 16 were used. In the last layer, a filter of size 3 × 3 × 16 is used to get the m f e a t u r e . Compared to downsampling the map directly, the m f e a t u r e obtained by CNN f e a t u r e contains more map information.
(2) Encoder for m a p f e a t u r e : The encoding process of m f e a t u r e is shown in Figure 3. The algorithm of m f e a t u r e encoder based on RLE and Huffman is shown in Algorithm 1. Firstly, the values that appear in the m f e a t u r e are recorded V = { v 1 , v 2 , . . . , v n } , and the probability of each value in m f e a t u r e are calculated. We order the probability of occurrence of each value from highest to lowest to obtain the set P = { p 1 , p 2 , . . . , p n } , where p i represents the probability of the i-th value in m f e a t u r e . N = { n 1 , n 2 , . . . , n n } , which are initialized according to P = { p 1 , p 2 , . . . , p n } , and n i corresponds to p i . Then, the two smallest values in the set N are merged into a new node, so N and P are updated to { p 1 , . . . , p n 2 , p m e r g e d } , and { n 1 , . . . , n n 2 , n m e r g e d } . We repeat this step until only one node remains in N . According to the process of node emergence, we can get the Huffman coding tree, e.g., Figure 4, where the grey nodes represent the initial nodes and the white nodes are the new nodes generated by merging the initial nodes. As a result, we can encode each value in m f e a t u r e to obtain the code table T = { code 1 : v 1 , . . . , code n : v n } . Then, according to the code table T , m f e a t u r e is encoded into the binary code. Finally, the binary code is converted into hexadecimal code h to compress the length of the code. In order to further enhance compression efficiency and reduce communication pressure, consecutive repeated characters in hexadecimal encoding are subsequently processed. We replace the consecutively repeated code segments in h with the repeated characters themselves and the number of times they are repeated to obtain e . Therefore, m f e a t u r e is compressed into a code table T and a string e .
Algorithm 1  m f e a t u r e compression algorithm
Input: 
m f e a t u r e
   1:
record the values in the m f e a t u r e V = { v 1 , v 2 , . . . , v n } , count the probability of different value P = { p 1 , p 2 , . . . , p n } , initialize nodes N = { n 1 , n 2 , . . . , n n }
   2:
repeat
   3:
   merge the two nodes with the lowest probability to get the new node and update P , N
   4:
until there is only one node left in the set N
   5:
encode each value according to the Huffman encoding tree to get code table T : { code 1 : v 1 , . . . , code n : v n }
   6:
rewrite m f e a t u r e data into binary code according to the code table
   7:
convert binary code to hexadecimal code h
   8:
Initialize e as an empty string and count as 1.
   9:
For index from 1 to length( h ) − 1:
    a.
Set currentChar as the character at the current index .
    b.
Set previousChar as the character at the previous index .
    c.
If currentChar equals previousChar, count + 1.
    d.
If currentChar does not equal previousChar, append count and previousChar to e , then reset count to 1.
Output:  
T , e
(3) Decoder for m f e a t u r e : The decoding process of m f e a t u r e is shown in Figure 3. The algorithm of the decoder is shown in Algorithm 2. First, the e obtained by the compression algorithm is restored to hexadecimal code d . Then, the hexadecimal code d is converted into the binary code. According to the code table T , we can replace the corresponding code with the corresponding value in m f e a t u r e . Finally, m f e a t u r e is recovered according to T  and  e .
Algorithm 2  m f e a t u r e decompression algorithm
Input:  
T , e
   1:
Initialize d as an empty string.
   2:
While there are characters remaining in e :
    a.
Extract the next count -value pair from e .
    b.
Append count occurrences of the associated character to d .
   3:
convert hexadecimal code d to binary code
   4:
Recover m f e a t u r e from binary code according to the code table T
Output:  
m f e a t u r e
(4) CNN r e c o v e r y : In Figure 5, CNN r e c o v e r y consists of six weighting layers, and there are three types, namely Convolution + ReLU, Convolution + BN + ReLU, and Convolution. The purpose of CNN r e c o v e r y is to recover the full map from m f e a t u r e , and a residual block is employed in CNN r e c o v e r y . The jump connections of the residual block can help the network learn the difference between the input and the output. This allows CNN r e c o v e r y to better recover the full map, thus preserving the detailed information of m f e a t u r e . A total of 16 filters of size 3 × 3 are used to generate 16 feature mappings with ReLU in the first layer. For the second to fifth layers, 16 filters of size 3 × 3 × 16 are utilized, and BN is inserted between Convolution and ReLU. The last layer uses a filter of size 3 × 3 × 16 to generate the final single-channel output, i.e., the full map.
(5) Learning algorithm: The entire network consists of CNN f e a t u r e , the encoder, the decoder, and CNN r e c o v e r y . The optimization function for the network is designed as follows:
( θ 1 ¯ , θ 2 ¯ ) = arg min θ 1 , θ 2 R ( θ 2 , C o d ( F ( θ 1 , m ) ) ) m 2
where R ( · ) and F ( · ) represent CNN r e c o v e r y and CNN f e a t u r e , respectively. C o d ( · ) represents the encoder and decoder. m is the full map that is the input of the network. θ 1 and θ 2 are the parameters of the optimization function.
Therefore, we calculate the mean square error between the map of network output and the original full map as the loss and the gradient of the whole network parameters concerning loss by backpropagation. Then, we use the Adaptive Moment Estimation (Adam) optimization algorithm to update the parameters θ 1 and θ 2 to minimize the optimization function.
(6) Code transmission based on UWB: When m f e a t u r e is processed by the encoder, it becomes a string. In order to facilitate the transmission, we split the string and transmitted it to other robots using UWB segment by segment. After a string has been transmitted, the integrity of the received string is judged. If any data are missing, the receiver commands the transmitter to resend the string.

2.3. Multi-Occupancy Grid Maps Fusion

We consider two cases of multi-occupancy grid map fusion:
(1) The initial positions of the robots are known: In this case, the rigid transformation ( R , t ) between the maps can be computed directly from the initial position of the robot (the x , y co-ordinates and the y a w angle) so that accurate fusion results can be obtained.
R = c o s θ s i n θ s i n θ c o s θ , t = t x t y
The ( R , t ) of the two maps are known; then, the two maps are fused as follows:
m a p f u s i o n = m a p 1 + ( m a p 2 × R + t )
(2) The initial positions of the robots are unknown: We used the approach in [31] to estimate the transformation relationship between maps using feature matching when the initial position is unknown. This requires that the maps have enough overlapping regions for reliable matching. The approach detects the features in each map, computes a match for the map pair, estimates the transform using RANSAC, and computes a confidence score for each match. Then, the matches with sufficient confidence are selected for map fusion. In this way, fusion can be performed in the case of overlapping maps, even if the initial positions of the robots are not known.

3. Experiments

In this section, we build the robot platform in Figure 6 and deploy the trained network model to the robots. The performance of the trained model was verified in Figure 7 and Figure 8a. Figure 7 shows the multi-robot SLAM process with and without the compressed communication method, demonstrating that compressed communication has little effect on map quality. Then, multi-robot co-operative SLAM experiments based on the method in this paper are conducted to validate the method further.

3.1. Datasets and Model Training

We used the Gmapping algorithm to build maps of different environments. A total of 2,800 occupancy grid maps were saved and used for model training. Model training based on the Pytorch framework was implemented on a computer with AMD Ryzen 5 2600X Six-Core Processor 3.60 GHz (AMD, Santa Clara, CA, USA), NVIDIA GeForce GTX 1050 Ti (NVIDIA, Santa Clara, CA, USA), and 16 G RAM. It takes about 10 h to train the network proposed in this paper.

3.2. Robotic Platform

As shown in Figure 6, the robots for collaborative SLAM are mainly composed of a Kobuki drive chassis, an NUC (a microcomputer), Rplidar, and a UWB module. The Gmapping algorithm utilizes information from the chassis’ odometer and Rplidar for full-map creation; then, the full maps are processed by the model deployed in the NUC to get the string. Finally, the string is transmitted via the UWB module. Similarly, the robot receives the string through the UWB module, and it is processed by NUC to get the recovery map. The recovered map is fused with the locally created map to obtain the global map.

3.3. Experimental Result

In order to validate the compressed communication frame in Figure 1, we used two robots to collaboratively build maps of a hallway and obtained the map data in Figure 7. In Figure 7, the whole process of the global maps obtained with and without compressed communication for the two robots is shown. The first row of Figure 7 shows the global map obtained without compressed communication; the second row shows the global map obtained with compressed communication; from left to right, the initial state of mapping to the end state of mapping can be seen, and each state of mapping is separated by 30 s.
We evaluated the method of this paper by calculating the metrics for the correspondence maps in Figure 7. The metrics Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index Measure (SSIM) were calculated, and the changes in these two metrics with the running time of multi-robot SLAM are plotted in the upper figure of Figure 8a. PSNR is a metric used to quantify the quality of a map, where higher PSNR values typically indicate less distortion introduced by compression and higher map quality. A PSNR above 40 dB indicates excellent map quality (i.e., very close to the original map). On the other hand, SSIM takes values in the range of [−1, 1], with a value closer to 1 indicating more similarity between two maps and a value closer to −1 suggesting greater dissimilarity. The objective is to measure the structural and content similarity between the recovered and original maps. As global map quality depends not only on pixel-level differences but also on factors such as structure and texture, SSIM provides a more accurate assessment of the recovered map’s quality. As shown in the upper figure of Figure 8a, the mean values of PSNR and SSIM are 50.649 dB and 0.975. Therefore, the difference between maps obtained with and without compressed communication is very small, and the quality of maps obtained with compressed communication can be guaranteed. As shown in the lower figure of Figure 8a, the map compression rate of up to 99 % significantly reduces the communication burden. Based on the data in Figure 8a, we can conclude that the compression method in this paper has a high compression rate while maintaining good recovery quality. The recovered map can be well utilized to construct the global map.
The upper figure of Figure 8b illustrates the time consumed with compressed communication and without compressed communication. The blue and red lines represent the time for map transmission and decompression, respectively. The black line depicts the total time for a map transmission without compressed communication, while the green line shows the time for a map transmission with compression. As shown in the lower figure of Figure 8b, the blue line (12.5 kb/s) is the bandwidth that can be stably achieved by the UWB model in this framework, and the red (124.7 kb/s) and green (Average 720.4 bit/s) lines represent the bandwidth required for transmission of the map without and with compression, respectively. By using compressed communication, the bandwidth required to transmit the maps is greatly reduced. From Figure 8b, we can conclude that the compressed communication in this paper can satisfy the UWB hardware bandwidth and allow for the fast transmission of maps.
In order to verify the effect of the maps obtained from the compressed communication method in robot localization, we used the occupancy grid map obtained with and without compressed communication to provide localization to the robot, respectively. We selected 28 locations in a realistic environment to cover the entire occupancy grid map. Each location was used separately with compressed and uncompressed maps to provide localization to the robot. We used two sets of data and calculated their errors, as in Figure 9. Figure 9a shows the error variation of the robot in the 28 sets of data, and Figure 9b shows the distribution of the error. The mean values of errors e r r o r x , e r r o r y , and e r r o r y a w are 3.59 cm, 3.43 cm, and 0.72°, respectively. This result shows that the compressed communication method in this paper can be applied in a multi-robot SLAM and that the subtle changes produced by compressing the map have little effect on robot localization.

4. Conclusions

The compressed communication method proposed in this paper consists of the following three main components: (1) full-map creation, (2) full-map compression and transmission, and (3) full-map fusion. We designed a lightweight map-feature extraction CNN and a map recovery CNN that can process occupancy grid maps in real time when processing units with limited arithmetic power. Meanwhile, the encoder and decoder are designed by combining the Huffman and RLE algorithms, which greatly reduce the code length and make the bandwidth required for the transmission code fully satisfy the hardware UWB. The compressed communication framework is validated in a multi-robot system, employing two robots for the collaborative mapping of an unknown environment. One robot compresses and transmits its full map, and the other robot receives and recovers it. Subsequently, the recovery map is fused with its map to generate the global map. The experimental results demonstrate that the method achieves a high compression ratio ( 99 % ) and maintains recovered map quality with a PSNR of 50.649 dB and SSIM of 0.975. In summary, the compressed communication method proposed in this paper satisfies the bandwidth-constrained multi-robot SLAM system.

Author Contributions

Conceptualization, L.Z. and J.D.; Data curation, J.D.; Formal analysis, L.Z. and J.D.; Funding acquisition, L.Z.; Investigation, L.Z.; Methodology, L.Z. and J.D.; Project administration, L.Z.; Software, L.Z. and J.D.; Supervision, L.Z.; Validation, J.D.; Visualization, J.D. Writing—original draft, L.Z. and J.D.; Writing—review & editing, L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under grant No. 62303002 and the Key Program of Natural Science Foundation of Anhui Higher Education Institutions of China under grant No. 2022AH050068.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SLAMSimultaneous Localization and Mapping
CNNConvolutional Neural Network
RLERun-Length Encoding
UWBUltra-Wide Band
PSNRPeak Signal-to-Noise Ratio
SSIMStructural Similarity Index Measure

References

  1. Durrant-Whyte, H.; Bailey, T. Simultaneous localization and mapping: Part I. IEEE Robot. Autom. Mag. 2006, 13, 99–110. [Google Scholar] [CrossRef]
  2. Cadena, C.; Carlone, L.; Carrillo, H.; Latif, Y.; Scaramuzza, D.; Neira, J.; Reid, I.; Leonard, J.J. Past, Present, and Future of Simultaneous Localization and Mapping: Toward the Robust-Perception Age. IEEE Trans. Robot. 2016, 32, 1309–1332. [Google Scholar] [CrossRef]
  3. Tian, Y.; Liu, K.; Ok, K.; Tran, L.; Allen, D.; Roy, N.; How, J.P. Search and rescue under the forest canopy using multiple UAVs. Int. J. Robot. Res. 2020, 39, 1201–1221. [Google Scholar] [CrossRef]
  4. Bonin-Font, F.; Burguera, A. Towards multi-robot visual graph-SLAM for autonomous marine vehicles. J. Mar. Sci. Eng. 2020, 8, 437. [Google Scholar] [CrossRef]
  5. Zhang, L.; Zhang, Z.; Siegwart, R.; Chung, J.J. Probabilistic network topology prediction for active planning: An adaptive algorithm and application. IEEE Trans. Robot. 2022, 39, 147–164. [Google Scholar] [CrossRef]
  6. Schmuck, P.; Chli, M. CCM-SLAM: Robust and efficient centralized collaborative monocular simultaneous localization and mapping for robotic teams. J. Field Robot. 2019, 36, 763–781. [Google Scholar] [CrossRef]
  7. Schmuck, P.; Ziegler, T.; Karrer, M.; Perraudin, J.; Chli, M. Covins: Visual-inertial slam for centralized collaboration. In Proceedings of the 2021 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), Bari, Italy, 4–8 October 2021; pp. 171–176. [Google Scholar]
  8. Patel, M.; Karrer, M.; Bänninger, P.; Chli, M. COVINS-G: A Generic Back-end for Collaborative Visual-Inertial SLAM. arXiv 2023, arXiv:2301.07147. [Google Scholar]
  9. Zhang, T.; Zhang, L.; Chen, Y.; Zhou, Y. CVIDS: A Collaborative Localization and Dense Mapping Framework for Multi-Agent Based Visual-Inertial SLAM. IEEE Trans. Image Process. 2022, 31, 6562–6576. [Google Scholar] [CrossRef] [PubMed]
  10. Chang, Y.; Ebadi, K.; Denniston, C.E.; Ginting, M.F.; Rosinol, A.; Reinke, A.; Palieri, M.; Shi, J.; Chatterjee, A.; Morrell, B.; et al. LAMP 2.0: A robust multi-robot SLAM system for operation in challenging large-scale underground environments. IEEE Robot. Autom. Lett. 2022, 7, 9175–9182. [Google Scholar] [CrossRef]
  11. Ebadi, K.; Chang, Y.; Palieri, M.; Stephens, A.; Hatteland, A.; Heiden, E.; Thakur, A.; Funabiki, N.; Morrell, B.; Wood, S.; et al. LAMP: Large-scale autonomous mapping and positioning for exploration of perceptually-degraded subterranean environments. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 80–86. [Google Scholar]
  12. Tian, Y.; Chang, Y.; Arias, F.H.; Nieto-Granda, C.; How, J.P.; Carlone, L. Kimera-multi: Robust, distributed, dense metric-semantic slam for multi-robot systems. IEEE Trans. Robot. 2022, 38, 2022–2038. [Google Scholar] [CrossRef]
  13. Tian, Y.; Chang, Y.; Quang, L.; Schang, A.; Nieto-Granda, C.; How, J.P.; Carlone, L. Resilient and distributed multi-robot visual slam: Datasets, experiments, and lessons learned. arXiv 2023, arXiv:2304.04362. [Google Scholar]
  14. Chang, Y.; Tian, Y.; How, J.P.; Carlone, L. Kimera-multi: A system for distributed multi-robot metric-semantic simultaneous localization and mapping. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 11210–11218. [Google Scholar]
  15. Lajoie, P.Y.; Ramtoula, B.; Chang, Y.; Carlone, L.; Beltrame, G. DOOR-SLAM: Distributed, online, and outlier resilient SLAM for robotic teams. IEEE Robot. Autom. Lett. 2020, 5, 1656–1663. [Google Scholar] [CrossRef]
  16. Huang, Y.; Shan, T.; Chen, F.; Englot, B. DiSCo-SLAM: Distributed scan context-enabled multi-robot lidar slam with two-stage global-local graph optimization. IEEE Robot. Autom. Lett. 2021, 7, 1150–1157. [Google Scholar] [CrossRef]
  17. Zhong, S.; Qi, Y.; Chen, Z.; Wu, J.; Chen, H.; Liu, M. Dcl-slam: A distributed collaborative lidar slam framework for a robotic swarm. arXiv 2022, arXiv:2210.11978. [Google Scholar] [CrossRef]
  18. Caccavale, A.; Schwager, M. Wireframe mapping for resource-constrained robots. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1–9. [Google Scholar]
  19. Saeedi, S.; Paull, L.; Trentini, M.; Li, H. Multiple robot simultaneous localization and mapping. In Proceedings of the 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems, San Francico, CA, USA, 25–30 September 2011; pp. 853–858. [Google Scholar]
  20. Benedettelli, D.; Garulli, A.; Giannitrapani, A. Multi-robot SLAM using M-Space feature representation. In Proceedings of the 49th IEEE Conference on Decision and Control (CDC), Atlanta, GA, USA, 15–17 December 2010; pp. 3826–3831. [Google Scholar]
  21. Chang, H.J.; Lee, C.G.; Hu, Y.C.; Lu, Y.H. Multi-robot SLAM with topological/metric maps. In Proceedings of the 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems, San Diego, CA, USA, 29 October–2 November 2007; pp. 1467–1472. [Google Scholar]
  22. Saeedi, S.; Paull, L.; Trentini, M.; Seto, M.; Li, H. Group mapping: A topological approach to map merging for multiple robots. IEEE Robot. Autom. Mag. 2014, 21, 60–72. [Google Scholar] [CrossRef]
  23. Choudhary, S.; Carlone, L.; Nieto, C.; Rogers, J.; Liu, Z.; Christensen, H.I.; Dellaert, F. Multi robot object-based slam. In Proceedings of the 2016 International Symposium on Experimental Robotics, Nagasaki, Japan, 3–8 October 2016; pp. 729–741. [Google Scholar]
  24. Zhang, H.; Chen, X.; Lu, H.; Xiao, J. Distributed and collaborative monocular simultaneous localization and mapping for multi-robot systems in large-scale environments. Int. J. Adv. Robot. Syst. 2018, 15, 1729881418780178. [Google Scholar] [CrossRef]
  25. Tardioli, D.; Montijano, E.; Mosteo, A.R. Visual data association in narrow-bandwidth networks. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 2572–2577. [Google Scholar]
  26. Dubé, R.; Gawel, A.; Sommer, H.; Nieto, J.; Siegwart, R.; Cadena, C. An online multi-robot SLAM system for 3D LiDARs. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 1004–1011. [Google Scholar] [CrossRef]
  27. Karrer, M.; Schmuck, P.; Chli, M. CVI-SLAM—Collaborative Visual-Inertial SLAM. IEEE Robot. Autom. Lett. 2018, 3, 2762–2769. [Google Scholar] [CrossRef]
  28. Xu, H.; Liu, P.; Chen, X.; Shen, S. D2 SLAM: Decentralized and Distributed Collaborative Visual-inertial SLAM System for Aerial Swarm. arXiv 2022, arXiv:2211.01538. [Google Scholar]
  29. Ramtoula, B.; De Azambuja, R.; Beltrame, G. CAPRICORN: Communication aware place recognition using interpretable constellations of objects in robot networks. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 8761–8768. [Google Scholar]
  30. Grisetti, G.; Stachniss, C.; Burgard, W. Improved techniques for grid mapping with rao-blackwellized particle filters. IEEE Trans. Robot. 2007, 23, 34–46. [Google Scholar] [CrossRef]
  31. Hörner, J. Map-Merging for Multi-Robot System; Charles University: Prague, Czech Republic, 2016. [Google Scholar]
Figure 1. The framework for compressed communication application in multi-robot SLAM.
Figure 1. The framework for compressed communication application in multi-robot SLAM.
Sensors 24 03154 g001
Figure 2. Map feature CNN.
Figure 2. Map feature CNN.
Sensors 24 03154 g002
Figure 3. Encoding and decoding process.
Figure 3. Encoding and decoding process.
Sensors 24 03154 g003
Figure 4. The Huffman coding tree.
Figure 4. The Huffman coding tree.
Sensors 24 03154 g004
Figure 5. Map recovery CNN.
Figure 5. Map recovery CNN.
Sensors 24 03154 g005
Figure 6. Robot hardware platform.
Figure 6. Robot hardware platform.
Sensors 24 03154 g006
Figure 7. Compressed communication in multi-robot SLAM.
Figure 7. Compressed communication in multi-robot SLAM.
Sensors 24 03154 g007
Figure 8. Performance of the proposed methodologies; (a) chage in map quality (upper) with change in map compression rate (lower) during the operation of multi-robot SLAM based on compressed communication; (b) comparison of map transfer time required with and without compressed communication (upper) versus bandwidth required with and without compressed communication (lower). T and T C denote the time to transmit the map using UWB with compressed communication and without compressed communication, respectively, where T C is the sum of T C t r a n s (time to transmit the compressed map) and T C d e (time to decompress the compressed map).
Figure 8. Performance of the proposed methodologies; (a) chage in map quality (upper) with change in map compression rate (lower) during the operation of multi-robot SLAM based on compressed communication; (b) comparison of map transfer time required with and without compressed communication (upper) versus bandwidth required with and without compressed communication (lower). T and T C denote the time to transmit the map using UWB with compressed communication and without compressed communication, respectively, where T C is the sum of T C t r a n s (time to transmit the compressed map) and T C d e (time to decompress the compressed map).
Sensors 24 03154 g008
Figure 9. Robot’s localization error using the merged map with compressed communication; (a) Performance of localization errors regarding the robot’s position ( x , y ) and its yaw angle θ ; (b) Distribution of the localization error over time.
Figure 9. Robot’s localization error using the merged map with compressed communication; (a) Performance of localization errors regarding the robot’s position ( x , y ) and its yaw angle θ ; (b) Distribution of the localization error over time.
Sensors 24 03154 g009
Table 1. Comparison of the types of data transferred with the related methods.
Table 1. Comparison of the types of data transferred with the related methods.
MethodType of Data TransferMap Density
Dubé et al. [26]Segmentation descriptionSparse
CVIDS [9]Descriptor extractionSparse
CVI-SLAM [27]Keyframe and map pointSparse
D 2 SLAM [28]Landmarks and descriptorsSparse
Zhang et al. [24]Descriptor extractionSparse
DCL-SLAM [17]DescriptorSparse
Door-SLAM [15]NetVlAD descriptorSparse
Kimera-Multi [12]Keypoints and feature descriptorsDense
Our methoed2D occupancy full mapDense
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, L.; Deng, J. Deep Compressed Communication and Application in Multi-Robot 2D-Lidar SLAM: An Intelligent Huffman Algorithm. Sensors 2024, 24, 3154. https://doi.org/10.3390/s24103154

AMA Style

Zhang L, Deng J. Deep Compressed Communication and Application in Multi-Robot 2D-Lidar SLAM: An Intelligent Huffman Algorithm. Sensors. 2024; 24(10):3154. https://doi.org/10.3390/s24103154

Chicago/Turabian Style

Zhang, Liang, and Jinghui Deng. 2024. "Deep Compressed Communication and Application in Multi-Robot 2D-Lidar SLAM: An Intelligent Huffman Algorithm" Sensors 24, no. 10: 3154. https://doi.org/10.3390/s24103154

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop