Next Article in Journal
Collaborative Task Allocation and Optimization Solution for Unmanned Aerial Vehicles in Search and Rescue
Previous Article in Journal
Online Predictive Visual Servo Control for Constrained Target Tracking of Fixed-Wing Unmanned Aerial Vehicles
Previous Article in Special Issue
IMUC: Edge–End–Cloud Integrated Multi-Unmanned System Payload Management and Computing Platform
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Easy Rocap: A Low-Cost and Easy-to-Use Motion Capture System for Drones

1
State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan 430072, China
2
Engineering Research Centre for Spatio-Temporal Data Acquisition and Smart Application (STSA), Ministry of Education in China, Wuhan 430072, China
3
Institute of Artificial Intelligence in Geomatics, Wuhan University, Wuhan 430072, China
4
Institute of Artificial Intelligence, School of Computer Science, Wuhan University, Wuhan 430072, China
*
Author to whom correspondence should be addressed.
Drones 2024, 8(4), 137; https://doi.org/10.3390/drones8040137
Submission received: 4 February 2024 / Revised: 27 March 2024 / Accepted: 28 March 2024 / Published: 2 April 2024
(This article belongs to the Special Issue Resilient UAV Autonomy and Remote Sensing)

Abstract

:
Fast and accurate pose estimation is essential for the local motion control of robots such as drones. At present, camera-based motion capture (Mocap) systems are mostly used by robots. However, this kind of Mocap system is easily affected by light noise and camera occlusion, and the cost of common commercial Mocap systems is high. To address these challenges, we propose Easy Rocap, a low-cost, open-source robot motion capture system, which can quickly and robustly capture the accurate position and orientation of the robot. Firstly, based on training a real-time object detector, an object-filtering algorithm using class and confidence is designed to eliminate false detections. Secondly, multiple-object tracking (MOT) is applied to maintain the continuity of the trajectories, and the epipolar constraint is applied to multi-view correspondences. Finally, the calibrated multi-view cameras are used to calculate the 3D coordinates of the markers and effectively estimate the 3D pose of the target robot. Our system takes in real-time multi-camera data streams, making it easy to integrate into the robot system. In the simulation scenario experiment, the average position estimation error of the method is less than 0.008 m, and the average orientation error is less than 0.65 degrees. In the real scenario experiment, we compared the localization results of our method with the advanced LiDAR-Inertial Simultaneous Localization and Mapping (SLAM) algorithm. According to the experimental results, SLAM generates drifts during turns, while our method can overcome the drifts and accumulated errors of SLAM, making the trajectory more stable and accurate. In addition, the pose estimation speed of our system can reach 30 Hz.

1. Introduction

The rapid technological advances in robotic systems have inspired the development of unmanned platforms, including unmanned ground vehicles (UGVs), unmanned aerial vehicles (UAVs), and quadruped robots [1,2]. Pose estimation plays an important role in robotic applications. Without accurate estimation of position and velocity, a robot is unable to build the correct spatial representation of the environment, which leads to the impossibility of trajectory planning and execution [3]. In open outdoor environments, GPS and inertial measurement unit (IMU)-integrated navigation technology can provide pose information effectively [4,5]. However, in indoor environments, GPS signals attenuate severely, leading to inaccuracies in the provided location information [6,7]. Therefore, real-time pose information remains particularly crucial for most mobile robots.
There is a large amount of research exploring indoor positioning technology for robots, including visible light positioning (VLP), computer vision [8], ultrasonic [9], Bluetooth [10], and so on. Among them, both VLP and computer vision approaches reach high accuracy at a low cost [7]. Previous research has achieved speed improvements in VLP. However, this approach requires the unmanned device to carry heavy industrial cameras and high-performance processors [11], which is a challenge for payload-limited drones and other small robots [12,13]. Some computer-vision-based approaches utilize depth cameras to estimate the position changes of drones [14], but they are unsuitable for small-scale experiments with high precision requirements because of the limited positioning range and the significant dependency on the target detection result of drones [15]. Recently, with the rapid technological advances in computer hardware, the cost of visual positioning has been significantly reduced, and the rapid progress in visual algorithms has made real-time positioning possible. However, these positioning methods need to integrate other data like IMU data to obtain the position and orientation, which makes the approaches complex.
Simultaneous localization and mapping (SLAM) provides a framework for positioning and mapping, and a variety of technical methods have emerged in recent years [16,17]. Xu et al. [18] proposed a method of tightly coupling multiple LiDARs with IMU data for odometry, significantly improving the accuracy of pose estimation. Sun et al. [19] achieved self-pose estimation in weak GNSS environments by integrating GNSS, RGB camera, and IMU data on the drone. However, there are several challenges for SLAM on lightweight UAV platforms: due to the limited load, it is difficult to have sensors and computers of the same quality as those on ground robots [20]; UAVs have six degrees of freedom (DOF) poses, and cannot make some simplified assumptions like wheeled robots. At the same time, autonomous UAVs require high-frequency and real-time pose information to ensure stable control of the platform. Given these challenges, as well as the need for true pose data for indoor SLAM accuracy evaluation, it is essential to study a low-cost and high-precision method for indoor UAV pose estimation.
Motion capture (Mocap) systems based on cameras have better performance in robotic applications [21,22]. Most of these systems are commercial Mocap systems, such as Vicon, OptiTrack, and so on. However, commercial Mocap systems in the market often have a high cost and lack open-source software and hardware, making it difficult to deploy them on robots. The open-source Mocap toolbox Easy Mocap (https://github.com/zju3dv/EasyMocap (accessed on 29 March 2021)) provides the technology for capturing human body motions using multiple RGB cameras. However, mobile unmanned platforms such as UGV and UAV do not have structural key points with fixed connections like human joints. Therefore, markerless Mocap methods based on deep learning are inefficient, and it is difficult to directly apply Easy Mocap to robots. Utilizing infrared (IR) camera-based Mocap systems enables the position triangulation of markers on the target robot, which can be used for pose estimation of rigid bodies [23]. However, unlike the typical Mocap experimental environment with a single background and controlled lighting, the actual working environment of unmanned platforms is complex. IR Mocap systems are susceptible to environmental light noise, and camera occlusion is the most challenging problem.
In response to the above challenges of ambient light noise and the lack of access to the source code of high-precision Mocap systems, we propose a simple, yet open-source motion capture system named Easy Rocap, which uses markers of special material as the tracking objects of the mobile robot and uses multiple fixed cameras to perform 3D intersection to accurately estimate the position and orientation of drones and other robots in real time. Diverging from Mocap systems that rely on infrared cameras, our system takes indoor potential noise and obstacles into consideration, employing object filtering and multi-object tracking (MOT) algorithms that fuse object detection technology to ensure trajectories remain continuous and precise. Our main contributions are as follows.
(1)
A Multi-view Correspondence method that combines Object Detection and MOT is proposed. The dual-layer detector achieves robust object detection in complex environments, and the MOT method can stably track markers under short-term occlusion. Multi-view correspondences use geometric constraints to correctly match corresponding image points.
(2)
A high-precision 3D robotic pose estimation system for complex dynamic scenes is open-sourced. When there is no obstruction in the simulation environment, the average positioning accuracy reaches 0.008 m, the average orientation accuracy reaches 0.65 degrees, and the solving speed reaches 30 frames per second (FPS).
The motion capture system proposed in this paper offers real-time pose information for robots in the experimental area. To assess its performance, extensive experiments were conducted using multiple cameras, including simulated environments and real scenes. The structure of the paper is outlined as follows. Section 2 presents the related works. Section 3 describes the system comprehensively, detailing keypoint detection, multi-view correspondence, and multi-view camera triangulation, respectively. Section 4 presents the experimental settings, results, and discussion. Section 5 provides a summary of the approach and offers a prospect for future work.

2. Related Works

This section provides an in-depth analysis of different motion capture techniques, comparing their distinct features. Additionally, algorithms for visual MOT and relative positioning are categorized and reviewed.

2.1. Motion Capture Methods

Motion capture (Mocap) serves as a digital method to track and record the spatial movements of targets [23], often employed in capturing and reconstructing the human body posture. The essential of Mocap is acquiring the 3D coordinates of each key point [24,25]. Mocap technology is widely applied in human activities [26,27] such as sports and entertainment, as well as robotic exploration fields for trajectory tracking and controlling [28].
In recent years, several types of Mocap systems have emerged. Depth cameras, for instance, can calculate depth by measuring the time delay between emitted light and the detection of its backscatter [29]. There are other systems such as inertial Mocap [30], mechanical Mocap, acoustic Mocap, and electromagnetic Mocap systems [24,25], but they are not as widespread as traditional Mocap systems, mainly due to deficiencies in accuracy or convenience.
Predominantly, traditional mocap technologies rely on optical approaches. Optical Mocap systems use high-speed cameras to capture reflective markers on the target object and triangulate the positions of these markers to reconstruct the object’s spatial posture with high precision. In the field of motion capture, optical Mocap systems are often referred to as the benchmark for accuracy [24,25] and perform better in robotics applications. In scenarios such as human–robot interaction and local multi-robot collaboration [31,32], visual Mocap systems can address the common issue of drifts in inertial sensors. However, camera-based Mocap systems face challenges such as occlusions in certain camera views [23]. To overcome these problems, some researchers have designed hybrid systems that combine different Mocap technologies [33,34] to improve accuracy and reduce camera occlusions. As previously mentioned, autonomous control of drones requires high-frequency position and orientation information, so the structure should be as simple and efficient as possible.
Currently, marker-based optical motion capture systems are pretty mature in the market, with Vicon being a representative Mocap system [35]. However, it is difficult to apply and popularize it in daily use since its hardware cost is high and the technology is not open source. Recently, the open-source release of the EasyMocap toolkit has made low-cost human motion capture possible. At the same time, advancements in the precision and efficiency of deep learning algorithms have demonstrated their ability to improve the capabilities of Mocap systems, particularly in areas like object detection [36].

2.2. Multi-Object Tracking

Multiple Object Tracking (MOT) aims to identify and track multiple targets simultaneously over a period of time in processing sequential data [37]. Recently, a large number of MOT approaches have emerged, among which deep learning approaches are increasingly prevalent [38].
Predominantly, the methods employed in MOT comply with a principle known as the “tracking by detection” paradigm [36,39,40]. These approaches typically execute object detection on every single frame of a video sequence and then convert the task into object association between adjacent frames. Simple Online and Real-time Tracking (SORT) [39] utilized the Kalman filter [41] and Hungarian Algorithm [42] for object association. Because SORT disregards the appearance characteristics of detected objects, it performs well only when the certainty of object state estimation is high. Wojke et al. [36] introduced cosine similarity as appearance similarity for tracking association. As a result, this method can track through longer periods of occlusion but struggles with challenges like motion blur. MOT approaches based on joint detection and embedding [43,44,45] allow sharing detection and embedding information in the model, and cross-scale association is feasible. Previous studies [46,47] represent pioneering efforts in applying Transformer to MOT, achieving notable improvements in tracking accuracy. However, they encounter significant challenges in addressing computational bottlenecks during deployment. Meanwhile, Chu et al. [48] introduced a cascade association framework to handle challenging tracking scenarios, thereby improving calculation efficiency. Zeng et al. [49] developed Transformer-based frameworks for MOT, enabling end-to-end tracking, but its performance in detecting newborn objects is limited.
It is appropriate to select the corresponding approach according to the specific task requirements since these MOT approaches have different advantages. In our work, we have access solely to the preceding frame and the frame currently in view since pose estimation requires a real-time online tracker. Therefore, we use the mature and representative detection-based MOT approach OC-SORT [50], a tracking methodology that demonstrated top-tier effectiveness when tested on the KITTI dataset [51].

2.3. Direct Pose Estimation

We classify pose estimation into direct and indirect methods. Direct pose estimation directly observes the target to determine its pose, relying solely on raw sensor data without external infrastructure (e.g., GPS, UWB with anchors) or environmental feature matching. Although direct pose estimation techniques typically require customized hardware, their autonomous operation, reliability, and precision continue to draw substantial interest. Cutler et al. [52] proposed a straightforward yet efficient method by designing positioning markers composed of infrared LEDs, aiming to estimate the position and velocity of known markers. Faessler et al. [53] proposed an infrared-LED-based pose estimation system that utilizes the Perspective-n-Point (PnP) algorithm to compute the relative position and orientation between the UAV and UGV, enhancing the collaborative working efficiency between aerial and ground robots. However, to maintain the distinctness of the LED spots, these methods often only work at a short distance.
Ultra-Wideband (UWB) technology presents a promising solution. Fishberg and How [54] developed a 3-DOF relative 2D pose estimation system among robots, achieving improvements in mean position error solely based on UWB measurements. Cossette et al. [55] introduced a method for estimating relative orientations using UWB distance measurements, demonstrating that estimation accuracy can be increased by optimizing the formation geometry of robot formation. To reduce the impact of robot formation geometry and UWB measurement errors on pose estimation, Jones et al. [56] modified the state vector using a kinematic bicycle model, thereby facilitating the direct estimation of the lead vehicle’s longitudinal speed and effective steering angle. Jin and Jiang [57] proposed a multi-vehicle mapping system that combines LiDAR, IMU, and UWB technologies to improve the robustness of positioning and mapping in degraded environments through multi-metric weights LiDAR-inertial odometry and pose estimation correction from the degeneration direction. Hao et al. [58] proposed the KD-EKF algorithm, which enhanced the accuracy and consistency of cooperative localization in multi-robot systems by addressing the observability issues in the standard EKF approach.
In addition to integrating UWB technology, Pritzl et al. [59] introduced a cooperative guidance method for UAVs in GNSS-denied environments, leveraging the fusion of LiDAR and Visual-Inertial Odometry (VIO) data for precise relative localization and trajectory tracking. In a diverse team with a major UAV carrying LiDAR and a minor UAV equipped with a camera, the method combines LiDAR relative positioning data with the VIO output on the major UAV to accurately determine the minor UAV’s pose.

3. Methods

To demonstrate our Easy RoCap system, the overall motion capture workflow is shown in Figure 1. The system is built using consumer-grade commercial IR RealSense (https://www.intelrealsense.com/depth-camera-d455/ (accessed on 17 June 2020)) cameras and consists of three major components: two-stage key points detection model (TKDM), Multi-view correspondences, and Multi-view Camera Triangulation.
Firstly, a trained object detector is employed to generate bounding boxes of robots and markers in each view. Considering that there is noise in 2D detection, a filtering algorithm based on confidence and class is proposed to eliminate inaccurate detections. The multi-view correspondences module employs an MOT algorithm to address occlusion issues and assigns matched marker IDs. Utilizing the calibrated multi-view camera parameters, the 3D coordinates of markers are calculated, and the position and orientation of the robot are ultimately estimated by the coordinates of multiple markers.

3.1. Two-Stage Keypoints Detection Model (TKDM)

Accurate single-camera-based marker detection is essential for achieving precise multi-view localization. To minimize the ambient light interference while meeting real-time pose requirements for robots, we employ the YOLOX detector [60] known for its rapid detection speed to implement efficient positioning of markers. The decoupled head strategy [61] effectively resolves the common conflict between classification and regression tasks by dedicating separate branches within the detector architecture to each specific task. By utilizing the decoupled head strategy, the detection speed of YOLOX is significantly enhanced without compromising accuracy. Furthermore, the multi-scale feature fusion technique in YOLOX is able to effectively improve detection performance in the presence of small objects and minor occlusions. Trained on a specific dataset, the detector produces detection boxes of different categories on image data.
Addressing the demand for simultaneously rapid inference across data from multiple views, the batch size is adjusted to satisfy that the quantity of images for inference aligns with the number of cameras, thereby accelerating image inference speed while aligning timestamps across multiple views.
Algorithm 1 Filtering of Object Detection Results
Input: frame f k ; object detector D e t ; detection score threshold τ
Output: Filtered marker detection boxes
1. D r e m a i n
2. D k D e t ( f k )
3. marker class boxes D m , robot class boxes D r
4. for d in D m do
5.  Check if d . s c o r e > τ and d is contained by a certain a robot box
6.  if Failed then
7.    delete d
8.  end
9.   D r e m a i n D r e m a i n d
10. return: D r e m a i n
Although it is already the most advanced detector in performance, there are still redundant or false detections in the real-world scene. To reduce the interference, detection results are filtered based on a strategy of taking the detection boxes of the robot as a constraint and filtering false detections. Considering that the markers are fixed on the robot target, the specific implementation is to use confidence to filter the object detection results and then take the robot detection boxes as a class label constraint to filter out false detections caused by environmental noise, enhancing the robustness of the detector. The specific algorithm is described in Algorithm 1, which is used to filter out invalid marker detection boxes.
d e t f i l t e r e d . s c o r e = d e t . s c o r e ,   if   d e t . s c o r e > thre   &   isincluded ( d e t ) 0   , else  
where d e t represents the marker detection box; d e t . s c o r e is the score of the corresponding detection box; d e t f i l t e r e d . s c o r e represents the score of the filtered detection box. For each marker detection box d e t , the confidence is first checked to see if it exceeds the threshold value t h r e . The function isincluded ( d e t ) indicates whether the detection box is contained within the robot detection box, which is used to filter out false detections.

3.2. Multi-View Correspondences

Before computing the 3D pose of a robot, 2D object bounding boxes across multiple views should be appropriately matched, i.e., the bounding boxes of the same marker across all views are needed. Due to occlusion and motion blur, there might be missing detection, which is challenging. To address this, a tracking algorithm assigns IDs to objects and saves status information, preserving the trajectory continuously. Matching of IDs across multiple views is then facilitated based on this.
The task of MOT is to estimate the trajectories of multiple objects across successive video frames. In real-time online MOT, only the detection results from the latest few frames are available, making tracking by detection the leading paradigm in MOT. An increasing number of MOT methods are utilizing more powerful detectors to achieve better tracking performance, with the YOLO series being widely applied due to its balance of speed and accuracy [40]. This paper builds upon the YOLO series detectors, with the tracking algorithm improved based on OC-SORT. Its characteristics include a fast tracking speed and robust handling of occlusions and non-linear motion [50]. Due to overlapping targets and non-linear motion, the Kalman filter fails to re-match with the current detection results, resulting in frequent track interruptions. OC-SORT proposes Observation-Centric Recovery (OCR) to recover the track based on the detection results rather than incorrect estimations. As shown in Figure 2, when a target with a lost track is redetected, the track is restored by associating the observations at the time of loss with the current observations. Experimental testing in this paper demonstrates that it is simple and online. In addition, to prevent errors in subsequent 3D calculations caused by incorrect detections, we establish additional conditions for tracker trajectory establishment.
As shown in Algorithm 2, for the matching of object detections and trajectories, the number of trajectories is constrained to be consistent with the number of IDs and markers. For example, if n target markers are used, the IDs can only be 1, 2, n …,. Additionally, when there are gaps in trajectory IDs, a new trajectory is established for targets that are successfully detected in three successive frames. Markers that are not updated temporarily do not participate in subsequent trajectory calculations until the next update.
Algorithm 2 Trajectory Establishment
Input: Track T ; Filtered detection boxes D r e m a i n
1. Associate T and D r e m a i n using Similarity
2. D r e r e m a i n remaining object boxes from D r e m a i n
3. T r e m a i n remaining tracks from T
4. for d e t in D r e r e m a i n  do
5.  Check if there is an available IDtraj  nummarket
6.  if True then
7.    T r e m a i n T r e m a i n d e t
8.  end
9.  else
10.   delete d e t
11.  end
12. return: T r e m a i n
An important cue to associate multiple objects across multiple views is that their associated joints should satisfy geometric constraints. Suppose p R N × 2 denotes the 2D distribution of N object points, the geometric constraints between view p i and another view p j can be quantified using the following function.
E g p i , p j = 1 2 N n = 1 N d g p i n , L i j p j n + d g p j n , L j i p i n
where p i n denotes the position of the n -th marker in view i and L i j p j n denotes the epipolar line (a fundamental concept within epipolar geometry, see [61] for definition) associated with p j n from the other view. d g a , l signifies the distance from the point a to the straight line l . The scale of the error function E g is used to measure the geometric similarity scores corresponding to multiple views.
A m a t c h = arg min A 0 i < j m E g p i , p j
Suppose there are m cameras in the scene; A records the matching relationship of multi-view marker IDs. A m a t c h is obtained by minimizing the sum of functions E g , so as to calculate the 3D coordinates of different markers directly.

3.3. Multi-View Camera Triangulation

After the cameras are correctly placed in the 3D space, the Easy Rocap system is then calibrated. Based on the detection and tracking results mentioned previously, this section mainly describes how to calibrate multi-view cameras and perform triangulation on the objects.
To reconstruct 3D positions using multi-view cameras, it is essential to rectify distortions produced by the camera lens and establish a geometric model for multiple cameras to denote their relative positions. The primary goal of calibration is to obtain both the intrinsic and extrinsic parameters of the cameras. The methodology adopted in this experiment is based on Zhang’s camera calibration method [62]. The camera calibration yields intrinsic and extrinsic parameters, allowing us to establish a camera model. This means we can find the relationship between the spatial points’ 3D coordinates and the camera image coordinates.
A X = 0
Here, X is the augmented 3D coordinates of a spatial point and A represents the matrix combination of image coordinates and camera parameters [63].
While conventional triangulation methods utilize linear algebra [63] to estimate 3D coordinates from 2D coordinates across multiple views, challenges arise when certain views do not provide reliable 2D estimations due to occlusions or out of frame. To address this issue, this study draws inspiration from methods employed in multi-camera multi-person motion capture systems [64]. The 2D positions together with the confidences of markers obtained by multiple cameras are passed to the algebraic triangulation module, which solves the triangulation problem in the form of a weighted linear equation:
w j A j X j = 0
where w j denotes the confidence vector of the j -th target marker across multiple views; A j is a matrix combination of the image coordinates and camera parameters for the j -th object point across different views; and X j is the 3D position of the j -th target marker.
By employing the Singular Value Decomposition (SVD) method to solve the system of triangulation linear equations, optimal 3D coordinates for the markers are obtained. The rigid body formed by the markers fixed on the robot aligns with the robot’s position and orientation. Through the calculation of the centroid translation and rotation matrix of the 3D point coordinates, the robot’s pose variations can be accurately represented.

4. Experiments

In this section, we showcase several experiments to illustrate the precision and rapidity of our Easy Rocap system. We designed a series of experiments using UAVs and UGVs and employed the evo toolbox to assess the pose accuracy. The experimental procedure is separated into two segments: (1) UAV trajectory accuracy verification in the simulation environment and (2) UGV and UAV trajectory accuracy verification in real scenarios.

4.1. Experimental Settings

4.1.1. Dataset Description

Due to the limited availability of datasets for evaluating robot poses, the experimental data were collected from carefully crafted simulation scenarios as well as real-world situations. Figure 3 shows the experimental sites with optical markers arranged on the surface of the unmanned drone/vehicle. The simulation scene dataset includes fixed multi-camera data with timestamps and ground truth trajectory data. The real-scene dataset includes fixed camera data with timestamps, as well as LiDAR data, which are used for SLAM methods to obtain trajectories for comparison.
We collected a total of 1200 sample images in the simulation and real scenarios as the training dataset, including drones, UGVs, and markers. On the basis of recording the calibration board data to obtain camera intrinsic and extrinsic parameters, we recorded data from different scenes and motion states, called Seq 01–06. Seq 01–03 data are recorded in the simulation scenario, where drones flew with and without obstacles, performing motions such as spirals. The average velocity is 0.4 m/s and the total length of the trajectory is 55.18 m. Seq04–05 data are recorded in the real scenario, where UGV performs curved and square movements. The average velocity is 0.26 m/s and the total length of the trajectory is 12.5 m. Seq06 data are recorded in the real scenario where the drone is performing curved movements. Table 1 shows the scene type, motion pattern, data length, data frequency, and average speed of the data.

4.1.2. Experimental Scenes

To validate the feasibility of our robotic motion capture system, a simulation experiment was first designed. As shown in Figure 4, the simulation experiment was conducted using the robot simulation software Gazebo, which provides high-fidelity physical simulation of the drone and the markers. A simulation experimental site of 10 m × 10 m was set up. Utilizing the Pinhole Camera Model, we customized camera parameters including the field of view and resolution. The cameras were positioned at a height of 3 m, as illustrated in Figure 4, with each camera tilted at an angle of 20°, facing towards the center of the venue. The drone was manually controlled for flight, and the software was able to record its ground truth trajectory.
In an effort to verify the stability of our Easy Rocap system, a respective experiment was organized in real scenarios (Figure 5) in comparison with the LiDAR-Inertial SLAM algorithm. The camera system consists of ix infrared lenses from Intel RealSense D455. The experimental site consisted of UGV and UAV platforms with markers, both equipped with a small-sized Livox Mid-360 LiDAR sensor.
For multiple object tracking of markers, the detector is YOLOX [60], and YOLOX-s was used as the backbone. The pretrained model was downloaded, and the categories were modified to “UGV”, “UAV”, and “marker”. The training schedule is 1000, the number of training data is approximately 1000, and the input image size is 640 × 480. The process of data augmentation encompasses techniques such as Mosaic [65] and Mixup [66]. These strategies serve the purpose of improving the model’s generalization ability. Regarding the technical setup of the model’s training phase, this process was conducted on an NVIDIA GeForce RTX 2070 graphics processing unit (GPU), with a batch size of 4 and a total training time of approximately 4 h.

4.1.3. Evaluation Metrics

In assessing the precision of our Easy Rocap system, we use an error evaluation method (https://github.com/MichaelGrupp/evo (accessed on 14 September 2017)). The reference in the simulation scenario is the ground truth trajectory recorded in Gazebo, and the reference in the real scenario is the pose estimation result of Easy Rocap. For the evaluation of trajectory accuracy, the global consistency of the trajectories was assessed utilizing absolute pose error (APE) as a metric. As illustrated in Equation (6), the calculation of APE was made based on the estimated trajectory posture P e s t , i S E ( 3 ) at timestamp i , and the ground truth trajectory P r e f , i S E ( 3 ) at timestamp i .
E i = P e s t , i Θ P r e f , i = P r e f , i 1 P e s t , i S E ( 3 )
where the operator Θ , deriving the relative pose when given two separate poses, signifies the inverse compositional process.
In order to show the accuracy of our system in translation and rotation, we used the corresponding two indicators, A P E t r a n s and A P E r o t , respectively. The translation accuracy indicator and rotation accuracy indicator are calculated by Equations (7) and (8), respectively. On this basis, we used the APE of each timestamp, i.e., A P E i to calculate the average A P E , A P E m e a n , and the root mean square error (RMSE) A P E R M S E of all timestamps to measure the accuracy of the overall trajectory, as shown in Equations (9) and (10), where the total count of timestamps is represented by N.
A P E t r a n s , i = t r a n s E i
A P E r o t , i = a n g l e log SO 3 r o t E i
A P E m e a n = 1 N i = 1 N A P E i
A P E R M S E = 1 N i = 1 N A P E i 2

4.2. Experimental Results

In order to showcase the effectiveness of our approach, we undertook experimental tests on the dataset. First, we considered the calculation time: the detector’s detection time for each frame is 0.02 ms, and the multi-view triangulation time for each frame is 0.01 ms.
Next, we executed simulation experiments to assess the precision of our technique. As shown in Figure 6 and Table 2, the outcomes of the experiment indicate our technique’s ability to supply consistent and real-time position data with great precision. Data Seq01 and Data Seq02 are collected in the unobstructed environment. The average translational accuracy ranges from 0.69 to 0.79 cm, achieving millimeter-level precision. The trajectory orientation error is less than 0.65°, validating the correctness of our algorithm. Data Seq03 is collected during fast drone flight around the obstacle, with an average position estimation error of 1.32 cm.
The experimental results of Seq01–03 are compared in Figure 7. In an unobstructed environment, the system achieves millimeter-level accuracy in 3D pose estimation. However, the precision of the trajectory is impacted by any occlusions on specific markers. It is evident that the displacement of detection bounding boxes for the object can influence the accuracy of the 3D pose estimation. This is also the reason why the overall trajectory APE of Seq03 is larger than that of Seq01–02.
In order to show the effectiveness of our system in real-world scenarios, we carried out an experiment comparing it with LiDAR-Inertial SLAM. As shown in Figure 8a,b, using Data Seq04–06, the pose estimation results obtained from Easy Rocap were compared with those from FAST LIO2 [67]. In the test on Seq04, when the trajectory length of curved motion is short, the average trajectory error estimated by the two methods is 4.6 cm. When compared to the accurate distance measured by a physical ruler, from the starting to the endpoint, the difference in distance noted by our system is 4.7 cm, as opposed to LiDAR-Inertial SLAM’s significantly larger difference of 14.1 cm. This also verifies the fact that LiDAR-Inertial SLAM systems may experience drifts during indoor turning. As shown in Figure 8c,d, Data Seq05 is a square trajectory for a UGV. By setting the result of Easy Rocap as the reference trajectory, it is clear that our method generates a trajectory that is smoother and more continuous in turns, while the SLAM method is more likely to experience drift when making sharp turns and accumulate significant errors. Figure 8e presents the actual height (Z-axis) of the trajectories of the ground vehicle, derived from LiDAR-based SLAM and our Easy Rocap, respectively. To explain, the discrepancy in starting positions is due to the UGV moving a certain distance before the initiation of SLAM. The SLAM method exhibits significant jumping in the z-axis direction, which is not in alignment with the real scenario. In contrast, our system’s calculated trajectory remains stable in the z-axis direction. Figure 8f,g show the flight trajectories of the drone from side and top views in the real-world scenario. The trajectory obtained by our system remains smooth and continuous, while the trajectory obtained by LiDAR-Inertial SLAM is more prone to drift when making sharp turns.

4.3. Discussion

This paper presents a cost-effective and easy-to-use robotic pose capture system. Our hardware system consists solely of a computer equipped with a graphics processing unit (GPU), three or more Intel RealSense D455 cameras, and several infrared marker balls. Compared to commercial high-precision optical motion capture systems on the market, which can cost upwards of tens or even hundreds of thousands of dollars, this setup significantly reduces cost. The system’s ease of use is mainly reflected in its robustness, eliminating the need for users to have a strong professional background.
A broad and varied dataset used in the experiments showcases the precision and flexibility of the proposed method. Firstly, the system’s accuracy and speed were effectively demonstrated in the experiments conducted using simulated scenarios. The average translational error between the trajectories of this method and the ground truth is less than 0.8 cm, and the calculation speed can reach 30FPS. Experiments in the simulated scenario have also verified that our method can accurately calculate poses even in the presence of non-linear motion or short-term occlusions. The strong performance of the methodology is closely linked to the combination of techniques like object detection and MOT. Secondly, compared to the real distance between the start and end points, our method achieves higher localization accuracy than SLAM. In the end, our system’s robustness was demonstrated in comparative experiments conducted in real-world scenarios. Compared to advanced SLAM methods, the trajectory obtained by the method is smoother, more continuous, and less prone to drifting. Additionally, there is no accumulation of errors in our method.
However, in order to improve the speed, the chosen MOT algorithm relies on a kinematic model for frame-to-frame correlation, and it does not record the appearance information of the objects. Therefore, our method may not be able to handle long-term occlusions. Nevertheless, we have implemented a response mechanism for long-term occlusions, where multi-view correspondences are repeated when the lost target reappears. This response mechanism may take less than 0.5 s. This mechanism improves the robustness of the robot’s pose estimation. In addition, there is a trade-off between camera resolution and transmission speed, and camera resolution also affects the effective area of the Rocap system. One of the limitations associated with cameras is the potential distortion at the lens edge, which can slightly affect the accuracy of the 3D coordinates when the robot moves to the edge of most cameras’ field of view. In the future, the research will combine high-performance computers with high-resolution cameras to further expand the effective area of the Rocap system.

5. Conclusions

This study proposes an affordable and easy-to-use robotic motion capture system, which can precisely capture the location and orientation of an unmanned platform in complex scenes. The project’s code will be open-sourced on Github (https://github.com/DCSI2022/EasyRocap (access after 1 June 2024)). To address the interference of environmental light noise on visual detection, a TKDM detection model with fine filtering is proposed. Unlike traditional tracking by detection paradigm MOT algorithms that may use wrong estimates when associating tracks, the MOT method used in this paper relies more on observations. To reduce the effects of occlusion from a certain view on the 3D pose estimation results, weights are added to the matrix coefficients of different views based on confidence, ensuring that the contributions of each view are reasonable. The system, composed of commercial-grade cameras, was tested under various situations and motion types, demonstrating high precision and fast speed. The average translation accuracy for pose estimation in the simulation environment is less than 0.8 cm, and the orientation accuracy is less than 0.65 degrees, with a solving speed of 30 FPS. In real scenarios, compared with advanced SLAM algorithms, the practical applicability is verified. The results show that our system can overcome the drift and accumulated errors of the SLAM method. The integrated system only requires data streams from calibrated cameras to estimate robot poses, making it convenient to deploy in robot systems. In addition, there are potential enhancements in future work. In one respect, we will focus on improving the accuracy of our system while maintaining speed by utilizing high-performance computers and high-resolution cameras. In another respect, we will explore marker-less methods for estimating robot 3D poses using 2D feature maps.

Author Contributions

Conceptualization, H.W. and C.C.; methodology, H.W. and C.C.; software, H.W.; validation, H.W.; formal analysis, H.W. and Y.H.; investigation, H.W. and Y.H.; resources, H.W., C.C., S.S. and L.L.; data curation, H.W., S.S. and L.L.; writing—original draft preparation, H.W.; writing—review and editing, H.W., Y.H. and Y.X.; visualization, H.W. and Y.H.; supervision, C.C.; project administration, C.C.; funding acquisition, C.C. and B.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No. U22A20568), the National Key RESEARCH and Development Program (No. 2022YFB3904101), the National Natural Science Foundation of China (No. 42071451), the Natural Science Foundation of Hubei, China (No. 2022CFB007), the Key RESEARCH and Development Program of Hubei Province (No. 2023BAB146), the National Natural Science Foundation of China (No. 42130105), the National Key RESEARCH and Development Program-Key Special Projects for International Cooperation in Science and Technology Innovation between Governments (No. 2019YFE0123300), the Fundamental Research Funds for the Central Universities, and the European Union’s Horizon 2020 Research and Innovation Program (No. 871149).

Data Availability Statement

The experimental data and the corresponding code of the Easy Rocap system will be available at https://github.com/DCSI2022/EasyRocap (access after 1 June 2024).

Acknowledgments

The authors would like to thank the Supercomputing Center of Wuhan University for providing the supercomputing system for the numerical calculations in this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Dinelli, C.; Racette, J.; Escarcega, M.; Lotero, S.; Gordon, J.; Montoya, J.; Dunaway, C.; Androulakis, V.; Khaniani, H.; Shao, S.J.D. Configurations and Applications of Multi-Agent Hybrid Drone/Unmanned Ground Vehicle for Underground Environments: A Review. Drones 2023, 7, 136. [Google Scholar] [CrossRef]
  2. Mohsan, S.A.H.; Khan, M.A.; Noor, F.; Ullah, I.; Alsharif, M.H. Towards the unmanned aerial vehicles (UAVs): A comprehensive review. Drones 2022, 6, 147. [Google Scholar] [CrossRef]
  3. Wisth, D.; Camurri, M.; Fallon, M. VILENS: Visual, inertial, lidar, and leg odometry for all-terrain legged robots. IEEE Trans. Robot. 2022, 39, 309–326. [Google Scholar] [CrossRef]
  4. Schneider, J.; Eling, C.; Klingbeil, L.; Kuhlmann, H.; Förstner, W.; Stachniss, C. Fast and effective online pose estimation and mapping for UAVs. In Proceedings of the 2016 IEEE International Conference on Robotics and Automation (ICRA), Stockholm, Sweden, 16–21 May 2016; pp. 4784–4791. [Google Scholar]
  5. Jiang, N.; Xu, Y.; Xu, T.; Xu, G.; Sun, Z.; Schuh, H. GPS/BDS short-term ISB modelling and prediction. GPS Solut. 2017, 21, 163–175. [Google Scholar] [CrossRef]
  6. Sefidgar, M.; Landry, R.J.D., Jr. Unstable landing platform pose estimation based on Camera and Range Sensor Homogeneous Fusion (CRHF). Drones 2022, 6, 60. [Google Scholar] [CrossRef]
  7. Brena, R.F.; García-Vázquez, J.P.; Galván-Tejada, C.E.; Muñoz-Rodriguez, D.; Vargas-Rosales, C.; Fangmeyer, J. Evolution of indoor positioning technologies: A survey. J. Sens. 2017, 2017, 2530413. [Google Scholar] [CrossRef]
  8. Van Opdenbosch, D.; Schroth, G.; Huitl, R.; Hilsenbeck, S.; Garcea, A.; Steinbach, E. Camera-based indoor positioning using scalable streaming of compressed binary image signatures. In Proceedings of the 2014 IEEE International Conference on Image Processing (ICIP), Paris, France, 27–30 October 2014; pp. 2804–2808. [Google Scholar]
  9. Schweinzer, H.; Syafrudin, M. LOSNUS: An ultrasonic system enabling high accuracy and secure TDoA locating of numerous devices. In Proceedings of the 2010 International Conference on Indoor Positioning and Indoor Navigation, Zurich, Switzerland, 15–17 September 2010; pp. 1–8. [Google Scholar]
  10. Bencak, P.; Hercog, D.; Lerher, T. Indoor positioning system based on bluetooth low energy technology and a nature-inspired optimization algorithm. Electronics 2022, 11, 308. [Google Scholar] [CrossRef]
  11. Guan, W.; Chen, S.; Wen, S.; Tan, Z.; Song, H.; Hou, W. High-accuracy robot indoor localization scheme based on robot operating system using visible light positioning. IEEE Photonics J. 2020, 12, 1–16. [Google Scholar] [CrossRef]
  12. Chen, C.; Jin, A.; Wang, Z.; Zheng, Y.; Yang, B.; Zhou, J.; Xu, Y.; Tu, Z. SGSR-Net: Structure Semantics Guided LiDAR Super-Resolution Network for Indoor LiDAR SLAM. IEEE Trans. Multimed. 2023, 26, 1842–1854. [Google Scholar] [CrossRef]
  13. Masiero, A.; Fissore, F.; Antonello, R.; Cenedese, A.; Vettore, A. A comparison of UWB and motion capture UAV indoor positioning. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2019, 42, 1695–1699. [Google Scholar] [CrossRef]
  14. Munawar, H.S.; Ullah, F.; Heravi, A.; Thaheem, M.J.; Maqsoom, A. Inspecting buildings using drones and computer vision: A machine learning approach to detect cracks and damages. Drones 2021, 6, 5. [Google Scholar] [CrossRef]
  15. Nenchoo, B.; Tantrairatn, S. Real-Time 3D UAV pose estimation by visualization. Multidiscip. Digit. Publ. Inst. Proc. 2020, 39, 18. [Google Scholar]
  16. Zhou, B.; Li, C.; Chen, S.; Xie, D.; Yu, M.; Li, Q. ASL-SLAM: A LiDAR SLAM with Activity Semantics-Based Loop Closure. IEEE Sens. J. 2023, 23, 13499–13510. [Google Scholar] [CrossRef]
  17. Zhou, B.; Mo, H.; Tang, S.; Zhang, X.; Li, Q. Backpack LiDAR-Based SLAM With Multiple Ground Constraints for Multistory Indoor Mapping. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–16. [Google Scholar] [CrossRef]
  18. Xu, Y.; Chen, C.; Wang, Z.; Yang, B.; Wu, W.; Li, L.; Wu, J.; Zhao, L. PMLIO: Panoramic Tightly-Coupled Multi-LiDAR-Inertial Odometry and Mapping. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2023, 10, 703–708. [Google Scholar] [CrossRef]
  19. Sun, S.; Chen, C.; Wang, Z.; Zhou, J.; Li, L.; Yang, B.; Cong, Y.; Wang, H. Real-Time UAV 3D Image Point Clouds Mapping. ISPRS Ann. Photogramm. Remote Sens. Spat. Inf. Sci. 2023, 10, 1097–1104. [Google Scholar] [CrossRef]
  20. Krul, S.; Pantos, C.; Frangulea, M.; Valente, J. Visual SLAM for indoor livestock and farming using a small drone with a monocular camera: A feasibility study. Drones 2021, 5, 41. [Google Scholar] [CrossRef]
  21. Jospin, L.; Stoven-Dubois, A.; Cucci, D.A. Photometric long-range positioning of LED targets for cooperative navigation in UAVs. Drones 2019, 3, 69. [Google Scholar] [CrossRef]
  22. Memon, S.A.; Son, H.; Kim, W.-G.; Khan, A.M.; Shahzad, M.; Khan, U. Tracking Multiple Unmanned Aerial Vehicles through Occlusion in Low-Altitude Airspace. Drones 2023, 7, 241. [Google Scholar] [CrossRef]
  23. Menolotto, M.; Komaris, D.-S.; Tedesco, S.; O’Flynn, B.; Walsh, M. Motion capture technology in industrial applications: A systematic review. Sensors 2020, 20, 5687. [Google Scholar] [CrossRef]
  24. Li, J.; Liu, X.; Wang, Z.; Zhao, H.; Zhang, T.; Qiu, S.; Zhou, X.; Cai, H.; Ni, R.; Cangelosi, A. Real-time human motion capture based on wearable inertial sensor networks. IEEE Internet Things J. 2021, 9, 8953–8966. [Google Scholar] [CrossRef]
  25. Tuli, T.B.; Manns, M. Real-time motion tracking for humans and robots in a collaborative assembly task. In Proceedings of the 6th International Electronic Conference on Sensors and Applications, Online, 15–30 November 2019; p. 48. [Google Scholar]
  26. Salisu, S.; Ruhaiyem, N.I.R.; Eisa, T.A.E.; Nasser, M.; Saeed, F.; Younis, H.A. Motion Capture Technologies for Ergonomics: A Systematic Literature Review. Diagnostics 2023, 13, 2593. [Google Scholar] [CrossRef] [PubMed]
  27. Gu, C.; Lin, W.; He, X.; Zhang, L.; Zhang, M. IMU-based Mocap system for rehabilitation applications: A systematic review. Biomim. Intell. Robot. 2023, 3, 100097. [Google Scholar]
  28. Luque-Vega, L.F.; Lopez-Neri, E.; Arellano-Muro, C.A.; González-Jiménez, L.E.; Ghommam, J.; Saad, M.; Carrasco-Navarro, R.; Ruíz-Cruz, R.; Guerrero-Osuna, H.A. UAV-based smart educational mechatronics system using a MoCap laboratory and hardware-in-the-loop. Sensors 2022, 22, 5707. [Google Scholar] [CrossRef] [PubMed]
  29. Zhang, Z. Microsoft kinect sensor and its effect. IEEE Multimed. 2012, 19, 4–10. [Google Scholar] [CrossRef]
  30. Roetenberg, D.; Luinge, H.; Slycke, P. Full 6dof human motion tracking using miniature inertial sensors. Xsens Motion Technol. BV Tech. Rep. 2013, 1, 1–17. [Google Scholar]
  31. Jha, A.; Chiddarwar, S.S.; Bhute, R.Y.; Alakshendra, V.; Nikhade, G.; Khandekar, P.M. Imitation learning in industrial robots: A kinematics based trajectory generation framework. In Proceedings of the 2017 3rd International Conference on Advances in Robotics, New Delhi, India, 28 June–2 July 2017; pp. 1–6. [Google Scholar]
  32. Mueller, F.; Deuerlein, C.; Koch, M. Intuitive welding robot programming via motion capture and augmented reality. IFAC-PapersOnLine 2019, 52, 294–299. [Google Scholar] [CrossRef]
  33. Shi, Y.; Zhang, Y.; Li, Z.; Yuan, S.; Zhu, S. IMU/UWB Fusion Method Using a Complementary Filter and a Kalman Filter for Hybrid Upper Limb Motion Estimation. Sensors 2023, 23, 6700. [Google Scholar] [CrossRef] [PubMed]
  34. Ren, Y.; Zhao, C.; He, Y.; Cong, P.; Liang, H.; Yu, J.; Xu, L.; Ma, Y. LiDAR-aid Inertial Poser: Large-scale Human Motion Capture by Sparse Inertial and LiDAR Sensors. IEEE Trans. Vis. Comput. Graph. 2023, 29, 2337–2347. [Google Scholar] [CrossRef]
  35. Hii, C.S.T.; Gan, K.B.; Zainal, N.; Mohamed Ibrahim, N.; Azmin, S.; Mat Desa, S.H.; van de Warrenburg, B.; You, H.W. Automated Gait Analysis Based on a Marker-Free Pose Estimation Model. Sensors 2023, 23, 6489. [Google Scholar] [CrossRef]
  36. Wojke, N.; Bewley, A.; Paulus, D. Simple online and realtime tracking with a deep association metric. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Bejing, China, 17–20 September 2017; pp. 3645–3649. [Google Scholar]
  37. Luo, W.; Xing, J.; Milan, A.; Zhang, X.; Liu, W.; Kim, T.-K. Multiple object tracking: A literature review. Artif. Intell. 2021, 293, 103448. [Google Scholar] [CrossRef]
  38. Ciaparrone, G.; Sánchez, F.L.; Tabik, S.; Troiano, L.; Tagliaferri, R.; Herrera, F. Deep learning in video multi-object tracking: A survey. Neurocomputing 2020, 381, 61–88. [Google Scholar] [CrossRef]
  39. Bewley, A.; Ge, Z.; Ott, L.; Ramos, F.; Upcroft, B. Simple online and realtime tracking. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 3464–3468. [Google Scholar]
  40. Zhang, Y.; Sun, P.; Jiang, Y.; Yu, D.; Weng, F.; Yuan, Z.; Luo, P.; Liu, W.; Wang, X. Bytetrack: Multi-object tracking by associating every detection box. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 1–21. [Google Scholar]
  41. Kalman, R.E. A new approach to linear filtering and prediction problems. J. Basic Eng. 1960, 82, 35–45. [Google Scholar] [CrossRef]
  42. Kuhn, H.W. The Hungarian method for the assignment problem. Nav. Res. Logist. Q. 1955, 2, 83–97. [Google Scholar] [CrossRef]
  43. Wang, Z.; Zheng, L.; Liu, Y.; Li, Y.; Wang, S. Towards real-time multi-object tracking. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 107–122. [Google Scholar]
  44. Yifu, Z.; Chunyu, W.; Xinggang, W.; Wenjun, Z.; Wenyu, L. A Simple Baseline for Multi-Object Tracking. arXiv 2020, arXiv:2004.01888. [Google Scholar]
  45. Zhou, X.; Koltun, V.; Krähenbühl, P. Tracking objects as points. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 474–490. [Google Scholar]
  46. Sun, P.; Cao, J.; Jiang, Y.; Zhang, R.; Xie, E.; Yuan, Z.; Wang, C.; Luo, P. Transtrack: Multiple object tracking with transformer. arXiv 2020, arXiv:2012.15460. [Google Scholar]
  47. Meinhardt, T.; Kirillov, A.; Leal-Taixe, L.; Feichtenhofer, C. Trackformer: Multi-object tracking with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 8844–8854. [Google Scholar]
  48. Chu, P.; Wang, J.; You, Q.; Ling, H.; Liu, Z. Transmot: Spatial-temporal graph transformer for multiple object tracking. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 4870–4880. [Google Scholar]
  49. Zeng, F.; Dong, B.; Zhang, Y.; Wang, T.; Zhang, X.; Wei, Y. Motr: End-to-end multiple-object tracking with transformer. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; pp. 659–675. [Google Scholar]
  50. Cao, J.; Pang, J.; Weng, X.; Khirodkar, R.; Kitani, K. Observation-centric sort: Rethinking sort for robust multi-object tracking. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 9686–9696. [Google Scholar]
  51. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
  52. Cutler, M.; Michini, B.; How, J.P. Lightweight infrared sensing for relative navigation of quadrotors. In Proceedings of the 2013 International Conference on Unmanned Aircraft Systems (ICUAS), Atlanta, GA, USA, 28–31 May 2013; pp. 1156–1164. [Google Scholar]
  53. Faessler, M.; Mueggler, E.; Schwabe, K.; Scaramuzza, D. A monocular pose estimation system based on infrared leds. In Proceedings of the 2014 IEEE International Conference on Robotics and Automation (ICRA), Hong Kong, China, 31 May–7 June 2014; pp. 907–913. [Google Scholar]
  54. Fishberg, A.; How, J.P. Multi-agent relative pose estimation with UWB and constrained communications. In Proceedings of the 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Kyoto, Japan, 23–27 October 2022; pp. 778–785. [Google Scholar]
  55. Cossette, C.C.; Shalaby, M.A.; Saussié, D.; Le Ny, J.; Forbes, J.R. Optimal multi-robot formations for relative pose estimation using range measurements. In Proceedings of the 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Kyoto, Japan, 23–27 October 2022; pp. 2431–2437. [Google Scholar]
  56. Jones, B.; Thompson, K.; Pierce, D.; Martin, S.; Bevly, D. Ground-Vehicle Relative Position Estimation with UWB Ranges and a Vehicle Dynamics Model. IFAC-PapersOnLine 2022, 55, 681–687. [Google Scholar] [CrossRef]
  57. Jin, Z.; Jiang, C. Range-Aided LiDAR-Inertial Multi-Vehicle Mapping in Degenerate Environment. arXiv 2023, arXiv:2303.08454. [Google Scholar] [CrossRef]
  58. Hao, N.; He, F.; Tian, C.; Yao, Y.; Mou, S. KD-EKF: A Kalman Decomposition Based Extended Kalman Filter for Multi-Robot Cooperative Localization. arXiv 2022, arXiv:2210.16086. [Google Scholar]
  59. Pritzl, V.; Vrba, M.; Štěpán, P.; Saska, M. Fusion of Visual-Inertial Odometry with LiDAR Relative Localization for Cooperative Guidance of a Micro-Scale Aerial Vehicle. arXiv 2023, arXiv:2306.17544. [Google Scholar]
  60. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  61. Zhang, Z.; Deriche, R.; Faugeras, O.; Luong, Q.-T. A robust technique for matching two uncalibrated images through the recovery of the unknown epipolar geometry. Artif. Intell. 1995, 78, 87–119. [Google Scholar] [CrossRef]
  62. Zhang, Z. A flexible new technique for camera calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
  63. Hartley, R.; Zisserman, A. Multiple View Geometry in Computer Vision; Cambridge University Press: Cambridge, UK, 2003. [Google Scholar]
  64. Iskakov, K.; Burkov, E.; Lempitsky, V.; Malkov, Y. Learnable triangulation of human pose. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–28 October 2019; pp. 7718–7727. [Google Scholar]
  65. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  66. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. Mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
  67. Xu, W.; Cai, Y.; He, D.; Lin, J.; Zhang, F. Fast-lio2: Fast direct lidar-inertial odometry. IEEE Trans. Robot. 2022, 38, 2053–2073. [Google Scholar] [CrossRef]
Figure 1. The workflow of Easy Rocap.
Figure 1. The workflow of Easy Rocap.
Drones 08 00137 g001
Figure 2. Instance of our MOT method based on observation. (a) The original detection results including confidence scores. (b) The trajectories of the algorithm before the improvement, which is only associated with the high-confidence detection boxes. The identical box color signifies the specific individual. (c) The utilization of low-confidence bounding boxes during object occlusion, and the red dashed box indicates the predicted position of the occluded target. Low-confidence detection boxes are successfully matched to previous unmatched trajectories.
Figure 2. Instance of our MOT method based on observation. (a) The original detection results including confidence scores. (b) The trajectories of the algorithm before the improvement, which is only associated with the high-confidence detection boxes. The identical box color signifies the specific individual. (c) The utilization of low-confidence bounding boxes during object occlusion, and the red dashed box indicates the predicted position of the occluded target. Low-confidence detection boxes are successfully matched to previous unmatched trajectories.
Drones 08 00137 g002aDrones 08 00137 g002b
Figure 3. Test robot platforms and scenarios. (a) Simulated drone with markers. (b) Drone in simulation scenario (occlusion). (c) UGV with markers. (d) UGV in real scenarios. (e) Drone with markers. (f) Drone in real scenarios.
Figure 3. Test robot platforms and scenarios. (a) Simulated drone with markers. (b) Drone in simulation scenario (occlusion). (c) UGV with markers. (d) UGV in real scenarios. (e) Drone with markers. (f) Drone in real scenarios.
Drones 08 00137 g003
Figure 4. Simulation scenario.
Figure 4. Simulation scenario.
Drones 08 00137 g004
Figure 5. Real scenario.
Figure 5. Real scenario.
Drones 08 00137 g005
Figure 6. Results of Seq01−03 trajectory evaluation. (a) Trajectory Plot of Seq01. (b) APE Plot of Seq01. (c) Trajectory Plot of Seq02. (d) APE Plot of Seq02. (e) Trajectory Plot of Seq03. (f) APE Plot of Seq03.
Figure 6. Results of Seq01−03 trajectory evaluation. (a) Trajectory Plot of Seq01. (b) APE Plot of Seq01. (c) Trajectory Plot of Seq02. (d) APE Plot of Seq02. (e) Trajectory Plot of Seq03. (f) APE Plot of Seq03.
Drones 08 00137 g006aDrones 08 00137 g006b
Figure 7. Statistics results for APE on Seq01−03. (a) Comparison histogram of APE for Seq01−03. (b) Box plot of APE for Seq01−03.
Figure 7. Statistics results for APE on Seq01−03. (a) Comparison histogram of APE for Seq01−03. (b) Box plot of APE for Seq01−03.
Drones 08 00137 g007
Figure 8. Results of Seq04–06 trajectory evaluation. (a) Trajectory Plot of Seq04. (b) APE Plot of Seq04. (c) Trajectory Plot of Seq05. (d) APE Plot of Seq05. (e) Comparison Plot of Z-Axis Trajectories for Seq05. (f) Trajectory Plot of Seq06. (g) APE Plot of Seq06.
Figure 8. Results of Seq04–06 trajectory evaluation. (a) Trajectory Plot of Seq04. (b) APE Plot of Seq04. (c) Trajectory Plot of Seq05. (d) APE Plot of Seq05. (e) Comparison Plot of Z-Axis Trajectories for Seq05. (f) Trajectory Plot of Seq06. (g) APE Plot of Seq06.
Drones 08 00137 g008aDrones 08 00137 g008b
Table 1. Data descriptions.
Table 1. Data descriptions.
PlatformSceneMotion
Pattern
Data Length/SecondsCamera Frequency/FPSTruth Traj Frequency
/Hz
LiDAR
Frequency
/Hz
Seq01UAVsimulationhorizontal60.71151000
Seq02UAVsimulationhelical44.81151000
Seq03UAVsimulationhorizontal46.33151000
Seq04UGVreal-worldcurvilinear6.0190 10
Seq05UGVreal-worldsquare42.8090 10
Seq06UAVreal-worldcurvilinear27.0990 10
Table 2. Results of simulation experiments.
Table 2. Results of simulation experiments.
SceneMotion
Pattern
Trajectory Length/mDuration
/s
RMSE(APE)
/m
Mean (APE)
/m
Seq01simulationhorizontal23.7360.710.00800.0069
Seq02simulationhelical12.7943.910.00900.0079
Seq03simulationhorizontal
(occlusion)
18.6746.330.01500.0132
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, H.; Chen, C.; He, Y.; Sun, S.; Li, L.; Xu, Y.; Yang, B. Easy Rocap: A Low-Cost and Easy-to-Use Motion Capture System for Drones. Drones 2024, 8, 137. https://doi.org/10.3390/drones8040137

AMA Style

Wang H, Chen C, He Y, Sun S, Li L, Xu Y, Yang B. Easy Rocap: A Low-Cost and Easy-to-Use Motion Capture System for Drones. Drones. 2024; 8(4):137. https://doi.org/10.3390/drones8040137

Chicago/Turabian Style

Wang, Haoyu, Chi Chen, Yong He, Shangzhe Sun, Liuchun Li, Yuhang Xu, and Bisheng Yang. 2024. "Easy Rocap: A Low-Cost and Easy-to-Use Motion Capture System for Drones" Drones 8, no. 4: 137. https://doi.org/10.3390/drones8040137

Article Metrics

Back to TopTop