2.1. Mobile Intelligent Construction
As mobile robot technology continues to improve, increased researchers are working on developing mobile robots suitable for different tasks. For example, the Fraunhofer Institute developed a mobile robot [
15], KUKA iisy, developed by KUKA, and the Spark robot, developed by NXROBO [
16]. Robots suitable for the construction industry have not only demonstrated the potential to increase construction productivity but have also shown the ability to address issues such as labor shortages and safety risks in the construction sector [
17,
18]. Although robots have had enormous success in intelligent construction, this has only been for small scale components. When the goal of construction is large-scale components, robots can only build several small-sized components individually and then assemble them into large-scale components. Since the fabrication of large architectural structures exceeds the workspace of a stationary robot, this requires that the robot be mobile during the construction process and be able to orientate itself according to the workspace [
19,
20]. In order to solve this problem, professionals in the industry came up with the idea of combining a robotic arm with a mobile platform and started to investigate the coordination of the robotic arm and the mobile platform [
21,
22,
23,
24]. Yamamoto et al. [
25] compensated for the dynamic interactions between the robotic arm and the mobile platform by using a nonlinear feedback control to give better co-ordination. The robot’s positioning technique is particularly important in the mobile intelligent construction process. Therefore, more people have started to investigate different positioning methods. Norman et al. [
26] used indoor GPS technology to achieve high-accuracy absolute positioning and motion of collaborative robots. Jiang et al. [
27] developed a visual positioning system which consists of a camera and a video tracker. Susemihl et al. [
28] proposed a new mobile robot system in order to process large aircraft parts. A laser tracker and a stereo camera are used for positioning, where the laser tracker is used to locate the position of the moving platform and the stereo camera system tracks the position of the robot end-effector.
In order to better apply mobile robotics to on-site construction, the ETH Zurich team has conducted a lot of attempts and experiments. Helm et al. [
29] mounted an ABB IRB 4600 robot on a mobile platform to create a “dimRob” robot, which combines the constructability of the arm and the mobility of the mobile platform for 1:1 on-site manufacturing. Sandy et al. [
10] proposed a repositioning system in order to address the limitations of robotics on construction sites, which uses only on-board sensors and allows the mobile robot to maintain a high degree of accuracy at multiple position in the construction task, demonstrating sub-centimeter repositioning accuracy by constructing vertical stacks of bricks in four different directions. Kathrin et al. [
30] proposed the concept of an in situ fabricator (IF) based on the “dimRob” robot, where a 3D point cloud of the surrounding environment is constructed by mounting a laser range finder on the end-effector, which is accurately aligned with the initial point cloud to infer the relative position of the robot. A 6.5 m long and 2 m high double-leaf brick wall was constructed using the robot, and 14 repositionings were performed throughout the process. Due to the extreme flexibility of the camera, Lussi et al. [
31] moved on to a vision-sensor-based solution, where the camera sensor was mounted on the robot’s moving platform, scanning the AprilTag affixed to the floor for the robot’s repositioning, and ultimately achieving millimeter-level positioning accuracy and replacing the laser scanner originally mounted on the end tooling with the camera to provide real-time feedback on the build tolerances, culminating in the fabrication of a lattice mold in the form of an S-shape that was 12 m in length. Gawel et al. [
32] proposed a fully integrated sensor and control system for high-accuracy mobile robotic building construction, making it possible to enable robots to perform construction tasks with millimeter-level accuracy using LIDAR, IMUs, wheel encoders, and laser scanners. Hamner et al. [
33] used a stereo camera as well as various sensors to accomplish autonomous navigation and operation, enabling it to move and operate independently during assembly tasks. Chai et al. [
34] used a depth camera to identify AprilTags, place sticks, and apply glue at specific positions to enable the operation of a mobile robot for on-site construction. If tags are used for positional recognition, it takes a long time to apply the tags and the camera is susceptible to the influence of the surrounding light environment, resulting in recognition errors. Wang et al. [
35] verified the potential of integrating parametric design and robotic construction into wood building construction projects using the robotic construction of wood buildings as an orientation. In their study, a stationary robot was used to construct each unit horizontally on the ground, and then a crane was utilized to erect each unit and ultimately assemble them together for the construction of the entire building. However, this approach is limited by the working range of the robotic arm, as well as the inability to build directly on-site in the construction environment. Therefore, this paper is oriented towards large-scale constructions, using MCP to construct wooden cabins continuously in the vertical direction.
Dependable high-accuracy positioning is essential for mobile intelligent construction scenarios. Available positioning methods in intelligent construction, such as QR code-assisted positioning, are low-cost and easy to deploy. Still, the QR code on the ground is easy to wear out, requires regular maintenance, and has high requirements for ground flatness, so it is only suitable for use on factory floors. Laser scanning positioning technology involves much computing, and the surrounding environment must remain unchanged. Visual positioning uses a vision camera to capture images of the surrounding area, then calculates the position and direction of the surrounding environment, builds a map of the unknown territory, and then realizes the robot’s positioning. Cameras are widely used because they are inexpensive and easy to operate. However, cameras are susceptible to light effects and produce erroneous images; for example, it takes work to identify the environment in dark environments. Therefore, to eliminate the above positioning limitations, this paper investigates various other positioning techniques for mobile robots to better apply them to mobile robot positioning.
2.2. Mobile Robot Positioning
Positioning techniques for mobile robots are crucial in performing tasks [
36]. They ensure that the robot knows exactly where it is in the environment for task execution and navigation. Accurate positioning helps the robot to avoid obstacles, find the shortest path, and navigate in complex environments. In addition, positioning also motivates the robot to perceive objects and structures in the environment, identify and locate target objects, and interact with them. Overall, positioning technology for mobile robots is crucial for navigation, task execution, environment awareness, collaboration, and safety [
37]. It improves the efficiency, reliability, and safety of robots, and enables them to better adapt to various environments. Therefore, more researchers are working on robot positioning techniques to improve positioning accuracy. On the one hand, the positioning accuracy of a single sensor is improved by improving the positioning algorithm to meet one’s task requirements; on the other hand, the data from two or more sensors are fused to improve the robot’s positioning accuracy.
Gentner et al. [
38] used WIFI-round-trip-time for distance estimation of a target robot to achieve precise positioning. Xu et al. [
39] achieved stable indoor positioning by extracting ceiling features with a camera mounted on an end tool. Jung et al. [
40] proposed a SLAM approach where, by using a ceiling feature map, the A robot fitted with a monocular camera searches upwards for ceiling features and estimates the robot position using a Monte Carlo positioning method. Among the positioning techniques, UWB ranging systems stand out for their strong anti-jamming performance, high transmission rate, good multipath resistance, and high security [
41], and they can achieve centimeter-level accuracy under line-of-sight conditions. Also, because UWB is deployed to achieve the positioning of the target node through the deployment of the base station, it is suitable for both indoor and outdoor applications, and thus has been favored by a large number of researchers. Richardson et al. [
42] analyzed the ranging accuracy of UWB and concluded that the ranging accuracy can reach up to 3 cm in the open outdoor environment. Li et al. [
43] proposed a method to enhance UWB positioning using neural networks, and the experimental results show that neural networks can make UWB maintain stable positioning in both line-of-sight and non-line-of-sight scenarios. Takahara et al. [
44] applied UWB positioning technology to the management of prison inmates, which can determine the position of the inmates in real time, and can improve the supervisory efficiency to a large extent. Ershadh et al. [
45] proposed a design scheme to improve the antenna of UWB sensor, which enhances the comprehensive performance of UWB positioning under the unchanged spatial environment, but its high cost makes it unsuitable for large-scale popular application. Since single sensor positioning results may not be accurate enough in scenarios with extremely high accuracy requirements, increased research is focusing on multi-sensor fusion positioning techniques. Yang et al. [
46] proposed a GA-BP neural network fusion UWB/IMU-based positioning method to address the problem of large error of a single positioning technique in complex indoor environments. The experimental results show that the GA-BP neural network fusion positioning method has a significant improvement in positioning accuracy over the single positioning method.
In multi-sensor fusion methods, Kalman filter (KF) is a commonly used tool for fusing data from different sensors. Since the KF algorithm was proposed in 1960 [
47], researchers have proposed various improved algorithms based on KF [
48], such as Extended KF (EKF) [
49], Unscented KF (UKF) [
50], Cubature KF (CKF) [
51], etc. Woosik et al. [
52] proposed a positioning method for data fusion of multiple low-cost GPS positioning systems using the EKF, thus improving the positioning accuracy of the system. Navarro et al. [
53] used a tracking method based on EKF fusion of time-of-arrival and direction-of-arrival measurements of position. Liu et al. [
54] proposed a method of fusion of UWB and IMU measurements using a CKF that corrects position, velocity, and orientation errors to improve orientation estimation and enhance the positioning accuracy of UWB. Agrawal et al. [
55] proposed a low-cost positioning system consisting of stereo vision, IMU, and GPS, which achieved outdoor dm-level positioning accuracy by fusing information from these sensors through a KF. When the visual odometer fails, the IMU populates the motion estimation.
In summary, dependable high-accuracy positioning technology in mobile robotics is crucial for robot motion and navigation. When a single sensor cannot meet the requirements of a specific application scenario, multi-sensor data fusion positioning can be used to improve the robot’s positioning accuracy. For different application scenarios, selecting appropriate sensors and positioning algorithms is crucial to ensure that the robot can achieve accurate positioning in various complex environments.