Next Article in Journal
Contract-Optimization Approach (COA): A New Approach for Optimizing Service Caching, Computation Offloading, and Resource Allocation in Mobile Edge Computing Network
Next Article in Special Issue
Adaptive Output Feedback Control for Nonholonomic Chained Systems with Integral Input State Stability Inverse Dynamics
Previous Article in Journal
Electrocardiogram Monitoring Wearable Devices and Artificial-Intelligence-Enabled Diagnostic Capabilities: A Review
Previous Article in Special Issue
TeCVP: A Time-Efficient Control Method for a Hexapod Wheel-Legged Robot Based on Velocity Planning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep-Reinforcement-Learning-Based Object Transportation Using Task Space Decomposition

Department of Mechatronics Engineering, Tech University of Korea, 237 Sangidaehak-ro, Siheung-si 15073, Gyeonggi-do, Republic of Korea
Sensors 2023, 23(10), 4807; https://doi.org/10.3390/s23104807
Submission received: 12 April 2023 / Revised: 12 May 2023 / Accepted: 13 May 2023 / Published: 16 May 2023
(This article belongs to the Special Issue Advances in Mobile Robot Perceptions, Planning, Control and Learning)

Abstract

:
This paper presents a novel object transportation method using deep reinforcement learning (DRL) and the task space decomposition (TSD) method. Most previous studies on DRL-based object transportation worked well only in the specific environment where a robot learned how to transport an object. Another drawback was that DRL only converged in relatively small environments. This is because the existing DRL-based object transportation methods are highly dependent on learning conditions and training environments; they cannot be applied to large and complicated environments. Therefore, we propose a new DRL-based object transportation that decomposes a difficult task space to be transported into simple multiple sub-task spaces using the TSD method. First, a robot sufficiently learned how to transport an object in a standard learning environment (SLE) that has small and symmetric structures. Then, a whole-task space was decomposed into several sub-task spaces by considering the size of the SLE, and we created sub-goals for each sub-task space. Finally, the robot transported an object by sequentially occupying the sub-goals. The proposed method can be extended to a large and complicated new environment as well as the training environment without additional learning or re-learning. Simulations in different environments are presented to verify the proposed method, such as a long corridor, polygons, and a maze.

1. Introduction

Object transportation is where a robot manipulates an object to a goal, and it is applied to various fields, such as logistics [1,2], exploration [3,4], retrieval tasks [5], and foraging [6].
Representative methods of transporting an object are prehensile and non-prehensile manipulation [7]. In prehensile manipulation, a robot first firmly grasps or captures an object using an equipped tool such as a gripper [8] or a dexterous robotic hand [9]. The robot and the object then move toward a goal as one body. The advantage of prehensile manipulation is that the object remains stationary during the transportation process, which allows the robot to precisely control the object. However, the disadvantage is that a pre-gripping or capturing motion is required before the object is transported. In addition, a robot should carefully determine how to grasp or where to apply force depending on the shape and material of the object. In contrast to prehensile manipulation, in non-prehensile manipulation, a robot does not use any tools and only pushes an object with its body [10,11,12]; preliminary grasping actions for object transportation are not required. In addition, the control of the robot is easier than that of prehensile manipulation because the robot is not tied to an object. However, it is difficult to predict the motion of the object, and thus a robot should reposition its pushing point every second to transport an object to a goal. Because of these characteristics, prehensile manipulation is an appropriate method for structured objects (i.e., known objects) [13] and non-prehensile manipulation is applicable to unstructured objects (i.e., unknown objects) [14,15].
Recently, reinforcement learning (RL)-based non-prehensile manipulation methods have been studied by many researchers [16,17,18,19]. These methods have performed well by learning how to transport an object without external intervention, but they only worked in limited environments (i.e., trained environments). In other words, their methods did not work well when the test and learning environments were different. In order to operate in a new environment, a robot would have to be retrained; this is an inefficient method due to time and space constraints.
Therefore, we propose a novel object transportation using deep reinforcement learning (DRL) and task space decomposition (TSD). First, a robot learns how to transport an object in a standard learning environment (SLE). The SLE is a learning space with a symmetric structure, which is usually smaller than a test environment, that allows a robot to learn a basic policy required for object transportation. In this phase, the robot becomes proficient in transporting an object using DRL. Second, a test environment (i.e., a whole task space) is divided into multiple sub-task spaces by considering the maximum traveled distance of a robot, which is called the TSD method. The TSD method divides a large and complex environment into many small and simple environments, making it easier to transport objects step-by-step. Finally, the object is transported to the goal by following the policy learned from the SLE in sub-task spaces. Each sub-task space computed from the TSD is smaller than the SLE, and thus the robot can apply the policy learned in the SLE. In other words, we simplified the problem by extending the learning results in the sub-task space to the whole-task space, as opposed to the existing DRL-based object transportation in the whole environment at once (i.e., end-to-end learning); this is the main contribution of this paper.
The proposed method has retained the advantage of prehensile manipulation, where a robot can learn how to transport without the complexity of control. At the same time, it trains only in a small environment (i.e., SLE), where it is easy to learn by DRL, and solves problems in large environments using the results in the SLE. This is similar to the divide and conquer strategy in computer science [20], and it has the advantage of being applicable to a wide variety of environments.
The contributions of this paper are summarized as follows:
  • We present a novel object transportation method using DRL and TSD.
  • We propose a DRL-based object transportation training method within an SLE.
  • A TSD approach is introduced that can be applied to arbitrary object transportation environments.
  • We verify the proposed method by performing simulations in a variety of environments, such as a long corridor, a polygon, and a maze.
This paper is organized as follows. Section 2 describes DRL-based non-prehensile object transportation methods and task decomposition methods. The object transportation problem is defined in Section 3, and the system overview is presented in Section 4. Section 5 and Section 6 explain the proposed object transportation method by separating the training and TSD steps. Simulations are shown in Section 7, and the conclusion is presented in Section 9.

2. Related Work

Traditionally, RL has been used as a tool to solve the object transportation problem and has been studied by many researchers. Manko et al. [21] presented a Q-learning-based large-object transportation method in a complex environment. Wang and De Silva [22] compared a single-agent Q-learning with a team Q-learning. They showed that single-agent Q-learning is better than the team Q-learning in a complicated environment because of the local-minima. Similarly, Rahimi et al. [23] compared different Q-learning-based box-pushing algorithms, and Wang and de Silva [24] proposed sequential Q-learning with the Kalman filtering method.
The above-mentioned RL-based object transportation methods showed limited performance under certain conditions because traditional Q-learning algorithms (e.g., the tabular method) have finite possible states. In the real-world, the possible states are very large, and thus a robot cannot sufficiently learn how to transport due to the limited size of the Q-table. Therefore, Mnih et al. [25] and Silver et al. [26] presented a deep Q-network (DQN) algorithm that shows a good performance by approximating the Q-table with a deep neural network.
Based on the improved performance of DQN, many DRL-based object transportation studies have been proposed. Zhang et al. [27] proposed multi-robot object transportation based on DRL. A large rod is carried to an exit by two robots using DQN. Xiao et al. [28] presented a decentralized macro-action-based policy via a centralized Q-net. Each decentralized Q-net is trained with the help of a centralized Q-net. Eoh and Park [29] proposed a curriculum-learning-based object transportation method using difficulty map generation and an adaptive determination of the episode size. Shibata et al. [30] presented a DRL-based multi-robot transportation method using an event-triggered communication and consensus-based control. The proposed multi-robot team can balance the control performance and communication savings even when the number of agents is different from that in the learning environment. In addition, various manipulation skills (e.g., pushing, pulling, and moving objects) based on RL have been studied in the fields of animation and virtual reality [31,32]. Their proposed methods allow a producer to create transport animations without complicated and tricky editing.
Meanwhile, various task decomposition methods have been proposed by many researchers to simplify complex navigation tasks. Kawano [33] proposed an automatic hierarchical sub-task decomposition method for a multi-robot delivery mission. They decomposed a task into multiple sub-tasks and learned in a stepwise manner by expanding the activity radius of the robot. This is a similar method to our proposed method, but their method only works in a grid world; robots have limited movements and action policies, such as pushing, non-pushing, and waiting. Mesesan et al. [34] proposed a hierarchical path planner that consists of two stages: a global planner and several local planners. It solves a complicated global path-planning problem by decomposing it into simpler local path-planning problems. Zhang et al. [35] proposed task-space-decomposed motion planning. The constrained manifold can be easily solved by a dual-resolution motion planning framework consisting of a global planner and a local planner. Fosong et al. [36] presented a multi-agent teamwork-learning method using a sub-task curriculum. A multi-agent learns simple specific sub-tasks, and then the policy of the team is merged and transferred to the target task. The above-mentioned task decomposition methods have different methodologies, but they all have one thing in common: they solve problems by breaking down a difficult problem into multiple easy problems. Following this principle, we propose a new object transportation approach that applies the learning results in a small and simple SLE to large and complex environments.

3. Problem Description

The problem of object transportation is manipulating an object to a goal within the goal boundary ϵ s u c c e s s using a robot as follows:
| | p t o p g | | 2 < ϵ s u c c e s s
where p t o is the position of the object at time t and p g is the position of the goal. If an object is within ϵ s u c c e s s , the object transport is considered complete. The robot is able to manipulate an object by pushing under its own power. We assumed that the object follows the quasi-static model, which is a reasonable assumption except for when the object is spherical. The environment was assumed to be polygonal in shape and, if it is not, it can be approximated as a polygon. We also assumed that there are no obstacles in the environment in order to focus on how an object is transported.

4. System Overview

The proposed object transportation method is based on DRL and TSD. This process is performed in three steps: (1) training to transport an object in an SLE, (2) TSD and sub-goal generation, and (3) transporting an object via sub-goals in the whole-task space.
First, we created the SLE in a small and square shape to facilitate learning, as shown in Figure 1a. In the training phase, we initialized the robot’s position randomly but ensured that the goal position is always in the center of the environment for symmetry. The size of the SLE can vary, but there are tradeoffs, such as the following. If the SLE is small, it is easy to converge and fast to learn, but the whole-task space should be divided into many small pieces in the transportation phase. On the other hand, if the SLE is large, then the number of sub-goals will decrease because the space of the sub-tasks will also be large. However, it is prone to divergence and the learning speed is slow.
Second, a whole-task space was decomposed into the spaces of sub-tasks, taking into account the size of the SLE in the first phase, as shown in Figure 1b. The size of the sub-tasks cannot exceed that of the SLE because a robot cannot learn how to transport an object beyond the size of the SLE. Several sub-goals were then computed, taking into account the shape of the environment and the size of the sub-task spaces.
Finally, a robot transported an object to a goal via sub-goals that were generated in the previous phase, as shown in Figure 1c. For each space of sub-tasks, the robot takes an action by applying the inference result of the DQN, which was pre-trained in the SLE. The robot can iteratively use the pre-trained results by sequentially changing a goal from the initial to the final sub-goal. For example, in Figure 1c, the robot selects a goal as sub-goal 1 at the beginning. When the robot has successfully transported an object to sub-goal 1, a goal is changed to sub-goal 2. The robot then transports an object to sub-goal 2. This continues until the object is successfully transported to the final goal.

5. Training in a Standard Learning Environment

In this section, we describe the training process for object transportation in an SLE. First, a Markov decision process (MDP) was introduced to solve the RL problem. Next, the DRL framework for object transportation is presented and each component of the framework is described. Finally, an object transportation training method based on DQN is proposed for efficient learning.

5.1. Markov Decision Process

The RL problem can be solved by finding sequential decisions under uncertainty, which was formulated by the MDP. The MDP consists of four tuples: state space ( S ), action space ( A ), state transition probability ( T ), and reward function ( R ). A robot observes the state s t S at time t and takes an action a t A according to a policy function π : s t P r o b ( A | s t ) . The policy maps states to actions and determines how the robot interacts with the environment. The state transition probability T ( s , a ) = P r o b ( s t + 1 = s | s t = s , a t = a ) is the probability when a robot takes the action a at the state s. A robot receives a reward r t R : S × A R if a robot performs an action a t in a given state s t . The goal of RL is to find a policy function π that maximizes the sum of all expected rewards over time.

5.2. Reinforcement Learning Framework for Object Transportation

An RL framework for object transportation is presented in Figure 2. A square environment ( L e n v × L e n v ) was created for standardized learning. The initial positions of a robot and an object were chosen randomly. However, the goal position should be fixed in the center (i.e., origin) of the SLE to preserve symmetry: p g = [ 0 , 0 ] and p t r , p t o [ L e n v 2 , L e n v 2 ] × [ L e n v 2 , L e n v 2 ] .
A state s t consists of a six-dimensional array as follows:
s t = [ d t r , o , d t r , g , cos θ t r , o , sin θ t r , o , cos θ t r , g , sin θ t r , g ] ,
where d t i , j and θ t i , j are the distance and the angle between i and j, respectively. The indices i and j can represent a robot, an object, and a goal: i , j { r o b o t , o b j e c t , g o a l } . In order to represent the robot’s pushing motion, we defined a new state by concatenating the last three consecutive frames similar to [25]:
s ˜ t = c o n c a t e n a t e { s t 2 , s t 1 , s t } .
The actions of the robot are represented by eight motions using translational and rotational velocities ( [ v t r a n s , v r o t ] ) as follows:
a t = [ v t r a n s , 0 ] B a c k w a r d ( B ) [ v t r a n s , 0 ] F o r w a r d ( F ) [ 0 , v r o t 2 ] L e f t T u r n ( LT ) [ 0 , v r o t 2 ] R i g h t T u r n ( RT ) [ v t r a n s , v r o t ] F o r w a r d L e f t ( FL ) [ v t r a n s , v r o t ] F o r w a r d R i g h t ( FR ) [ v t r a n s , v r o t ] B a c k w a r d L e f t ( BL ) [ v t r a n s , v r o t ] B a c k w a r d R i g h t ( BR )
A robot receives a reward according to action a t given the state s ˜ t as follows:
r t ( s ˜ t , a t ) = r s u c c e s s | | p t o p g | | 2 < ϵ s u c c e s s 2 r c o l l i s i o n | | p t o p g | | 1 > L 2 ϵ m a r g i n r c o l l i s i o n | | p t r p g | | 1 > L 2 ϵ m a r g i n w 1 Δ d t g , o + w 2 Δ d t r , o otherwise
where r s u c c e s s > 0 and r c o l l i s i o n < 0 are rewards according to the success of object transportation and the collision with respect to the wall, respectively. The margin parameter ϵ m a r g i n is determined by considering the size of the SLE, a robot, and an object. The indicator function Δ d t i , j is how close an agent i is getting to an agent j ( i , j { r o b o t , o b j e c t , g o a l } ). Since it is most important that the object reaches its goal, we set w 1 and w 2 to satisfy the following conditions: w 1 = 10 w 2 .

5.3. Object Transportation Learning Using Deep Q-Network

Q-learning is a model-free RL that uses the Bellman equation to iteratively update estimates of the expected rewards of each action, allowing an agent to learn an optimal policy that maximizes the long-term expected return [37]. This process is formulated as a Q-learning algorithm:
Q ( s t , a t ) Q ( s t , a t ) + α [ r t + 1 + γ max a t + 1 Q ( s t + 1 , a t + 1 ) Q ( s t , a t ) ] ,
where α ( 0 , 1 ] is a learning rate that determines how fast a robot will learn. If α 0 , a robot will not learn anything. On the contrary, a robot will ignore previous information if α = 1 . Parameter γ [ 0 , 1 ] is the discount factor that determines how much future rewards will be considered. If γ = 1 , a robot will consider a long-term reward. Conversely, a robot will only consider a short-term (i.e., immediate) reward if γ = 0 .
The Q-table generated by the Q-learning algorithm can be parameterized by a neural network denoted as θ t . We can rewrite Equation (6) as follows:
Q ( s t , a t | θ t ) Q ( s t , a t | θ t ) + α [ r t + 1 + γ max a t + 1 Q ( s t + 1 , a t + 1 | θ t ) Q ( s t , a t | θ t ) ] ,
where θ t and θ t are the weights and biases in the active and target Q-networks, respectively. The active and target Q-networks are illustrated in Figure 2.
The loss function L ( s t , a t | θ t ) is defined as follows:
L ( s t , a t | θ t ) [ r t + 1 + γ max a t + 1 Q ( s t + 1 , a t + 1 | θ t ) Q ( s t , a t | θ t ) ] 2 ,
and the weights and biases are updated to minimize a differentiable loss function:
θ t + 1 θ t + α θ L ( θ t ) .
A robot explores a new environment or exploits the existing information in the learning phase. This is called the exploration–exploitation tradeoff [38]. The ϵ -greedy strategy strikes a balance between exploration and exploitation by gradually decreasing the ϵ value from 1.0 to 0.1. The action of a robot is chosen by the ϵ -greedy algorithm using the active Q-network, as shown in the upper part of Figure 2.
Meanwhile, if only a single Q-network is used for learning, there may be oscillation or divergence problems during updating. Therefore, we separated the active (top) and target (bottom) Q-networks, as shown in Figure 2. The learning system is stable by periodically copying the parameters of the active Q-network to the target Q-network during the learning phase [25].
Finally, an experience replay memory is introduced to uncorrelate between information [39]; this records experiences e t = ( s ˜ t , a t , r t , s ˜ t + 1 ) in a replay memory D , and randomly extracts samples for each learning phase.
The pseudo-code for training in an SLE is given in Algorithm 1. First, we initialized the replay memory D , the weights and biases of two neural networks θ t and θ t with identical values, and the episode size n e p (line 9–11). Second, we initialized the positions of a robot and an object in L e n v × L e n v space for each episode (line 14). A concatenated new state s ˜ t { s t 2 , s t 1 , s t } was also initialized with the identical states (line 15). The ϵ decreases linearly with the ϵ -greedy algorithm (line 16). During the episode, a robot takes a random action a t using the ϵ -greedy algorithm, observes a new state s t + 1 , and then receives a reward r t + 1 (line 18–19). A new concatenated state s ˜ t + 1 at time t + 1 is generated from three consecutive frames, and an experience ( s ˜ t , a t , r t + 1 , s ˜ t + 1 ) is recorded in the experience replay memory D (line 20–21). This process is repeated T s t e p times until the end of the episode (line 17). In the training phase, we extracted random mini-batch samples from D and it trains N i t e r times to minimize the loss function L ( s t , a t | θ t ) (line 23–25). The weights of the target Q-network θ t are substituted with those of the active Q-network θ t every T K episodes (line 26). Then, we measured the success rate of object transportation to check the current performance (line 27–29). If a robot succeeds in transporting an object to a goal with the success rate P s u c c e s s , it stops training at the current level. On the contrary, if the success rate is less than P s u c c e s s , it trains more by increasing the unit size of episodes ( N e p ) (line 30–33). When using DRL, there are cases where overtraining can actually hinder learning, and this can be prevented by stopping training after P s u c c e s s .
Algorithm 1: Training in a standard learning environment
Sensors 23 04807 i001
Sensors 23 04807 i006

6. Object Transportation Using a Task Space Decomposition Method

In the previous section, we presented how a robot could learn how to transport an object in the SLE. While this worked well in the SLE, it was difficult to generalize to larger or more complex environments. In this section, we present a method for transporting objects in complex and large environments using the training results of the SLE. To achieve this, we first generated midpoints, taking into account the inner and outer vertices of the polygonal environment (Section 6.1). Then, we divided a whole task space into several sub-task spaces considering the midpoints and the size of the SLE (Section 6.2). Finally, we computed sub-goals of object transportation for each sub-task space, and the robot can transport the object via the sub-goals (Section 6.3).

6.1. Midpoint Generation

In order to divide a whole-task space into sub-task spaces, we should first compute the midpoints of the transportation path using sets of the inner and outer vertices in the test environment. Algorithm 2 shows the process of generating midpoints. First, we declared two arrays to hold the distances ( dist [ ] [ ] ) and the indices ( A [ ] ) between the inner and outer vertices (line 1–2). Second, we computed the Euclidean distances between all inner and outer vertices, and inserted the results into the dist [ ] [ ] (line 3–5). Third, we computed the outer vertex indices of the minimum distances for each inner vertex and appended them to the A [ ] (line 6–8). Finally, we computed the midpoints of the inner and outer vertices using the minimum distance pair as p m i d k (line 9–11). A list of all midpoints, P m i d , is returned as the final result (line 12). Figure 3 shows two examples of midpoint generation. A list of midpoints displayed in red dots ( P m i d = [ p m i d 1 , p m i d 2 , , p m i d N m i d ] ) is created by calculating the midpoints of the shortest line segment between the inner and outer vertices.
Algorithm 2: Midpoint generation
Sensors 23 04807 i002

6.2. Task Space Decomposition

In general, a whole-task space is larger than the SLE, as already mentioned in Section 5. If the robot is located beyond the size of its SLE, the robot will not know how to transport an object because its state (e.g., the distance from the robot to the object) has not been used for training. Therefore, the whole-task space (i.e., the test environment) should be divided into several sub-task spaces with sizes that are smaller than that of the SLE. This allows a robot to transport an object by following connected lines through midpoints for each sub-task space. However, if a connected line is longer than the size of the SLE, the line should be split into multiple shorter lines. This TSD method is described in detail in Algorithm 3.
Algorithm 3: Task space decomposition
Sensors 23 04807 i003
First, the recursive flag r f l a g is initialized to f a l s e and the Euclidean distance between the start and the end points, d ( p s t a r t , p e n d ) , is calculated (line 2–3). The maximum distance traveled between sub-goals L max w is equal to L e n v 2 in the SLE because a robot has learned to transport only by the length of L e n v 2 . If the distance d is greater than L max w , an intermediate point p v i a is calculated by dividing p s t a r t and p e n d by [ d ( p s t a r t , p e n d ) L max w ] : L max w (line 5). In this case, the recursive flag is set to t r u e for recursive calculations with respect to the remaining line (line 6). On the other hand, if the distance d is less than L max w , the intermediate point p v i a ( = p s t a r t ) and the recursive flag r f l a g are returned unchanged (line 8–9).

6.3. Sub-Goal Generation and Object Transportation

In the proposed method, a robot can transport an object by sequentially traversing multiple sub-goals. Algorithm 4 shows how to generate a list of sub-goals using the results of Algorithms 2 and 3. Using the SplitTaskspace function in Algorithm 3, we divided a sub-task space for successive midpoints by considering the size of the SLE, L max w (line 2–3). The starting midpoint p m i d i d is inserted unconditionally into the list P s u b (line 4). If the distance between p v i a s t a r t and p m i d i d n e w is greater than L max w , the TSD method is performed recursively until the remaining line is shorter than L max w (line 5–8). The final output is a list of sub-goals, P s u b (line 9).
Now, a robot can transport an object using the Q-network inference result from Algorithm 1 and a list of sub-goals from Algorithm 4. Algorithm 5 shows the object transportation process via sub-goals. First, the index of the sub-goals and the flag are initialized as initial values (line 1–2). A robot observes the state, takes an action to maximize the Q-value from the deep Q-network inference (line 4–6). When an object reaches the sub-goal by robot pushing, the sub-goal index increases by one (line 7–8). When a robot reaches all sub-goals, we consider the object transportation to have succeeded and set the flag to true (line 9–10). Finally, the success flag s f l a g is returned.
Algorithm 4: Sub-goal generation for object transportation
Sensors 23 04807 i004
Algorithm 5: Object transportation via sub-goals
Sensors 23 04807 i005

7. Simulations

7.1. Simulation Environment

We verified the proposed method using a Gazebo simulator based on the ROS [40]. The Gazebo simulator ran 60–80 times faster, and we trained the object transportation model on the Nvidia Geforce GTX-3090 with AMD Ryzen-5950X. In the simulation, the robot was TurtleBot3-waffle [41], and a pallet (1.2 m (W) × 0.8 m (H)) was used as an object. The margins of the goal ( ϵ s u c c e s s ) and the collision ( ϵ m a r g i n ) in Equation (5) are 0.3 m and 0.1 m, respectively.
The state and action parameters are presented in Table 1. The transitional velocity of 0.3 m/s is the speed at which the robot moves steadily but not too slowly, and the rotational velocity of 1.0 rad/s is the speed at which the robot can adjust its heading in the target direction without turning too sharply when turning from one side to the other, thus compromising the stability of the robot’s motion. The values of the rotational and translational velocities were chosen for the free movement of the robot. The reward of r s u c c e s s is 20 times larger than r c o l l i s i o n in absolute value. We also gave more weight to the distance difference between the goal–object than the robot–object. This is because completing the object transportation is the most important mission; it does not matter if the robot arrives at the goal or not.
The hyperparameters of the training phase are shown in Table 2. These hyperparameters were chosen to quickly and efficiently learn the object transportation method in the SLE. In particular, we varied the length of one side of the SLE to 6, 8, and 12 m. In order to compare the performance of different training space sizes, we ran the training in different SLE. In addition, the threshold of the success rate P s u c c e s s was set to 0.98 to guarantee the minimum performance of object transportation. The learning capacity of the Q-network is related to the depth and width of the layers, and thus we chose a depth and width of 4 and 256, respectively.

7.2. Training Results in Standard Learning Environments

We trained the object transportation method based on DRL in SLEs of different sizes: 6 × 6 m, 8 × 8 m, and 12 × 12 m. The goal position was fixed to the origin, and the position of the robot was randomly initialized for each episode. If the current success rate p s u c c e s s is lower than the success threshold ( P s u c c e s s = 0.98 ), we increased the episode by one unit step ( N e p = 500 ) until the current success rate reached the threshold, as already mentioned in Algorithm 1. The final training results are described in Table 3. The success rates were higher than the success rate threshold for all environments. In addition, the number of episodes needed to reach the success rate threshold increased as the size of the environment increased. This is because a robot should learn more in the large environment.
Figure 4 shows 100 trajectories of an object in different SLEs. The robot successfully learned how to transport an object in SLEs regardless of the initial positions of the robot and an object. This result will be used for transporting an object in different test environments, which will be discussed in the next section.

7.3. Test Results in Various-Shaped Environments

In the training phase, the SLEs are square environments of different sizes. These environments are designed to facilitate optimization and to improve convergence in the training phase. However, in the real world, there are many different sizes and shapes of environments. Therefore, we built our test environment with the following two aspects in mind: (1) sub-goal changes and (2) path length. For example, if a robot transports along a straight line, the path length is long, but the sub-goal changes are infrequent. Conversely, in a maze, the length of the segmented path is typically short, but the sub-goals change frequently and sequentially as the object is transported. In the real world, sub-goal changes and path length changes occur frequently, and thus the following environments are constructed to reflect this, as shown in Figure 5: (a) a long corridor; (b) a simple polygon; (c) a complex polygon; (d) a maze-shaped environment.
Table 4 shows the end-to-end DRL (i.e., the existing method) results in the differently shaped environments. The robot can be trained without any pre-processing, such as task decomposition, which is called end-to-end DRL. While the end-to-end DRL-based object transportation methods are easy to adapt to a wide variety of environments, they perform poorly in complex environments, as shown in Table 4; there are some successes with relatively simple environments, such as long corridors, but the robot cannot transport objects in complex shapes, such as polygons and mazes. However, in the case of end-to-end learning, we found that the transport distance is relatively short because we did not specify any waypoints.
Meanwhile, Figure 6, Figure 7, Figure 8 and Figure 9 show representative trajectories of an object and a robot using the proposed method for each test environment; only 10 trajectories out of a total of 200 trials are shown for easy comparison between trajectories. The left and right sides of each figure show the trajectories of an object and a robot, respectively. By comparing (a), (b), (d) or (c), (e) for each figure, we can see that the larger the size of the SLE, the less volatile the desired trajectory of the robot and the object. This means that if a robot learns how to transport in a large learning environment, the motion variations in the test environment are small. Meanwhile, for the same SLE size, the smaller the maximum distance traveled between sub-goals, the more likely the robot is to travel directly to the goal without motion variation. This is because the larger the maximum distance between sub-goals, the longer the object will stay in its current pushing direction once it starts pushing, making it harder to change the pushing direction; we can derive these conclusions by comparing (b), (c) or (d), (e), (f) for each figure. In general, it is a good result if the trajectory is plotted consistently and without deviation over several trials. For example, we can see that the case of Figure 6d has less path variability than that of (f). This indicates that Figure 6d performs better than (f) in that it is able to transport the object reliably every time.
Table 5 shows the test results in differently shaped environments using the proposed method. We tested 200 trials in four test environments using the deep Q-network learned from the differently sized SLEs. The average distance traveled by the robot and the object was recorded only if the object transportation was succeeded.
From the table, we can see the following results. First, we can see that the success rate is generally higher for 12 × 12 m. This is because the robot can also transport an object in a small environment if it has learned on a large environment. However, as shown in Table 3, the larger the training environment, the more episodes needed for training, which could lead to long training times or show poor convergence. Therefore, we should design the size of the SLE appropriately by considering the trade-off between the performance and training time. Second, the success rate increases as the maximum distance traveled between sub-goals increases. Conversely, the smaller the distance between sub-goals, the greater the risk that the object transportation will fail. This is because the greater the number of sub-goals, the more motion changes that are required to reach the sub-goals; frequent motion changes make it difficult for a robot to control precisely.

8. Discussion

In general, DRL-based object transportation has the advantage of it being easy to learn pushing motions, but it has the disadvantage of it being difficult to apply in various environments; this is because the DRL algorithm only works well within the specific conditions and environment in which it has been trained. Therefore, we solved this problem by using the task space decomposition method while maintaining the advantages of the DRL algorithm’s ease of learning how to transport an object.
Furthermore, the proposed method has shown that a robot can transport an object not only in simple environments but also in complex environments. A robot learns how to transport an object on short paths in simple environments (e.g., a square environment) and extends these results to set up multiple sub-goals to apply to larger, more complex environments, such as a long corridor, polygon, and maze. Finally, the robot transports the object by passing through the set sub-goals in sequence. In this way, it is possible to transport objects in unknown and arbitrary environments, not just the environment in which they were trained; the proposed method will be useful in the real world, where complex structures and environments exist.
On the other hand, the proposed method has some limitations. First, objects have to be transported via sub-goals, which may lead to inefficient routes. Inefficient energy consumption may occur because the minimum distance is not considered in the process of setting sub-goals. Therefore, it is necessary to consider the efficiency of the route at the sub-goal selection stage for optimal object transport. Second, depending on the environmental conditions, it may not be possible to transport with a single robot. Due to the shape of the object and the surrounding environment, there is a possibility of falling into a local minimum, so a transport solution may not exist with a single robot. In this case, a multi-robot object transport method should be considered.

9. Conclusions

This paper proposed a novel DRL-based object transportation using a task space decomposition method. First, a robot learned how to transport an object in the SLE based on DRL. Second, a whole-task space was divided into several sub-task spaces using the task space decomposition method. The learning results from the SLE can be applied to each sub-task space without modification. Finally, the robot can transport an object in arbitrary environments as well as in the SLE; the performance of the existing DRL-based object transportation method dropped dramatically when the test environment was different from the training environment or in large and complex environments. On the other hand, since our proposed method decomposes the whole environment into simple and easy environment units to transport objects, we found that it works well in arbitrary environments such as a long corridor, polygon, and maze. In the future work, we will consider various environments with static and dynamic obstacles, such as a logistics and airport environment. We will also verify the performance of the proposed algorithm on a real robot.

Funding

This research was funded by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (No. NRF-2022R1F1A1069515).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (No. NRF-2022R1F1A1069515) and the Academic Promotion System of Tech University of Korea (2022).

Conflicts of Interest

The author declares no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
RLReinforcement learning
DRLDeep reinforcement learning
SLEStandard learning environment
TSDTask space decomposition
DQNDeep Q-network

References

  1. Matsuo, I.; Shimizu, T.; Nakai, Y.; Kakimoto, M.; Sawasaki, Y.; Mori, Y.; Sugano, T.; Ikemoto, S.; Miyamoto, T. Q-bot: Heavy object carriage robot for in-house logistics based on universal vacuum gripper. Adv. Robot. 2020, 34, 173–188. [Google Scholar] [CrossRef]
  2. Koung, D.; Kermorgant, O.; Fantoni, I.; Belouaer, L. Cooperative multi-robot object transportation system based on hierarchical quadratic programming. IEEE Robot. Autom. Lett. 2021, 6, 6466–6472. [Google Scholar] [CrossRef]
  3. Hu, J.; Niu, H.; Carrasco, J.; Lennox, B.; Arvin, F. Voronoi-based multi-robot autonomous exploration in unknown environments via deep reinforcement learning. IEEE Trans. Veh. Technol. 2020, 69, 14413–14423. [Google Scholar] [CrossRef]
  4. Choi, D.; Kim, D. Intelligent multi-robot system for collaborative object transportation tasks in rough terrains. Electronics 2021, 10, 1499. [Google Scholar] [CrossRef]
  5. Eoh, G.; Choi, J.S.; Lee, B.H. Faulty robot rescue by multi-robot cooperation. Robotica 2013, 31, 1239. [Google Scholar] [CrossRef]
  6. Ordaz-Rivas, E.; Rodriguez-Liñan, A.; Torres-Treviño, L. Autonomous foraging with a pack of robots based on repulsion, attraction and influence. Auton. Robot. 2021, 45, 919–935. [Google Scholar] [CrossRef]
  7. Imtiaz, M.B.; Qiao, Y.; Lee, B. Prehensile and Non-Prehensile Robotic Pick-and-Place of Objects in Clutter Using Deep Reinforcement Learning. Sensors 2023, 23, 1513. [Google Scholar] [CrossRef]
  8. Appius, A.X.; Bauer, E.; Blöchlinger, M.; Kalra, A.; Oberson, R.; Raayatsanati, A.; Strauch, P.; Suresh, S.; von Salis, M.; Katzschmann, R.K. Raptor: Rapid aerial pickup and transport of objects by robots. In Proceedings of the 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Kyoto, Japan, 23–27 October 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 349–355. [Google Scholar]
  9. Gupta, A.; Eppner, C.; Levine, S.; Abbeel, P. Learning dexterous manipulation for a soft robotic hand from human demonstrations. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 3786–3793. [Google Scholar]
  10. Ebel, H.; Eberhard, P. Non-prehensile cooperative object transportation with omnidirectional mobile robots: Organization, control, simulation, and experimentation. In Proceedings of the 2021 International Symposium on Multi-Robot and Multi-Agent Systems (MRS), Cambridge, UK, 4–5 November 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–10. [Google Scholar]
  11. Stüber, J.; Zito, C.; Stolkin, R. Let’s Push Things Forward: A Survey on Robot Pushing. Front. Robot. AI 2020, 7, 8. [Google Scholar] [CrossRef]
  12. De Berg, M.; Gerrits, D.H. Computing push plans for disk-shaped robots. Int. J. Comput. Geom. Appl. 2013, 23, 29–48. [Google Scholar] [CrossRef]
  13. Lamiraux, F.; Mirabel, J. Prehensile manipulation planning: Modeling, algorithms and implementation. IEEE Trans. Robot. 2021, 38, 2370–2388. [Google Scholar] [CrossRef]
  14. Pasricha, A.; Tung, Y.S.; Hayes, B.; Roncone, A. PokeRRT: Poking as a Skill and Failure Recovery Tactic for Planar Non-Prehensile Manipulation. IEEE Robot. Autom. Lett. 2022, 7, 4480–4487. [Google Scholar] [CrossRef]
  15. Lee, J.; Nam, C.; Park, J.; Kim, C. Tree search-based task and motion planning with prehensile and non-prehensile manipulation for obstacle rearrangement in clutter. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 8516–8522. [Google Scholar]
  16. Wu, J.; Sun, X.; Zeng, A.; Song, S.; Rusinkiewicz, S.; Funkhouser, T. Learning pneumatic non-prehensile manipulation with a mobile blower. IEEE Robot. Autom. Lett. 2022, 7, 8471–8478. [Google Scholar] [CrossRef]
  17. Eoh, G.; Park, T.H. Cooperative Object Transportation Using Curriculum-Based Deep Reinforcement Learning. Sensors 2021, 21, 4780. [Google Scholar] [CrossRef]
  18. Kalashnikov, D.; Irpan, A.; Pastor, P.; Ibarz, J.; Herzog, A.; Jang, E.; Quillen, D.; Holly, E.; Kalakrishnan, M.; Vanhoucke, V.; et al. Scalable deep reinforcement learning for vision-based robotic manipulation. In Proceedings of the Conference on Robot Learning, Zurich, Switzerland, 29–31 October 2018; pp. 651–673. [Google Scholar]
  19. Lowrey, K.; Kolev, S.; Dao, J.; Rajeswaran, A.; Todorov, E. Reinforcement learning for non-prehensile manipulation: Transfer from simulation to physical system. In Proceedings of the 2018 IEEE International Conference on Simulation, Modeling, and Programming for Autonomous Robots (SIMPAR), Brisbane, Australia, 16–19 May 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 35–42. [Google Scholar]
  20. Smith, D.R. The design of divide and conquer algorithms. Sci. Comput. Program. 1985, 5, 37–58. [Google Scholar] [CrossRef]
  21. Manko, S.V.; Diane, S.A.; Krivoshatskiy, A.E.; Margolin, I.D.; Slepynina, E.A. Adaptive control of a multi-robot system for transportation of large-sized objects based on reinforcement learning. In Proceedings of the 2018 IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (EIConRus), Moscow and St. Petersburg, Russia, 29 January–1 February 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 923–927. [Google Scholar]
  22. Wang, Y.; De Silva, C.W. Multi-robot box-pushing: Single-agent q-learning vs. team q-learning. In Proceedings of the 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems, Beijing, China, 9–13 October 2006; IEEE: Piscataway, NJ, USA, 2006; pp. 3694–3699. [Google Scholar]
  23. Rahimi, M.; Gibb, S.; Shen, Y.; La, H.M. A comparison of various approaches to reinforcement learning algorithms for multi-robot box pushing. In Proceedings of the International Conference on Engineering Research and Applications, Tokyo, Japan, 22–24 August 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 16–30. [Google Scholar]
  24. Wang, Y.; de Silva, C.W. Sequential Q-Learning with Kalman Filtering for Multirobot Cooperative Transportation. IEEE/ASME Trans. Mechatron. 2009, 15, 261–268. [Google Scholar] [CrossRef]
  25. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef]
  26. Silver, D.; Huang, A.; Maddison, C.J.; Guez, A.; Sifre, L.; Van Den Driessche, G.; Schrittwieser, J.; Antonoglou, I.; Panneershelvam, V.; Lanctot, M.; et al. Mastering the game of Go with deep neural networks and tree search. Nature 2016, 529, 484–489. [Google Scholar] [CrossRef]
  27. Zhang, L.; Sun, Y.; Barth, A.; Ma, O. Decentralized Control of Multi-Robot System in Cooperative Object Transportation Using Deep Reinforcement Learning. IEEE Access 2020, 8, 184109–184119. [Google Scholar] [CrossRef]
  28. Xiao, Y.; Hoffman, J.; Xia, T.; Amato, C. Learning Multi-Robot Decentralized Macro-Action-Based Policies via a Centralized Q-Net. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 10695–10701. [Google Scholar]
  29. Eoh, G.; Park, T.H. Automatic Curriculum Design for Object Transportation Based on Deep Reinforcement Learning. IEEE Access 2021, 9, 137281–137294. [Google Scholar] [CrossRef]
  30. Shibata, K.; Jimbo, T.; Matsubara, T. Deep reinforcement learning of event-triggered communication and consensus-based control for distributed cooperative transport. Robot. Auton. Syst. 2023, 159, 104307. [Google Scholar] [CrossRef]
  31. Yang, H.Y.; Wong, S.K. Agent-based cooperative animation for box-manipulation using reinforcement learning. Proc. ACM Comput. Graph. Interact. Tech. 2019, 2, 1–18. [Google Scholar] [CrossRef]
  32. Rybak, L.; Behera, L.; Averbukh, M.; Sapryka, A. Development of an algorithm for managing a multi-robot system for cargo transportation based on reinforcement learning in a virtual environment. In Proceedings of the IOP Conference Series: Materials Science and Engineering, Chennai, India, 16–17 September 2020; IOP Publishing: Bristol, UK, 2020; Volume 945, p. 012083. [Google Scholar]
  33. Kawano, H. Hierarchical sub-task decomposition for reinforcement learning of multi-robot delivery mission. In Proceedings of the 2013 IEEE International Conference on Robotics and Automation, Karlsruhe, Germany, 6–10 May 2013; IEEE: Piscataway, NJ, USA, 2013; pp. 828–835. [Google Scholar]
  34. Mesesan, G.; Roa, M.A.; Icer, E.; Althoff, M. Hierarchical path planner using workspace decomposition and parallel task-space rrts. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–9. [Google Scholar]
  35. Zhang, X.; Yan, L.; Lam, T.L.; Vijayakumar, S. Task-space decomposed motion planning framework for multi-robot loco-manipulation. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 8158–8164. [Google Scholar]
  36. Fosong, E.; Rahman, A.; Carlucho, I.; Albrecht, S.V. Learning Complex Teamwork Tasks using a Sub-task Curriculum. arXiv 2023, arXiv:2302.04944. [Google Scholar]
  37. Watkins, C.J.; Dayan, P. Q-learning. Mach. Learn. 1992, 8, 279–292. [Google Scholar] [CrossRef]
  38. Yogeswaran, M.; Ponnambalam, S. Reinforcement learning: Exploration–exploitation dilemma in multi-agent foraging task. Opsearch 2012, 49, 223–236. [Google Scholar] [CrossRef]
  39. Lin, L.J. Self-improving reactive agents based on reinforcement learning, planning and teaching. Mach. Learn. 1992, 8, 293–321. [Google Scholar] [CrossRef]
  40. Takaya, K.; Asai, T.; Kroumov, V.; Smarandache, F. Simulation environment for mobile robots testing using ROS and Gazebo. In Proceedings of the 2016 20th International Conference on System Theory, Control and Computing (ICSTCC), Sinaia, Romania, 13–15 October 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 96–101. [Google Scholar]
  41. Amsters, R.; Slaets, P. Turtlebot 3 as a robotics education platform. In Proceedings of the Robotics in Education: Current Research and Innovations 10, Vienna, Austria, 10–12 April 2019; Springer: Berlin/Heidelberg, Germany, 2020; pp. 170–181. [Google Scholar]
Figure 1. Overview of DRL-based object transport using the TSD method. (a) A robot learns how to transport an object to a goal in an SLE. A deep Q-network (DQN) is used for training. (b) A whole-task space is decomposed into sub-task spaces by considering the size of the SLE. The size of the sub-task spaces cannot exceed the size of the SLE. For each sub-task space, sub-goals are generated by considering the maximum traveled distance of a robot. (c) The object is transported via the sub-goals by the pushing motion of a robot.
Figure 1. Overview of DRL-based object transport using the TSD method. (a) A robot learns how to transport an object to a goal in an SLE. A deep Q-network (DQN) is used for training. (b) A whole-task space is decomposed into sub-task spaces by considering the size of the SLE. The size of the sub-task spaces cannot exceed the size of the SLE. For each sub-task space, sub-goals are generated by considering the maximum traveled distance of a robot. (c) The object is transported via the sub-goals by the pushing motion of a robot.
Sensors 23 04807 g001
Figure 2. Object transport training process in a square SLE.
Figure 2. Object transport training process in a square SLE.
Sensors 23 04807 g002
Figure 3. (a,b) Two examples of midpoints and sub-goal generation using the task space decomposition method.
Figure 3. (a,b) Two examples of midpoints and sub-goal generation using the task space decomposition method.
Sensors 23 04807 g003
Figure 4. One hundred transportation trajectories of an object in different sizes of SLEs: (a) 6 × 6 m environment; (b) 8 × 8 m environment; (c) 12 × 12 m environment.
Figure 4. One hundred transportation trajectories of an object in different sizes of SLEs: (a) 6 × 6 m environment; (b) 8 × 8 m environment; (c) 12 × 12 m environment.
Sensors 23 04807 g004
Figure 5. Various simulation environments for the testing of the proposed method: (a) long corridor; (b) simple polygon; (c) complex polygon; (d) maze.
Figure 5. Various simulation environments for the testing of the proposed method: (a) long corridor; (b) simple polygon; (c) complex polygon; (d) maze.
Sensors 23 04807 g005
Figure 6. Ten trajectories of object transportation in a long corridor environment (Figure 5a). The left and right sides of each figure represent the trajectory of an object and a robot, respectively. (a) L e n v = 6 m and L max w = 2 m; (b) L e n v = 8 m and L max w = 2 m; (c) L e n v = 8 m and L max w = 4 m; (d) L e n v = 12 m and L max w = 2 m; (e) L e n v = 12 m and L max w = 4 m; (f) L e n v = 12 m and L max w = 6 m.
Figure 6. Ten trajectories of object transportation in a long corridor environment (Figure 5a). The left and right sides of each figure represent the trajectory of an object and a robot, respectively. (a) L e n v = 6 m and L max w = 2 m; (b) L e n v = 8 m and L max w = 2 m; (c) L e n v = 8 m and L max w = 4 m; (d) L e n v = 12 m and L max w = 2 m; (e) L e n v = 12 m and L max w = 4 m; (f) L e n v = 12 m and L max w = 6 m.
Sensors 23 04807 g006
Figure 7. Ten trajectories of object transportation in a simple polygon environment (Figure 5b). The left and right sides of each figure represent the trajectory of an object and a robot, respectively. (a) L e n v = 6 m and L max w = 2 m; (b) L e n v = 8 m and L max w = 2 m; (c) L e n v = 8 m and L max w = 4 m; (d) L e n v = 12 m and L max w = 2 m; (e) L e n v = 12 m and L max w = 4 m; (f) L e n v = 12 m and L max w = 6 m.
Figure 7. Ten trajectories of object transportation in a simple polygon environment (Figure 5b). The left and right sides of each figure represent the trajectory of an object and a robot, respectively. (a) L e n v = 6 m and L max w = 2 m; (b) L e n v = 8 m and L max w = 2 m; (c) L e n v = 8 m and L max w = 4 m; (d) L e n v = 12 m and L max w = 2 m; (e) L e n v = 12 m and L max w = 4 m; (f) L e n v = 12 m and L max w = 6 m.
Sensors 23 04807 g007
Figure 8. Ten trajectories of object transportation in a complex polygon environment (Figure 5c). The left and right sides of each figure represent the trajectory of an object and a robot, respectively. (a) L e n v = 6 m and L max w = 2 m; (b) L e n v = 8 m and L max w = 2 m; (c) L e n v = 8 m and L max w = 4 m; (d) L e n v = 12 m and L max w = 2 m; (e) L e n v = 12 m and L max w = 4 m; (f) L e n v = 12 m and L max w = 6 m.
Figure 8. Ten trajectories of object transportation in a complex polygon environment (Figure 5c). The left and right sides of each figure represent the trajectory of an object and a robot, respectively. (a) L e n v = 6 m and L max w = 2 m; (b) L e n v = 8 m and L max w = 2 m; (c) L e n v = 8 m and L max w = 4 m; (d) L e n v = 12 m and L max w = 2 m; (e) L e n v = 12 m and L max w = 4 m; (f) L e n v = 12 m and L max w = 6 m.
Sensors 23 04807 g008
Figure 9. Ten trajectories of object transportation in a maze environment (Figure 5d). The left and right sides of each figure represent the trajectory of an object and a robot, respectively. (a) L e n v = 6 m and L max w = 2 m; (b) L e n v = 8 m and L max w = 2 m; (c) L e n v = 8 m and L max w = 4 m; (d) L e n v = 12 m and L max w = 2 m; (e) L e n v = 12 m and L max w = 4 m; (f) L e n v = 12 m and L max w = 6 m.
Figure 9. Ten trajectories of object transportation in a maze environment (Figure 5d). The left and right sides of each figure represent the trajectory of an object and a robot, respectively. (a) L e n v = 6 m and L max w = 2 m; (b) L e n v = 8 m and L max w = 2 m; (c) L e n v = 8 m and L max w = 4 m; (d) L e n v = 12 m and L max w = 2 m; (e) L e n v = 12 m and L max w = 4 m; (f) L e n v = 12 m and L max w = 6 m.
Sensors 23 04807 g009
Table 1. State and action parameters of the proposed object transportation.
Table 1. State and action parameters of the proposed object transportation.
CategoryDescriptionIndexValue
Action
(Equation (4))
Rotational velocity v r o t 1.0 rad/s
Translational velocity v t r a n s 0.3 m/s
Reward
(Equation (5))
Success reward r s u c c e s s 1.0
Collision reward r c o l l i s i o n −0.05
Weight w 1 1.0
w 2 0.1
Table 2. Hyperparameters of the training phase.
Table 2. Hyperparameters of the training phase.
SourceDescriptionIndexValue
Training
(Algorithm 1)
Unit size of episode N e p 500
Length of one side of the SLE L e n v 6/8/12 m
Number of attempts for calculating success rate N t e s t 30
Step size for each episode T s t e p 1000
Number of training iterations per episode N i t e r 200
Update period of target Q-network T K 32
Threshold of the success rate P s u c c e s s 0.98
Q-Network
(Figure 2)
Depth of Q-network-4
Width of each layer-256
Activation function-ReLU
Learning rate α 0.001
Discount factor γ 0.99
Batch size-512
Replay memory size# of D 1.0 × 10 7
Initial exploration probability in ϵ -greedy-1.0
Final exploration probability in ϵ -greedy-0.1
Table 3. Results in the standard learning environment.
Table 3. Results in the standard learning environment.
Environment Size
( L env × L env )
The Number of Success/TrialSuccess Rate
( p success )
The Number of
Training Episodes
( n ep )
6 × 6 m100/100100%2000
8 × 8 m99/10099%4000
12 × 12 m100/100100%8000
Table 4. Results in the end-to-end deep reinforcement learning method (the existing method).
Table 4. Results in the end-to-end deep reinforcement learning method (the existing method).
EnvironmentThe Number of
Success/Trial
(Success Rate)
Avg. Travelling Distance of an ObjectAvg. Travelling Distance of a Robot
Long corridor
(Figure 5a)
133/200 (0.67)23.59 m23.21 m
Simple polygon
(Figure 5b)
0/200 (0.0)--
Complex polygon
(Figure 5c)
0/200 (0.0)--
Maze (Figure 5d)0/200 (0.0)--
Table 5. Test results in differently shaped environments.
Table 5. Test results in differently shaped environments.
EnvironmentSize of Standard Learning Environment ( L env × L env )Maximum Distance Traveled between Sub-Goals ( L max w )The Number of Success/Trial (Success Rate)Avg. Travelling Distance of an ObjectAvg. Travelling Distance of a Robot
Long corridor
(Figure 5a)
6 m × 6 m2 m127/200 (0.64)37.47 m36.42 m
8 m × 8 m2 m143/200 (0.71)29.55 m28.86 m
4 m193/200 (0.96)26.94 m26.02 m
12 m × 12 m2 m131/200 (0.66)27.82 m27.82 m
4 m183/200 (0.92)20.20 m19.75 m
6 m200/200 (1.00)19.93 m19.33 m
Simple polygon
(Figure 5b)
6 m × 6 m2 m196/200 (0.98)27.45 m27.00 m
8 m × 8 m2 m196/200 (0.98)22.61 m22.25 m
4 m146/200 (0.73)22.25 m22.39 m
12 m × 12 m2 m190/200 (0.95)21.32 m21.11 m
4 m195/200 (0.97)17.15 m16.75 m
6 m191/200 (0.95)18.08 m17.88 m
Complex polygon
(Figure 5c)
6 m×6 m2 m150/200 (0.75)59.97 m61.67 m
8 m × 8 m2 m166/200 (0.83)59.53 m61.10 m
4 m166/200 (0.83)56.72 m57.46 m
12 m × 12 m2 m179/200 (0.90)54.33 m55.47 m
4 m177/200 (0.89)47.75 m48.78 m
6 m157/200 (0.79)56.71 m57.14 m
Maze
(Figure 5d)
6 m × 6 m2 m192/200 (0.96)151.06 m150.05 m
8 m × 8 m2 m198/200 (0.99)125.18 m124.08 m
4 m135/200 (0.68)134.40 m133.41 m
12 m × 12 m2 m200/200 (1.00)116.70 m115.58 m
4 m200/200 (1.00)102.38 m101.46 m
6 m196/200 (0.98)98.57 m97.91 m
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Eoh, G. Deep-Reinforcement-Learning-Based Object Transportation Using Task Space Decomposition. Sensors 2023, 23, 4807. https://doi.org/10.3390/s23104807

AMA Style

Eoh G. Deep-Reinforcement-Learning-Based Object Transportation Using Task Space Decomposition. Sensors. 2023; 23(10):4807. https://doi.org/10.3390/s23104807

Chicago/Turabian Style

Eoh, Gyuho. 2023. "Deep-Reinforcement-Learning-Based Object Transportation Using Task Space Decomposition" Sensors 23, no. 10: 4807. https://doi.org/10.3390/s23104807

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop