Next Article in Journal
Analytical Modeling of Metamaterial Differential Transmission Line Using Corrugated Ground Planes in High-Speed Printed Circuit Boards
Previous Article in Journal
Distributed Magnetic Flux Density on the Cross-Section of a Transformer Core
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cooperative Carrying Control for Multi-Evolutionary Mobile Robots in Unknown Environments

1
Institute of Electrical and Control Engineering, National Chiao Tung University, Hsinchu 300, Taiwan
2
Department of Computer Science and Information Engineering, National Chin-Yi University of Technology, Taichung 411, Taiwan
*
Author to whom correspondence should be addressed.
Electronics 2019, 8(3), 298; https://doi.org/10.3390/electronics8030298
Submission received: 13 February 2019 / Revised: 1 March 2019 / Accepted: 1 March 2019 / Published: 6 March 2019
(This article belongs to the Section Artificial Intelligence)

Abstract

:
This study provides an effective cooperative carrying and navigation control method for mobile robots in an unknown environment. The manager mode switches between two behavioral control modes—wall-following mode (WFM) and toward-goal mode (TGM)—based on the relationship between the mobile robot and the unknown environment. An interval type-2 fuzzy neural controller (IT2FNC) based on a dynamic group differential evolution (DGDE) is proposed to realize the carrying control and WFM control for mobile robots. The proposed DGDE uses a hybrid method that involves a group concept and an improved differential evolution to overcome the drawbacks of the traditional differential evolution algorithm. A reinforcement learning strategy was adopted to develop an adaptive WFM control and achieve cooperative carrying control for mobile robots. The experimental results demonstrated that the proposed DGDE is superior to other algorithms at using WFM control. Moreover, the experimental results demonstrate that the proposed method can complete the task of cooperative carrying, and can realize navigation control to enable the robot to reach the target location.

1. Introduction

Mobile robot control has been widely used in several applications, such as navigation, obstacle avoidance, path planning, and cooperative transport. To enhance the robot control quality, Cupertino [1] adopted a fuzzy controller. The fuzzy controller possesses robustness and an anti-noise ability; the controller can identify and calculate signals with uncertainties. However, to design an applicable fuzzy controller, designers must spend a considerable amount of time analyzing the experimental input and output data of a mobile robot. Thus, machine-learning technology has gradually attracted considerable research attention. Zhu and Yang [2] and Rusu et al. [3] used supervised learning methods to adjust the parameters of the if–then rules in fuzzy neural networks by training data. The disadvantages of supervised learning [2,3] are that it is difficult to collect training data in advance and obtain precise training data.
Recently, reinforcement learning [4] has been widely used in control applications for mobile robots. The method not only can automatically construct a complete fuzzy neural network in the absence of precise training data but also adjust the parameters of the system through the machine leaning algorithm to complete the navigation task. Therefore, designing a mobile robot using evolutionary computing in unknown environments has become a topic of interest. Hsu and Juang [5] implemented a wall-following control using reinforcement learning. A fitness function was used to evaluate a robot’s movement in an unknown environment, and an optimal fitness value was adopted as the reinforcement signal. Anish and Parhi [6] used sensors to measure the distance between a robot and an obstacle. In [6], measured distances were used as inputs, and steering angle and speed were used as outputs for the adaptive network-based fuzzy inference system controller. Juang and Chang [7] adopted four sonar sensors and used the data from these sensors as input signals. In this application, the speeds of the left and right wheel are the outputs of the fuzzy neural network controller. The mobile robot was trained in the training environment by adjusting the parameters of the fuzzy neural network controller. Although the aforementioned methods can be used to successfully complete the navigation task, their performances are not optimal. Because the robot is in a real environment, the input signal contains uncertainties due to noise interferences from the sensors.
Although some researchers [8,9] have used the type-1 fuzzy set to solve uncertain problems, the control performance in the real environment is not optimal. Therefore, Kim and Chwa [10] used an interval type-2 fuzzy set for solving uncertain problems. The membership values of fuzzy sets provide a footprint of uncertainty (FOU) to deal with uncertainties. Thus, the controller design is more flexible and has improved error tolerance. Castillo and Melin [11] reduced the computational complexity of an interval type-2 fuzzy system.
Recently, researchers have used evolutionary algorithms for solving the parameter optimization problem, such as particle swarm optimization (PSO) [12], ant colony optimization (ACO) [13], differential evolution (DE) [14], the artificial bee colony (ABC) algorithm [15], and bacterial foraging optimization (BFO) [16]. In recent years, DE has been widely used in various fields [17,18] and has the advantages of a simple structure, reduced parameter setting requirements, and superior problem solving ability. However, the traditional DE method has a disadvantage, in that it can easily become trapped in a local optimal solution. To eliminate this disadvantage, an improved DE is proposed for solving mobile robot control problems.
The major contributions of this paper are described as follows. First, an efficient interval type-2 fuzzy neural controller (IT2FNC) based on dynamic group differential evolution (DGDE) was designed to implement the carrying control and wall-following mode (WFM) control for mobile robots. In the proposed IT2FNC, a functional link neural network (FLNN) with a nonlinear combination input was added in consequent part of a fuzzy rule. Second, the proposed DGDE used a hybrid method that involves a group concept and an improved DE to overcome the drawbacks of easily trapped into local optimal in the traditional DE algorithm. Second, a manager mode was developed to assist mobile robots in navigation control. Third, the manager mode switched to WFM or toward-goal mode (TGM), based on the relationship between the mobile robot and the unknown environment. Fourth, the proposed control method can complete the task of cooperative carrying and can realize navigation control to enable the robot to reach a target location. Moreover, the experimental results demonstrated that the proposed DGDE learning algorithm is superior to other algorithms at using WFM control.

2. Mobile Robot Specifications

In this study, the e-puck mobile robot developed by the Ecole Polytechnique Fédérale de Lausanne, was adopted, as displayed in Figure 1a. The mobile robot has been widely used in studies, such as in embedded computing, signal processing, inter-robot communication, robot control, feature extraction from images and sounds, swarm intelligence, and cooperative behavior in robotics.
The e-puck mobile robot is a two-wheeled mobile robot with an axle diameter of 4.12 cm and a maximum speed of 15 cm/s. The infrared sensor has a sensing range of approximately 360° and is symmetrical. The sensor performs tasks such as object detection, distance detection, and obstacle avoidance. The sensors S0–S3 that are on the right side of the robot are mounted at 10°, 45°, 90°, and 135°, respectively. Each sensor can detect a distance between 1 and 6 cm, as shown in Figure 1b.

3. Proposed Type-2 Fuzzy Controller Based on an Evolutionary Algorithm

An IT2FNC was proposed to realize wall-following control. The associated DGDE learning algorithm can be used to adjust the parameters of the IT2FNC.

3.1. Interval Type-2 Fuzzy Neural Controller

In this section, the structure of the IT2FNC is introduced. Figure 2 displays the structure of the IT2FNC. X 1 , , X n represents the input of IT2FNC, whereas Y L e f t and Y R i g h t represent the left and right wheel speed of the robot, respectively. To reduce the computational complexity of the order reduction during defuzzification, this study adopted the centers of sets (COS) [11,19] to conduct the order reduction process. A functional link neural network (FLNN) with a nonlinear combination input was added in consequent part of a fuzzy rule [20]. Figure 2 presents the five-layer structure of an IT2FNC. The IT2FNC consists of an input layer, a membership function layer, a firing layer, a consequent layer, and an output layer. The if–then rule can be expressed as follows:
Rule   j :   IF   x 1   is   A ˜ 1 j   and   x 2   is   A ˜ 2 j   and   x i   is   A ˜ i j   and   x n   is   A ˜ n j THEN   y j   =   k = 1 M ω k j k , =   ω 1 j 1 + ω 2 j 2 + + ω M j M ,
where n is the number of inputs, x i represents the ith input, y j denotes the output of the jth fuzzy rule, A ˜ 1 j , A ˜ 2 j , , A ˜ n j represent the interval type-2 fuzzy sets, ω k j is the link weight, k represents the basis trigonometric function, and M is the number of basis functions.
The five-layer structure of the IT2FNC is described as follows:
Layer 1 (input layer): This layer only imports the input data into the next layer:
u i ( 1 ) = x i .
Layer 2 (membership function layer): This layer performs the fuzzification. Each node in this layer defines an interval type-2 fuzzy set, as displayed in Figure 3. The Gaussian primary membership function has an uncertainty mean [ m i j 1 ,   m i j 2 ] and standard deviation σ i j and is expressed as follows:
u i j ( 2 ) = e x p ( [ u i ( 1 ) m i j ] 2 σ i j 2 ) N ( m i j ,   σ i j ;   u i ( 1 ) ) ,   m i j [ m i j 1 ,   m i j 2 ] .
The membership degree of the Gaussian primary membership function u i j ( 2 ) is called the footprint of uncertainty (FOU) and is expressed as the upper bound u ¯ i j ( 2 ) and the lower bound u _ i j ( 2 ) . The membership degree is expressed as follows:
u ¯ i j ( 2 ) ( u i ( 1 ) ) = { N ( m i j 1 ,   σ i j ;   u i ( 1 ) ) , i f      u i ( 1 ) < m i j 1 1 , i f      m i j 1 u i ( 1 ) m i j 2 , N ( m i j 2 ,   σ i j ;   u i ( 1 ) ) , i f       u i ( 1 ) > m i j 2
and
u _ i j ( 2 ) ( u i ( 1 ) ) = { N ( m i j 2 ,   σ i j ;   u i ( 1 ) ) ,   i f       u i ( 1 ) m i j 1 + m i j 2 2 N ( m i j 1 ,   σ i j ;   u i ( 1 ) ) ,   i f        u i ( 1 ) > m i j 1 + m i j 2 2 .
The output of each node is represented by the following interval: [ u ¯ i j ( 2 ) , u _ i j ( 2 ) ] .
Layer 3 (firing layer): Each node is a rule node and uses an algebraic product operation to obtain the firing strengths u ¯ j ( 3 )   and   u _ j ( 3 ) of each rule node. The firing strength of each rule node is defined as follows:
u ¯ j ( 3 ) = i u ¯ i j ( 2 )   and   u _ j ( 3 ) = i u _ i j ( 2 ) ,
where i u ¯ i j ( 2 ) and i u _ i j ( 2 ) represent the firing strength of the interval’s upper bound and lower bound, respectively.
Layer 4 (consequent layer): Interval type-2 fuzzy sets are reduced to interval type-1 fuzzy sets through a type-reduction operation. The traditional type-2 order reduction method is based on highly complex calculation. Therefore, the centers of sets (COS) method [11] is adopted for implementing the reduction process and is described as follows:
Y ( x ) = [ y l , y r ] = a 1   a M     f 1 [ f _ 1 , f ¯ 1 ]   f M [ f _ M , f ¯ M ]     1 i = 1 M f i a i i = 1 M f i ,
u _   ( 4 ) = y l = j = 1 R u _ j ( 3 ) ( k = 1 M ω k j ϕ k ) j = 1 R u _ j ( 3 ) ,
and
u ¯   ( 4 ) = y r = j = 1 R u ¯ j ( 3 ) ( k = 1 M ω k j ϕ k ) j = 1 R u ¯ j ( 3 ) ,
where k = 1 M ω k j ϕ k represents a nonlinear combination of FLNN inputs, ω k j represents the link weight, and k represents the functional expansion of FLNN inputs. The functional expansion is based on basis trigonometric functions and defined as follows:
[ ϕ 1 , ϕ 2 , , ϕ M ] = [ x 1 , sin ( π x 1 ) , cos ( π x 1 ) , , x n , sin ( π x n ) , cos ( π x n ) ] ,
where M = 3 × n is the number of basis functions and n is the number of inputs.
Layer 5 (output layer): The node output of layer 5 is defuzzified by computing the average of u ¯   ( 4 ) and u _   ( 4 ) . The crisp value y is obtained as follows:
y = u ¯ ( 4 ) + u _ ( 4 ) 2 = u ( 5 ) .

3.2. Proposed DGDE

DE has the advantages of fast convergence and simple implementation. However, some problems, such as low precision and becoming easily trapped into local optima, are encountered when complex problems are solved. Therefore, an efficient DGDE algorithm is proposed to overcome the shortcomings of traditional DE. The steps of DGDE are described as follows:
Step 1: Initialization and coding
By setting the parameter vectors of DE and randomly initializing the target vector X i , G , the mathematical model is expressed as follows:
X i , G = [ x i , G 1 ,   x i , G 2 , ,   x i , G D ]
where i = 1 , 2 , , N P ; N P is the number of population; X i , G represents the ith parameter vector in Gth generation, and D is the number of dimensions.
The proposed DGDE is used to adjust the parameters of the IT2FNC. All the parameters in the IT2FNC are coded into one vector. Each vector represents an IT2FNC. The adjustable parameters in each IT2FNC are the uncertainty mean m i j , standard deviation σ i j , displacement value of the uncertainty mean d i j , and link weight ω k j . Where the d i j = m i j 2 m i j 1 > 0 . The coding format is presented in Figure 4.
Step 2: Vector group
The fitness values of all parameter vectors are sorted in descending order. The initial group number of all vectors is set to zero (Figure 5).
The computation of the fitness values for the individuals in the population will be discussed later, see Equations (24)–(30). The highest fitness value of the vectors is set as the new group leader. This implies that the group number is updated from zero to one, as shown in Figure 6. On the basis of the average distance difference and the average fitness difference between these ungrouped vectors (i.e., group number 0) and the group leader, the threshold value of similarity comprises the threshold values of fitness and distance:
D I S g =   i = 1 N P j = 1 D ( L j g X j i ) 2   ,   i f   X   i   is   ungrouped ,
F I T g = i = 1 N P | F i t ( L   g ) F i t ( X   i ) |   ,   i f   X   i   is   ungrouped ,
A v e r a g e _ D i s t a n c e ( A D I S g ) = D I S g N I ,
A v e r a g e _ F i t n e s s   ( A F I T g ) = F I T g N I ,
where D represents the encoded dimension, N P denotes the number of parameter vectors, L j g is the jth dimension of the gth group leader, N I is the total number of ungrouped vectors, and A D I S g and A F I T g represent the distance threshold value and the fitness threshold value of the gth group, respectively.
The difference of the distance values ( D i s i ) and the difference of the fitness values F i t i between the ungrouped vectors and the leader vectors are calculated as follows:
D i s i =   j = 1 D ( L j g X j i ) 2 ,
F i t i = | F i t ( L g ) F i t ( X i ) |
If the D i s i <   A D I S g and F i t i <   A F I T g conditions are satisfied for a vector, then the vector is similar to the gth group leader. Therefore, these vectors are grouped together and the group number is updated to g . Otherwise, the vectors are not grouped together. Figure 7 shows that the vectors X2,G and Xi,G are similar to the 1st group leader.
Next, if any ungrouped vectors exist, the ungrouped vector with the highest fitness value is set as the group leader in a new group. Figure 8 presents the vector X3,G from the ungrouped vectors that has the highest fitness value and is set as a new group leader (i.e., the 2nd group leader). The grouping process is completed when no ungroup vectors exist.
Step 3: Mutation
Two new mutation methods are proposed for improving the disadvantages of the traditional DE algorithms. The modified formula is expressed as follows:
Mutation method 1:
U i , G + 1 = X r L , G + F ( X r 1 , G X r 2 , G )  
Mutation method 2:
U i , G + 1 = X b e s t ,   G + F ( X r L , G X r 1 , G ) + F ( X r 2 , G X r 3 , G )
where U i , G + 1 = [ u i , G + 1 1 ,   u i , G + 1 2 , ,   u i , G + 1 D ] , F is the mutation weight factor. Moreover, the weight factor is set to 0.5 as a general rule, X b e s t ,   G is the best fitness vector, and X r L , G is a random leader selected from all the group leaders. Because the traditional DE method easily falls into local optima, this study employed the random leader as the base vector to increase the search ability effectively, as presented in Equation (19). In Equation (20), the best vector is set as the base vector, and three random vectors and one random leader vector are used. The mutated vector in Equation (20) revolves around the best vector and enhances the search ability in the solution space.
Step 4: Recombination
The recombination operation crosses the mutation vector with the target vector and generates a new vector V i , G + 1 . The definition of V i , G + 1 is expressed as follows:
V i , G + 1 = [ v i , G + 1 1 ,   v i , G + 1 2 , ,   v i , G + 1 D ]
v i ,   G + 1 j = { u i , G + 1 j ,           i f   r a n d j ( 0 , 1 ) C R x i , G + 1 j ,     o t h e r w i s e   j = 1 D
where r a n d j ( 0 , 1 ) represents random values between zero and one in jth dimension, and CR is the crossover rate. The higher CR value represents the higher similarity between the vector and the mutation vector.
Step 5: Selection
The fitness values of vectors are evaluated for selecting the target vectors in the next generation. If the fitness value of a vector is less favorable than the current target vector, the target vector will remain in the next generation. The selection operation is described as follows:
X i , G + 1 = { V i , G + 1   ,   i f   F i t ( V i , G + 1 ) > F i t ( X i , G ) X i , G , o t h e r w i s e
The flowchart of DGDE is presented in Figure 9.

3.3. Wall-Following Control of Mobile Robots

Recently, some researchers have proposed the wall-following control of mobile robots for using reinforcement learning. Jhang et al. [21] used an interval type-2 recurrent fuzzy cerebellar model articulation controller (IT2RFCMAC). They adopted a Takagi–Sugeno–Kang (TSK) in the consequent part in IT2RFCMAC. The TSK is a linear function and simple implementation. However, in complex problems, nonlinear functions are better in terms of performance than linear functions. Therefore, in this study a nonlinear functional link neural network of the proposed IT2FNC was used as the consequent part to improve control performance. The proposed IT2FNC that based on DGDE is demonstrated. Wall-following control was utilized by the IT2FNC to control the mobile robot. A reinforcement learning strategy was used to adjust the controller parameters of the proposed IT2FNC. The block diagram for the wall-following control of the mobile robot is presented in Figure 10. The four input signals of the proposed IT2FNC are the S 0 , S 1 , S 2 , and S 3   distances , which are measured by the infrared sensor. The outputs of the IT2FNC are the rotational speeds VL and VR of the two wheels.
To allow mobile robots to be used in different environments, the training environment in this study featured straight lines, corners, right-angled corners, and slope lines. Figure 11 presents the 1.7 × 1.6 m2 training environment.
To avoid collision with obstacles and deviation from the wall during the wall-following control learning process, three terminal conditions of wall-following control learning were specified:
  • If the total moving distance of the mobile robot was larger than the predefined maximal distance of the training environment, the mobile robot successfully moved in a circular path in an unknown environment.
  • The mobile robot collided with the wall when the measured distance from any infrared sensor was less than 1 cm, as displayed in Figure 12a.
  • The mobile robot deviated from the wall when the measured distance S 2 was greater than 6 cm, as displayed in Figure 12b.
The proposed DGDE was used to train the parameters of the IT2FNC. Each vector in the DGDE represents a solution of an IT2FNC. When any terminal condition during the wall-following control learning process is satisfied, a fitness function was used to evaluate the performance of the mobile robot. The proposed fitness function comprises three sub-fitness functions— S F 1 , S F 2 , and S F 3 . S F 1 , S F 2 , and S F 3 represent the total moving distance, the distance between the robot and the wall, and the degree of parallelism between the robot and the wall, respectively. The three sub-fitness functions are defined as follows:
(1)
S F 1 : If the moving distance R d i s was greater than the predefined value R s t o p , the robot successfully moved around a circular path in the training environment and set R d i s = R s t o p . The sub-fitness function S F 1 is defined as follows:
S F 1 = R s t o p R d i s
(2)
S F 2 : The goal of the wall-following control was to maintain a fixed distance between the side of the robot and the wall. Therefore, the sub-fitness function S F 2 is defined as the average of W D ( t ) , where W D ( t ) represents the distance between the side of the robot and the wall at each time step, and is defined as follows:
W D ( t ) = | S 2 ( t ) d w a l l |
S F 2 = t = 1 T s t o p W D ( t ) T s t o p
where d w a l l is the pre-defined fixed distance ( i . e . ,   d w a l l = 4 cm), as presented in Figure 13a. T s t o p is the total number of time steps in a learning process. If the robot remains at a fixed distance from the wall, the S F 2 value is equal to zero.
(3)
S F 3 : This sub-fitness function was used for evaluating the degree of parallelism between the robot and the wall. If the robot was parallel to the wall, the angle θ between the robot and wall was 90°. On the basis of the law of cosines, x ( t ) must have the same value as that of R S 2 , as presented in Figure 13b.
R S 1 = r + δ 1   , R S 2 = r + δ 2  
x ( t ) =   R S 1 2 +   R S 2 2   2 R S 1 R S 2 cos ( 45 ° )
where r is the radius of the robot; δ 1 and δ 2 represent the distance between the sensor 1 and the wall and that between sensor 2 and the wall, respectively, and S F 3 represents the average value of the degree of parallelism during movement. If the robot is parallel to the wall, S F 3 is equal to zero.
S F 3 = t = 1 T s t o p | R S 2 x ( t ) | T s t o p
Therefore, the proposed fitness function is defined as follows:
F ( · ) = 1 1 + ( S F 1 + S F 2 + S F 3 )
Figure 14 shows the block diagram of learning process of wall-following control. Each solution represents an IT2FNC controller, which is evaluated by the fitness function. Additionally, it automatically adjust parameters of IT2FNC using evolutionary strategies. The best solution will be replaced when a better solution exits in each generation.

3.4. Experimental Results of the Wall-Following Control

To verify the effectiveness of the proposed method, the performance of the WFM controller while using the proposed DGDE-1 (mutation method 1) and DGDE-2 (mutation method 2) were compared with the performance of the WFM control while using other methods. Each method was evaluated 10 times to verify the stability of each algorithm.
The initial parameters of the DGDE are the number of the population (NP), crossover rate, generation, weighting factor of mutation, and number of fuzzy rules, as presented in Table 1. Moreover, we considered different fuzzy rule numbers for performance evaluation. Table 2 presents the performance evaluation results of different fuzzy rule numbers. The IT2FNC with six fuzzy rules was more efficient than those with five and seven fuzzy rules.
Table 3 presents the performance evaluations of different algorithms. In this table, the performance indexes include the best fitness function, the worst fitness function, the average fitness function, the standard deviation (STD), the number of successful runs, and computation time for one training. The number of successful runs is the number of times the robot moved successfully around a circular path in the training environment. Figure 15 presents the learning curves of the WFM control when various evolutionary algorithms are used. The proposed DGDE achieved superior fitness values and successful runs than other methods. On the other hand, the proposed method also compares with the IT2RFCMAC controller [21]. Table 4 shows that the proposed IT2FNC with nonlinear functional link neural network performs better than IT2RFCMAC with linear TSK architecture [21].
To verify the WFM control performance of different learning algorithms, a training environment and two unknown testing environments were created, as presented in Figure 16, Figure 17 and Figure 18. The testing environment in Figure 17 focuses on many difficult and large curves, whereas the other testing environment in Figure 18 focuses on many right angle curves. The fitness value was used to evaluate the WFM, and the detailed comparison results are presented in Table 5.

4. Cooperative Carrying and Navigation Control of Multi-Evolutionary Mobile Robots

In this section, the cooperative carrying and navigation control of multi-evolutionary mobile robots is discussed. Figure 19 shows that the distance between two robots R d was set to 15 cm and a rectangular object was placed on the two robots. The front and rear robots represent the leader and follower, respectively. In the experiments, the leader explored the front environment, and the follower assisted the leader in lifting objects to achieve obstacle avoidance and prevent objects being dropped during cooperative carrying.

4.1. Wall-Following Control of the Cooperative Carrying Method

A dual controller is proposed for the cooperative carrying of two mobile robots. An auxiliary controller was incorporated in the follower robot to learn the WFM of cooperative carrying. The block diagram is presented in Figure 20. The auxiliary controller contained five input signals and two output signals. The inputs are the sensed distances by the follower’s sensors ( S 0 , S 1 , S 2 , and S 3 ) and the distances between two robots ( R d ). The outputs are the rotational speeds VL and VR of the two wheels.
To implement cooperative carrying in an unknown environment, the training environment featured straight lines, smooth curves, U-shaped curves, and continuous curves to train the follower’s auxiliary controller. Figure 21 displays the 1.5 × 1.4 m2 training environment.
To avoid collision with obstacles and objects being dropped during the cooperative carrying learning process, five terminal conditions were specified:
(1)
If the measured distance from one of the sensors in the follower robot is less than 1 cm, the follower robot collides with the obstacles.
(2)
If the measured distance of the sensor S 2 is higher than 6 cm, the follower robot deviates from the wall.
(3)
If the measured distance between the leader robot and follower robot R d is less than 10 cm or higher than 20 cm, the leader and follower robots are inferred to be very close or very far.
(4)
If the measured distance of the sensor is less than the height R h of the robot, the object is dropped by robots.
(5)
If the measured distance S w a l l of the sensor is less than 1 cm or greater than 7.5 cm, the object approaches the wall or deviates from the wall.
When one of the aforementioned conditions is satisfied, the cooperative carrying of the robots has failed. If the robots engaging in cooperative carrying can successfully move around a circular path in the training environment, the auxiliary controller completes the training process. The time step of the robot moving in the training environment is defined as the fitness function F ( · ) :
F ( · ) = T s t o p

4.2. Navigation Control of Cooperative Carrying

This paper also proposes an effective navigation control method for cooperative carrying in unknown environments. The manager mode automatically selects between the TGM and WFM on the basis of the relative position of the mobile robot and the target location.
(1)Toward-Goal Mode
In the navigation control of the unknown environment, the robot used infrared sensors to detect the object. In order to turn towards the goal position, the mobile robot calculates the angle difference θ T G between the current direction of the robot and the target direction, as presented in Figure 22:
θ T G =   θ R o b o t   θ G o a l
where θ R o b o t is the angle between the mobile robot and the x axis and θ G o a l is the angle between the goal and the x axis in inertial coordinate system.
To avoid objects being dropped during the cooperative carrying process, both the follower and leader move in the same direction and at the same speed, as displayed in Figure 23.
(2)Manager Mode
The robot is divided into three zones— O 1 , O 2 , and O 3 , as displayed in Figure 24. In the process of TGM, the manager mode switches to the left WFM or the right WFM on the basis of the zone of the robot O i and on the basis of which sensor S i (i = 0, 2, 3, …,7) detected an obstacle.
  • Left wall-following control:
    (i)
    The goal direction is located at O 1 , and S 7 or S 6 detects obstacles.
    (ii)
    The goal direction is located at O 2 , and S 7 , S 6 , and S 5 detect obstacles.
  • Right wall-following control:
    (i)
    The goal direction is located at O 1 , and S 0 or S 1 detects obstacles.
    (ii)
    The goal direction is located at O 3 , and S 0 , S 1 , and S 2 detect obstacles.
Before switching to the WFM, the manager mode determines which robot is closest to the obstacle. If the leader robot encounters an obstacle, it will execute a prerotation process. This aim of this process is to enable the two robots to approach the wall for a smooth navigational control, as displayed in Figure 25.
The leader robot turns in order to be positioned parallel to the wall and records the angle of rotation θ L . Then, this message is sent to the follower robot. The follower robot turns to π θ L °, maintains a fixed distance from the leader robot, and moves toward the obstacle. The manager mode of the two robots switch to WFM until the pre-rotation process of the follower robot is completed. The pre-rotation process is presented in Figure 26.
If the follower robot encounters an obstacle, the manager mode switches to the WFM; this process is displayed in Figure 27. If the target direction is located in O 1 of the follower robot and the distance between the sensor on right side S 1 (the sensor of left side S 6 ) and the wall is greater than 6 cm, the object being held by the two robots is inferred to have passed the obstacle. Then, the manager mode switches to the TGM, as displayed in Figure 28.

4.3. Experimental Results of Cooperative Carrying Control

In this subsection, the proposed method is used to verify the success of cooperative carrying control in unknown environments. Figure 29 demonstrates that the robots complete the wall-following control of cooperative carrying in the training environment. Moreover, to verify the performance of navigation control, two different test environments were created for testing whether the robots successfully accomplished cooperative carrying and navigation control. The experimental results of the two test environments are presented in Figure 30. In this experiment, the average distance (RD) between the two robots and the average distance (FWD) between the follower robot and the wall were evaluated. The results are presented in Table 6. If the RD is large, the two robots (i.e., the leader robot and the follower robot) are not at a suitable distance during the cooperative transport control, and the object drops easily. However, if the FWD is very large or very small, the robots pass the curves with poor efficiency, and the object is easily moved and dropped.

5. Conclusions

Aiming at the navigation control for cooperative carrying in an unknown environment, this study proposed an IT2FNC based on a dynamic group differential evolution to realize the carrying control and WFM control for mobile robots. At the same time, the developed DGDE learning algorithm adopts dynamic grouping and local search methods, which enhance the search ability and convergence stability of the traditional DE method, and is used to adjust IT2FNC parameters. On this basis, manager mode is established to assist mobile robots in navigation control. The manager mode automatically selects between the WFM or the TGM base on the relative position between the mobile robot and the target location. In addition, the pre-rotation mechanism was employed to accomplish cooperative carrying control. In the training process, the best fitness function, the worst fitness function, the average fitness function, the standard deviation (STD), the number of successful runs, and the computation time of the proposed DGDE were 0.962, 0.919, 0.942, 0.009, 10, and 4:38:56, respectively. Although DE and ABC are shorter than the proposed method at the computation time, DE and ABC had only eight successful runs during 10 runs. Experimental results revealed that the proposed method achieved a superior WFM performance than other methods and successfully accomplished the navigation control of cooperative carrying to target locations in unknown environments. Since the trained controller using reinforcement learning needs to take a long time, future research work needs to implement the proposed learning algorithm on chip to improve the learning speed.

Author Contributions

Revised manuscript, K.-Y.Y. and J.-Y.J.; Conceptualization and Methodology, J.-Y.J., K.-Y.Y. and C.-J.L.; Software, J.-Y.J.; Writing—Original Draft Preparation, C.-J.L.

Funding

This research was funded by Ministry of Science and Technology of the Republic of China, Taiwan (No. MOST 107-2221-E-167-023).

Conflicts of Interest

The authors declare that there is no conflict of interests regarding the publication of this manuscript.

References

  1. Cupertino, F.; Giordano, V.; Naso, D.; Delfine, L. Fuzzy control of a mobile robot. IEEE Robot. Autom. Mag. 2006, 13, 74–81. [Google Scholar] [CrossRef]
  2. Zhu, A.; Yang, S.X. Neurofuzzy-based approach to mobile robot navigation in unknown environments. IEEE Trans. Syst. Man Cybern. C Appl. Rev. 2007, 37, 610–621. [Google Scholar] [CrossRef]
  3. Rusu, P.; Petriu, E.M.; Whalen, T.E.; Cornell, A.; Spoelder, H.J.W. Behavior-based neuro-fuzzy controller for mobile robot navigation. IEEE Trans. Instrum. Meas. 2003, 52, 1335–1340. [Google Scholar] [CrossRef]
  4. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 1998. [Google Scholar]
  5. Hsu, C.H.; Juang, C.F. Multi-objective continuous-ant-colony-optimized FC for robot wall-following control. IEEE Comput. Intell. Mag. 2013, 8, 28–40. [Google Scholar] [CrossRef]
  6. Anish, P.; Parhi, D.R. Multiple mobile robots navigation and obstacle avoidance using minimum rule based ANFIS network controller in the cluttered environment. Int. J. Adv. Robot. Automn. 2016, 1, 1–11. [Google Scholar]
  7. Juang, C.F.; Chang, Y.C. Evolutionary group-based particle swarm-optimized fuzzy controller with application to mobile robot navigation in unknown environments. IEEE Trans. Fuzzy Syst. 2011, 19, 379–392. [Google Scholar] [CrossRef]
  8. Mendel, J.M. Advances in type-2 fuzzy sets and systems. Inf. Sci. 2007, 177, 84–110. [Google Scholar] [CrossRef]
  9. Mendel, J.M. Type-2 fuzzy sets and systems: An overview. IEEE Comput. Intell. Mag. 2007, 2, 20–29. [Google Scholar] [CrossRef]
  10. Kim, C.J.; Chwa, D. Obstacle avoidance method for wheeled mobile robots using interval type-2 fuzzy neural network. IEEE Trans. Fuzzy Syst. 2015, 23, 677–687. [Google Scholar] [CrossRef]
  11. Castillo, O.; Melin, P. A review on the design and optimization of interval type-2 fuzzy controllers. Appl. Soft. Comput. 2012, 12, 1267–1278. [Google Scholar] [CrossRef]
  12. Kennedy, J.; Eberhart, R. Particle swarm optimization. IEEE Int. Conf. Neural Netw. 1995, 4, 1942–1948. [Google Scholar]
  13. Dorigo, M.; Caro, G.D. Ant colony optimization: A new meta-heuristic. In Proceedings of the 1999 Congress on Evolutionary Computation, Washington, DC, USA, 6–9 July 1999; Volume 2, pp. 1470–1477. [Google Scholar]
  14. Storn, R.; Price, K. Differential evolution—A simple and efficient heuristic for global optimization over continuous spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  15. Karaboga, D.; Basturk, B. A powerful and efficient algorithm for numerical function optimization: Artificial bee colony (ABC) algorithm. J. Glob. Optim. 2007, 39, 459–471. [Google Scholar] [CrossRef]
  16. Passino, K.M. Biomimicry of bacterial foraging for distributed optimization and control. IEEE Control Syst. Mag. 2002, 22, 52–67. [Google Scholar]
  17. Zhou, Y.; Wang, J.; Zhou, Y.; Qiu, Z.; Bi, Z.; Cai, Y. Differential evolution with guiding archive for global numerical optimization. Appl. Soft Comput. 2016, 43, 424–440. [Google Scholar] [CrossRef]
  18. Das, S.; Mullick, S.S.; Suganthan, P.N. Recent advances in differential evolution—An updated survey. Swarm Evol. Comput. 2016, 27, 1–30. [Google Scholar] [CrossRef]
  19. Chang, J.Y.; Lin, Y.Y.; Han, M.F.; Lin, C.T. A functional-link based interval type-2 compensatory fuzzy neural network for nonlinear system modeling. In Proceedings of the 2011 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE 2011), Taipei, Taiwan, 27–30 July 2011; pp. 939–943. [Google Scholar]
  20. Chen, C.H.; Su, M.T.; Lin, C.J.; Lin, C.T. A hybrid of bacterial foraging optimization and particle swarm optimization for evolutionary neural fuzzy classifier design. Int. J. Fuzzy Syst. 2014, 16, 422–433. [Google Scholar]
  21. Jhang, J.Y.; Lin, C.J.; Lin, T.C.; Chen, C.C.; Young, K.Y. Using Interval Type-2 Recurrent Fuzzy Cerebellar Model Articulation Controller Based on Improved Differential Evolution for Cooperative Carrying Control of Mobile Robots. Sens. Mater. 2018, 30, 2499–2516. [Google Scholar] [CrossRef]
  22. Zhang, J.; Sanderson, A.C. JADE: Adaptive differential evolution with optional external archive. IEEE Trans. Evol. Comput. 2009, 13, 945–958. [Google Scholar] [CrossRef]
  23. Gong, W.; Cai, Z. Differential evolution with ranking-based mutation operators. IEEE Trans. Cybern. 2013, 43, 2066–2081. [Google Scholar] [CrossRef] [PubMed]
Figure 1. E-puck mobile robot architecture (a) E-puck mobile robot, (b) Infrared sensor position.
Figure 1. E-puck mobile robot architecture (a) E-puck mobile robot, (b) Infrared sensor position.
Electronics 08 00298 g001
Figure 2. Structure of an interval type-2 fuzzy neural controller (IT2FNC).
Figure 2. Structure of an interval type-2 fuzzy neural controller (IT2FNC).
Electronics 08 00298 g002
Figure 3. Interval type-2 fuzzy sets.
Figure 3. Interval type-2 fuzzy sets.
Electronics 08 00298 g003
Figure 4. Coding format of the vector.
Figure 4. Coding format of the vector.
Electronics 08 00298 g004
Figure 5. Fitness values of the parameter vectors in descending order.
Figure 5. Fitness values of the parameter vectors in descending order.
Electronics 08 00298 g005
Figure 6. The first group leader with the highest fitness value.
Figure 6. The first group leader with the highest fitness value.
Electronics 08 00298 g006
Figure 7. Generation of the first group members.
Figure 7. Generation of the first group members.
Electronics 08 00298 g007
Figure 8. Generation of the second group leader.
Figure 8. Generation of the second group leader.
Electronics 08 00298 g008
Figure 9. Flowchart of the proposed dynamic group differential evolution (DGDE).
Figure 9. Flowchart of the proposed dynamic group differential evolution (DGDE).
Electronics 08 00298 g009
Figure 10. Block diagram of the mobile robot wall-following control.
Figure 10. Block diagram of the mobile robot wall-following control.
Electronics 08 00298 g010
Figure 11. Training environment of the mobile robot wall-following control.
Figure 11. Training environment of the mobile robot wall-following control.
Electronics 08 00298 g011
Figure 12. Mobile robot (a) collides with the wall (b) deviates from the wall.
Figure 12. Mobile robot (a) collides with the wall (b) deviates from the wall.
Electronics 08 00298 g012
Figure 13. Definition of (a) d w a l l and (b) degree of parallelism.
Figure 13. Definition of (a) d w a l l and (b) degree of parallelism.
Electronics 08 00298 g013
Figure 14. Block diagram of learning process of wall-following control.
Figure 14. Block diagram of learning process of wall-following control.
Electronics 08 00298 g014
Figure 15. Learning curves of the wall-following control when various evolutionary algorithms are employed.
Figure 15. Learning curves of the wall-following control when various evolutionary algorithms are employed.
Electronics 08 00298 g015
Figure 16. Moving path of the robot in the training environment: (a) DGDE-1 and (b) DGDE-2.
Figure 16. Moving path of the robot in the training environment: (a) DGDE-1 and (b) DGDE-2.
Electronics 08 00298 g016
Figure 17. Moving path of the robot in the testing environment 1: (a) DGDE-1 and (b) DGDE-2.
Figure 17. Moving path of the robot in the testing environment 1: (a) DGDE-1 and (b) DGDE-2.
Electronics 08 00298 g017
Figure 18. Moving path of the robot in the testing environment 2: (a) DGDE-1 and (b) DGDE-2.
Figure 18. Moving path of the robot in the testing environment 2: (a) DGDE-1 and (b) DGDE-2.
Electronics 08 00298 g018
Figure 19. Cooperative carrying of two mobile robots.
Figure 19. Cooperative carrying of two mobile robots.
Electronics 08 00298 g019
Figure 20. Block diagram of the cooperative carrying of mobile robots.
Figure 20. Block diagram of the cooperative carrying of mobile robots.
Electronics 08 00298 g020
Figure 21. Training environment for cooperative carrying.
Figure 21. Training environment for cooperative carrying.
Electronics 08 00298 g021
Figure 22. Angle between the robot and the target location.
Figure 22. Angle between the robot and the target location.
Electronics 08 00298 g022
Figure 23. Follower and leader move in the same direction and at the same speed.
Figure 23. Follower and leader move in the same direction and at the same speed.
Electronics 08 00298 g023
Figure 24. Divided zones of the mobile robot.
Figure 24. Divided zones of the mobile robot.
Electronics 08 00298 g024
Figure 25. Leader robot encounters an obstacle.
Figure 25. Leader robot encounters an obstacle.
Electronics 08 00298 g025
Figure 26. (a) Follower robot turns π θ L °; (b) Pre-rotation process; (c) Manager mode of the two robots switch to the wall-following mode (WFM).
Figure 26. (a) Follower robot turns π θ L °; (b) Pre-rotation process; (c) Manager mode of the two robots switch to the wall-following mode (WFM).
Electronics 08 00298 g026
Figure 27. Follower robot encounters an obstacle.
Figure 27. Follower robot encounters an obstacle.
Electronics 08 00298 g027
Figure 28. Object on the two robots passes the obstacle.
Figure 28. Object on the two robots passes the obstacle.
Electronics 08 00298 g028
Figure 29. Wall-following control of cooperative carrying in the training environment.
Figure 29. Wall-following control of cooperative carrying in the training environment.
Electronics 08 00298 g029
Figure 30. Navigation control of cooperative carrying in (a) test environment 1 and (b) test environment 2.
Figure 30. Navigation control of cooperative carrying in (a) test environment 1 and (b) test environment 2.
Electronics 08 00298 g030
Table 1. Initial parameters of DGDE.
Table 1. Initial parameters of DGDE.
NPCRFGenerationRule
300.90.530005,6,7
Table 2. Performance evaluation of different fuzzy rule numbers. STD: standard deviation.
Table 2. Performance evaluation of different fuzzy rule numbers. STD: standard deviation.
Number of RulesDGDE-1DGDE-2
Fitness Value 567567
Best0.9320.9610.9480.9490.9620.953
Worst0.8650.8910.8210.9080.9190.891
Average0.9030.9330.9110.9230.9420.913
STD0.0170.0120.2000.0120.0090.016
Number of successful runs101010101010
Table 3. Performance comparison of various algorithms in the wall-following control behavior.
Table 3. Performance comparison of various algorithms in the wall-following control behavior.
Evaluation ItemsFitness ValueNumber of Success RunsComputation Time
(H:M:S)
Algorithms BestWorstAverageSTD
DGDE-10.9610.8910.9330.012105:01:39
DGDE-20.9620.9190.9420.009104:38:56
JADE [22]0.9500.8600.9110.0291010:11:03
Rank-DE [23]0.9580.8670.9220.0251018:21:05
DE [14]0.9410.2620.7860.18481:03:38
PSO [12]0.9470.2060.7380.25775:49:45
ABC [15]0.9320.3540.7350.14982:57:23
Table 4. Performance comparison of different network architecture.
Table 4. Performance comparison of different network architecture.
Evaluation ItemsIT2FNCIT2RFCMAC [23]
Algorithms DGDE-1DGDE-2DGDE
Best0.9610.9620.925
Worst0.8910.9190.868
Average0.9330.9420.906
Table 5. Fitness value of various algorithms in the testing environments.
Table 5. Fitness value of various algorithms in the testing environments.
Evaluation ItemsFitness Value
Algorithms Training EnvironmentTesting Environment 1Testing Environment 2
DGDE-10.9610.9010.864
DGDE-20.9620.8990.872
JADE [22]0.9500.8950.862
Rank-DE [23]0.9580.8740.789
DE [14]0.941failfail
PSO [12]0.9470.8280.721
ABC [15]0.932failfail
Table 6. Evaluation of the cooperative carrying performance.
Table 6. Evaluation of the cooperative carrying performance.
Evaluation ItemsTraining EnvironmentTesting Environment 1Testing Environment 2
Algorithms RD (cm)FWD (cm)RD (cm)FWD (cm)RD (cm)FWD (cm)
DGDE-116.183.8116.543.6117.033.42
DGDE-215.433.9615.763.8917.164.23

Share and Cite

MDPI and ACS Style

Jhang, J.-Y.; Lin, C.-J.; Young, K.-Y. Cooperative Carrying Control for Multi-Evolutionary Mobile Robots in Unknown Environments. Electronics 2019, 8, 298. https://doi.org/10.3390/electronics8030298

AMA Style

Jhang J-Y, Lin C-J, Young K-Y. Cooperative Carrying Control for Multi-Evolutionary Mobile Robots in Unknown Environments. Electronics. 2019; 8(3):298. https://doi.org/10.3390/electronics8030298

Chicago/Turabian Style

Jhang, Jyun-Yu, Cheng-Jian Lin, and Kuu-Young Young. 2019. "Cooperative Carrying Control for Multi-Evolutionary Mobile Robots in Unknown Environments" Electronics 8, no. 3: 298. https://doi.org/10.3390/electronics8030298

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop