Next Article in Journal / Special Issue
Segment-Tube: Spatio-Temporal Action Localization in Untrimmed Videos with Per-Frame Segmentation
Previous Article in Journal
Suspended Carbon Nanotubes for Humidity Sensing
Previous Article in Special Issue
Lane Marking Detection and Reconstruction with Line-Scan Imaging Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automated Field-of-View, Illumination, and Recognition Algorithm Design of a Vision System for Pick-and-Place Considering Colour Information in Illumination and Images

1
Department of Precision Engineering, the University of Tokyo, Tokyo 113-8656, Japan
2
Research into Artifacts, Center for Engineering (RACE), The University of Tokyo, Chiba 277-8568, Japan
3
Denso Wave Incorporated, Aichi 470-2298, Japan
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(5), 1656; https://doi.org/10.3390/s18051656
Submission received: 28 March 2018 / Revised: 15 May 2018 / Accepted: 18 May 2018 / Published: 22 May 2018
(This article belongs to the Special Issue Visual Sensors)

Abstract

:
Machine vision is playing an increasingly important role in industrial applications, and the automated design of image recognition systems has been a subject of intense research. This study has proposed a system for automatically designing the field-of-view (FOV) of a camera, the illumination strength and the parameters in a recognition algorithm. We formulated the design problem as an optimisation problem and used an experiment based on a hierarchical algorithm to solve it. The evaluation experiments using translucent plastics objects showed that the use of the proposed system resulted in an effective solution with a wide FOV, recognition of all objects and 0.32 mm and 0.4° maximal positional and angular errors when all the RGB (red, green and blue) for illumination and R channel image for recognition were used. Though all the RGB illumination and grey scale images also provided recognition of all the objects, only a narrow FOV was selected. Moreover, full recognition was not achieved by using only G illumination and a grey-scale image. The results showed that the proposed method can automatically design the FOV, illumination and parameters in the recognition algorithm and that tuning all the RGB illumination is desirable even when single-channel or grey-scale images are used for recognition.

1. Introduction

Machine vision technologies have been widely applied in the industrial field for automated visual inspection, process control, parts identification, and robotic guidance [1,2]. Designers have been attempting to tune the parameters for a variety of vision systems. A vision system is usually composed of a camera and an illumination and recognition algorithm [3], which are also known as the main design factors of a vision system. In the object recognition system of a pick-and-place robot, for example, the camera position needs to be set to obtain a suitable Field-of-View (hereinafter referred to as FOV), the illumination requires to be changed to strengthen features in targets, and the image recognition process needs to be optimised through parameter tuning. As this creates a number of conflicting variables, the design process must be reiterated until acceptable results are obtained. This is a time-consuming task even when carried out by experts, and even a simple pick-and-place vision system usually takes several days to design.
Previous studies have addressed the automated design of sensor locations [4,5,6,7,8,9,10,11,12], illumination levels [13,14,15,16,17,18], and recognition algorithms [19,20,21,22,23,24,25]. Some studies proposed a method to automatically determine the place to set a vision sensor for specific features of recognition targets to satisfy the specific constraints of recognition requirements [4,5,6]. Some other studies focused on the sensing strategies for recognition and localisation of targets with the help of 3D models [7,8,9]. Moreover, several sensor planning methods were designed respectively based on the vision tasks in [10,11,12].
Researches on automated planning of illumination parameters have also been carried out. Experiment-based approaches have been proposed to optimise illumination with a set of images of an object captured under different illumination conditions [13,14]. Besides, illumination planning methods based on mathematical models of illumination were proposed [15,16]. More recently, with the help of rendering techniques, illumination planning approaches based on computer simulation were reported [17,18].
Some studies have attempted to automate the image processing procedures. Automated image pre-processing techniques were proposed in [19,20,21]. Some other studies investigated the automated design for feature extraction [22,23]. Automated generation of discriminators were discussed in [24,25]. Especially, an approach was proposed for automatically designing an image recognition procedure from the aspect of pre-processing, feature extraction, and discriminator [25].
It is clear from the former studies that an overall design approach to vision systems could hardly be found. One reason is probably the interactions among the different design factors. Therefore, in the case of an overall design, the situation becomes more and more complex because design factors influence each other in unpredictable ways. To the best of our knowledge, a design approach to deal with various design factors has been presented only in [26,27]. Experiment-based methods were applied to achieve an automated design of a vision system on the basis of illumination and a recognition algorithm in [26]. By adding FOV, Chen Y., et al. [27] provided a more comprehensive vision system design method. The problem is that in both the studies, the recognition tasks were far from a being practical task because only one or two objects were considered.
Another obstacle in taking out an overall design approach consists of the uncertainties of the real world. Colour is known as one of the uncertainties in image recognition. Because objects’ colours change with illumination, colour- and illumination-invariant recognition methods have been postulated [28,29,30,31]. The greyscale process transforms colourful multi-channel images into grey and single-channel ones, which could be more easily understood by vision systems. Such multi-channel image encoding approaches were presented in [32,33,34], while it was also pointed out that greyscale approaches could influence the recognition performance to a great extent [35]. In this study, we have mainly focused on the uncertainties caused by colour information contained in both illumination and grabbed images.
This study transferred a vision system design problem into an optimisation problem and proposed an experiment-based approach to realise an automated vision system design. It was proved in the study that the proposed design could provide vision systems that were effective in pick-and-place tasks with suitable parameters of the FOV, illumination and recognition algorithm. Moreover, we studied one kind of uncertainties from the real world, that is, colour information illuminate from the light source that is absorbed by the camera sensor. Thus, we conducted an experiment of automated designs using our proposed method by changing the colour channels that were utilised for both illumination and recognition. By this experiment, we investigated: (1) whether or not providing colourful illumination improves recognition accuracy when even the vision system reads only the greyscale images and (2) whether or not single-channel images like R-channel images provide better performance in recognition than the greyscale ones.

2. Problem Formulation

2.1. Preconditions

The vision systems applied to a pick-and-place robot are set to the design target in this study. In order to pick objects and place them in the right positions, the vision systems are required to provide the following information:
● Types
Before picking up an object, the system must know which kind of object to select. For instance, in some sorting tasks, type refers to information that describes targets’ appearance, such as the shape, colour, and which side is facing upwards. By using the type information, the robot is able to distinguish the target objects into several categories. Therefore, a dictionary which contains type information must be available to vision systems for pick-and-place.
● Position
For a pick-and-place robot, definitely ‘pick’ is one of the most important quests. To pick objects up, position information, in other words, the centre of gravity of each target object should be identified. A vision system captures positional information in pixels. In this study, the coordinate origin is set to the left-top of the image, the x-axis forward direction to the right, and the y-axial forward direction is downward.
● Orientations
For both ‘pick’ and ‘place’ quests, the orientation information is important. That is to say, the vision system must also provide angular information about each object. In this study, we assumed the measurement range to be [0, 360).
To clarify the problem, the working environment for the proposed system is set up based on the following three requirements: choice of camera, assumed scenes of recognition, and image processing software. The details for each of these requirements are given as follows:
● Camera
Compared to binocular cameras, monocular cameras are more widely used in pick-and-place tasks. As a result, we used a monocular CMOS (Complementary Metal Oxide Semiconductor) camera in our system. Since the camera is mounted on the end effector of an industrial manipulator, its viewpoint is held perpendicular to the workspace on which the recognition targets are arranged; the FOV is therefore of a 2D type.
● Scenes
Based on where to pick objects, pick-and-place tasks could be categorised into two types: tasks in dynamic systems, for example, a moving conveyor and tasks in static systems such as a tray. In this study, we chose the latter one as the option for the proposed system. In this case, all the recognition targets are placed in a limited space. This means that the camera distance at which all objects can be captured in a single image can be specified in advance. Moreover, just like most situations in industrial applications, the recognition targets are placed on the same plane surface, without overlaps. This is also true for industrial applications such as picking objects from a conveyor.
● Software
In this study, the proposed system was tested with a commercially available image processing library: MVTec HALCON (MVTec, Seeshaupt, Germany).

2.2. Design Variables

By describing the settings and the working environment, the basic information on the target vision system to be designed was provided. In order to arrive at a proper design of the described vision system, several parameters, or we can call them design variables, are required to be optimised. To further clarify the problem, such design variables are determined in this section.
In general, a vision system could be established by considering three design factors, which are: illumination condition; camera FOV; and the recognition algorithm. The illumination is usually designed for its strength and colour, illuminating the workpieces and repressing the reflections at the same time. Camera FOV determines the resolution with which the targets are recognised and the size of the recognition area. By tuning FOV, accuracy and efficiency of the vision system could be balanced. Besides, in order to maximise the performance of the chosen recognition algorithm, some parameters inside the algorithm also require optimisation.
The design variables of the vision system could also be taken as parameters of the optimisation problem that were defined in the previous section. Categorised by the three design factors, the design variables of this study are addressed as follows (details are given in Table 1):
● FOV
FOV is the extent of the observable world that is seen at any given moment. In the case of a camera set up for pick-and-place tasks, the FOV directly determines the number of objects that can be captured in a single image. To maximise the efficiency of recognition, the FOV must be maximised while meeting the required accuracy tolerances. In this study, FOV was balanced from the viewpoint of shoot time and camera distance.
Shoot time means the time taken by the camera to capture figures within the current camera distance. Obviously, with a limited FOV size, the vision system could not comprehend the intricate details of the workspace. Thus, it requires a system based on a moving camera which captures images several times.
On the other hand, camera distance refers to the distance from the camera lens to the plane on which the recognition targets are placed. As mentioned in the preconditions, the camera’s viewpoint is held perpendicular to the workspace; the distance therefore reflects the actual FOV size.
● Illumination
The illumination variables include the strength of the red, green, and blue components. Increasing the strength may produce reflections, whereas at low strength, some details of the target objects may not be captured. Both will reduce the recognition accuracy. Additionally, some details may be enhanced by selecting the specific colour of illumination. We therefore allowed the strength of each RGB component to be controlled individually. The illumination strength ranges from 0 to 255, and is searched by an increment.
● Recognition Algorithm
Not only the recognition algorithm but also the parameters inside the chosen algorithm influence the performance of a vision system. We just focus on the latter to optimise the inner parameters of a given recognition algorithm.
The inner parameters, for example, image pre-processing parameters or parameters for making proper templates, could more or less influence the performance of a recognition algorithm. Since different parameters may have their own properties, the optimisation method should be designed individually.
Moreover, no matter what recognition algorithm is used, a discriminator to classify correct and incorrect detections by the recognition process is required. The discriminators should also be considered as one of the design variables.

2.3. Inputs and Outputs

2.3.1. Inputs

Aiming to turn the vision system design process into a fully-automated one, manual operations during the process of design must be minimised. Hence, the inputs to the automated design system should be considered from many aspects which are preparation data for both scenes and templates, ground truth data, and camera calibration data.
  • Preparation Data for Scenes:
    S = ( S 1 , ,   S i , ,   S n ) ,
    here, Si denotes the i-th coordinate on the work plane of the position where the corresponding scene was set, and n the total number of scenes. The number of images required to capture one scene depends on the camera distance.
    S i = ( x i , ,   y i , ,   z i ) ,
    Each Si contains the locations of x, y, and z directions such that the manipulator can hold the camera and capture images of the existing scene. zi describes the distance from the camera to the plane where the recognition targets are arranged.
  • Preparation Data for Templates:
    T = ( T 1 , ,   T l , ,   T n T ) ,
    where nT represents the total number of recognition target kinds.
    T l = ( x ,   y ,   z ,   x l ,   y l ,   w l ,   h l ) ,
    the l-th template is prepared by automatically cutting the object image from the original image which was obtained by holding the camera at the position (x, y, z). By using the position of the objects in the acquired image, namely xl and yl, as well as the predetermined width and height, wl and hl, the template could be obtained.
  • Ground Truth Data:
    G = ( G S 1 , ,   G S i , ,   G S n ) ,
    where G S i denotes the ground truth data for the i-th scene.
    G S i = ( G S i ,   1 , ,   G S i ,   k , ,   G S i ,   m i ) ;
    however, the scene contains many recognition targets; the ground truth data always include information on each recognition target, from the 1st to the mi-th.
    G S i ,   k = ( T y p e i , k ,   x i , k ,   y i , k ,   θ i , k ) ,
    the ground truth data for each object includes the object type, the x and y position in captured images and the orientation angle.
  • Camera Calibration Data:
    C = ( C 1 , ,   C i , ,   C n ) ,
    where Ci denotes the i-th image for calibration and n the total number of images required for a calibration.

2.3.2. Outputs

The system output is the optimal solution to the set of design variables.
  • Optimal solution:
    S o l u t i o n = ( R ,   G ,   B , P r e c o g n i t i o n ) ,
    where R, G, and B denote the light strength of red, green and blue, and Precognition the set of parameters related to the chosen recognition algorithm. Especially, Precognition consists of:
    P r e c o g n i t i o n = ( P 1 , P 2 ,   , P n )
    the entire number of parameters n is determined by the chosen recognition algorithm.

2.4. Evaluation Function and Constraints

2.4.1. Evaluation Function

The evaluation uses four values which are the FOV size, Fmeasure, positional error, and angular error.
The shoot time describes the FOV and largely determines the computing speed, as the time cost increases in line with the number of images and camera movements.
The Fmeasure is used to describe the accuracy of recognition. It considers both the Precision and Recall, and the definition is given by Equation (1):
F m e a s u r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l .
In this study, the Precision and Recall values were given by the following equations:
P r e c i s i o n = I i I m c i I i I m i
R e c a l l = I i I m c i I i I m d i .
Here, mi, mci, and mdi refer to the total number of targets, correctly recognised targets, and targets detected by the recognition process for the i-th learning image set, respectively. The Fmeasure value ranges from [0, 1]. A value closer to 1 indicates greater accuracy.
Positional errors (PosErr) were defined as follows:
P o s E r r = max { P o s E r r 1 , , P o s E r r i , , P o s E r r n }
P o s E r r i = ( x i x g t i ) 2 + ( y i y g t i ) 2 .
The maximum positional error among n targets was used in the evaluation, and each positional error was calculated from the difference between the points detected by the recognition system (xi, yi) and the ground truth (xgti, ygti). As the proposed system used a moveable camera, we first transformed the positional results from the camera coordinates to world coordinates and then measured the error in millimetres.
The angular errors ( A g E r r ) were defined as follows:
A g E r r = max { A g E r r 1 , , A g E r r i , , A g E r r n }
A g E r r i = | ( θ i θ g t i ) ( m o d   360 ) | .
The maximum angular error among n targets was used in the evaluation. As the detection range was from [0, 360), the angular error was given by the difference between the angles detected by the recognition system θi and the ground truth θgti. This was given by Equation (7).
We set the following order for evaluation: first, the camera distance; second, the Fmeasure; third, the positional error; and finally, the angular error. Accuracy was determined from the minimum Fmeasure values and maximum positional error and angular error values. The system would therefore choose the solution based on the FOV size, Fmeasure, positional error, and angular error, successively.
For a vision system in pick-and-place tasks, it is important to do recognition as efficient as possible with the guarantee of accuracy. The positional and angular errors can be often tolerated to some extent by the selection of the manipulator though we did not discuss the type of the manipulator in this paper. If the recognition accuracy is high enough, the higher the image capturing efficiency is, the better. Therefore, FOV size and Fmeasure take the first two priorities for evaluation. If the positional error is too large, the manipulator cannot pick the objects. Therefore, we decided to prioritize positional error over angular error.

2.4.2. Constraints

For the designed vision system to be applied to a pick-and-place task, it is necessary to ensure the minimal performance. In other words, at least the designed vision system could pick up and place the objects without any failure. The constraints are therefore set to guarantee the minimal performance of designed system.

3. Methodology

3.1. Algorithm Overview

Figure 1 shows the algorithm we proposed to solve the problem formulated in Section 2. In general, we prepared respective optimisations for parameters of the three design factors and arranged them hierarchically.
The system first set the FOV size to its maximum, so that all the target objects could be captured into one image. Based on the multi-start nearest neighbour search, which is discussed in more detail in Section 3.3, the illumination search centre was set randomly to (Redi, Greeni, Bluei), and the parameters in recognition algorithm were then designed. After the recognition algorithm design, the system obtained the local solution (Redi, Greeni, Bluei, Heighti, ParametersBest) and its accuracy evaluation (Fmeasurei, PosErri, AgErri) for the corresponding illumination condition. Evaluation was performed repeatedly under neighbour illumination conditions around the selected search centre. After all neighbours were searched, the search centre was moved to its best neighbour until it became the best design. Illumination optimisation was repeated N times, yielding N local optimal solutions. If solutions meeting the design criteria were found, the system chose the optimal solution among N candidates. Otherwise, the system returned to its initial state and narrowed the FOV size by decreasing the camera distance and increasing the shoot time. The methods to apply narrow FOV and estimate FOV size by camera distance are presented in Section 3.2.

3.2. FOV Design

The FOV is applied to the vision system in the following two ways: first, carry out recognition once with an FOV size and just fit the size of the recognition area and second, carry out recognition by scanning the entire area n2 times with a FOV of a specific size. Figure 2 shows an example of taking an image of an object placed in the area for recognition. Since the angle between the viewpoint of the camera and the work plane is fixed, which is stated in the preconditions, the FOV size could be easily estimated from the distance between the camera and work plane.
The steps to estimate the FOV size by camera distance are:
(1) Obtain the mathematical relation between the width of a taken image and camera distance.
Several images are captured under different camera distances. By adding camera calibrations, the FOVwidth, or in other words, the distance of y direction in the taken images, can be measured in millimetres. Repeating this operation several times, the relations between FOVwidth and camera distance could be fitted to a linear one:
F O V w i d t h ( C d i s t a n c e ) = a w i d t h × C d i s t a n c e + b w i d t h .
The Cdistance denotes the camera distance, a and b are coefficients calculated by experimental data.
(2) Obtain mathematical relation between the length of a taken image and camera distance.
Similar to FOV width, relations between FOVlength and camera distance are found using the following expression:
F O V l e n g t h ( C d i s t a n c e ) = a l e n g t h × C d i s t a n c e + b l e n g t h .
(3) Choose either length or width to represent the FOVsize based on length-width ratios of the recognition area and captured image.
F O V l e n g t h ( C d i s t a n c e ) F O V w i d t h ( C d i s t a n c e ) > R l e n g t h R w i d t h .
Rlength and Rwidth denote length and width of the recognition area, respectively. Equation (10) is the criterion to judge whether to use length or width to represent the size of FOV. Based on the condition of inequality applied in this study, if the length-width ratio of FOV is larger than the ratio of the recognition area, then the width should be selected for calculations in later steps. Otherwise, the length should be chosen.
(4) Calculate desired FOV size.
Using either length or width to stand for the size, the desired FOVsize could be calculated in addition to the size margin and the scan time.
F O V s i z e = R s i z e + ( S t i m e 1 ) × M a r g i n S t i m e .
Rsize denotes the length or width of the recognition area, Margin the margin of the FOV size decided by the maximum size of chosen recognition targets and Stime the total scan time. Here the square root of Stime is used to present the scan time in either the x or y direction.
(5) Estimate corresponding camera distance.
By substitution of the calculated desired FOV size into either Equation (8) or Equation (9), the corresponding camera distance for the desired FOV size could be obtained:
C d i s t a n c e = F O V s i z e b s i z e a s i z e .
Here, asize and bsize denotes awidth and bwidth in Equation (8) or alength and blength in Equation (9) depend on the truth or false of Equation (10).

3.3. Illumination Design

We selected a random multi-start nearest neighbour search, which is one of metaheuristic method, for optimisation of the illumination strength of red, green, and blue. Due to find constraint satisfaction solutions in limited time, we allowed the system choose search centres randomly, even that may result in different optimums in a fixed condition.
The neighbours were generated by changing the value (adding or subtracting the increments shown in Table 1) of one variable, while holding the others constant. The system then created six neighbours for RGB strength in illumination:
N e i g h b o r s = { ( R + I n c r e m e n t , G , B ) , ( R I n c r e m e n t , G , B ) ,   ( R , G + I n c r e m e n t , B ) , ( R , G I n c r e m e n t , B ) ,   ( R , G , B + I n c r e m e n t ) , ( R , G , B I n c r e m e n t ) } .

3.4. Recognition Algorithm Design

The proposed system is capable of automatically selecting threshold values as discriminators for all kinds of recognition objects.
Suppose that a recognition method has a value E to evaluate its detections, the higher E value indicates the detection is more likely to be a correct one. For a given recognition object, a series of E values are used for n detections after one recognition.
D = { E D 1 , E D 2 , , E D n } .
Set D was then categorised into two sets with the help of ground truth data; T for correctly detected results, F for incorrectly detected results:
T = { E T 1 , E T 2 , , E T m } ,
F = { E F 1 , E F 2 , , E F l } ,
l + m = n .
The threshold Th was then generated as follows:
T h = max { E T i | E T i T E T i < m i n { E F j | E F j F } } .
The threshold represents the maximal evaluation in the correctly recognised results, and is smaller than the minimal evaluation of the incorrectly recognised results.
As stated before, the optimisation target and corresponding approach rely on the chosen algorithm for recognition. A contour matching method in HALCON library called shape-based matching was utilised as the recognition algorithm in this study.
The target parameter of the shape-based matching algorithm to be designed is the contrast value to extract contour models from the templates. Figure 3 illustrates two contour models extracted from different contrast values. A too large contrast value decreases the number of contours in the obtained model to a great extent. Matching with less contours therefore yields more possible candidates, and finally results in longer matching time. If the contour model is decreased to just a short line, the matching time can be infinity.
Based on this principle, the design method for the contrast value in shape-based matching is set to traverse all the possible values from the minimum to the maximum and end if the detection number reaches a threshold (Figure 4).

4. Evaluation Experiment

4.1. Experimental Setup

The experimental environment was an industrial manipulator with six DoFs, a ring-shaped illumination device and an industrial monocular camera (Figure 5). The camera and illumination were mounted on the tips of the manipulator using a 3D-printed joint. The processor was an Intel Core [email protected] GHz.
To reflect potential applications, we chose the two sides of a semi-transparent plastic part (Figure 5) as the recognition target. Different from its side at the rear, the face side had a convex structure in the middle. The following constraints were applied: an Fmeasure score no less than 1; a positional error no more than 3 mm; and an angular error no more than 5°.
Three scenes with different functions were arranged on a piece of black cloth below the manipulator (Figure 5). In order to prevent overfitting, two scenes were prepared for recognition. Every time the FOV or illumination changed, the templates were updated; a scene for updating the templates was therefore required. Scene 1 and Scene 2, with two face and two rear side objects in each, were set up for recognition. Scene 3, with a face side and a rear side object, was set up to create templates.
Two different plans for FOV were given to our system: one was to shoot once with a wide FOV at a camera distance of 158 mm, and the other was to shoot four times with narrow FOVs at a camera distance of 105 mm.
We controlled the colour channels utilised in both recognition and illumination, and created three experimental conditions. Recognitions were conducted with greyscale images in Condition I and II, while R-channel images were used in Condition III. The details are listed below. On the other hand, illuminations were changed from only G channel in Condition I, and changed from RGB three channels in Condition II and III. Details of the conditions based on which each experiment was conducted are listed in Table 2.
The reason why G illumination was chosen in Condition I is that it is considered to influence the brightness in the obtained images to the greatest extent. Therefore, the dimension of illumination was reduced to a great extent, and the increment of illumination strength was set to 1 in Condition I.
For all aforementioned conditions, we manually measured the ground truth data. For the illumination variables, 16 local optimisation searches were performed. The maximum detection to end contrast value search was set to 4.

4.2. Results

The best-three solutions and their evaluations of the three conditions are presented in Table 3, Table 4 and Table 5 and the images taken under the optimal parameter sets are shown in Figure 6.
Proper design could not be achieved with either with one shot or four shots when tuning illumination from only G component in Condition I. The optimal design was realised with a 79 in green illumination, four shots and a contrast value of 3, which provided a 0.93 Fmeasure, about 0.6 mm maximum positional error, and 2.1° maximum angular error.
The optimal design of Condition II corresponded to a (195, 120, 75) illumination RGB strength, four shoots and contour models generated by a contrast value of 4. This set up resulted in an Fmeasure value of 1, a maximum positional error of about 0.6 mm, and a 3.1° maximum angular error.
Replacing the greyscale images with R-channel images, suitable designs were found only with one shoot. The optimal design was (195, 120, and 75) in illumination RGB, 11 in contrast value, and with one shoot. Its evaluation showed an Fmeasure of 1, about 0.3 mm in maximum positional error, and 0.4° in maximum angular error.

5. Discussion

Generally speaking, designs under the accuracy constraints, that is, an Fmeasure of 1 and no more than 3 mm and 5° in positional and angular errors were found in both conditions of illumination tuned from RGB channels. This finding proved that our system is capable of tuning parameters for a vision system used in pick-and-place tasks.
Comparing the results of the first two conditions, designs under accuracy constraints were found when illumination was tuned from RGB, while no proper design was found with an Fmeasure of 1 with illumination tuned only from G. In both conditions, the input images were of the greyscale type, which indicated that although a vision system finally converts colour images into grey, it is still essential to tune the illumination based on the three RGB channels.
On the other hand, from the results of Condition II and Condition III, it was found that using R-channel images could provide better performance in recognition than greyscale ones. Designs with one shot and a wide FOV were found in Condition III, while a narrow FOV was designed with four shoots in Condition II.
In order to further discuss the effects of R-channel images, the R-channel images for the two scenes under the optimal design of Condition III (illumination RGB equals to 15, 225, and 240, 1 shoot) were extracted. We processed the two figures with greyscale; both R-channel and greyscale figures are shown in Figure 7. Moreover, to confirm that the R-channel images perform better than greyscale images under the same situation, an additional design was implemented with greyscale images, as shown in Figure 7. Results showed that the optimal design with the greyscale images could only provide an Fmeasure of 0.93.
To our human eye, it is obvious that the greyscale images are easier to recognise. However, in a vision system, R-channel images are recognised with a higher recognition accuracy. The probable reason might be that in a sufficiently bright image, the noise is also enlarged to a great extent. Vision systems do not detect a picture as humans do; these systems read the limited features in the form of mathematical values in matrixes instead. When the noise is so large that it obscures the useful information indicated in these matrixes, judgements made by the system could be flawed. From this point of view, the key to a ‘clear’ image for vision systems is that these images must contain little but important information. As an example, though the R-channel images in Figure 7 were really dark, the contours of each object could still be seen clearly. The great contrast between contours and background therefore make the images ‘clear’. In some ways, image pre-processing is just a method to serve the vision systems with ‘clearer’ images.
Moreover, illumination in the designs with high evaluations showed no relations to each other with greyscale images input, while a clear pattern was discovered in the circumstance of R-channel images. Based on Table 4, illuminations of the best-three designs were found with low red illumination (under 50), high green illumination (near 225), and relatively high blue illumination (from 150 to 240). Generally speaking, tuning green and blue illuminations is not effective when the image can only be seen using a red channel. However, Figure 8 shows that even with no red illumination, the objects are visible in the R-channel image. The probable reason may be that the RGB tuned from the illumination side is not the same as the RGB information contained in an image. Because of the wavelength of the illumination device or some reflections, the G and B components could still influence the R-channel image to some extent. Actually, such an influence eventually resulted in ‘clearer’ R-channel images compared with the greyscale ones. The illumination pattern found in Condition III also confirmed the importance of green and blue illumination. In addition, patterns of illumination indicated that relations might exist between the recognition performance and its illumination conditions, which give rise to possibilities for the application of other optimisation methods.
Nevertheless, the experiment was limited under the environment we prepared. We could only state that R-channel image could provide better recognition accuracy under the experimental settings. We cannot affirm that whether this phenomenon could be discovered with other recognition targets, or by recognition with other algorithms. To better explain it, further experiments will be required.

6. Conclusions

In this study, we proposed an automated design approach for vision systems in pick-and-place tasks. The vision system design was first formulated as a parameter optimisation problem and then solved in an experiment-based approach with a hierarchical algorithm. Rather than seeking a suitable parameter set randomly in the solution space, the proposed algorithm separates and sets hierarchies for each optimisation based on the design factors. As one of the uncertainties from the real world, the influence of colour on the recognition performance of the designed vision systems was also investigated through experiments in this research.
It could be seen through the experiments that the proposed system was able to design a vision system with a 100% recognition rate, and a positional and angular error of 0.32 mm and 0.4°, respectively. When using greyscale images for recognition, G illumination resulted in an Fmeasure of only 0.93, which proved the necessity for colourful illumination. Consequently, when RGB illumination was used, designs with R-channel images used only one shot, which indicates that R-channel images provide better recognition accuracy than the greyscale ones.
In future work, from the viewpoint of robustness, it is necessary to improve the prevention against overfitting by increasing the number of scenes for recognition and include the measurement of overfitting in the evaluation of the designed vision system. Aiming to take out better solutions, the selection of recognition algorithm should also be included into the design process. Additionally, further research could be conducted on searching more appropriate optimisation methods, for example, neural networks or genetic algorithms, to provide better solutions that are less time-consuming for the vision system design problem.

Author Contributions

Y.C. and T.O. conceived and designed the study; Y.C. performed the experiments; Y.C. analysed the data; T.U. and T.T. contributed experimental platform; Y.C. wrote the manuscript; J.O. and T.O. reviewed the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Newman, T.S.; Jain, A.K. A survey of automated visual inspection. Comput. Vis. Image Underst. 1995, 61, 231–262. [Google Scholar] [CrossRef]
  2. Golnabi, H.; Asadpour, A. Design and application of industrial machine vision systems. Robot. Comput. Integr. Manuf. 2007, 23, 630–637. [Google Scholar] [CrossRef]
  3. Malamas, E.N.; Petrakis, E.G.M.; Zervakis, M.; Petit, L.; Legat, J.D. A survey on industrial vision systems, applications and tools. Image Vis. Comput. 2003, 21, 171–188. [Google Scholar] [CrossRef]
  4. Cowan, C.K.; Kovesi, P.D. Automatic sensor placement from vision task requirements. IEEE Trans. Pattern Anal. Mach. Intell. 1988, 10, 407–416. [Google Scholar] [CrossRef]
  5. Tarabanis, K.; Tsai, R.Y.; Allen, P.K. Automated sensor planning for robotic vision tasks. In Proceedings of the IEEE International Conference on Robotics and Automation, Sacramento, CA, USA, 9–11 April 1991; pp. 76–82. [Google Scholar]
  6. Tarabanis, K.; Tsai, R.Y. Computing viewpoints that satisfy optical constraints. In Proceedings of the Computer Vision and Pattern Recognition, Maui, HI, USA, 3–6 June 1991; pp. 152–158. [Google Scholar]
  7. Hutchinson, S.A.; Kak, A.C. Planning sensing strategies in a robot work cell with multi-sensor capabilities. IEEE Trans. Robot. Autom. 1989, 5, 407–416. [Google Scholar] [CrossRef]
  8. Cameron, A.; Wu, H.L. Identifying and localizing electrical components: A case study of adaptive goal-directed sensing. In Proceedings of the IEEE International Symposium on Intelligent Control, Arlington, VA, USA, 13–15 August 1991; pp. 495–500. [Google Scholar]
  9. Thuilot, B.; Martinet, P.; Cordesses, L.; Gallice, J. Position based visual servoing: Keeping the object in the field of vision. In Proceedings of the IEEE International Conference on Robotics and Automation, Washington, DC, USA, 11–15 May 2002. [Google Scholar]
  10. Rahimian, P.; Kearney, K.J. Optimal camera placement for motion capture systems. IEEE Trans. Vis. Comput. Graph. 2017, 3, 1209–1221. [Google Scholar] [CrossRef] [PubMed]
  11. Ito, A.; Tsujiuchi, N.; Okada, Y. Multipurpose optimization of camera placement and application to random bin-picking. In Proceedings of the 41st Annual Conference of the Industrial Electronics Society, Yokohama, Japan, 9–12 November 2015; pp. 528–533. [Google Scholar]
  12. Foix, S.; Alenyà, G.; Torras, C. 3D Sensor planning framework for leaf probing. In Proceedings of the Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 6501–6506. [Google Scholar]
  13. Murase, H.; Nayar, S.K. Illumination planning for object recognition using parametric eigenspaces. IEEE Trans. Pattern Anal. Mach. Intell. 1994, 6, 1219–1227. [Google Scholar] [CrossRef]
  14. Pfeifer, T.; Wiegers, L. Reliable tool wear monitoring by optimized image and illumination control in machine vision. Measurement 2000, 28, 209–218. [Google Scholar] [CrossRef]
  15. Yi, S.; Haralick, R.M.; Shapiro, L.G. Automatic sensor and light source positioning for machine vision. In Proceedings of the 10th International Conference on Pattern Recognition, Atlantic City, NJ, USA, 16–21 June 1990; pp. 55–59. [Google Scholar]
  16. Eltoft, T.; deFigueiredo, R.J.P. Illumination control as a means of enhancing image features in active vision systems. IEEE Trans. Pattern Anal. Mach. Intell. 1995, 4, 1520–1530. [Google Scholar] [CrossRef] [PubMed]
  17. Slusallek, P.; Seidel, H.P. Vision—An architecture for global illumination calculations. IEEE Trans. Vis. Comput. Graph. 1995, 1, 77–96. [Google Scholar] [CrossRef]
  18. Cang, N.Y.C.; Wu, C.C. Automatic optimal lighting adjustment and control for vision recognition. In Proceedings of the 14th IFToMM World Congress, Taipei, Taiwan, 25–30 October 2015. [Google Scholar]
  19. Aoki, S.; Nagao, T. Automatic construction of tree-structural image transformation using genetic programming. In Proceedings of the IEEE 10th International Conference on Image Analysis and Processing, Venice, Italy, 27–29 September 1999. [Google Scholar]
  20. Shirakawa, S.; Nagao, T. Genetic Image Network (GIN): Automatically construction of image processing algorithm. In Proceedings of the International Workshop on Advanced Image Technology (IWAIT), Bangkok, Thailand, 8–9 January 2007. [Google Scholar]
  21. Bai, H.; Yata, N.; Nagao, T. Automatic finding of optimal image processing for extracting concrete image cracks using features ACTIT. IEEJ Trans. Electr. Electron. 2012, 7, 308–315. [Google Scholar] [CrossRef]
  22. Lillywhite, K.; Tippetts, B.; Lee, D. Self-tuned evolution-constructed features for general object recognition. Pattern Recognit. 2012, 45, 241–251. [Google Scholar] [CrossRef]
  23. Lillywhite, K.; Lee, D.; Tippetts, B.; Archibald, J. A feature construction method for general object recognition. Pattern Recognit. 2013, 46, 3300–3314. [Google Scholar] [CrossRef]
  24. Kumar, R.; Lal, S.; Kumar, S.; Chand, P. Object detection and recognition for a pick and place robot. In Proceedings of the 2014 Asia-Pacific World Congress on Computer Science and Engineering (APWC on CSE), Nadi, Fiji, 4–5 November 2014. [Google Scholar]
  25. Ogata, T.; Tsujimoto, K.; Yukisawa, T.; Huang, Y.J.; Arai, T.; Ueyama, T.; Takada, T.; Ota, J. Automated design of image recognition process for picking system. Int. J. Autom. Technol. 2016, 10, 737–752. [Google Scholar] [CrossRef]
  26. Ogata, T.; Yukisawa, T.; Arai, T.; Ueyama, T.; Takada, T.; Ota, J. Automated design of image recognition in capturing environment. IEEJ Trans. Electr. Electron. 2017, 12, S49–S55. [Google Scholar] [CrossRef]
  27. Chen, Y.; Ogata, T.; Ueyama, T.; Takada, T.; Ota, J. Automated design of the field-of-view, illumination, and image pre-processing parameters of an image recognition system. In Proceedings of the 13th IEEE Conference on Automation Science and Engineering (CASE), Xi’an, China, 20–23 August 2017; pp. 1079–1084. [Google Scholar]
  28. Gevers, T.; Smeulders, A.W.M. Color-based object recognition. Pattern Recognit. 1999, 32, 453–464. [Google Scholar] [CrossRef]
  29. Drew, M.S.; Wei, J.; Li, Z.N. Illumination-invariant color object recognition via compressed chromaticity histograms of color-channel-normalized images. In Proceedings of the 6th International Conference on Computer Vision, Bombay, India, 4–7 January 1998; pp. 533–540. [Google Scholar]
  30. Alferez, R.; Wang, Y.F. Geometric and illumination invariants for object recognition. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 505–536. [Google Scholar] [CrossRef]
  31. Diplaros, A.; Gevers, T.; Patras, I. Combining color and shape information for illumination-viewpoint invariant object recognition. IEEE Trans. Image Process. 2006, 15, 1–11. [Google Scholar] [CrossRef] [PubMed]
  32. Bala, R.; Eschbach, R. Spatial color-to-grayscale transform preserving chrominance edge information. In Proceedings of the 12th Color and Imaging Conference, Scottsdale, AZ, USA, 9–12 November 2004; pp. 82–86. [Google Scholar]
  33. Grundland, M.; Dodgson, N. Decolorize: Fast, contrast enhancing, color to grayscale conversion. Pattern Recognit. 2007, 40, 2891–2896. [Google Scholar] [CrossRef]
  34. Gooch, A.A.; Olsen, S.C.; Tumblin, J.; Gooch, B. Color2Gray: Salience-preserving color removal. ACM Trans. Graph. 2005, 24, 634–639. [Google Scholar] [CrossRef]
  35. Kanan, C.; Cottrell, G.W. Color-to-grayscale: Does the method matter in image recognition? PLoS ONE 2012, 7, e29740. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Proposed algorithm to design FOV, illumination and image pre-processing parameters for recognition system. *: The procedures will be skipped if the current selected point has been searched before.
Figure 1. Proposed algorithm to design FOV, illumination and image pre-processing parameters for recognition system. *: The procedures will be skipped if the current selected point has been searched before.
Sensors 18 01656 g001
Figure 2. Illustration of the estimation of the FOV size. The positional relation between camera, object(s) and the work plane are shown on left side. The right side shows the length-width ratio of the recognition area and FOV. In the given example, the ratio of FOV is larger than that of the recognition area, which suggests FOV width to represent FOV size.
Figure 2. Illustration of the estimation of the FOV size. The positional relation between camera, object(s) and the work plane are shown on left side. The right side shows the length-width ratio of the recognition area and FOV. In the given example, the ratio of FOV is larger than that of the recognition area, which suggests FOV width to represent FOV size.
Sensors 18 01656 g002
Figure 3. Contour models (marked with red lines) made by different contrast values; the number of contours decrease as the contrast increases. Less contours in a model increase the number of detections that have to be matched and thus result in a longer time taken for matching.
Figure 3. Contour models (marked with red lines) made by different contrast values; the number of contours decrease as the contrast increases. Less contours in a model increase the number of detections that have to be matched and thus result in a longer time taken for matching.
Sensors 18 01656 g003
Figure 4. Contour models (marked with red lines) made by different contrast values, the number of contours decrease as the contrast increases. Less contours in a model increase the number of detections that should be matched and thus result in a longer time taken for matching. Algorithm proposed for contrast value design of the chosen recognition algorithm [HALCON (an image processing library of MVTec Company) shape-based matching].
Figure 4. Contour models (marked with red lines) made by different contrast values, the number of contours decrease as the contrast increases. Less contours in a model increase the number of detections that should be matched and thus result in a longer time taken for matching. Algorithm proposed for contrast value design of the chosen recognition algorithm [HALCON (an image processing library of MVTec Company) shape-based matching].
Sensors 18 01656 g004
Figure 5. Experimental devices, scenes and recognition targets. Monocular camera and ring-shaped illumination were attached to the end effector of a six-DoF manipulator. Three scenes were prepared on a piece of black cloth; Scene 1 and Scene 2 were for recognition; Scene 3 was for making templates. Two sides of a semi-transparent plastic part (20 mm in length, 20 mm in width, and 8 mm in height) were chosen as the recognition targets.
Figure 5. Experimental devices, scenes and recognition targets. Monocular camera and ring-shaped illumination were attached to the end effector of a six-DoF manipulator. Three scenes were prepared on a piece of black cloth; Scene 1 and Scene 2 were for recognition; Scene 3 was for making templates. Two sides of a semi-transparent plastic part (20 mm in length, 20 mm in width, and 8 mm in height) were chosen as the recognition targets.
Sensors 18 01656 g005
Figure 6. Optimal illumination and FOV conditions designed for the three conditions. Condition I: 1 shoot under strong green illumination. Condition II: four shoots under illumination with red component relatively higher. Condition III: 1 shoot under strong green and blue illumination.
Figure 6. Optimal illumination and FOV conditions designed for the three conditions. Condition I: 1 shoot under strong green illumination. Condition II: four shoots under illumination with red component relatively higher. Condition III: 1 shoot under strong green and blue illumination.
Sensors 18 01656 g006
Figure 7. R-channel images, greyscale images for the two scenes under the optimal design of Condition III [illumination RGB (red, green and blue) equals to 15, 225 and 240, with a wide FOV], and their recognition results. Objects labelled in green were those that could be correctly recognised, while the red one could not be recognised.
Figure 7. R-channel images, greyscale images for the two scenes under the optimal design of Condition III [illumination RGB (red, green and blue) equals to 15, 225 and 240, with a wide FOV], and their recognition results. Objects labelled in green were those that could be correctly recognised, while the red one could not be recognised.
Sensors 18 01656 g007
Figure 8. An image taken with no red component in illumination (illumination RGB were set to 0, 225, and 255, respectively), and its R-channel image.
Figure 8. An image taken with no red component in illumination (illumination RGB were set to 0, 225, and 255, respectively), and its R-channel image.
Sensors 18 01656 g008
Table 1. List of design variables.
Table 1. List of design variables.
Design FactorNameDescriptionRange
FOVShoot timeNumber of images required in one recognition for the entire area1, 4, …, n2
Camera distanceRepresents FOV sizeDetermined by shoot time
IlluminationLight strength (Red)Strength of red component in illumination[0, 255]
Light strength (Green)Strength of green component in illumination[0, 255]
Light strength (Blue)Strength of blue component in illumination[0, 255]
Recognition algorithmDiscriminatorThresholds for classifying different kinds of recognition objects(0, 1)
ContrastContrast value to extract contour model from template[0, 255]
Table 2. Experimental conditions.
Table 2. Experimental conditions.
ConditionIllumination Channel(s)Increment of Illumination Parameter(s)Recognition Image(s)
IG only1Greyscale
IIRGB15Greyscale
IIIRGB15R-channel
Table 3. Best-three designs of Condition I.
Table 3. Best-three designs of Condition I.
RankRGBFOVContrastFmeasurePositional Error (mm)Angular Error (°)
102320wide10.930.642.1
20790narrow30.930.583.4
30840narrow30.860.353.6
The results were ranked by their evaluations; higher rank represents better evaluation. A wide FOV denotes one shot at a camera distance of 158 mm, and a narrow FOV denotes four shots at a camera distance of 105 mm.
Table 4. Best-three designs of Condition II.
Table 4. Best-three designs of Condition II.
RankRGBFOVContrastFmeasurePositional Error (mm)Angular Error (°)
119512075narrow41.000.603.1
22404575narrow31.000.923.0
310530150narrow41.001.152.9
The results were ranked by their evaluations; higher rank represents better evaluation. A wide FOV denotes one shot at a camera distance of 158 mm, and a narrow FOV denotes four shots at a camera distance of 105 mm.
Table 5. Best-three designs of Condition III.
Table 5. Best-three designs of Condition III.
RankRGBFOVContrastFmeasurePositional Error (mm)Angular Error (°)
115225240wide111.000.320.4
20225150wide111.000.500.4
345225210wide91.000.620.4
The results were ranked by their evaluations; higher rank represents better evaluation. A wide FOV denotes one shot at a camera distance of 158 mm, and a narrow FOV denotes four shots at a camera distance of 105 mm.

Share and Cite

MDPI and ACS Style

Chen, Y.; Ogata, T.; Ueyama, T.; Takada, T.; Ota, J. Automated Field-of-View, Illumination, and Recognition Algorithm Design of a Vision System for Pick-and-Place Considering Colour Information in Illumination and Images. Sensors 2018, 18, 1656. https://doi.org/10.3390/s18051656

AMA Style

Chen Y, Ogata T, Ueyama T, Takada T, Ota J. Automated Field-of-View, Illumination, and Recognition Algorithm Design of a Vision System for Pick-and-Place Considering Colour Information in Illumination and Images. Sensors. 2018; 18(5):1656. https://doi.org/10.3390/s18051656

Chicago/Turabian Style

Chen, Yibing, Taiki Ogata, Tsuyoshi Ueyama, Toshiyuki Takada, and Jun Ota. 2018. "Automated Field-of-View, Illumination, and Recognition Algorithm Design of a Vision System for Pick-and-Place Considering Colour Information in Illumination and Images" Sensors 18, no. 5: 1656. https://doi.org/10.3390/s18051656

APA Style

Chen, Y., Ogata, T., Ueyama, T., Takada, T., & Ota, J. (2018). Automated Field-of-View, Illumination, and Recognition Algorithm Design of a Vision System for Pick-and-Place Considering Colour Information in Illumination and Images. Sensors, 18(5), 1656. https://doi.org/10.3390/s18051656

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop