Next Article in Journal
Deep Learning of Fuzzy Weighted Multi-Resolution Depth Motion Maps with Spatial Feature Fusion for Action Recognition
Previous Article in Journal
Tip Crack Imaging on Transparent Materials by Digital Holographic Microscopy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Inspection of Aeronautical Mechanical Assemblies by Matching the 3D CAD Model and Real 2D Images

by
Hamdi Ben Abdallah
1,2,
Igor Jovančević
2,
Jean-José Orteu
1,* and
Ludovic Brèthes
2
1
Institut Clément Ader (ICA), Université de Toulouse, CNRS, IMT Mines Albi, INSA, UPS, ISAE, Campus Jarlard, 81013 Albi, France
2
DIOTASOFT, 201 Rue Pierre et Marie Curie, 31670 Labège, France
*
Author to whom correspondence should be addressed.
J. Imaging 2019, 5(10), 81; https://doi.org/10.3390/jimaging5100081
Submission received: 27 June 2019 / Revised: 13 September 2019 / Accepted: 14 October 2019 / Published: 19 October 2019

Abstract

:
In the aviation industry, automated inspection is essential for ensuring quality of production. It allows acceleration of procedures for quality control of parts or mechanical assemblies. As a result, the demand of intelligent visual inspection systems aimed at ensuring high quality in production lines is increasing. In this work, we address a very common problem in quality control. The problem is verification of presence of the correct part and verification of its position. We address the problem in two parts: first, automatic selection of informative viewpoints before the inspection process is started (offline preparation of the inspection) and, second, automatic treatment of the acquired images from said viewpoints by matching them with information in 3D CAD models is launched. We apply this inspection system for detecting defects on aeronautical mechanical assemblies with the aim of checking whether all the subparts are present and correctly mounted. The system can be used during manufacturing or maintenance operations. The accuracy of the system is evaluated on two kinds of platform. One is an autonomous navigation robot, and the other one is a handheld tablet. The experimental results show that our proposed approach is accurate and promising for industrial applications with possibility for real-time inspection.

1. Introduction

Industrial quality control is often realized as an automated end-of-line inspection system that checks certain features on the products. In the aeronautical field, many industrial companies use computer vision to automate tiresome operations of quality control in mechanical assembly for several reasons. On one hand, the human operators have implicit limits like overseeing of the missing parts and low repeatability rate [1]. On the other hand, the inspection task must be carried out in a cycle time established for a specific production process which is hard to achieve with manual inspection [2].
In modern manufacturing, industry robots are employed to conduct the inspection tasks with machine vision systems. These robots provide the physical link between intelligence and action. Usually, a robot inspection system has three modules: (1) mechanical: links, joints, end-effector, base, and mobile platform; (2) vision-based sensor system mounted on its end-effector; and (3) the perception, planning, and control unit, which interprets the sensory input and plans and controls the robot’s actions [3].
In this paper, we present a robust approach for detecting a common type of defect on aeronautical mechanical assemblies whose CAD model we dispose and exploit extensively. Our system aims to verify the presence of the correct part within the assembly. It also checks that the part is well mounted. The parts we are checking are metallic supports whose color and texture are often similar to the one of the background, making our task challenging. More details as well as illustrations will be given further in the article.
We evaluated our algorithm using two different systems: robotic and tablet-based handheld interface. Our approach consists of two processes.
First, automatic selection of informative viewpoints is started, given a priori 3D CAD models of the object and of the the whole assembly. The method proposed in this paper aims to find the best viewpoint according to designed scoring function.
Second, after the favorable viewpoints have been selected, our system verifies that the needed elements of the assembly are present and that they have been mounted in the correct position. The CAD model is used as a reference and it describes the assembly as it should be.
The proposed method has the following advantages; it (1) combines the precise information of CAD model with fast and inexpensive 2D sensors, (2) inspects parts with arbitrary geometrical shapes, (3) processes in real-time, and (4) fully automates the inspection process.
This work has been conducted as a successful collaboration between the Institut Clément Ader research laboratory and the Diota © company (www.diota.com). More precisely, within the Institut Clément Ader research laboratory, the “Metrology, Instrumentation, Control and Monitoring” research group carries out research work on vision-based inspection for manufacturing or maintenance applications. The Diota © company has been recently focusing on vision-based solutions for industrial control, relying on its core technology in CAD-based tracking and augmented reality.
The remainder of this paper is organized as follows. Section 2 introduces related works. The overview of the proposed inspection system is given in Section 3. The proposed method for the matching between the set of projected edgelets and the set of real edges is explained in Section 4. The overview of viewpoint selection scheme and some considerations about generating a set of candidate local viewpoints and the scoring function are detailed in Section 5. Section 6 shows a few empirical experiments and discusses about experimental results. Finally, in Section 7, some future directions are presented and the paper is concluded.

2. Related Work

In the aviation industry, the product inspection requirements have become stricter and more challenging, as products become smaller and more complex and the assembling process becomes faster. Consequently, an automated visual inspection system is one of the most important pieces of equipment for quality assurance and production efficiency with objectivity, endurance, and repeatability equal to and even surpassing human perception.
The systems that we are interested in are based on standard industrial robots that use a vision-based sensor system mounted on their end-effectors, such as those described in [4] for visual weld seam inspection, or the multi-axis Degree of Freedom (DoF) robot arm mounting an optical 3D scanner for inspection of mechanical components. We are also interested in the works aiming to find the optimal viewpoints as the one described in [5]. Further, our focus moves to the developed algorithms for mechanical assembly inspection. We also give some references to the works in the domain of CAD-based tracking and localization, which are not our domains of contribution, but whose results we use as inputs to our inspection method.
In this work, the starting point is a 3D CAD model of an assembly to be inspected. The first step is offline viewpoint selection, which is performed per object to be inspected. In the computer vision community, the viewpoint selection has been investigated. Examples are optimal 3D object recognition [6], grasp detection [7], human actions recognition [8], visual search [9], or global 3D reconstruction [10]. The problem of viewpoint selection is defined as generating candidate viewpoints and constructing scoring function for finding the best viewpoints with the highest score as presented by Liu et al. [11]. For reducing the search space, many authors use an approximate sphere [12] or a cylinder to generate the candidate viewpoints [11]. In our work, the candidate viewpoints are sampled over a sphere, therefore each viewpoint is at a safe distance from the center of mass of the element to be inspected, so as to comply with industrial security standards for ensuring the security of the inspected element as well as of the inspecting robot.
After the offline process of viewpoint selection is completed, the robotic inspection system is ready for operation. Comparison of acquired 2D or 3D data with the CAD model is a fast-growing type of inspection. In fact, it is the only effective and precise method of inspection, because the model contains an exact specification of the part [13] and provides a rich and useful source of information. In early works, Newman et al. [14] presented automated visual inspection (AVI) and techniques that have been reported in the literature from 1988 to 1993. They especially focused on the expanded role of CAD models in many systems of inspection. Jovancevic et al. [15] show that utilizing CAD model whenever possible is beneficial for their results in the task of exterior aircraft inspection. Viana et al. [16] have presented a CAD-based inspection of mechanical parts by means of comparison between a reference and a test image. The CAD-based matching method compares an image generated from a CAD model with an image acquired with the 2D camera, using primitives extracted from the contours. Berndt et al. [17] present a new technique for the detection of assembly errors using virtual images created from the CAD model. Their software compares a virtual image with the image of the real component. VisionLib © , developed by the Fraunhofer IGD, combines the strengths of real-time object localization, live comparison against CAD data, and augmented reality for inspections and virtual measurements. It offers augmented reality-based stationary and mobile inspection systems for quality control. Image processing and a CAD-based approach have been also applied to the shipbuilding inspection task [18].
Recently, systems for aircraft inspection have been developed as well, such as the mobile robot named Air-Cobot [1,15].
Our work strongly relies on a result of CAD-based tracking and localization by combining 3D data with 2D images. Using model-based tracking technologies, CAD data is registered in real-time to objects captured in camera images. De Crescenzio et al. [19] present a prototype system for aircraft maintenance and inspection. The system includes a marker-less camera pose estimation to track mechanical parts by using the SURF features [20] based on an offline and an online process. In the offline process, they first acquire a reference image of the object to be analyzed and they extract local invariant features. These extracted features are stored in the system for later use. In the online process, the system continuously processes the camera’s video stream, processing each incoming frame by extracting the SURF features from the frame and matching with the local features extracted from the reference image.
To the best of our knowledge, the existing methods for assembly/CAD conformity check are based on comparison of rendered synthetic images with real images [16,17]. Those images are further compared in terms of their respective sets of extracted geometrical primitives such as lines and circles/ellipses. Our method proposes three main novelties:
  • Offline phase of 3D edgelets extraction enables the extraction of meaningful 3D edges, which are prone to produce corresponding 2D edges.
  • Offline preparation phase: used for selecting the most favorable viewpoints for our inspection type.
  • Matching edgelets one by one and full control of comparison between expected shape and observed shape by taking into account the environment of the element (context) in both cases, when the element is present and absent.

3. The Proposed Inspection Systems

3.1. Overview of the Proposed Systems

We are aiming to support two modes of operation: one based on utilizing multiple sensors (two cameras and a 3D scanner) moved by a robot, and another relying on a single camera linked to a handheld tablet.

3.1.1. Robot-Based Inspection

The robot-based inspection platform is made of a mobile robot (Figure 1) equipped with three sensors (two cameras and a 3D scanner) mounted on the robot arm effector. Note that the sensor DiotaSensor ONE is a product commercialized by Diota © . The same holds for a whole robotic solution we are describing, which is commercialized under the name SIRIS © .
The sensors mounted on the robot (see Figure 2) allow several features:
  • Inspection: A high-resolution camera with a reduced field of view allows capturing details and observing the elements very finely with 7 DoF. This is a high-resolution camera with large sensor, which, together with 25 mm focal length lens, enables high sensitivity and observability of fine details in the working zone. Our typical working distance is ∼600 mm. At this distance, our image covers a zone 311 mm × 245 mm, and each pixel covers a square of dimensions 0.12 mm × 0.12 mm. Therefore, one-pixel error means an error of ∼1/10 mm. The specifications of the inspection camera are shown in Table 1.
    As shown in Figure 2, our vision-based sensor is equipped with a ring-type Effilux lighting system positioned around the inspection camera. This light source is only activated in the moment of image acquisition, when the robot is not moving.
  • Localization: A wide field tracking camera allows precise localization of the effector with respect to the part it controls. The choice of the camera is the result of extensive and permanent evaluation and testing of all significant industrial cameras available on the market (uEye © (https://en.ids-imaging.com), Basler © (www.baslerweb.com), etc). Our augmented reality solution (see Section 4.2) is well established and has been accepted by many industrial partners. The specifications of the tracking camera are shown in Table 2.
  • 3D scan: A structured light stereo sensor complements the sensor’s capabilities by digitizing the areas observed in 3D.
  • Controlling: The whole is controlled by an industrial PC which manages data acquisition and localization of the effector.
In this paper, we focus on the inspection based on 2D image analysis. The preliminary results on the automatic inspection method based on matching the 3D CAD model and real 3D point clouds have been presented in our recent conference paper [21]. These two methods treat two different inspection cases. Although the 2D inspection method presented in this paper puts attention on verifying the presence of certain rigid elements, the 3D inspection method focuses on measuring a 3D distance between nonrigid elements and checking conformity of this distance with predefined security standards.

3.1.2. Tablet-Based Inspection

The tablet-based inspection platform is made of a tablet with a mounted single camera. This set-up has been commercialized since many years by Diota © under the name Diota AR Tablet © . The tablet-based interface (Figure 3) is equipped with a HD high-speed industrial camera and ensures a consistent high-quality experience in various work contexts. The device will display 3D model superimposed on a real image of the corresponding item of interest (augmented reality). While the industrial assembly is scanned by an operator, our algorithmic approach automatically checks whether or not all the items have been present and correctly mounted. At the end of the inspection, a report is automatically generated. It includes details of any nonconforming parts that should be replaced or repaired.

3.2. Data Acquisition

We use two kinds of platform to collect our data: an autonomous navigation robot (Figure 1) and a handheld tablet (Figure 3). The handheld platform is a tablet computer that utilizes a single camera, whereas the robotic platform utilizes two cameras: one for tracking which has a wide field of view, and one for inspection which has a smaller field of view. By independently and separately using both platforms, to test the robustness of our approach, we collected data indoors with different points of view (angle, scale) and with different lighting conditions.
The data collected from the tablet is meant for real-time use, therefore images as well as videos are collected. The data from the robotic platform is meant for the analysis with no strong time constraints, as it can be done during and after the data collection routine is completed by the robot. A few examples of our CAD models and data samples are shown in Figure 4 and Figure 5, respectively.

4. Inspection Process for the Defect Detection

Our inspection process for the defect detection, illustrated in Figure 6 and Figure 7, consists of four main steps. First, from the CAD model of the observed object, 3D contour points (edgelets) are extracted. Further, the edgelets with their direction vectors are projected onto the image using a pinhole camera model (extrinsic and intrinsic parameters). Note that the pose of the effector is due to the localization module. Finally, we perform the matching between the set of projected edgelets and the set of real edges. We explain each building block of our strategy in the following sections.

4.1. Step 1 (Edgelet Generation)

An edgelet, E i , is defined as a 3D point, p i = ( x i , y i , z i ) , and a normalized direction vector, d i , of the edgelet. These points, p i , represent the 3D contour points of an object, see Figure 8. The edgelets are extracted offline by rendering a CAD model, as explained in [22]. The edgelets has been used in other works, see, e.g., in [23,24,25].
The edgelet generation process is well explained in [22]. Loesch et al. [22] introduce an improvement of this process through “dynamic edgelets” extracted online for each point of view. It should be noted that we have opted for using “static edgelets” defined in [22], as edgelets extracted offline using the dihedral angle criteria. This is because, as mentioned by [22], the generation of dynamic online edgelets induces an additional computing cost. Moreover, we use the extracted edgelets for generating optimal views in an offline process.
  • In the case of robot, there is no interest for dynamic edgelets, as the point of view is known in advance and there is neither video stream nor tracking as in the case of tablet.
  • In the case of tablet, we have more severe limitations in computing power so the static edgelets are our first choice for the moment.
The output of this step is a set of edgelets that are evenly distributed.

4.2. Step 2 (Camera Pose Estimation)

Advantage of our use case is that we dispose of the CAD model of the assembly. Further, we rely on an in-house developed system for initial pose estimation based on 2D-3D alignment, our inspection camera being then considered as moving in a partially known environment. The mentioned algorithm accurately estimates the camera’s pose with respect to the object by including the geometric constraints of its CAD model in an optimization process. In other words, the object is tracked in real time.

4.3. Step 3 (3D/2D Projection)

In this phase, we project 3D points of edgelets onto the image using known extrinsic and intrinsic camera parameters (see Figure 9). The input of this step are an image, the camera pose with respect to the assembly (CAD), and a set of evenly distributed edgelets of the element to be inspected.
To acquire the camera intrinsic parameters, we used an offline calibration method based on the observation of a classical pattern. Based on various observations of the pattern under different arm configurations, we obtained the intrinsic parameters of all the cameras used on the system along with the extrinsic parameters (relative localization) and their absolute localization relatively to the arm effector reference frame. A detailed description of the calibration procedure is beyond the scope of this paper.
Having the pose and camera intrinsic parameters, we can project the 3D point p i (see Figure 9) and the direction vector d i . By rotating the projected direction vector d i p by + 90 and 90 , we obtain the inward-pointing normal, n i + , (pointing towards the interior of the surface) and outward-pointing normal, n i , (pointing towards the exterior of the surface). Therefore, following is true; ( d i p , n i + ) = + 90 and ( d i p , n i ) = 90 . By concatenating n i + and n i , we obtain search line l i (see Figure 9b).

4.4. Step 4 (Matching Projected Edgelets with Real Edges)

The goal of this step is the matching between the set of projected edgelets and the set of real edges according to the direction of the normal vectors calculated in step 3. We calculate real edges by employing known Canny edge detector [26]. If at least one real edge is found on the search line, the edgelet is considered matched; otherwise, it is considered not matched.

4.4.1. Parasitic Edges Handling

In this section, we describe our principal contribution, which was introduced to remove as much as possible irrelevant edges. Namely, we anticipate the existence of some portion of parasitic edges coming from other elements mounted in the vicinity of the inspected element. Indeed, often there are edges very close to the external boundary of the element to be inspected. We call them parasitic edges or unwanted (irrelevant) edges. To solve this kind of problem, we introduce a new notion called context image of an element to be inspected (see Figure 10). From the CAD data and the particular camera position, we form a synthetic view (context image), which has exactly the same perspective as the real image, only that the inspected element is excluded from the assembly (see Figure 11).
The process of decreasing number of parasitic edges is illustrated in Figure 11. First, we project the edgelets of the element of interest onto the context image, as explained in the step 3.
Further, we form search lines (see Figure 11d) and look for the intensity change in this virtual context image. Intensity changes are shown in Figure 11d. If such a change in context image is found, it is very probable that this edge will be present in the real image as well. Therefore, we do not consider this edgelet for matching. These edgelets are shown in red in Figure 11d. Other edgelets, shown in green in Figure 11d, are considered in matching phase.

4.4.2. Edges Weighting

For each edgelet in the 3D CAD projected onto the image plane, we are interested in finding an image location that contains a real change in image intensity (a real edge). For this, we first produce an edge image and search for candidate locations along a search line, l i , that is obtained as described in step 3. Each edge location along the search line that is within a prespecified distance from the location of the projected edgelet is weighted according to a Gaussian function of the distance of the edge point from the location of the projected edgelet. W i is a weight of the edge e i , and W j is a weight of the edge e j (see Figure 12). Instead of simply counting matched edgelets, these weights will be added up in the final score. We do this to favorize the edgelets matched with very close real edges and penalize those matched with real edges that are far away.

4.4.3. Gradient Orientations

When searching for an edge in the image, we may encounter edges because of image noise or geometric features that arise from the parts different from the inspection element. To reject such candidates, at each encountered edge location, we compute gradient of the edge and find the angle it makes with the search line θ i (see Figure 13). If this angle is greater than a threshold ( θ i > θ t h ), we reject the candidate and continue our search. If no edges are found within the search limits, the orientation criteria are satisfied, and we decide that no edge corresponding to the inspection element of interest is present. In our experiments, we have set θ t h = 20 .

4.4.4. Occlusion Handling Using CRI (Color Rendering Index)

Image locations where edgelets of an inspection element are projected can belong to a different element, due to the inspection element being occluded by it in the real assembly. Searching for real image edges corresponding to the projected edgelets at such image locations is thus incorrect. An example of an occlusion problem is shown in Figure 14. In Figure 14, we find that the red element is behind the transparent green element. This means that part of the red element should be occluded by the transparent green element.
Occlusions are managed by determining what objects are in front of others in a scene, which allows determining whether each rendered object appears in front of any other object in the scene, on a per-pixel basis.
To ignore such locations, we first render the inspection element along with the surrounding context by assigning to each element a different random color using the Color Rendering Index (CRI). To filter projected edgelets occluded by another elements, we generate a mask using the colored render, so that all areas in the render that have a color different from that of the inspection element are ignored. This way, only non-occluded edgelets are considered for matching.
In computer graphics, there are two popular rendering techniques for handling occlusions: Z-buffer technique and classic ray tracing.
The Z-buffer (or depth Buffer) is intrinsically used by the classical graphic pipeline, moreover the Z-Buffer only offers a finite amount of precision to store the depth value for each pixel, problems can occur when multiple objects resolve to the same depth value. In addition, the available precision is affected by the choice of the near and far clipping planes for the 3D scene [27].
Ray tracing is also an effective rendering technique, but it is very slow and is massively used for collision checking on some sparse pixels and knowing which the object are collided (we need all the image here and do not need to know which object is hided by the one in front). Moreover, none of the two mentioned technique will add quality to the mask rendering.
The proposed CRI technique is perfectly adapted to the objectives: knowing which object can be seen at which pixel of the image. In addition, classical graphic rendering with using flat color as index allows to do it very quickly.
The operations described from Section 4.4.1, Section 4.4.2, Section 4.4.3 and Section 4.4.4 provide the so-called filtering of projected edgelets.

4.5. Step 5 (Making Decision on State of the Element)

Producing an inspection decision based on matching filtered edgelets independently, with edges in the real image, can lead to false positives, as the matched edges in the real image may also arise from an element with a different shape than the inspected one. To take the global shape of the element into account, we propose to characterize the filtered edgelets and the edges detected in the real image using the Shape Context framework [28]. The Shape Context (represented by a set of points sampled from the edges of the object) is a shape descriptor that captures the relative positions of points on the shape contours. This gives a globally discriminative characterization of the shape, and it is not just a localized descriptor. These are then used to measure similarities between shape context representations of the two sets of edge locations.
For example, Figure 15 shows all projected edgelets (black), matched edges (green), and filtered projected edgelets (red). In this example, the computation of the distance between the green shape and the red shape provides a similarity score = 0.01.
The workflow of the making decision step is shown in Figure 16. The decision is taken using two indicators: the similarity score ( similarity score ) and the matched ratio ( matched ratio ), represented by the relationship between matched and not matched edgelets.
In Figure 17, we show the matched ratio and matched th computed on two different cases (element OK and element NOK).
  • If similarity score < similarity th , the element is considered NOK (absent or incorrectly mounted). In our experiments, we set similarity th = 0.7 , based on an extensive comparison between different shapes.
  • If similarity score > similarity th , we compute the ratio between the number of matched edgelets and the number of not matched edgelets ( matched ratio ).
  • If matched ratio > matched th , the element is considered conform; otherwise, the element is considered as defective (missing or badly mounted). The value of matched th has been determined experimentally (see Section 6).
  • In conclusion, the decision is taken using two indicators: the similarity score ( similarity score ) and the matched ratio ( matched ratio ).

5. Viewpoint Selection

The initial set-up of camera viewpoints cannot be done manually, because a human operator cannot define with sufficient accuracy the camera position that will allow acquiring the best viewpoint of the element to be inspected. Therefore, we need a a (semi-)automatic offline configuration process that is used to compute the best viewpoints, which can help to improve the quality and the efficiency of inspection [29].
The strategy proposed to find the best viewpoint (computing the 6D location (position and orientation) of the camera with respect to the scene to be observed) for the aeronautical mechanical assembly inspection, based on the 3D model of the assembly, is illustrated in Figure 18. A scoring function is constructed which combines three factors including the amount of the information, the viewing quality, and the visibility of the information to evaluate the quality of the candidate viewpoint [11]. Finally, the best viewpoint can be selected from all of candidate viewpoints according to the scoring function.

5.1. Generate a Set of Candidate Viewpoints

A viewpoint, V p o i n t , is defined by a point, V ( x , y , z ) , around an approximate sphere and the sensor positioning (distance and orientation) relatively to the part surface. This sphere is defined as a set of views.
The radius, ρ , of the visibility sphere (Figure 19) is chosen such that each viewpoint is at a safe distance from the center of mass of the element to be inspected ( c x , c y , c z ) . This is to comply with industrial security standards for ensuring security of the inspected element as well as of the inspecting robot.
The sphere (Figure 20) is defined as a set of viewpoints, and the candidate viewpoints are evenly distributed on the sphere surface according the range of longitude, θ , and colatitude, φ .

5.2. Selection of the Best Viewpoint from Candidate Viewpoints

The best viewpoint is selected according to the value of the scoring function. The scoring function f s c o r e quantifies the quality of the candidate viewpoint. The best viewpoint with highest score is selected. Our scoring function combines three criteria:
Visibility of the information, f v i s i b i l i t y .
Ambiguity of observed shape due to parasite edges, f p a r a s i t e .
Similarity of observed shape to expected shape, f s h a p e .

5.2.1. Visibility of the Information f v i s i b i l i t y

We evaluate each generated viewpoint for visibility of the inspection element by taking into account the occlusion using CRI as explained in Section 4.4.4.
The viewpoint that yields a render with the greatest number of pixels belonging to the inspection element is chosen to be the viewpoint with best visibility, and therefore the best point of view for performing the inspection task according to this “visibility” criteria (see Figure 21).

5.2.2. Ambiguity of Observed Shape Due to Parasite Edges f p a r a s i t e

Each candidate viewpoint is also evaluated by computing the number of parasite edges and edgelets of the inspection element in a render from the viewpoint as described in Section 4.4.1.
The viewpoint that maximizes the ratio of the number of edgelets after filtering to the number of parasite edges is chosen as the best inspection viewpoint according to this criteria. In Figure 22, the difference between a viewpoint that yields many parasite edges and a viewpoint that yields few is presented.
Note that there is no guarantee that any viewpoint will be chosen for a given element, namely, if too few edges are left, we announce not being able to perform inspection.

5.2.3. Similarity of Observed Shape to Expected Shape f s h a p e

Finally, each candidate viewpoint is alsot evaluated by computing the similarity between the shape of filtered projected edgelets of the inspection element (obtained by rejecting those prone to parasitic edges and occlusion) and the unfiltered projected edgelets of the same, i.e., the expected shape (see Figure 23).

5.2.4. The Scoring Function f s c o r e

The three criteria presented before for evaluating each viewpoint are combined using a scoring function f s c o r e detailed below,
f s c o r e = w v × f v i s i b i l i t y + w p × f p a r a s i t e + w s × f s h a p e
w v + w p + w s = 1
w i ’s are weights that assign importances to the different viewpoint evaluation criteria, and they can be adjusted by the user depending on the importance of the criteria for a specific inspection application. f v i s i b i l i t y , f p a r a s i t e , and f s h a p e are normalized.

5.2.5. Viewpoint Selection in the Handheld Tablet Mode Based on Augmented Reality

Viewpoints are generated in an offline process by using the CAD model of the assembly. While in the robot mode, the robot is programmed to reach as close to those points as possible (path planning); in the tablet mode, there is an augmented reality-based graphical interface which guides the operator to reach the desired viewpoints or at least their neighborhoods (see Figure 24). Note that our CAD-based tracking module enables the augmented reality experience and constant real-time pose estimation of the camera with respect to the inspected assembly.

6. Experiments and Discussion

6.1. Illumination Description

Our work took inside a building/factory, which allows a quite stable environment/stable light conditions. Moreover, for the inspection camera, we can adjust the exposure depending on the object to inspect. Either we give a constant exposure time or choose a level of exposure, which will be set by an automatic exposure algorithm. Our robot is not moving when acquiring images. Therefore, we are taking more than one acquisition for each element. We change the exposure time between those acquisitions. The goal is to improve as much as we can the contrast of the image by choosing the one which has the most favorable histogram within region of interest. Depending on the element and the inspection task, this region of interest can be either the whole image either the part of it. Additionally, to the exposure time of the camera, we are also controlling the additional light source that we added around our inspection camera. We are activating this light source ONLY in the moment of image acquisition, when the robot is not moving.
Our CAD-based tracking module, which exploits the tracking camera images, is a robust stable solution already industrialized for years. We do not use any additional light source for this task, as artificial light can cause problems in performing tracking. We are relying on the ambient light coming from the context of the factory.

6.2. Performance of Tracking

In terms of performance, the reprojection error of the assembly edges is 1.5 pixels. Our use case is conformity of the assembly with CAD (missing elements), and not precise measures of the surface defects, etc. Therefore, 1.5 pixels is sufficient for our task.
Usually the elements which are inspected make a minority of the elements observed by the tracking camera. The tracking module works with at least 40 % of the CAD present in reality. In our use case (typically assemblies close to the end of the production line), this is always satisfied.

6.3. Study of Some Parameters

First, we conducted experiments to determine the effect of parameters on the inspection accuracy and also to obtain the best parameters for each type of data. Parameters used in this exploration are (a) threshold ( matched th ) for making decision on state of element, and (b) the high t h and low t l threshold used by the hysteresis procedure of Canny edge detection algorithm.

6.3.1. Determination of Optimal matched th

To find the optimal decision threshold, we conducted some experiments to compute the optimal parameter of our algorithm and evaluate its effectiveness on six inspected elements with a total number of 541 images acquired indoors at different scale, angle, and lighting conditions.
According to the results obtained in Figure 25 and Table 3, we can use matched th in the range of {30%, 60%}. In most cases, we set matched th = 40 % .

6.3.2. Determination of Optimal Canny Edge Detection Parameters

Canny edge detection contains two adjustable parameters—high, t h , and low, t l , thresholds—for the gradient magnitude image, which can affect the effectiveness of algorithm. The number of edge points is dependent on the choice of these thresholds. Figure 26 shows the influence of these values on edge detection. In our experiments, we used t l in the range {20, 40}. Following Canny’s recommendation, the high threshold should be three times the lower threshold ( t h = 3 × t l ). In most cases, we set ( t l , t h ) = ( 30 , 90 ) . When we reduce the lower threshold, the algorithm gives more edge points, which produces more noise (see Figure 23b). We reject noise edges by employing our gradient orientation condition described in Section 4.4.3.
After the optimal parameters of the algorithm have been determined, we can evaluate our algorithm. The proposed inspection method has been tested on two kinds of platforms: a handheld platform and a robotized platform.
The framework used for the evaluations is first provided in Section 6.4.

6.4. Accuracy, Sensitivity, Specificity

Table 5 outlines definitions of the true positive (TP), the false positive (FP), the true negative (TN), and the false negative (FN) in defect detection [30].
Based on these notations, the detection success rate (known also as detection accuracy) is defined as the ratio of correct assessments and number of all assessments. Sensitivity is defined as the correct detection of defective samples. Similarly, the specificity can be defined as the correct detection of defect-free samples. Finally, the overall accuracy could be calculated based on the metrics of sensitivity and specificity. The accuracy, sensitivity, and specificity can be defined as
accuracy = TP + TN TP + FN + TN + FP
sensitivity = TP TP + FN
specificity = TN TN + FP

Precision and Recall

Precision and recall metrics are used to validate the performance of defect detection. Precision and recall can be defined as
precision = TP TP + FP
recall = TP TP + FN
F s c o r e = 2 × precision × recall precision + recall
The F s c o r e is calculated to measure the overall performance of a defect detection method. The highest value of the F s c o r e is 100%, whereas its lowest value is 0%.

6.5. Experiments Using the Handheld Inspection Platform

The performance of the proposed method was evaluated by considering the performance metrics such as accuracy, sensitivity, specificity, precision and recall defined in Section 6.4.
We performed 4 experiments with a total number of 74 inspected elements in 7256 images. A few examples of our dataset are shown in Figure 27. The experiments show that our method can run in real-time on a video-stream with HD resolution ( 1920 × 1080 ) . The values of specificity, accuracy, sensitivity, precision and recall of our method are shown in Table 6 and Table 7.

6.6. Experiments Using the Robotized Inspection Platform

The robotic platform utilizes two cameras: one for tracking which has a wide field of view, and one for inspection with a reduced field of view, allowing capturing the details and observing the elements very finely. With a displacement speed of 250 mm/s, we can control 2 elements every 30 s. The distance between camera and inspected element has been set at 600 mm with a tolerance of ± 6 mm.
The approach was tested on a dataset acquired in a factory with a total number of 43 inspected elements and 643 images. Each element is acquired once at a time and decision is made on one acquired image. A few examples of our dataset are shown in Figure 28.
The overall performance of our method is represented by F s c o r e = 76.72 % . The experimental results show that our proposed approach is accurate and promising for industrial deployment. The values of specificity, accuracy, sensitivity, precision, and recall of our method are shown in Table 8 and Table 9.
Note that the results obtained with the handheld mode are better than those obtained with the robotic platform. There are three main reasons for this:
  • The robotic platform dataset is acquired in a factory, on a real and very complex assembly at the very end of the production line (almost all the elements are mounted), see Figure 28. Therefore, the environment around inspected element is very cluttered and the matching task becomes much more challenging. On the other hand, the “handheld” dataset has been acquired on quite simplified assemblies in laboratory conditions, see Figure 27. In those examples, inspected elements are mainly placed in the middle of uniform regions and parasitic edges are not many.
  • In this moment, multiview inspection is possible with handheld platform only. This means that we make a final decision on the element presence by fusing results obtained from multiple views; whereas, in robotic mode, one-shot (one view) inspection is performed. This is a feature to be added to our robotic platform as further work.
  • Mobility of the handheld system enables much more precise tracking and pose estimation which is a crucial precondition of our method.

6.7. Viewpoint Selection

In Figure 29 and Table 10, we show the result of the viewpoint selection method on an example where the weighting coefficients have been set to w v = 0.5 , w p = 0.25 , and w s = 0.25 . The definition of additional rules for selecting automatically the weighting coefficients is under study.

7. Conclusions and Future Work

In this paper, we proposed an automated inspection system for mechanical assemblies in an aeronautical context. We address the problem of inspection in two parts: the first being automatic selection of informative viewpoints before the inspection process is started, and the second being automatic treatment of the acquired images from said viewpoints by matching them with information in 3D CAD model. The system is not limited to aeronautical parts; it could also be applied in a wide range of industries. The developed system shows that the precise state-of-the-art localization methods can support an industrial process. Furthermore, the experimental results show that our proposed method is highly accurate in detecting missing or badly mounted mechanical parts.
We are currently working on (1) utilizing the 3D cloud of points provided by the 3D scanner (robotized platform) in order to solve some inspection problems that cannot be solved using 2D images (for instance when we need to check that 2 cables are at a minimum distance) [21], (2) considering how the combination of 3D point clouds and 2D images can improve the detection results, (3) concerning the viewpoint selection we are trying to define some rules for choosing the weighting factors automatically, and (4) implementing multiview inspection in the robotic inspection mode in order to take the decision by fusing results obtained from multiple views.
We are also investigating deep learning-based methods relying on synthetic data that will be presented in a forthcoming paper.

Author Contributions

Investigation: H.B.A., I.J., J.-J.O. and L.B.; Software: H.B.A. and I.J.; Supervision: I.J., J.-J.O. and L.B.; Writing—original draft: Hamdi Ben Abdallah, I.J. and J.-J.O.; Writing—review & editing: H.B.A., I.J., J.-J.O. and L.B.

Funding

This research was funded by the French “Région Occitanie”. The APC was funded by IMT Mines Albi.

Acknowledgments

This research work has been carried out within a PhD thesis co-funded by the French “Région Occitanie”. We would like to thank the Région Occitanie for its financial support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jovancevic, I.; Pham, H.H.; Orteu, J.J.; Gilblas, R.; Harvent, J.; Maurice, X.; Brèthes, L. 3D Point Cloud Analysis for Detection and Characterization of Defects on Airplane Exterior Surface. J. Non Destr. Eval. 2017, 36, 36–74. [Google Scholar]
  2. San Biagio, M.; Beltrán-González, C.; Giunta, S.; Del Bue, A.; Murino, V. Automatic inspection of aeronautic components. Mach. Vis. Appl. 2017, 28, 591–605. [Google Scholar] [CrossRef]
  3. Trivedi, M.M.; Chen, C.; Marapane, S.B. A Vision System for Robotic Inspection; Manipulation. Computer 1989, 22, 91–97. [Google Scholar] [CrossRef]
  4. Worn, H.; Langle, T.; Gauss, M. ARIKT: Adaptive Robot Based Visual Inspection. Kunstl. Intell. 2003, 17, 33–35. [Google Scholar]
  5. Raffaeli, R.; Mengoni, M.; Germani, M. Context Dependent Automatic View Planning: The Inspection of Mechanical Components. Comput.-Aided Des. Appl. 2013, 10, 111–127. [Google Scholar] [CrossRef]
  6. Deinzer, F.; Denzler, J.; Niemann, H. Viewpoint Selection—Planning Optimal Sequences of Views for Object Recognition. In Proceedings of the 10th International Conference on Computer Analysis of Images and Patterns, Groningen, The Netherlands, 25–27 August 2003; pp. 65–73. [Google Scholar]
  7. Gualtieri, M.; Patt, R. Viewpoint Selection for Grasp Detection. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems, Vancouver, BC, Canada, 24–28 September 2017; pp. 258–264. [Google Scholar]
  8. Rudoy, D.; Zelnik-Manor, L. Viewpoint Selection for Human Actions. Int. J. Comput. Vis. 2012, 97, 243–254. [Google Scholar] [CrossRef]
  9. Wixson, L. Viewpoint selection for visual search. In Proceedings of the 1994 Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 21–23 June 1994; pp. 800–805. [Google Scholar]
  10. Jonnalagadda, K.; Lumia, R.; Starr, G.; Wood, J. Viewpoint selection for object reconstruction using only local geometric features. In Proceedings of the 2003 IEEE International Conference on Robotics and Automation, Taipei, Taiwan, 4–19 September 2003; Volume 2, pp. 2116–2122. [Google Scholar]
  11. Liu, C.A.; Dong, R.F.; Wu, H. Flying Robot Based Viewpoint Selection for the Electricity Transmission Equipment Inspection. Math. Probl. Eng. 2014, 1–10. [Google Scholar] [CrossRef]
  12. Vasquez-Gomez, J.I.; Lopez-Damian, E.; Sucar, L.E. View planning for 3D object reconstruction. In Proceedings of the International Conference on Intelligent Robots and Systems, St Louis, MO, USA, 10–15 October 2009; pp. 4015–4020. [Google Scholar]
  13. Prieto, F.; Redarce, T.; Lepage, R.; Boulanger, P. Visual System for Fast; Automated Inspection of 3D Parts. Int. J. CAD/CAM Comput. Graph. 1998, 13, 211–227. [Google Scholar]
  14. Newman, T.; Jain, A. A survey of automated visual inspection. Comput. Vis. Image Underst. 1995, 61, 231–262. [Google Scholar] [CrossRef]
  15. Jovancevic, I.; Larnier, S.; Orteu, J.J.; Sentenac, T. Automated exterior inspection of an aircraft with a pan-tilt-zoom camera mounted on a mobile robot. J. Electron. Imaging 2017, 24, 061110. [Google Scholar] [CrossRef]
  16. Viana, I.; Orteu, J.J.; Cornille, N.; Bugarin, F. Inspection of aeronautical mechanical parts with a pan-tilt-zoom camera: An approach guided by the computer-aided design model. J. Electron. Imaging 2015, 24, 061118. [Google Scholar] [CrossRef]
  17. Berndt, D. Digital Assembly Inspection: Automatic Quality Control Even for Small Quantities. Available online: https://idw-online.de/de/news532655 (accessed on 10 December 2016).
  18. Kim, M.; Choi, W.; Kim, B.-C.; Kim, H.; Seol, J.; Woo, J.; Ko, K. A vision-based system for monitoring block assembly in shipbuildin. Comput.-Aided Des. 2015, 59, 98–108. [Google Scholar] [CrossRef]
  19. Cusano, C.; Napoletano, P. Visual recognition of aircraft mechanical parts for smart maintenance. Comput. Ind. 2017, 86, 26–33. [Google Scholar] [CrossRef]
  20. Bay, H.; Tuytelaars, T.; Van Gool, L. SURF: Speeded Up Robust Features. Comput. Vis. Image Underst. 2008, 110, 346–359. [Google Scholar] [CrossRef]
  21. Ben Abdallah, H.; Orteu, J.J.; Dolives, B.; Jovancevic, I. 3D Point Cloud Analysis for Automatic Inspection of Aeronautical Mechanical Assemblies. In Proceedings of the 14th International Conference on Quality Control by Artificial Vision (QCAV), Mulhouse, France, 15–17 May 2019. [Google Scholar]
  22. Loesch, A.; Bourgeois, S.; Gay-Bellile, V.; Dhome, M. Generic edgelet-based tracking of 3D objects in real-time. In Proceedings of the International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 6059–6066. [Google Scholar]
  23. Drummond, T.; Cipolla, R. Real-time visual tracking of complex structures. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 932–946. [Google Scholar] [CrossRef] [Green Version]
  24. Marchand, E.; Bouthemy, P.; Chaumette, F.; Moreau, V. Robust real-time visual tracking using a 2D-3D model-based approach. In Proceedings of the Seventh IEEE International Conference on Computer Vision, Kerkyra, Greece, 20–27 September 1999; pp. 262–268. [Google Scholar]
  25. Vacchetti, L.; Lepetit, V.; Fua, P. Combining edge and texture information for real-time accurate 3D camera tracking. In Proceedings of the Third IEEE and ACM International Symposium on Mixed and Augmented Reality, Arlington, VA, USA, 2–5 November 2004; pp. 48–56. [Google Scholar]
  26. Canny, J. A Computational Approach to Edge Detector. IEEE Trans. PAMI 1986, PAMI-8, 679–698. [Google Scholar]
  27. Stevenson, A.; Wood, D.; Lehenbauer, D. Automatically Generating Appropriate Near and Far Clipping Planes for a 3D Scene While Guaranteeing Minimum of Z-Buffer Precision. U.S. Patent 2007/0216710A1, 20 September 2007. [Google Scholar]
  28. Belongie, S.; Malik, J.; Puzicha, J. Shape matching; object recognition using shape contexts. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 509–522. [Google Scholar] [CrossRef]
  29. Eitzinger, C.; Zambal, S.; Thanner, P. Chapter 10: Robotic Inspection Systems. In Integrated Imaging and Vision Techniques for Industrial Inspection; Advances in Computer Vision and Pattern Recognition; Liu, Z., Ukida, H., Ramuhalli, P., Niel, K., Eds.; Springer: New York, NY, USA, 2015. [Google Scholar]
  30. Rebhi, A.; Abid, S.; Fnaeich, F. Texture Defect Detection Using Local Homogeneity and Discrete Cosine Transform. World Appl. Sci. J. 2014, 31, 1677–1683. [Google Scholar]
Figure 1. The robot-based inspection platform and an aircraft engine to be inspected.
Figure 1. The robot-based inspection platform and an aircraft engine to be inspected.
Jimaging 05 00081 g001
Figure 2. (a) Vision-based sensory system mounted on the robot; (b) additional lighting around the inspection camera.
Figure 2. (a) Vision-based sensory system mounted on the robot; (b) additional lighting around the inspection camera.
Jimaging 05 00081 g002
Figure 3. Tablet-based interface equipped with a camera for visual inspection. (a) the tablet-based interface equipped with a camera (b) the tablet-based handheld interface used by an operator for the inspection of an aircraft engine
Figure 3. Tablet-based interface equipped with a camera for visual inspection. (a) the tablet-based interface equipped with a camera (b) the tablet-based handheld interface used by an operator for the inspection of an aircraft engine
Jimaging 05 00081 g003
Figure 4. Some of our CAD models: (a,b) aircraft engine, (c,d) two testing plates with several elements.
Figure 4. Some of our CAD models: (a,b) aircraft engine, (c,d) two testing plates with several elements.
Jimaging 05 00081 g004
Figure 5. Examples of our dataset: (1st and 3rd rows) real images and (2nd and 4th rows) corresponding renders of CAD models.
Figure 5. Examples of our dataset: (1st and 3rd rows) real images and (2nd and 4th rows) corresponding renders of CAD models.
Jimaging 05 00081 g005
Figure 6. Overview of the detection phase during online processing using robot-based inspection.
Figure 6. Overview of the detection phase during online processing using robot-based inspection.
Jimaging 05 00081 g006
Figure 7. Overview of the detection phase during online processing using tablet-based inspection.
Figure 7. Overview of the detection phase during online processing using tablet-based inspection.
Jimaging 05 00081 g007
Figure 8. Edgelets extraction: (a) example of CAD part, and (b) edgelets extracted from this CAD part.
Figure 8. Edgelets extraction: (a) example of CAD part, and (b) edgelets extracted from this CAD part.
Jimaging 05 00081 g008
Figure 9. (a) Projection of edgelets (3D points p i ), and (b) inward-pointing normals, n i + , outward-pointing normals, n i , and generation of search line l i .
Figure 9. (a) Projection of edgelets (3D points p i ), and (b) inward-pointing normals, n i + , outward-pointing normals, n i , and generation of search line l i .
Jimaging 05 00081 g009
Figure 10. (a) Example of CAD model; (b) context image of the blue element.
Figure 10. (a) Example of CAD model; (b) context image of the blue element.
Jimaging 05 00081 g010
Figure 11. Illustration of the parasitic edges handling: (a) input image, (b) projection of the element to be inspected, (c) context image of the element to be inspected and its projected edgelets, and (d) gradient changes in the context image and considered edgelets (green) and rejected edgelets (red).
Figure 11. Illustration of the parasitic edges handling: (a) input image, (b) projection of the element to be inspected, (c) context image of the element to be inspected and its projected edgelets, and (d) gradient changes in the context image and considered edgelets (green) and rejected edgelets (red).
Jimaging 05 00081 g011
Figure 12. Illustration of edges weighting.
Figure 12. Illustration of edges weighting.
Jimaging 05 00081 g012
Figure 13. Challenges due to outlier edges, such as those due to image noise and edges arising from features that do not belong to the inspection element, when searching for an image edge corresponding to the projected edgelet.
Figure 13. Challenges due to outlier edges, such as those due to image noise and edges arising from features that do not belong to the inspection element, when searching for an image edge corresponding to the projected edgelet.
Jimaging 05 00081 g013
Figure 14. Occlusion handling by rendering using Color Rendering Index (CRI): (a) CAD model rendered with CATIA composer, (b) CAD model rendered using CRI, (c) mask generated using CRI render, white areas represent region to be inspected, and (d) filtered edgelets (Figure 11d) projected inside the region of interest specified by the generated mask.
Figure 14. Occlusion handling by rendering using Color Rendering Index (CRI): (a) CAD model rendered with CATIA composer, (b) CAD model rendered using CRI, (c) mask generated using CRI render, white areas represent region to be inspected, and (d) filtered edgelets (Figure 11d) projected inside the region of interest specified by the generated mask.
Jimaging 05 00081 g014
Figure 15. The shapes used in the shape context approach.
Figure 15. The shapes used in the shape context approach.
Jimaging 05 00081 g015
Figure 16. Decision-making.
Figure 16. Decision-making.
Jimaging 05 00081 g016
Figure 17. (a) CAD model; (b) input image; (c) result of inspection with green OK (element present) and red NOK (element absent or incorrectly mounted); (df) the matched ratio (curve) and threshold matched th (red horizontal line), respectively, computed on the three different cases: element A (absent) in red, element B (present) in green, and element C (incorrectly mounted) in red.
Figure 17. (a) CAD model; (b) input image; (c) result of inspection with green OK (element present) and red NOK (element absent or incorrectly mounted); (df) the matched ratio (curve) and threshold matched th (red horizontal line), respectively, computed on the three different cases: element A (absent) in red, element B (present) in green, and element C (incorrectly mounted) in red.
Jimaging 05 00081 g017
Figure 18. The overview of viewpoint selection scheme.
Figure 18. The overview of viewpoint selection scheme.
Jimaging 05 00081 g018
Figure 19. A viewpoint on the visibility sphere.
Figure 19. A viewpoint on the visibility sphere.
Jimaging 05 00081 g019
Figure 20. The candidate viewpoint region is defined as a part of sphere surface. The accessible viewpoints in green and the inaccessible viewpoints, due to occlusions related to the global CAD, in red (a) for a given element to be inspected and (b) for the whole assembly.
Figure 20. The candidate viewpoint region is defined as a part of sphere surface. The accessible viewpoints in green and the inaccessible viewpoints, due to occlusions related to the global CAD, in red (a) for a given element to be inspected and (b) for the whole assembly.
Jimaging 05 00081 g020
Figure 21. (1st row) CAD model and (2nd row) different candidate viewpoints for inspecting the element in the red rectangle, with the corresponding f s c o r e .
Figure 21. (1st row) CAD model and (2nd row) different candidate viewpoints for inspecting the element in the red rectangle, with the corresponding f s c o r e .
Jimaging 05 00081 g021
Figure 22. Two viewpoints with best-view and worst-view according to the number of parasite edges.
Figure 22. Two viewpoints with best-view and worst-view according to the number of parasite edges.
Jimaging 05 00081 g022
Figure 23. (a) “Best-view”, (b) all edgelets of best-view, (c) filtred edgelets of best-view, (d) “worst-view”, (e) all edgelets of worst-view, and (f) filtered edgelets of worst-view.
Figure 23. (a) “Best-view”, (b) all edgelets of best-view, (c) filtred edgelets of best-view, (d) “worst-view”, (e) all edgelets of worst-view, and (f) filtered edgelets of worst-view.
Jimaging 05 00081 g023
Figure 24. User guidance in the handheld tablet mode (see the blue arrows).
Figure 24. User guidance in the handheld tablet mode (see the blue arrows).
Jimaging 05 00081 g024
Figure 25. Experimental results for determining the optimal r a t i o t h . The red horizontal line represents the min matched ratio in the case OK (element in green) and the max matched ratio in the case NOK (element in red).
Figure 25. Experimental results for determining the optimal r a t i o t h . The red horizontal line represents the min matched ratio in the case OK (element in green) and the max matched ratio in the case NOK (element in red).
Jimaging 05 00081 g025
Figure 26. Images of Canny algorithm results (a). Images (bh) have been obtained using the high and low threshold values specified in Table 4.
Figure 26. Images of Canny algorithm results (a). Images (bh) have been obtained using the high and low threshold values specified in Table 4.
Jimaging 05 00081 g026
Figure 27. Dataset used to evaluate our algorithms: (a,c,e,g) real images and (b,d,f,h) corresponding inspection results with OK elements in green and NOK elements in red.
Figure 27. Dataset used to evaluate our algorithms: (a,c,e,g) real images and (b,d,f,h) corresponding inspection results with OK elements in green and NOK elements in red.
Jimaging 05 00081 g027
Figure 28. Some examples of our dataset used to evaluate our approach in a context of robotized inspection in conditions of very cluttered environment.
Figure 28. Some examples of our dataset used to evaluate our approach in a context of robotized inspection in conditions of very cluttered environment.
Jimaging 05 00081 g028
Figure 29. Images (bj) correspond to different viewpoints of the element shown on image (a). The f s c o r e of each viewpoint is provided in Table 10.
Figure 29. Images (bj) correspond to different viewpoints of the element shown on image (a). The f s c o r e of each viewpoint is provided in Table 10.
Jimaging 05 00081 g029
Table 1. Inspection camera specifications.
Table 1. Inspection camera specifications.
Camera nameUI-3180 (PYTHON 5000)
Frame rate73 fps
Resolution ( h × v ) 2592 × 2048
Optical Area 12.44  mm × 9.830 mm
Resolution 5.31 MPix
Pixel size4.80 μ m
Lens namevs- 2514 h 1
Focal length25 mm ( 25.42 )
Minimal object distance (M.O.D.)300 mm
Angel of view ( h × v ) 21.5 × 28.5
Table 2. Tracking camera specifications.
Table 2. Tracking camera specifications.
Camera nameIDS 3070
Frame rate123 fps
Resolution ( h × v ) 2056 × 1542
Optical Area7.093 mm × 5.0320 mm
Resolution 3.17 MPix
Pixel size3.45 μ m
Lens nameKowa 8 mm
Focal length8 mm
Minimal object distance (M.O.D.) 0.1  mm
Angel of view ( h × v ) 79.7 × 63.0
Table 3. Determination of optimal parameter for making decision on state of element.
Table 3. Determination of optimal parameter for making decision on state of element.
Inspecting ElementState of ElementMax Matched Edgelet RatioMin Matched Edgelet Ratio
Part ANOK25%0%
Part BOK100%68%
Part CNOK30%0%
Part DOK100%70%
Part EOK100%65%
Part FNOK25%0%
Table 4. The proportional coefficient of high and low thresholds of Canny algorithm in Figure 26a.
Table 4. The proportional coefficient of high and low thresholds of Canny algorithm in Figure 26a.
FigureFigure 26bFigure 26cFigure 26dFigure 26eFigure 26fFigure 26gFigure 26h
( t l , t h )(10, 30)(20, 60)(30, 90)(40, 120)(50, 150)(60, 180)(70, 210)
Table 5. Definitions of TP, FP, TN, and FN in defect detection.
Table 5. Definitions of TP, FP, TN, and FN in defect detection.
Actually DefectiveActually Non-Defective
Detected as defectiveTPFP
Detected as nondefectiveFNTN
Table 6. Result of the evaluation (TP, TN, FP, and FN).
Table 6. Result of the evaluation (TP, TN, FP, and FN).
ExperimentNumber of ElementsNumber of ImagesTPTNFPFN
1 (Figure 27a,b)324911413500
2 (Figure 27c,d)1615803051264110
3 (Figure 27e,f)32921939900
4 (Figure 27g,h)5251356074482460
Total74725612195980570
Table 7. Performance measures (Section 6.5).
Table 7. Performance measures (Section 6.5).
AccuracySensitivitySpecificityPrecisionRecall F score
99.12%100%99.01%95.53%100%97.71%
Table 8. Result of the evaluation (TP, TN, FP, and FN).
Table 8. Result of the evaluation (TP, TN, FP, and FN).
Number of ElementsNumber of ImagesTPTNFPFN
43643145410880
Table 9. Performance measures (Section 6.6).
Table 9. Performance measures (Section 6.6).
AccuracySensitivitySpecificityPrecisionRecall F score
86.31%100%82.32%62.23%100%76.72%
Table 10. The f s c o r e of viewpoints for inspecting the element of Figure 29a.
Table 10. The f s c o r e of viewpoints for inspecting the element of Figure 29a.
FigureFigure 29bFigure 29cFigure 29dFigure 29eFigure 29fFigure 29gFigure 29hFigure 29iFigure 29j
f s c o r e 0.037660.168940.274470.4698850.530090.556250.601780.738310.80842

Share and Cite

MDPI and ACS Style

Ben Abdallah, H.; Jovančević, I.; Orteu, J.-J.; Brèthes, L. Automatic Inspection of Aeronautical Mechanical Assemblies by Matching the 3D CAD Model and Real 2D Images. J. Imaging 2019, 5, 81. https://doi.org/10.3390/jimaging5100081

AMA Style

Ben Abdallah H, Jovančević I, Orteu J-J, Brèthes L. Automatic Inspection of Aeronautical Mechanical Assemblies by Matching the 3D CAD Model and Real 2D Images. Journal of Imaging. 2019; 5(10):81. https://doi.org/10.3390/jimaging5100081

Chicago/Turabian Style

Ben Abdallah, Hamdi, Igor Jovančević, Jean-José Orteu, and Ludovic Brèthes. 2019. "Automatic Inspection of Aeronautical Mechanical Assemblies by Matching the 3D CAD Model and Real 2D Images" Journal of Imaging 5, no. 10: 81. https://doi.org/10.3390/jimaging5100081

APA Style

Ben Abdallah, H., Jovančević, I., Orteu, J. -J., & Brèthes, L. (2019). Automatic Inspection of Aeronautical Mechanical Assemblies by Matching the 3D CAD Model and Real 2D Images. Journal of Imaging, 5(10), 81. https://doi.org/10.3390/jimaging5100081

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop