*Proceedings* **Data Extraction in Insurance Photo-Inspections Using Computer Vision †**

#### **Mateo Gende 1,2,\* , Marcos Ortega 1,2 and Manuel G. Penedo 1,2**


Published: 21 August 2020

**Abstract:** Recent advances in computer vision and artificial intelligence allow for a better processing of complex information in many fields of human activity. One such field is vehicle expertise and inspection. This paper presents the development of systems for the automatic reading of French and Spanish license plates, as well as odometer value reading in dashboard photographs. These were trained and validated with real examples of more than 4000 vehicles, while addressing typical problems with irregular data acquisition. The systems proposed have found use in a real environment and are employed as assistance in vehicle appraisal.

**Keywords:** computer vision; machine learning; deep learning; license plate; odometer

#### **1. Introduction**

The ever-advancing field of artificial intelligence is constantly finding new uses in many activities and businesses. Vehicle insurance and expertise companies involve activities such as automobile inspections in which many images are taken to record and verify not only possible damages but also identity, in the form of license plate and vehicle identification number, and usage, like travelled distance. Possible mistakes in acquiring these data can result in problems ranging from insurance fraud allegations to uninsured vehicles being involved in accidents. Experts performing routine activities such as manually typing data in the field are susceptible to these kinds of mistakes. Computer vision techniques, combined with machine learning models, can be used to produce systems that are able to extract this information from images directly. This can have the benefit of speeding up the task of data acquisition by skipping manual typing entirely or, at least, provide a layer of verification by notifying the expert when the data they introduced does not match the automatic reading, indicating there may have been a typing mistake or a problem with how the image was taken. Blurred, poorly illuminated images make for deficient expertise material.

#### **2. Methodology**

#### *2.1. License Plate Reading*

To read license plate numbers, a hybrid system was developed. First, the license plate is localised and segmented from the image using a novel convolutional neural network based on U-net [1] architecture. From the segmentation map, the four corner points of the license plate rectangle are obtained so a projection transformation can be applied. With this transformation we obtain a flattened, standardised image of just the license plate, from which we can extract the characters using computer

vision techniques. Figure 1 shows a summary of these steps. Each character image is then classified using a convolutional neural network. Finally, the ordered character list is filtered using regular expressions to fix possible errors and classified as a Spanish, French or incorrectly read license plate.

**Figure 1.** Summary of the license plate reading procedure: (**left**) Original image. (**center**) Segmentation and perspective estimation. (**right**) Cleaning, thresholding and single character segmentation.

#### *2.2. Odometer Reading*

For odometers, a different approach was taken due to the inherent variability of analysing many different dashboards from any car make or model. To read the kilometre count, the user is asked to mark a point in the image over the digits that they want read. This way the system can limit the surface of the image that is analysed, as well as know which transcription is the one of interest, avoiding unnecessary processing and returning only what was requested from an image that may contain irrelevant data from the trip counter to time and date. A standardised window is taken from the image around the marked point and fed to a YOLOv3 [2] network. This model segments and classifies the numbers found in the image fitting a bounding box for each, as seen on Figure 2. Next, the user-supplied point is used to determine the extent of the figure of interest, including similar numbers in close proximity. For this, the size and position of already included numbers is used, providing flexibility to compensate for much of the heterogeneity of these images.

**Figure 2.** Example of character recognition in a car odometer.

#### **3. Results**

A test dataset of 251 real images of varying illumination and quality was constructed in order to evaluate license plate reading. Of these, 236 were perfectly transcribed by the system, achieving an accuracy of 94.02%. For the remaining 15 incorrectly read, the system was able to report an image quality problem in 12 of them, while three mistakes, 1.20%, went unnoticed. In total, 100 dashboard images were used to test the odometer reading. The system read 75 of them correctly for an accuracy of 75%.

**Funding:** This research was co-funded by Ministerio de Ciencia, Innovación y Universidades and the European Union (European Regional Development Fund—ERDF) as part of the RTC-2017-5908-7 research project. We wish to acknowledge the support received from the Centro de Investigación de Galicia CITIC, funded by Consellería de Educación, Universidade e Formación Profesional from Xunta de Galicia and European Union. (European Regional Development Fund - FEDER Galicia 2014-2020 Program) by grant ED431G 2019/01.

**Acknowledgments:** This work was supported by and used material from project STEPS (RTC-2017-5908-7).

**Conflicts of Interest:** The authors declare no conflict of interest.

#### **References**


c 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
