Rulers2023: An Annotated Dataset of Synthetic and Real Images for Ruler Detection Using Deep Learning
Abstract
:1. Introduction
1.1. Ruler Detection in Images
1.2. Synthetic Image Generation for Deep Learning
- Data availability and scalability. Traditional methods for collecting real-world images are both time-consuming and expensive. Moreover, the acquisition of high-quality labeled data often requires additional human labor. Synthetic images, on the other hand, can be generated in large quantities with automated labeling, offering a scalable solution for data collection [54].
- Controlled experimental conditions. Among the challenges in using real-world data are the inherent variability and noise, which can introduce confounding factors into experiments. Synthetic images can be generated under controlled conditions, allowing for greater experimental rigor [55].
- Ethical and privacy concerns. Real-world images, especially those involving human subjects, often come with ethical and privacy concerns, including the need for informed consent. Synthetic images, being artificial constructs, bypass these issues, allowing for broader applicability in research [56,57].
1.3. The Novelty and Contributions of This Work
- Proposing a method to generate a dataset consisting of synthetic ruler image samples of high fidelity;
- Publicly providing the synthetic ruler image training dataset generated according to the presented method;
- Publicly providing the real ruler image training and test datasets;
- Presenting ruler detection benchmark results using six different CNN architectures;
- Experimentally showing that the created synthetic ruler image dataset is effective, sufficient, and superior to the real image dataset for the development of ruler detectors.
2. Materials and Methods
2.1. Generation of Synthetic Ruler Image Dataset
2.1.1. Requirements for Synthetic Ruler Images
- Variability and diversity: The dataset should include rulers of various lengths, colors, materials (e.g., plastic, metal, wood), and units of measurement (e.g., inches, centimeters). It should also cover a wide range of backgrounds, lighting conditions, and occlusions to mimic real-world scenarios:
- Ruler appearance: Generating rulers of different shapes, colors, materials. Ruler appearance can be simulated by randomly selecting width, height, color, transparency, and gradation marking type and placement from the predefined range of parameter values;
- Perspective and orientation: The dataset should represent rulers from multiple angles and perspectives, including top-down views and side views with varying degrees of perspective distortion. This can be achieved by randomly applying geometric transformations. Transformations (rotation, scaling, perspective distortions) can be introduced during dataset construction and at the time of training as data augmentation procedures;
- Contextual variety: Rulers should be placed in a variety of relevant contexts, such as on different surfaces, alongside common objects of measurement, and within complex scenes, to train the model for practical applications. Contexts can be simulated by using natural images as backgrounds;
- Occlusions: Including images where the ruler is partially obscured or interacting with other objects to simulate realistic use cases where the ruler is not always fully visible. Occlusions can be simulated during the training phase as a data augmentation procedure.
- Balanced image quality and resolution: High-resolution images are necessary to ensure that the measurement markings on the rulers are clear and discernible. On the other hand, unnecessarily large images would waste memory; moreover, image preprocessing for passing to neural networks would consume more time. The chosen input size of the neural network should be considered. Therefore, the resolution of images should be tuned while keeping in mind the maximum input size of the neural network and by leaving enough freedom for various augmentation transformations that enlarge the image regions.
- Realism: While synthetic images need to be diverse and comprehensive, they also need to be realistic enough not to introduce a domain gap when the model is applied to real-world images. The visual evaluation and performance of a ruler detector that is trained on a synthetic dataset provide feedback on realism.
- Consistency in synthetic generation parameters: Ensuring consistent application of synthetic image generation parameters such as lighting, textures, and noise models across the dataset to avoid biases in the training process. Controlled parameter sampling allows for the implementation of this requirement.
- Dataset size and balance: The size of the dataset must be large enough to train deep learning models effectively and should be balanced in terms of the variety of ruler types and scenarios presented.
- Legal considerations: The dataset creation process must adhere to intellectual property laws. It is especially relevant when selecting real background images for the dataset generation process; e.g., the COCO dataset [82] is licensed under a Creative Commons Attribution 4.0 license.
2.1.2. Method for Generating Synthetic Ruler Image Dataset
- Background preparation. The real image from the COCO dataset was randomly cropped and resized to px image size. The crop box was defined by randomly shifting the corner coordinates of the original image () by a maximum of 20% of the width/height of the image.
- Definition of the ruler area. The location of the ruler was designated in the pixel background area and centered. The length of the ruler, which is the longer dimension, was randomly chosen to be between 70% and 95% of the background width, while the width of the ruler, which is the shorter dimension, was randomly selected to be between 4% and 20% of the background height.
- Setup of the appearance of the rule. A set of random parameters that determined the visual attributes of the ruler was generated. The color of the ruler was defined in the HSV color space, where a triplet of values within the ranges of , , and was drawn randomly. The transparency level was established by selecting a "face alpha" value from the interval . The curvature at the corners of the ruler was set by choosing a value for the curvature parameter of the rectangle from the range . The areas for measurement markings, or gradations, were specified to be between 85% and 100% of the ruler’s length, with the width set at 30% to 45% for double-sided rulers or 40% to 60% for single-sided rulers. The decision to use a single-sided or double-sided ruler was made at random. The positions of the gradation areas were also randomly adjusted, with a shift of % from the edge of the ruler. The length of the ruler scale was chosen to fall within the range of measurement units. Subdivisions of the ruler scale, indicating tick positions, were selected from sets of , , or . For rulers with three-level subdivisions, the tick lengths were set to % of the width of the gradation area. For those with two-level subdivisions, tick lengths were determined to be %, %, or % of the width of the gradation area.
- Adding geometric distortions. The perspective transformation was simulated by randomly selecting ruler corner shifts and fitting the geometric transformation to those corner coordinate shifts. Each corner coordinate of the ruler was randomly shifted up/down by up to 20% of the ruler width and right/left by up to 10% of the ruler length. In addition, one of the ends of the ruler was widened by a value corresponding to the length up to the initial width of the ruler.
2.2. Synthetic Data Evaluation
- Training MobileNetV2 for 200 epochs using transfer learning or “from scratch”;
- Training MobileNetV2 for 200 epochs using transfer learning or “from scratch” with 5-fold cross-validation.
2.3. Data Collection
2.4. Software Used
- MATLAB programming and numeric computing platform (version R2022a; The Mathworks Inc., Natick, MA, USA) for the synthetic dataset creation workflow;
- Python (version 3.10.12) (https://www.python.org) (accessed on 21 August 2023) [88], an interpreted, high-level, general-purpose programming language. Used for machine learning applications and for data analysis and visualization;
- TensorFlow with Keras (version 2.10.1) (https://www.tensorflow.org (accessed on 21 August 2023)) [89], an open-source platform for machine learning. Used for the online data augmentation stage of the synthetic dataset creation workflow and for the training/testing of deep learning models;
- Albumentations (version 1.3.1) (https://albumentations.ai (accessed on 21 August 2023)) [90], a Python library for fast and flexible image augmentation. Used for the image augmentations during deep learning model training;
- OpenCV (version 4.8.0) (https://opencv.org/ (accessed on 21 August 2023)) [91], an open-source computer vision library. Used for image input/output and manipulations.
3. Results and Discussion
4. Conclusions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
ML | Machine learning |
DL | Deep learning |
AI | Artificial intelligence |
ANN | Artificial neural network |
CNN | Convolutional neural network |
DNN | Deep neural network |
GAN | Generative adversarial network |
DM | Diffusion model |
VAE | Variational autoencoder |
LR | Learning rate |
MSE | Mean squared error |
BBox | Bounding box |
Appendix A
References
- Deng, J.; Xuan, X.; Wang, W.; Li, Z.; Yao, H.; Wang, Z. A review of research on object detection based on deep learning. Proc. J. Phys. Conf. Ser. Iop Publ. 2020, 1684, 012028. [Google Scholar] [CrossRef]
- Diwan, T.; Anirudh, G.; Tembhurne, J.V. Object detection using YOLO: Challenges, architectural successors, datasets and applications. Multimed. Tools Appl. 2023, 82, 9243–9275. [Google Scholar] [CrossRef] [PubMed]
- Tamulionis, M.; Sledevič, T.; Abromavičius, V.; Kurpytė-Lipnickė, D.; Navakauskas, D.; Serackis, A.; Matuzevičius, D. Finding the Least Motion-Blurred Image by Reusing Early Features of Object Detection Network. Appl. Sci. 2023, 13, 1264. [Google Scholar] [CrossRef]
- Pathak, A.R.; Pandey, M.; Rautaray, S. Application of deep learning for object detection. Procedia Comput. Sci. 2018, 132, 1706–1717. [Google Scholar] [CrossRef]
- Plonis, D.; Katkevičius, A.; Mališauskas, V.; Serackis, A.; Matuzevičius, D. Investigation of New Algorithms for Estimation of Losses in Microwave Devices Based on a Waveguide or a Meander Line. Acta Phys. Pol. A 2016, 129, 414–424. [Google Scholar] [CrossRef]
- Žuraulis, V.; Matuzevičius, D.; Serackis, A. A method for automatic image rectification and stitching for vehicle yaw marks trajectory estimation. Promet-Traffic Transp. 2016, 28, 23–30. [Google Scholar] [CrossRef]
- Xiao, Y.; Tian, Z.; Yu, J.; Zhang, Y.; Liu, S.; Du, S.; Lan, X. A review of object detection based on deep learning. Multimed. Tools Appl. 2020, 79, 23729–23791. [Google Scholar] [CrossRef]
- Varna, D.; Abromavičius, V. A System for a Real-Time Electronic Component Detection and Classification on a Conveyor Belt. Appl. Sci. 2022, 12, 5608. [Google Scholar] [CrossRef]
- Sledevič, T.; Serackis, A.; Plonis, D. FPGA Implementation of a Convolutional Neural Network and Its Application for Pollen Detection upon Entrance to the Beehive. Agriculture 2022, 12, 1849. [Google Scholar] [CrossRef]
- Sharaff, A.; Ukey, K.S.; Choure, R.; Couradiya, A.; Sinha, G. Automatic dimension detection of fish images. In Proceedings of the Data Driven Approach Towards Disruptive Technologies: Proceedings of MIDAS 2020, Dehradun, India, 4–5 September 2020; Springer: Berlin/Heidelberg, Germany, 2021; pp. 49–59. [Google Scholar]
- Konovalov, D.A.; Saleh, A.; Efremova, D.B.; Domingos, J.A.; Jerry, D.R. Automatic weight estimation of harvested fish from images. In Proceedings of the 2019 Digital Image Computing: Techniques and Applications (DICTA), Perth, Australia, 2–4 December 2019; pp. 1–7. [Google Scholar]
- Tseng, C.H.; Hsieh, C.L.; Kuo, Y.F. Automatic measurement of the body length of harvested fish using convolutional neural networks. Biosyst. Eng. 2020, 189, 36–47. [Google Scholar] [CrossRef]
- Konovalov, D.; Domingos, J.; Bajema, C.; White, R.; Jerry, D. Ruler detection for automatic scaling of fish images. In Proceedings of the International Conference on Advances in Image Processing, Bangkok, Thailand, 25–27 August 2017; pp. 90–95. [Google Scholar]
- Liu, S.; Zheng, W.; Wang, X.; Xiong, H.; Cheng, J.; Yong, C.; Zhang, W.; Zou, X. A novel depth measurement method for urban flooding based on surveillance video images and a floating ruler. Natural Hazards 2023, 1–23. [Google Scholar] [CrossRef]
- Dou, G.; Chen, R.; Han, C.; Liu, Z.; Liu, J. Research on water-level recognition method based on image processing and convolutional neural networks. Water 2022, 14, 1890. [Google Scholar] [CrossRef]
- Qiu, R.; Cai, Z.; Chang, Z.; Liu, S.; Tu, G. A two-stage image process for water level recognition via dual-attention CornerNet and CTransformer. Vis. Comput. 2023, 39, 2933–2952. [Google Scholar] [CrossRef]
- Bai, G.; Hou, J.; Zhang, Y.; Li, B.; Han, H.; Wang, T.; Hinkelmann, R.; Zhang, D.; Guo, L. An intelligent water level monitoring method based on SSD algorithm. Measurement 2021, 185, 110047. [Google Scholar] [CrossRef]
- Lin, Y.T.; Lin, Y.C.; Han, J.Y. Automatic water-level detection using single-camera images with varied poses. Measurement 2018, 127, 167–174. [Google Scholar] [CrossRef]
- Zhang, Z.; Zhou, Y.; Liu, H.; Gao, H. In-situ water level measurement using NIR-imaging video camera. Flow Meas. Instrum. 2019, 67, 95–106. [Google Scholar] [CrossRef]
- Chen, G.; Bai, K.; Lin, Z.; Liao, X.; Liu, S.; Lin, Z.; Zhang, Q.; Jia, X. Method on water level ruler reading recognition based on image processing. Signal Image Video Process. 2021, 15, 33–41. [Google Scholar] [CrossRef]
- Xue, Z.; Yu, K.; Pearlman, P.C.; Pal, A.; Chen, T.C.; Hua, C.H.; Kang, C.J.; Chien, C.Y.; Tsai, M.H.; Wang, C.P.; et al. Automatic detection of oral lesion measurement ruler toward computer-aided image-based oral cancer screening. In Proceedings of the 2022 44th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Glasgow, PA, USA, 21–29 May 2022; pp. 3218–3221. [Google Scholar]
- Xue, Z.; Yu, K.; Pearlman, P.; Chen, T.C.; Hua, C.H.; Kang, C.J.; Chien, C.Y.; Tsai, M.H.; Wang, C.P.; Chaturvedi, A.; et al. Extraction of Ruler Markings For Estimating Physical Size of Oral Lesions. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; pp. 4241–4247. [Google Scholar]
- Lama, N.; Kasmi, R.; Hagerty, J.R.; Stanley, R.J.; Young, R.; Miinch, J.; Nepal, J.; Nambisan, A.; Stoecker, W.V. ChimeraNet: U-Net for hair detection in dermoscopic skin lesion images. J. Digit. Imaging 2023, 36, 526–535. [Google Scholar] [CrossRef]
- Kasmi, R.; Hagerty, J.; Young, R.; Lama, N.; Nepal, J.; Miinch, J.; Stoecker, W.; Stanley, R.J. SharpRazor: Automatic removal of hair and ruler marks from dermoscopy images. Ski. Res. Technol. 2023, 29, e13203. [Google Scholar] [CrossRef]
- Ünver, H.M.; Ayan, E. Skin lesion segmentation in dermoscopic images with combination of YOLO and grabcut algorithm. Diagnostics 2019, 9, 72. [Google Scholar] [CrossRef]
- Abbas, Q.; Fondón, I.; Rashid, M. Unsupervised skin lesions border detection via two-dimensional image analysis. Comput. Methods Programs Biomed. 2011, 104, e1–e15. [Google Scholar] [CrossRef] [PubMed]
- Jaisakthi, S.M.; Mirunalini, P.; Aravindan, C. Automated skin lesion segmentation of dermoscopic images using GrabCut and k-means algorithms. IET Comput. Vis. 2018, 12, 1088–1095. [Google Scholar] [CrossRef]
- Öztürk, Ş.; Özkaya, U. Skin lesion segmentation with improved convolutional neural network. J. Digit. Imaging 2020, 33, 958–970. [Google Scholar] [CrossRef] [PubMed]
- Hameed, A.; Umer, M.; Hafeez, U.; Mustafa, H.; Sohaib, A.; Siddique, M.A.; Madni, H.A. Skin lesion classification in dermoscopic images using stacked Convolutional Neural Network. J. Ambient. Intell. Humaniz. Comput. 2021, 14, 3551–3565. [Google Scholar] [CrossRef]
- Chen, C.; Kojcev, R.; Haschtmann, D.; Fekete, T.; Nolte, L.; Zheng, G. Ruler based automatic C-arm image stitching without overlapping constraint. J. Digit. Imaging 2015, 28, 474–480. [Google Scholar] [CrossRef] [PubMed]
- Gooßen, A.; Schlüter, M.; Hensel, M.; Pralow, T.; Grigat, R.R. Ruler-based automatic stitching of spatially overlapping radiographs. In Proceedings of the Bildverarbeitung für die Medizin 2008: Algorithmen—Systeme—Anwendungen Proceedings des Workshops vom 6. bis 8. April 2008 in Berlin; Springer: Berlin/Heidelberg, Germany, 2008; pp. 192–196. [Google Scholar]
- Jaworski, N.; Farmaha, I.; Marikutsa, U.; Farmaha, T.; Savchyn, V. Implementation features of wounds visual comparison subsystem. In Proceedings of the 2018 XIV-th International Conference on Perspective Technologies and Methods in MEMS Design (MEMSTECH), Lviv, Ukraine, 18–22 April 2018; pp. 114–117. [Google Scholar]
- Gertsovich, I.; Nilsson, M.; Bartunek, J.; Claesson, I. Automatic estimation of a scale resolution in forensic images. Forensic Sci. Int. 2018, 283, 58–71. [Google Scholar] [CrossRef]
- Bhalerao, A.; Reynolds, G. Ruler detection for autoscaling forensic images. Int. J. Digit. Crime Forensics (IJDCF) 2014, 6, 9–27. [Google Scholar] [CrossRef]
- Tian, F.; Zhao, Y.; Che, X.; Zhao, Y.; Xin, D. Concrete crack identification and image mosaic based on image processing. Appl. Sci. 2019, 9, 4826. [Google Scholar] [CrossRef]
- Lukashchuk, B. Method of automatic identification of ruler scales on images to solve the problem of estimating the geometric dimensions of certain objects. In Proceedings of the 2022 IEEE 3rd KhPI Week on Advanced Technology (KhPIWeek), Kharkiv, Ukraine, 3–7 October 2022; pp. 1–4. [Google Scholar]
- Jun, S. Research on Image Measuring method based on monocular camera and Ruler. In Proceedings of the 2015 Conference on Informatization in Education, Management and Business (IEMB-15), Guangzhou, China, 12–13 September 2015; Atlantis Press: Amsterdam, The Netherlands, 2015; pp. 150–155. [Google Scholar]
- Telahun, M.; Sierra-Sossa, D.; Elmaghraby, A.S. Heuristic Analysis for In-Plane Non-Contact Calibration of Rulers Using Mask R-CNN. Information 2020, 11, 259. [Google Scholar] [CrossRef]
- Herrmann, M.; Zambanini, S.; Kampel, M. Image based measurement of ancient coins. In Proceedings of the VAST, Washington, DC, USA, 30 October–1 November 2007; Volume 7, pp. 55–62. [Google Scholar]
- Dunn, K.W.; Fu, C.; Ho, D.J.; Lee, S.; Han, S.; Salama, P.; Delp, E.J. DeepSynth: Three-dimensional nuclear segmentation of biological images using neural networks trained with synthetic data. Sci. Rep. 2019, 9, 18295. [Google Scholar] [CrossRef]
- Kriese, J.; Hoeser, T.; Asam, S.; Kacic, P.; Da Ponte, E.; Gessner, U. Deep Learning on Synthetic Data Enables the Automatic Identification of Deficient Forested Windbreaks in the Paraguayan Chaco. Remote Sens. 2022, 14, 4327. [Google Scholar] [CrossRef]
- Ge, N.; Liu, Y.; Xu, X.; Zhang, X.; Jiang, M. A Fast Generative Adversarial Network for High-Fidelity Optical Coherence Tomography Image Synthesis. Photonics 2022, 9, 944. [Google Scholar] [CrossRef]
- Matuzevičius, D. Synthetic Data Generation for the Development of 2D Gel Electrophoresis Protein Spot Models. Appl. Sci. 2022, 12, 4393. [Google Scholar] [CrossRef]
- Sa, I.; Lim, J.Y.; Ahn, H.S.; MacDonald, B. deepNIR: Datasets for generating synthetic NIR images and improved fruit detection system using deep learning techniques. Sensors 2022, 22, 4721. [Google Scholar] [CrossRef] [PubMed]
- Park, S.W.; Ko, J.S.; Huh, J.H.; Kim, J.C. Review on generative adversarial networks: Focusing on computer vision and its applications. Electronics 2021, 10, 1216. [Google Scholar] [CrossRef]
- Figueira, A.; Vaz, B. Survey on synthetic data generation, evaluation methods and GANs. Mathematics 2022, 10, 2733. [Google Scholar] [CrossRef]
- Joshi, I.; Grimmer, M.; Rathgeb, C.; Busch, C.; Bremond, F.; Dantcheva, A. Synthetic data in human analysis: A survey. arXiv 2022, arXiv:2208.09191. [Google Scholar]
- Lu, Y.; Wang, H.; Wei, W. Machine Learning for Synthetic Data Generation: A Review. arXiv 2023, arXiv:2302.04062. [Google Scholar]
- Rasmussen, I.; Kvalsvik, S.; Andersen, P.A.; Aune, T.N.; Hagen, D. Development of a Novel Object Detection System Based on Synthetic Data Generated from Unreal Game Engine. Appl. Sci. 2022, 12, 8534. [Google Scholar] [CrossRef]
- El Emam, K.; Mosquera, L.; Hoptroff, R. Practical Synthetic Data Generation: Balancing Privacy and the Broad Availability of Data; O’Reilly Media: Sebastopol, CA, USA, 2020. [Google Scholar]
- Man, K.; Chahl, J. A Review of Synthetic Image Data and Its Use in Computer Vision. J. Imaging 2022, 8, 310. [Google Scholar] [CrossRef]
- Abufadda, M.; Mansour, K. A survey of synthetic data generation for machine learning. In Proceedings of the 2021 22nd International arab Conference on Information Technology (ACIT), Muscat, Oman, 21–23 December 2021; pp. 1–7. [Google Scholar]
- Yang, Z.; Zhan, F.; Liu, K.; Xu, M.; Lu, S. AI-Generated Images as Data Source: The Dawn of Synthetic Era. arXiv 2023, arXiv:2310.01830. [Google Scholar]
- Ding, K.; Zhou, M.; Wang, H.; Gevaert, O.; Metaxas, D.; Zhang, S. A large-scale synthetic pathological dataset for deep learning-enabled segmentation of breast cancer. Sci. Data 2023, 10, 231. [Google Scholar] [CrossRef] [PubMed]
- Sapkota, B.B.; Popescu, S.; Rajan, N.; Leon, R.G.; Reberg-Horton, C.; Mirsky, S.; Bagavathiannan, M.V. Use of synthetic images for training a deep learning model for weed detection and biomass estimation in cotton. Sci. Rep. 2022, 12, 19580. [Google Scholar] [CrossRef] [PubMed]
- Makrushin, A.; Mannam, V.S.; Dittmann, J. Privacy-Friendly Datasets of Synthetic Fingerprints for Evaluation of Biometric Algorithms. Appl. Sci. 2023, 13, 10000. [Google Scholar] [CrossRef]
- Grande-Barreto, J.; Polanco-Castro, E.; Peregrina-Barreto, H.; Rosas-Mialma, E.; Puig-Mar, C. Generation of Synthetic Images of Trabecular Bone Based on Micro-CT Scans. Information 2023, 14, 375. [Google Scholar] [CrossRef]
- Hartley, Z.K.; French, A.P. Domain adaptation of synthetic images for wheat head detection. Plants 2021, 10, 2633. [Google Scholar] [CrossRef] [PubMed]
- Atapour-Abarghouei, A.; Breckon, T.P. Real-time monocular depth estimation using synthetic data with domain adaptation via image style transfer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2800–2810. [Google Scholar]
- Chatterjee, S.; Hazra, D.; Byun, Y.C.; Kim, Y.W. Enhancement of Image Classification Using Transfer Learning and GAN-Based Synthetic Data Augmentation. Mathematics 2022, 10, 1541. [Google Scholar] [CrossRef]
- Nie, X.; Xu, Z.; Zhang, W.; Dong, X.; Liu, N.; Chen, Y. Foggy lane dataset synthesized from monocular images for lane detection algorithms. Sensors 2022, 22, 5210. [Google Scholar] [CrossRef]
- Ivanovs, M.; Ozols, K.; Dobrajs, A.; Kadikis, R. Improving semantic segmentation of urban scenes for self-driving cars with synthetic images. Sensors 2022, 22, 2252. [Google Scholar] [CrossRef]
- Dunphy, K.; Fekri, M.N.; Grolinger, K.; Sadhu, A. Data augmentation for deep-learning-based multiclass structural damage detection using limited information. Sensors 2022, 22, 6193. [Google Scholar] [CrossRef]
- Zhou, J.; Li, G.; Wang, R.; Chen, R.; Luo, S. A Novel Contrastive Self-Supervised Learning Framework for Solving Data Imbalance in Solder Joint Defect Detection. Entropy 2023, 25, 268. [Google Scholar] [CrossRef]
- Wu, X.; Liang, L.; Shi, Y.; Fomel, S. FaultSeg3D: Using synthetic data sets to train an end-to-end convolutional neural network for 3D seismic fault segmentation. Geophysics 2019, 84, IM35–IM45. [Google Scholar] [CrossRef]
- Rosado, P.; Fernández, R.; Reverter, F. GANs and artificial facial expressions in synthetic portraits. Big Data Cogn. Comput. 2021, 5, 63. [Google Scholar] [CrossRef]
- Aljohani, A.; Alharbe, N. Generating synthetic images for healthcare with novel deep pix2pix gan. Electronics 2022, 11, 3470. [Google Scholar] [CrossRef]
- Pan, Z.; Yu, W.; Yi, X.; Khan, A.; Yuan, F.; Zheng, Y. Recent progress on generative adversarial networks (GANs): A survey. IEEE Access 2019, 7, 36322–36333. [Google Scholar] [CrossRef]
- Pavez, V.; Hermosilla, G.; Silva, M.; Farias, G. Advanced Deep Learning Techniques for High-Quality Synthetic Thermal Image Generation. Mathematics 2023, 11, 4446. [Google Scholar] [CrossRef]
- Wu, W.; Zhao, Y.; Chen, H.; Gu, Y.; Zhao, R.; He, Y.; Zhou, H.; Shou, M.Z.; Shen, C. Datasetdm: Synthesizing data with perception annotations using diffusion models. arXiv 2023, arXiv:2308.06160. [Google Scholar]
- Katakis, S.; Barotsis, N.; Kakotaritis, A.; Tsiganos, P.; Economou, G.; Panagiotopoulos, E.; Panayiotakis, G. Generation of Musculoskeletal Ultrasound Images with Diffusion Models. BioMedInformatics 2023, 3, 405–421. [Google Scholar] [CrossRef]
- Elbattah, M.; Loughnane, C.; Guérin, J.L.; Carette, R.; Cilia, F.; Dequen, G. Variational autoencoder for image-based augmentation of eye-tracking data. J. Imaging 2021, 7, 83. [Google Scholar] [CrossRef]
- Han, Y.J.; Yu, H.J. Fabric defect detection system using stacked convolutional denoising auto-encoders trained with synthetic defect data. Appl. Sci. 2020, 10, 2511. [Google Scholar] [CrossRef]
- Kim, K.; Myung, H. Autoencoder-combined generative adversarial networks for synthetic image data generation and detection of jellyfish swarm. IEEE Access 2018, 6, 54207–54214. [Google Scholar] [CrossRef]
- Matuzevičius, D.; Serackis, A. Three-Dimensional Human Head Reconstruction Using Smartphone-Based Close-Range Video Photogrammetry. Appl. Sci. 2021, 12, 229. [Google Scholar] [CrossRef]
- Zou, Z.; Shi, T.; Li, W.; Zhang, Z.; Shi, Z. Do game data generalize well for remote sensing image segmentation? Remote Sens. 2020, 12, 275. [Google Scholar] [CrossRef]
- Matuzevičius, D.; Serackis, A.; Navakauskas, D. Mathematical models of oversaturated protein spots. Elektron. Elektrotechnika 2007, 73, 63–68. [Google Scholar]
- Berenguel-Baeta, B.; Bermudez-Cameo, J.; Guerrero, J.J. Omniscv: An omnidirectional synthetic image generator for computer vision. Sensors 2020, 20, 2066. [Google Scholar] [CrossRef] [PubMed]
- Li, X.; Wang, K.; Gu, X.; Deng, F.; Wang, F.Y. ParallelEye Pipeline: An Effective Method to Synthesize Images for Improving the Visual Intelligence of Intelligent Vehicles. IEEE Trans. Syst. Man Cybern. Syst. 2023, 53, 5545–5556. [Google Scholar] [CrossRef]
- Zdziebko, P.; Holak, K. Synthetic image generation using the finite element method and blender graphics program for modeling of vision-based measurement systems. Sensors 2021, 21, 6046. [Google Scholar] [CrossRef] [PubMed]
- Seong, H.; Baik, S.; Lee, Y.; Lee, S.; Kim, E. Content swapping: A new image synthesis for construction sign detection in autonomous vehicles. Sensors 2022, 22, 3494. [Google Scholar] [CrossRef] [PubMed]
- Lin, T.; Maire, M.; Belongie, S.J.; Bourdev, L.D.; Girshick, R.B.; Hays, J.; Perona, P.; Ramanan, D.; Doll’a r, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. arXiv 2014, arXiv:1405.0312. [Google Scholar]
- Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
- Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
- Zoph, B.; Vasudevan, V.; Shlens, J.; Le, Q.V. Learning transferable architectures for scalable image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8697–8710. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning. PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
- Van Rossum, G.; Drake, F.L. Python 3 Reference Manual; CreateSpace: Scotts Valley, CA, USA, 2009. [Google Scholar]
- Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. 2015. Available online: tensorflow.org (accessed on 21 August 2023).
- Buslaev, A.; Iglovikov, V.I.; Khvedchenya, E.; Parinov, A.; Druzhinin, M.; Kalinin, A.A. Albumentations: Fast and Flexible Image Augmentations. Information 2020, 11, 125. [Google Scholar] [CrossRef]
- Bradski, G. The OpenCV Library. Dr. Dobb’S J. Softw. Tools 2000, 120, 122–125. [Google Scholar]
- Bento, N.; Rebelo, J.; Barandas, M.; Carreiro, A.V.; Campagner, A.; Cabitza, F.; Gamboa, H. Comparing handcrafted features and deep neural representations for domain generalization in human activity recognition. Sensors 2022, 22, 7324. [Google Scholar] [CrossRef] [PubMed]
- El-Gayar, M.; Soliman, H.; Meky, N. A comparative study of image low level feature extraction algorithms. Egypt. Inform. J. 2013, 14, 175–181. [Google Scholar] [CrossRef]
- Alshazly, H.; Linse, C.; Barth, E.; Martinetz, T. Handcrafted versus CNN features for ear recognition. Symmetry 2019, 11, 1493. [Google Scholar] [CrossRef]
- Tsalera, E.; Papadakis, A.; Samarakou, M.; Voyiatzis, I. Feature extraction with handcrafted methods and convolutional neural networks for facial emotion recognition. Appl. Sci. 2022, 12, 8455. [Google Scholar] [CrossRef]
- Hamdi, M.; Senan, E.M.; Jadhav, M.E.; Olayah, F.; Awaji, B.; Alalayah, K.M. Hybrid Models Based on Fusion Features of a CNN and Handcrafted Features for Accurate Histopathological Image Analysis for Diagnosing Malignant Lymphomas. Diagnostics 2023, 13, 2258. [Google Scholar] [CrossRef]
Backbone | Parameters 1 | Depth 2 | Size (MB) 3 | Time (ms) CPU 4 | Time (ms) GPU 4 |
---|---|---|---|---|---|
MobileNetV2 (0.5) 5 | 706,224 | 52 | 3.06 | 233.6 | 15.5 |
MobileNetV2 (0.75) 5 | 1,382,064 | 52 | 5.64 | 363.2 | 19.2 |
MobileNet (0.5) 5 | 829,536 | 27 | 3.36 | 178.1 | 12.1 |
NASNetMobile | 4,269,716 | 196 | 17.9 | 585.4 | 44.1 |
ResNet50 | 23,587,712 | 53 | 90.3 | 1171 | 44.3 |
EfficientNetB0 | 4,049,571 | 81 | 15.9 | 729.6 | 32.5 |
Dataset | No. of Samples | Resolution(s) |
---|---|---|
Synthetic-train | 5000 | |
Real-train | 1901 | |
Real-test | 810 |
Synthetic-Train | Real-Train | Mixed (All) | ||||
---|---|---|---|---|---|---|
Experimental Scenarios | Mean | 95% CI | Mean | 95% CI | Mean | 95% CI |
MobileNetV2 (0.5), TL, E100 | 0.77 | 0.76–0.79 | 0.66 | 0.62–0.70 | 0.82 | 0.81–0.84 |
MobileNetV2 (0.75), TL, E100 | 0.77 | 0.76–0.78 | 0.74 | 0.70–0.78 | 0.84 | 0.83–0.86 |
MobileNet, TL, E100 | 0.83 | 0.81–0.84 | 0.68 | 0.67–0.70 | 0.86 | 0.84–0.87 |
NASNetMobile, TL, E100 | 0.82 | 0.80–0.83 | 0.67 | 0.63–0.72 | 0.78 | 0.76–0.80 |
ResNet50, TL, E100 | 0.81 | 0.80–0.82 | 0.61 | 0.57–0.65 | 0.84 | 0.81–0.86 |
EfficientNetB0, TL, E100 | 0.83 | 0.81–0.85 | 0.73 | 0.71–0.75 | 0.85 | 0.84–0.87 |
MobileNetV2 (0.5), TL, E200 | 0.84 | 0.83–0.84 | 0.72 | 0.68–0.75 | 0.86 | 0.85–0.88 |
MobileNetV2 (0.5), FS, E200 | 0.82 | 0.81–0.83 | 0.64 | 0.62–0.66 | 0.80 | 0.79–0.82 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Matuzevičius, D. Rulers2023: An Annotated Dataset of Synthetic and Real Images for Ruler Detection Using Deep Learning. Electronics 2023, 12, 4924. https://doi.org/10.3390/electronics12244924
Matuzevičius D. Rulers2023: An Annotated Dataset of Synthetic and Real Images for Ruler Detection Using Deep Learning. Electronics. 2023; 12(24):4924. https://doi.org/10.3390/electronics12244924
Chicago/Turabian StyleMatuzevičius, Dalius. 2023. "Rulers2023: An Annotated Dataset of Synthetic and Real Images for Ruler Detection Using Deep Learning" Electronics 12, no. 24: 4924. https://doi.org/10.3390/electronics12244924
APA StyleMatuzevičius, D. (2023). Rulers2023: An Annotated Dataset of Synthetic and Real Images for Ruler Detection Using Deep Learning. Electronics, 12(24), 4924. https://doi.org/10.3390/electronics12244924