Next Article in Journal
A Predictive Model for Benchmarking the Performance of Algorithms for Fake and Counterfeit News Classification in Global Networks
Previous Article in Journal
Towards the Instrumentation of Facemasks Used as Personal Protective Equipment for Unobtrusive Breathing Monitoring of Workers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Computer-Simulated Virtual Image Datasets to Train Machine Learning Models for Non-Invasive Fish Detection in Recirculating Aquaculture

The Conservation Fund Freshwater Institute, Shepherdstown, WV 25443, USA
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(17), 5816; https://doi.org/10.3390/s24175816 (registering DOI)
Submission received: 29 July 2024 / Revised: 28 August 2024 / Accepted: 2 September 2024 / Published: 7 September 2024
(This article belongs to the Section Smart Agriculture)

Abstract

:
Artificial Intelligence (AI) and Machine Learning (ML) can assist producers to better manage recirculating aquaculture systems (RASs). ML is a data-intensive process, and model performance primarily depends on the quality of training data. Relatively higher fish density and water turbidity in intensive RAS culture produce major challenges in acquiring high-quality underwater image data. Additionally, the manual image annotation involved in model training can be subjective, time-consuming, and labor-intensive. Therefore, the presented study aimed to simulate fish schooling behavior for RAS conditions and investigate the feasibility of using computer-simulated virtual images to train a robust fish detection model. Additionally, to expedite the model training and automate the virtual image annotation, a process flow was developed. The ‘virtual model’ performances were compared with models trained on real-world images and combinations of real and virtual images. The results of the study indicate that the virtual model trained solely with computer-simulated images could not perform satisfactorily (mAP = 62.8%, F1 score = 0.61) to detect fish in a real RAS environment; however, replacing a small number of the virtual images with real images in the training dataset significantly improved the model’s performance. The M6 mixed model trained with 630 virtual and 70 real images (virtual-to-real image ratio: 90:10) achieved mAP and F1 scores of 91.8% and 0.87, respectively. Furthermore, the training time cost for the M6 model was seven times shorter than that for the ‘real model’. Overall, the virtual simulation approach exhibited great promise in rapidly training a reliable fish detection model for RAS operations.

1. Introduction

Land-based intensive fish farming methods such as recirculating aquaculture systems (RASs) are emerging as a sustainable approach to produce premium quality fish near the consumers [1,2]. Appropriate RAS management strategies are critical to maintain an optimal rearing environment for cultured fish [3]. Currently, precision technologies such as sensor networks, computer vision, Artificial Intelligence (AI), and Internet of Things (IoT) are being adopted in the aquaculture industries to improve yield, profitability, and production sustainability [4,5,6]. The harsh operating conditions, high maintenance cost, and limited power and data connectivity, however, present major challenges in large-scale commercial adaptation and affect the economic viability of incorporating advanced technology into coastal sea cage farming [7]. Such technologies can be relatively easily adopted in RASs, as these systems enable farmers to rear fish in a controlled environment. Recent studies indicate several applications of Machine Learning (ML)-assisted image classification and object detection for non-invasive biomass estimation [8], fish health and welfare monitoring [9,10], disease prediction [11], feed optimization [12], behavior monitoring [13], water quality optimization [14], and mortality monitoring [15] in RASs.
Despite numerous research efforts toward the development of AI- and ML-aided solutions for efficient RAS management, the scalability of this approach is still questionable [16]. ML techniques are highly data-intensive, and the precision and accuracy of the models primarily depend on the quality of the input data [17,18]. A widely adopted technique used to increase the data size and incorporate variability in the training dataset is augmentation [19]. A series of transformations (e.g., rotation, flip, brightness, saturation, etc.) are applied to the images, and these images are then added to the original training dataset to enhance model performance; however, these techniques can only be effective when sufficient input data are available to perform augmentations. This may exclude scenarios where image acquisition is difficult, costly, or hazardous but are relevant to computer vision problems [20]. Alternatively, computer simulation techniques can be adapted to mimic real-world conditions and generate virtual videos and/or images for model training. The feasibility of using virtual images for ML model training has been explored over the past decade, with an emphasis on semantic segmentation and object detection [21]. Studies report the use of synthetic training data for traffic sign recognition [22,23], pedestrian detection [17], and vehicle detection [24]. Similar approaches have also been explored in agricultural and environmental applications such as crop yield estimation [25], habitat monitoring [20,26], semantic part localization [27], and underwater inspection [28], and they have produced promising results.
Acquiring underwater images to train ML models presents a unique challenge, as the image quality is adversely affected by high turbidity, high population density, moving targets (i.e., fish), and their proximity to the camera [29,30]. Recent research has explored techniques such as Principal Component Fusion of Foreground and Background, and Weighted Wavelet Visual Perception Fusion to enhance underwater image quality [31,32]. However, accurately restoring the true colors of underwater scenes remains challenging, particularly in instances of significant color degradation caused by water absorption and scattering. Furthermore, manual image annotation prior to model training is a time-consuming and labor-intensive process and may introduce subjectivity and inaccuracy into the model [33]. Additionally, the data acquisition time frame highly relies upon the occurrence of the conditions for which the model is being trained. For instance, if an ML model is being trained to detect infrequently occurring high-mortality events, creating conditions like this for model development is both challenging and undesirable [15].
Using a virtually generated dataset simulated for a RAS environment may present a solution to these problems. This approach can assist in generating various scenarios (e.g., high mortality, disease conditions, etc.) for RASs and theoretically provide unlimited data for model training. Imaging conditions (e.g., camera location and orientation, fish density, water quality, etc.) can also be adjusted with a click of the mouse. Additionally, the annotation process can be automated for rapid and robust model development. Under this hypothesis, the presented study was conducted to explore the feasibility of computer simulation to generate virtual images and utilize these images for training in-tank fish detection models. The specific objectives of the study were as follows:
  • To virtually simulate a RAS environment and optimize the fish schooling pattern to attain high-quality virtual image data suitable for training a robust in-tank fish detection model.
  • To analyze the performance of a virtual image-trained fish detection model and compare its performance with a model trained with real-world data.

2. Materials and Methods

2.1. Virtual Simulation

2.1.1. Fish Schooling

RAS-appropriate fish schooling was simulated in Blender (Blender Foundation, Amsterdam, The Netherlands), an open-source simulation and rendering software, using a Boids particle system. A Boids particle system can be used to simulate flocks, herds, swarms, and schools of various kinds of animals. The particles within the Boids system can be programmed to follow basic rules and behavior to emulate certain movement patterns [34].
The first step in the simulation process was to generate a virtual RAS tank. A hollow cylinder was created to mimic a fish tank so that fish schooling behavior and light conditions could be simulated within the tank. A 3D rainbow trout (Oncorhynchus mykiss) model (hereafter termed as ‘fish’, format: FBX; PoserWorld, Vanishing Point Inc., Orlando, FL, USA) with 5291 polygons was imported into the Blender environment. A unique fish was associated with each Boids particle to simulate the swimming pattern of fish in the virtual RAS tank (Figure 1). Collision avoidance was enabled so that collisions among the fish and between fish and tank walls/floor were avoided. Additionally, an invisible plane was used to cover the tank top to contain the fish within the virtual RAS tank. Two vortex forces were applied from the top and bottom of the tank, and their vortex intensities and orientations were adjusted to optimize fish movements and swimming patterns within the virtual RAS environment. The clockwise top vortex and counter-clockwise bottom vortex forces ensured the uniform vertical and horizontal distribution and circular motion of fish with reference to the tank axis. To solve the problem of fish swimming too close to the camera, a virtual ramp (represented by a cuboid in Figure 1) was introduced upstream to the camera near the tank wall. This barrier helped to divert the fish away from the camera, as illustrated in Figure 1.

2.1.2. Underwater Environment

The simulation aimed to mimic the swimming patterns of fish in RAS environment, realistically incorporating the effects of water quality and lighting on underwater visibility. Area light was adopted for the simulation to emulate the uniform lighting often seen in indoor RAS facilities. The Shading feature in Blender was applied to the virtual tank to simulate the turbid water quality seen in real-world RAS conditions. The Shader uses the material property and distance between the virtual camera and objects to evaluate the scattering or absorption of the light. As an object moves away from the camera, higher absorption and scattering of light create a darkening and blurring effect. The software allows users to adjust the distance associated with light scattering and absorption. This method was used to create the perception of turbidity (i.e., low or high) in the water to reflect real-world visibility conditions in the virtual RAS tank. The images generated for low-turbidity conditions had a higher number of fish in the foreground and well-defined object features compared to the images generated for high-turbidity conditions (Figure 2).

2.2. Validation Data Acquisition

The best-performing image dataset from our recent study [30] was utilized to analyze the performance of the virtual image-trained fish detection models in real-world scenarios. The images were collected in a 150 m3 fully recirculating, semi-commercial-scale growout tank located at The Conservation Fund’s Freshwater Institute (Shepherdstown, WV, USA). The tank was stocked with rainbow trout with a tank density at the time of data collection of 40 kg m−3. The image data were captured using a Raspberry Pi sensor (RPi, model: M23272M14, Arducam, Nanjing, China; focal Length: 2.72 mm, Aperture: F2.5, horizontal field of view (HFOV): 140°, and captured resolution: 1920 × 1080 px) in supplemental light conditions. The details of the data acquisition protocol are reported in [30].

2.3. Automated Image Annotation

The virtual simulation model rendered layered images in OpenEXR format, which allowed us to store an arbitrary number of attributes in an image file. OpenEXR images can contain a combination of image channels, for instance, red (R), green (G), blue (B), luminance, chroma, depth, surface normal directions, and/or motion vectors. For this study, RGB, depth, and semantic index map layers were included in the output image (Figure 3). While RGB features assisted in visualizing the object of interest (i.e., fish) in the image, the semantic index map provided information on whether an individual pixel belonged to the object of interest (i.e., fish) or background (i.e., tank or water). If an object was identified as fish, the number of pixels for an individual fish was used to determine whether to include or exclude a fish for annotation. The selected fish was further checked for its distance from the camera, and pixel depth information retrieved from the depth map was used to exclude the farther-located blurry fish. Finally, based on the semantic information of the fish, the center of the volume, depth, and box coordinates of each fish was determined, and annotations were created. The time cost required to annotate the virtual training image was logged and compared with the manual annotation of real images (Figure 4).
The training dataset containing the image and corresponding annotation metadata was stored in Common Object in Context (COCO) format [35]. COCO formats the data by information, licenses, categories, images, and annotations in a JavaScript Object Notation (JSON) file. This format enables the storage of various annotation types, such as bounding boxes, segmentation masks, key points, and natural language descriptions. The COCO dataset was then imported into Roboflow (Roboflow, Inc., Des Moines, Iowa, USA) to train the fish detection model.

2.4. Model Training

A single-stage object detection model, YOLOv8 [36], was adopted for in-tank fish detection. Three different types of training datasets were used to train various fish detection models. The first model (hereafter termed the ‘virtual model’) was trained with simulation-generated virtual training images only, whereas the second model used real-world underwater imagery data acquired under RAS conditions to train the fish detection model (hereafter termed the ‘real model’), as described in Section 2.2. The third type of model (hereafter termed the ‘mixed model’) was trained with datasets containing both virtual and real images in varying proportions. Eight mixed models (M1, M2, M3, M4, M5, M6, M7, and M8) were trained with datasets containing both virtual and real images in proportions of 99:1, 98:2, 96:4, 94:6, 92:8, 90:10, 75:25, and 50:50, respectively (Table 1). Only real images were used for validation and test datasets so that the feasibility of using virtual image datasets to train fish detection models for real-world scenarios could be evaluated. Model training and model validation were performed using two computer vision tools (Roboflow, Inc., Des Moines, Iowa, USA; Ultralytics, Los Angeles, CA, USA). A proportion of 70:20:10 was maintained for the training, validation, and test dataset, and each model was trained up to 100 epochs.

2.5. Data Analysis

The model performances were analyzed in terms of mean average precision (mAP) at an intersection over union (IoU) threshold of 0.5 (i.e., mAP0.5) and F1 score as described by Flach and Kull (2015) [37]. Mean average precision (mAP0.5) was the mean of the precision under different recall values at an IoU of 0.5, whereas the F1 score was the harmonic mean of precision and recall. The real, mixed (M1–M8), and virtual models were first trained with different dataset sizes (100–1000 images at a step of 100 images), and the effect of data size on model performance was analyzed. Additionally, the preprocessed training images were augmented (2X) with brightness [±25%], exposure [±25%], saturation [±25%], and blur [±5%], and the performances of the augmented models (i.e., With Aug) were compared with that of the non-augmented model (i.e., No Aug). Finally, the performances and total training time costs of the optimized real, mixed, and virtual models were compared. The total time cost was evaluated as the sum of the annotation time (i.e., time required for image annotation) and model training time (i.e., time required to train a model).

3. Results and Discussion

3.1. Model Optimization

3.1.1. Epoch and Data Size

While training the real and mixed models (M1–M8) for fish detection, a logistic increase in the mAP scores was observed up to 30 epochs. Further training the models up to 100 epochs slightly improved the mAP and reduced the losses; however, beyond 100 epochs, no improvement in model performance was observed (Figure 5a). Unlike real and mixed models, the virtual model training stopped after 60 epochs due to model overfitting and resulted in an increase in the loss values and degraded mAP scores for higher epochs. Similar mAP trends were observed while analyzing the performance of virtual, real, and mixed models trained with different data sizes. Increasing the data size beyond 1000 images failed to improve the model performance in terms of mAP. Additionally, training the model beyond the optimal data size and epochs may lead to model overfitting and result in higher computing and time costs [38]. Therefore, for further analysis, all models were trained with 700 images (70%), validated on 200 images (20%), and tested on 100 images (10%) up to 100 epochs.

3.1.2. Fish Detection Model

While analyzing the impact of data augmentation on the performance of real and mixed models, we found that the augmentation apparently does not considerably affect the model performance. The model trained with an augmented dataset showed better performance during the initial stage of training; however, the effect of augmentation diminishes for a higher number of epochs (Figure 6a,b). Similar results were reported in our prior study conducted to study the effect of image data quality on fish detection model performance [30]. The difference in mAP score between models trained with augmented and non-augmented datasets was less than 1%. Unlike real and mixed models, data augmentation assisted in improving the performance of the virtual model (Figure 6c). The maximum mAP and F1 scores of the virtual model trained with a non-augmented dataset were 51.4% and 0.53, respectively. The augmented training dataset enhanced model performance and achieved mAP and F1 scores of 62.8%, and 0.61, respectively. Overall, data augmentation either improved or had no impact on model performance; however, augmentation did not negatively affect model performance. Therefore, to maintain consistency in the analysis, all datasets were augmented prior to the training, and pertinent results are presented in the following section.

3.2. Model Performance

The real model trained with underwater images acquired by the Rpi sensor performed satisfactorily in terms of in-tank fish detection. The maximum mAP value for the optimized real model was above 95% (Figure 7a). The real model also achieved high precision and recall, with an F1 score of 0.91 (Figure 7b). When deploying the model on an independent image dataset captured in a real RAS environment, the model accurately detected the whole fish. Additionally, the model also successfully detected partial fish with more than 50% area of their body visible in the frame (Figure 8b).
Unlike the real model, the virtual model trained solely with computer-simulated virtual images did not perform satisfactorily in real-world underwater environments. The maximum mAP and F1 scores for this model were 62.8%, and 0.61, respectively. The virtual model accurately detected the whole fish; however, it failed to detect the partial fish in the frame (Figure 8c). This resulted in an underprediction of the fish population by the virtual model.
While the model trained with virtual images alone could not perform satisfactorily, replacing small numbers of virtual images with real images significantly improved the model performance. Out of the 700 images used to train the virtual model, replacing 1% of the virtual images with real images resulted in a substantial improvement in the performance of the M1 mixed model. Replacement of seven virtual images with real images resulted in improvements in the mAP score, increasing from 62.8% to 79.3%, and F1 score, which increased from 0.61 to 0.75. The performance of the mixed model further improved as the proportion of real images in the training dataset was increased. The M4 mixed model trained with 658 virtual and 42 real images (virtual-to-real image ratio: 94:6) surpassed an mAP value of 90%. Replacing 10% of the virtual images with real images helped the M6 model to achieve a satisfactory mAP of 91.8% and an F1 score of 0.87. Additionally, the M6 model performed on par with the real model in terms of fish detection in real-world conditions and precisely detected the whole and partial fish in the frame (Figure 8d). Therefore, the M6 mixed model was adopted for further comparative analysis. A study conducted by Jelic et al. (2022) [39] to develop advanced driver-assistance system algorithms for autonomous vehicles reported similar findings and concluded that synthetic data can contribute to better detector performance until a certain ratio of real-world and synthetic data is reached.

3.3. Model Comparison

Comparing performance, the real model performed best among the tested models (Figure 7c). Since the real model was trained with images captured in the actual RAS environment and had sufficient data from which to learn, the model performed well while deployed in similar conditions. Unlike the real model, the virtual model was trained solely with virtual images, and although these images were simulated to mimic fish schooling, it was difficult to match the exact body texture, orientation, fish-to-fish color, and shape variation. These differences may have contributed to the inferior performance of the virtual model. Despite the poor performance of the virtual model, this model was able to detect the whole fish in the frame. Therefore, for an application where partial fish detection is not important and whole fish alone is of interest (e.g., fish biomass estimation), virtual models could be utilized. Despite a 4% difference in the mAP score and a difference of 0.04 in the F1 score between the real and M6 mixed model (Figure 7c,d), the latter performed satisfactorily in terms of partial and whole fish detection, as discussed in Section 3.2. A representative training dataset attained by supplementing real images in the training set of mixed models assisted in improving model robustness.

3.4. Time Cost Analysis

This analysis indicated that the time cost required to train the virtual model was considerably shorter than that for the real and mixed models (Table 2). The time required to annotate 700 virtual images was around 216 times shorter than that for the manual annotation of the same number of real images. The automated annotation of the virtual images aided in achieving a substantial reduction in the annotation time. Additionally, the virtual model’s training was terminated at lower epochs, as described in Section 3.1.1. Therefore, the training time for the virtual model was also reduced by half as compared to the real model. Overall, the total time cost required to train the virtual model was around 42 times shorter than that for the real model. While the virtual model’s training was much faster and more labor-efficient than that of the real model, the model performance of the former was not satisfactory (Figure 7c,d). Further investigation indicated that the annotation time for the M6 mixed model was around 10 times shorter than that for the real model. Since 90% of the training images (i.e., 630 virtual images) were annotated automatically, the annotation time for the M6 model reduced substantially. Moreover, the training time for M6 was also marginally shorter than that for the real model. The total time cost required to train the mixed model was seven times shorter than that for the real model, with the two models showing nearly similar model performance, as described in Section 3.2.
Notably, the data acquisition time for real images and the time spent in generating the simulated datasets were not included in the time cost analysis. It was assumed that data acquisition/simulation is a one-time process and that it will be reused for future projects. Additionally, virtual models can be easily replicated for various species cultured in aquaculture tank systems. Overall, the models trained with virtual images substantially reduced the training time, which means they can assist in the rapid deployment of fish detection models for aquaculture applications.

4. Conclusions

This study was conducted to assess the feasibility of using computer-simulation-generated virtual images to train a fish detection model for a real RAS environment. The following are the major conclusions derived from the presented investigation:
  • The virtual model trained solely with simulated images did not perform satisfactorily in partial fish detection; however, replacing small numbers of virtual images from the training dataset with real images significantly improved model performance. The M6 mixed model trained with 630 virtual and 70 real images achieved a satisfactory mAP of 91.8% and an F1 score of 0.87, and it precisely detected whole and partial fish in an actual RAS environment.
  • The automated annotation considerably reduced the annotation time for virtual images. This resulted in a seven-fold reduction in total training time cost for the M6 mixed model. Overall, virtual simulation can assist in developing a rapid and robust fish detection model for aquaculture applications.
While this study focused specifically on underwater fish detection, a similar approach can be adapted to train Machine Learning models with the help of simulated data for various purposes, such as detecting disease conditions (e.g., fin erosion, musculoskeletal deformities, cataracts, etc.), tracking fish maturity, monitoring fish schooling behavior, and other welfare indicators. Moreover, the present study aimed to train a fish detection model for a real RAS environment; however, a similar approach can be also adopted for other aquaculture production systems. Recent advancements in Large Language Model (LLM)-assisted image generation tools have provided the capability to efficiently generate extensive image datasets for Machine Learning model training without sophisticated simulation software. Therefore, future research should explore the potential of AI-generated datasets to train ML models to address challenges in the aquaculture industry.

Author Contributions

Conceptualization, R.R.; methodology, S.R.S., R.R., K.S. and S.T.; software, S.R.S.; formal analysis, S.R.S. and R.R.; investigation, R.R.; resources, C.G.; writing—original draft preparation, S.R.S. and R.R.; writing—review and editing, K.S., C.G. and S.T.; supervision, R.R. and C.G. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the USDA Agricultural Research Service (agreement number 59-8082-0-001). The use of trade names is solely to provide accurate information and does not imply endorsement by the USDA. The Conservation Fund and USDA are equal opportunity providers and employers.

Institutional Review Board Statement

All experimental protocols and methods were approved by the Freshwater Institute’s Institutional Animal Care and Use Committee.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data will be made available on request.

Acknowledgments

The authors want to thank Brian Vinci for his assistance in idea generation and Shanen Cogan and JC Stanley for providing their operational and maintenance expertise.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Timmons, M.B.; Vinci, B.J. Recirculating Aquaculture Systems, 5th ed.; Ithaca Publishing Company: New York, NY, USA, 2022. [Google Scholar]
  2. Ahmed, N.; Turchini, G.M. Recirculating aquaculture systems (RAS): Environmental solution and climate change adaptation. J. Clean. Prod. 2021, 297, 126604. [Google Scholar] [CrossRef]
  3. Badiola, M.; Mediola, D.; Bostock, J. Recirculating Aquaculture Systems (RAS) analysis: Main issues on management and future challenges. Aquacult. Eng. 2012, 51, 26–35. [Google Scholar] [CrossRef]
  4. Fore, M.; Frank, K.; Norton, T.; Svendsen, E.; Alfredsen, J.A.; Dempster, T.; Eguiraun, H.; Watson, W.; Stahl, A.; Sunde, L.M.; et al. Precision fish farming: A new framework to improve production in aquaculture. Biosyst. Eng. 2018, 173, 176–193. [Google Scholar] [CrossRef]
  5. O’Donncha, F.; Grant, J. Precision Aquaculture. IEEE Internet Things Mag. 2019, 2, 26–30. [Google Scholar] [CrossRef]
  6. Antonucci, F.; Costa, C. Precision aquaculture: A short review on engineering innovations. Aquac. Int. 2020, 28, 41–57. [Google Scholar] [CrossRef]
  7. Ubina, N.A.; Cheng, S.C. A review of unmanned system technologies with its application to aquaculture farm monitoring and management. Drones 2022, 6, 12. [Google Scholar] [CrossRef]
  8. Zhao, S.; Zhang, S.; Liu, J.; Wang, H.; Zhu, J.; Li, D.; Zhao, R. Application of machine learning in intelligent fish aquaculture: A review. Aquaculture 2021, 540, 736724. [Google Scholar] [CrossRef]
  9. Barreto, M.O.; Rey Planellas, S.; Yang, Y.; Phillips, C.; Descovich, K. Emerging indicators of fish welfare in aquaculture. Rev. Aquacult. 2022, 14, 343–361. [Google Scholar] [CrossRef]
  10. Li, D.; Wang, G.; Du, L.; Zheng, Y.; Wang, Z. Recent advances in intelligent recognition methods for fish stress behavior. Aquacult. Eng. 2022, 96, 102222. [Google Scholar] [CrossRef]
  11. Yilmaz, M.; Çakir, M.; Oral, O.; Oral, M.A.; Arslan, T. Using machine learning technique for disease outbreak prediction in rainbow trout (Oncorhynchus mykiss) farms. Aquac. Res. 2022, 53, 6721–6732. [Google Scholar] [CrossRef]
  12. Li, D.; Wang, Z.; Wu, S.; Miao, Z.; Du, L.; Duan, Y. Automatic recognition methods of fish feeding behavior in aquaculture: A review. Aquaculture 2020, 528, 735508. [Google Scholar] [CrossRef]
  13. Sun, M.; Yang, X.; Xie, Y. Deep learning in aquaculture: A review. J. Comput. 2020, 31, 294–319. [Google Scholar] [CrossRef]
  14. Li, T.; Lu, J.; Wu, J.; Zhang, Z.; Chen, L. Predicting Aquaculture Water Quality Using Machine Learning Approaches. Water 2022, 14, 2836. [Google Scholar] [CrossRef]
  15. Ranjan, R.; Sharrer, K.; Tsukuda, S.; Good, C. MortCam: An Artificial Intelligence-aided fish mortality detection and alert system for recirculating aquaculture. Aquacult. Eng. 2023, 102, 102341. [Google Scholar] [CrossRef]
  16. Johnson-Roberson, M.; Barto, C.; Mehta, R.; Sridhar, S.N.; Rosaen, K.; Vasudevan, R. Driving in the matrix: Can virtual worlds replace human-generated annotations for real world tasks? arXiv 2016. [Google Scholar] [CrossRef]
  17. Marin, J.; Vázquez, D.; Gerónimo, D.; López, A.M. Learning appearance in virtual scenarios for pedestrian detection. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 137–144. [Google Scholar] [CrossRef]
  18. Gudivada, V.; Apon, A.; Ding, J. Data quality considerations for big data and machine learning: Going beyond data cleaning and transformations. Int. J. Softw. Innov. 2017, 10, 1–20. [Google Scholar]
  19. Maharana, K.; Mondal, S.; Nemade, B. A review: Data pre-processing and data augmentation techniques. Glob. Transit Proc. 2022, 3, 91–99. [Google Scholar] [CrossRef]
  20. Lin, X.; Jha, N.; Joshi, M.; Karapetvan, N.; Aloimonos, Y.; Yu, M. OysterSim: Underwater Simulation for Enhancing Oyster Reef Monitoring. In Proceedings of the OCEANS, Hampton Roads, VA, USA, 17–20 October 2022; pp. 1–6. [Google Scholar] [CrossRef]
  21. Tsirikoglou, A.; Eilertsen, G.; Unger, J. A survey of image synthesis methods for visual machine learning. Comput. Graph. Forum. 2020, 39, 426–451. [Google Scholar] [CrossRef]
  22. Shakhuro, V.; Faizov, B.; Konushin, A. Rare traffic sign recognition using synthetic training data. In Proceedings of the 3rd International Conference on Video and Image Processing, Shanghai China, 20–23 December 2019; pp. 23–26. [Google Scholar] [CrossRef]
  23. Dewi, C.; Chen, R.C.; Liu, Y.T.; Jiang, X.; Hartomo, K.D. Yolo V4 for advanced traffic sign recognition with synthetic training data generated by various GAN. IEEE Access 2021, 9, 97228–97242. [Google Scholar] [CrossRef]
  24. Tremblay, J.; Prakash, A.; Acuna, D.; Brophy, M.; Jampani, V.; Anil, C.; To, T.; Cameracci, E.; Boochoon, S.; Birchfield, S. Training deep networks with synthetic data: Bridging the reality gap by domain randomization. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–23 June 2018; pp. 969–977. [Google Scholar] [CrossRef]
  25. Rahnemoonfar, M.; Sheppard, C. Deep count: Fruit counting based on deep simulated learning. Sensors 2017, 17, 905. [Google Scholar] [CrossRef]
  26. Lin, X.; Sanket, N.J.; Karapetvan, N.; Aloimonos, Y. Oysternet: Enhanced oyster detection using simulation. In Proceedings of the IEEE International Conference on Robotics and Automation (ICRA), London, UK, 29 May–2 June 2023; pp. 5170–5176. [Google Scholar] [CrossRef]
  27. Barth, R.; IJsselmuiden, J.; Hemming, J.; Van Henten, E.J. Data synthesis methods for semantic segmentation in agriculture: A Capsicum annuum dataset. Comput. Electron. Agric. 2018, 144, 284–296. [Google Scholar] [CrossRef]
  28. O’Byrne, M.; Ghosh, B.; Schoefs, F.; Pakrashi, V. Applications of virtual data in subsea inspections. J. Mar. Sci. Eng. 2020, 8, 328. [Google Scholar] [CrossRef]
  29. Di Benedetto, M.; Meloni, E.; Amato, G.; Falchi, F.; Gennaro, C. Learning safety equipment detection using virtual worlds. In Proceedings of the International Conference on Content-Based Multimedia Indexing (CBMI), Dublin, Ireland, 4–6 September 2019; pp. 1–6. [Google Scholar] [CrossRef]
  30. Ranjan, R.; Sharrer, K.; Tsukuda, S.; Good, C. Effects of image data quality on a convolutional neural network trained in-tank fish detection model for recirculating aquaculture systems. Comput. Electron. Agric. 2023, 205, 107644. [Google Scholar] [CrossRef]
  31. Zhang, W.; Liu, Q.; Feng, Y.; Cai, L.; Zhuang, P. Underwater Image Enhancement via Principal Component Fusion of Foreground and Background. IEEE Trans. Circuits Syst. Video Technol. 2024. [Google Scholar] [CrossRef]
  32. Zhang, W.; Zhou, L.; Zhuang, P.; Li, G.; Pan, X.; Zhao, W.; Li, C. Underwater image enhancement via weighted wavelet visual perception fusion. IEEE Trans. Circuits Syst. Video Technol. 2014, 34, 2469–2483. [Google Scholar] [CrossRef]
  33. Rasmussen, C.B.; Kirk, K.; Moeslund, T.B. The challenge of data annotation in deep learning—A case study on whole plant corn silage. Sensors 2022, 22, 1596. [Google Scholar] [CrossRef]
  34. Reynolds, C.W. Flocks, herds and schools: A distributed behavioral model. In Proceedings of the 14th Annual Conference on Computer Graphics and Interactive Techniques, Anaheim, CA, USA, 27–31 July 1987; pp. 25–34. [Google Scholar] [CrossRef]
  35. Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. arXiv 2015. [Google Scholar]
  36. Jocher, G.; Chaurasia, A.; Qiu, J. YOLO by Ultralytics. 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 30 February 2023).
  37. Flach, P.; Kull, M. Precision-recall-gain curves: PR analysis done right. In Proceedings of the 28th International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 8–13 December 2014; Volume 1, pp. 838–846. [Google Scholar]
  38. Ying, X. An overview of overfitting and its solutions. J. Phys. Conf. Ser. 2019, 1168, 022022. [Google Scholar] [CrossRef]
  39. Jelic, B.; Grbic, R.; Vranjes, M.; Mijic, D. Can we replace real-world with synthetic data in deep learning-based ADAS algorithm development? IEEE Consum. Electron. Mag. 2021, 12, 32–38. [Google Scholar] [CrossRef]
Figure 1. The Boids particle system used to mimic the schooling behavior of fish in a RAS tank. A cuboid attached to the tank wall acted as a virtual barrier to divert the Boids particles (i.e., fish, in inset) away from the camera.
Figure 1. The Boids particle system used to mimic the schooling behavior of fish in a RAS tank. A cuboid attached to the tank wall acted as a virtual barrier to divert the Boids particles (i.e., fish, in inset) away from the camera.
Sensors 24 05816 g001
Figure 2. The training images generated for (a) low-turbidity and (b) high-turbidity conditions. The low-turbidity images had well-defined object features, whereas in turbid conditions, blurry object features can be observed.
Figure 2. The training images generated for (a) low-turbidity and (b) high-turbidity conditions. The low-turbidity images had well-defined object features, whereas in turbid conditions, blurry object features can be observed.
Sensors 24 05816 g002
Figure 3. Algorithm process flow developed for automated annotation of simulated images.
Figure 3. Algorithm process flow developed for automated annotation of simulated images.
Sensors 24 05816 g003
Figure 4. An automatically annotated (a) virtual image and (b) manually annotated real image, which were used to train the virtual model and real model, respectively. The rectangular boxes in the image represent annotated partial and whole fish in the image.
Figure 4. An automatically annotated (a) virtual image and (b) manually annotated real image, which were used to train the virtual model and real model, respectively. The rectangular boxes in the image represent annotated partial and whole fish in the image.
Sensors 24 05816 g004
Figure 5. The mean average precision (mAP0.5) scores attained by the M6 mixed model (training dataset consisting of 90% virtual and 10% real images) trained with different (a) epochs and (b) data sizes.
Figure 5. The mean average precision (mAP0.5) scores attained by the M6 mixed model (training dataset consisting of 90% virtual and 10% real images) trained with different (a) epochs and (b) data sizes.
Sensors 24 05816 g005
Figure 6. The effect of data augmentation on the mean average precision (mAP) score of (a) real, (b) mixed, and (c) virtual fish detection models.
Figure 6. The effect of data augmentation on the mean average precision (mAP) score of (a) real, (b) mixed, and (c) virtual fish detection models.
Sensors 24 05816 g006
Figure 7. The maximum (a) mean average precision (mAP0.5) and (b) F1 scores attained by virtual, mixed (M1–M8), and real fish detection models and (c,d) performance comparison of M6 model trained with 90% virtual and 10% real images against the virtual and real models.
Figure 7. The maximum (a) mean average precision (mAP0.5) and (b) F1 scores attained by virtual, mixed (M1–M8), and real fish detection models and (c,d) performance comparison of M6 model trained with 90% virtual and 10% real images against the virtual and real models.
Sensors 24 05816 g007
Figure 8. (a) A sample image acquired in a RAS environment and output images depicting the fish detected in the frame while deploying the (b) real model, (c) virtual model, and (d) mixed model (M6) to sample images.
Figure 8. (a) A sample image acquired in a RAS environment and output images depicting the fish detected in the frame while deploying the (b) real model, (c) virtual model, and (d) mixed model (M6) to sample images.
Sensors 24 05816 g008aSensors 24 05816 g008b
Table 1. Fish detection models trained with virtual-, real-, and mixed-image datasets.
Table 1. Fish detection models trained with virtual-, real-, and mixed-image datasets.
Training Dataset Model Name Virtual to Real Image Proportion Number of Training Images
VirtualReal
Virtual imagesVirtual100:07000
Virtual and real images* M199:16937
M298:268614
M396:467228
M494:665842
M592:864456
M690:1063070
M775:25525175
M850:50 350350
Real images Real0:1000500
* M represents mixed models trained with varying proportions of virtual and real images.
Table 2. The total time costs of virtual, mixed, and real model training.
Table 2. The total time costs of virtual, mixed, and real model training.
Model NameAnnotation Time (s)Training Time (s)Total Training Time Cost (s) *
Virtual ImageReal Image
Virtual330.00.01429.21759.2
Mixed (M6)297.87140.02750.410,187.4
Real0.071,400.0282674,226
* reported time for annotating and training 700 images.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Steele, S.R.; Ranjan, R.; Sharrer, K.; Tsukuda, S.; Good, C. Computer-Simulated Virtual Image Datasets to Train Machine Learning Models for Non-Invasive Fish Detection in Recirculating Aquaculture. Sensors 2024, 24, 5816. https://doi.org/10.3390/s24175816

AMA Style

Steele SR, Ranjan R, Sharrer K, Tsukuda S, Good C. Computer-Simulated Virtual Image Datasets to Train Machine Learning Models for Non-Invasive Fish Detection in Recirculating Aquaculture. Sensors. 2024; 24(17):5816. https://doi.org/10.3390/s24175816

Chicago/Turabian Style

Steele, Sullivan R., Rakesh Ranjan, Kata Sharrer, Scott Tsukuda, and Christopher Good. 2024. "Computer-Simulated Virtual Image Datasets to Train Machine Learning Models for Non-Invasive Fish Detection in Recirculating Aquaculture" Sensors 24, no. 17: 5816. https://doi.org/10.3390/s24175816

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop