KRID: A Large-Scale Nationwide Korean Road Infrastructure Dataset for Comprehensive Road Facility Recognition
Abstract
:1. Introduction
- We introduce a new large-scale road infrastructure dataset encompassing 34 facility categories, with annotations provided in both bounding-box and polygon segmentation formats;
- We design and implement a robust annotation workflow—including strict privacy compliance and rigorous quality-control measures—that ensures the high quality of the labels;
- We validate the dataset’s utility through extensive object detection and segmentation experiments, and release it as an open resource to promote further research and collaboration.
2. Materials and Methods
2.1. Data Collection Sites
2.2. Data Acquisition Tools
2.2.1. Imaging System
- Viewer and Control Program: A custom-developed application supported both single and multi-camera feeds, providing a real-time display to the operator and allowing for direct control over imaging parameters;
- GPS-Based Time Server: To synchronize camera timestamps, a GPS-based time server was incorporated. This infrastructure ensured that the camera’s capture time was accurately logged for each frame, at a rate exceeding 10 frames per second;
- Precise Timestamping: Each captured frame received a timestamp (in milliseconds) from the GPS server, enabling the accurate tagging of spatial and temporal data;
- Automatic Positional Extraction: By aligning each frame’s timestamp with the associated GPS log, the system automatically extracted precise location coordinates, facilitating robust georeferencing.
2.2.2. Vehicle System
- Vehicle Platforms: The data acquisition fleet included the UOK-LAND I (1 unit), UOK-LAND II (5 units), and 3 newly designed survey vehicles from Sodasystem—9 vehicles in total;
- Mounting Structure: When necessary, a custom frame or elevated mounting apparatus was installed on top of the vehicles to secure the cameras in positions at or above 1900 mm;
- Fixed Camera Angle: The camera pitch and yaw angles were held constant throughout data collection, minimizing perspective variation and streamlining subsequent processing steps.
2.2.3. Data Gathering Protocol
- Survey Planning: First, we identified weekly survey regions (index) and obtained local weather forecasts. We also verified whether any construction projects or restricted zones might interfere with data acquisition in the targeted areas.
- Preparation: All vehicles underwent mechanical inspections, and the camera equipment was tested to confirm functionality. These steps ensured that any technical issues were addressed before field deployment.
- Field Operation: A skilled team with extensive data acquisition experience was dispatched to conduct the surveys. To maintain steady imaging and consistent viewpoints, the team observed specific driving protocols, such as maintaining a constant speed to reduce motion blur and ensure uniform image quality, positioning the vehicle near the lane center where safety permitted, keeping an appropriate following distance to avoid abrupt stops or accelerations, and minimizing unnecessary lane changes. These measures helped to standardize the conditions under which each road segment was surveyed (see Figure A3 for a detailed workflow illustration).
- Reporting and Data Storage: Upon the completion of each run, the team prepared a concise survey report. Video data and GPS logs were uploaded to a centralized server, and backups were created to safeguard against potential data loss.
- Verification and Quality Check: Finally, the collected data underwent a verification process. Image quality (resolution, focus, exposure) was inspected, and any missing elements (e.g., GPS information or video segments) were identified. Duplicate data instances were also detected and removed, ensuring the dataset’s overall consistency and reliability.
2.3. Data Annotation and Management
2.3.1. Data Annotation Tools
- Task Logging: The system records all worker actions, enabling the traceability of labeling tasks;
- Online Operation: Runs on a web-based interface, allowing distributed annotators to collaborate remotely;
- Auto-Labeling: Provides an optional auto-labeling function, where baseline annotations are automatically generated and then refined by human annotators;
- Operating Institution: Developed and maintained in-house by Testworks, ensuring alignment with project-specific requirements.
2.3.2. Data Processing and Management
- Task Allocation: Distributes labeling work among annotators and supervisors;
- Project and Worker Management: Monitors overall status, individual performance, and assignments for each project;
- Quality Assurance Workflow: Tracks the review and acceptance process (e.g., rate of rejections, final approvals), allowing for the real-time monitoring of quality metrics;
- Collaborative Issue Tracking: Offers a bulletin board system, enabling annotators, reviewers, and administrators to share and resolve issues;
- Dashboard and Statistics: Presents real-time progress indicators such as completion rate, rejection percentage, and quality metrics, with the ability to download task data or generate charts for both worker-level and project-level reporting.
2.4. De-Identification
- 1.
- Initial Automated DetectionAn AI-based de-identification solution was employed to automatically detect potential privacy-sensitive regions in the images (e.g., faces, license plates). This first-pass detection generated candidate bounding boxes likely to contain personal data. In particular, we utilized an in-house de-identification tool, developed from well-known benchmark deep learning models and further refined with our own data. It served to set initial bounding-box candidates before human annotators performed final corrections;
- 2.
- Secondary Review and VerificationThe automatically detected regions were then provided to trained annotators for manual inspection and refinement. These annotators were rigorously selected from among individuals with proven experience in other vision-data processing projects, ensuring a high level of labeling proficiency. During this step, missing detections were added, inaccuracies were corrected, and all personal data regions were effectively blurred or obscured. This double-layered approach ensured that sensitive information was comprehensively removed while preserving the overall utility of the dataset for AI analysis.
2.5. Workforce Training
2.6. Data Description
3. Results
3.1. Object Detection Model (Bounding Box)
Object Detection Evaluation Results
3.2. Segmentation Model (Polygon)
Segmentation Evaluation Results
4. Discussion
4.1. Interpretation of Results
4.1.1. Balanced Coverage
4.1.2. Practical Utility
4.1.3. Comparison with Prior Work
4.2. Data Collection and Quality Assurance
4.3. Limitations
4.3.1. Class Imbalance
4.3.2. Environmental Constraints
4.3.3. Lack of Temporal Annotations
4.4. Practical and Theoretical Implications
4.5. Future Outlook
- Sensor Fusion: Integrating LiDAR, radar, or thermal imagery could expand detection accuracy and ensure greater reliability for autonomous vehicles;
- Rare Class Augmentation: Adopting GAN-based synthetic data generation or targeted data gathering may alleviate class imbalance issues;
- Real-Time Monitoring Systems: Building on the current segmentation and detection capabilities, the continuous monitoring of critical road infrastructure is an attainable next step, enabling prompt maintenance responses and safer traffic management.
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Appendix A. Dataset
Appendix A.1. Data Annotation Guidelines
- Proper tagging (a): All objects are correctly identified and precisely outlined.
- Wrong tagging (b): Bounding boxes are misplaced, either incorrectly aligned or mislabeled.
- Under tagging (c): Some objects remain unannotated, resulting in incomplete labeling.
- Over tagging (d): Bounding boxes are applied excessively or redundantly, potentially introducing confusion and inflating object counts.
Appendix A.2. Data Acquisition Detail
Appendix A.3. Additional Data Detail Information
Region | Object Count | Ratio |
---|---|---|
Gangwon-do | 428,170 | 16.00% |
Gyeonggi-do | 506,115 | 18.91% |
Gyeongsangnam-do (South) | 183,905 | 6.87% |
Gyeongsangbuk-do (North) | 352,906 | 13.18% |
Gwangju | 27,178 | 1.02% |
Daegu | 74,706 | 2.79% |
Daejeon | 47,347 | 1.77% |
Busan | 50,348 | 1.88% |
Seoul | 96,767 | 3.62% |
Sejong | 10,157 | 0.38% |
Ulsan | 18,811 | 0.70% |
Incheon | 20,416 | 0.76% |
Jeollanam-do (South) | 182,585 | 6.82% |
Jeollabuk-do (North) | 188,040 | 7.03% |
Chungcheongnam-do (South) | 273,799 | 10.23% |
Chungcheongbuk-do (North) | 215,333 | 8.05% |
Total | 2,676,583 | 100.00% |
Region Type | Road Type | Count | Ratio |
---|---|---|---|
City | National Road | 193,387 | 7.23% |
City | Local Road | 9,649 | 0.36% |
Non-City | Highway | 781,924 | 29.21% |
Non-City | National Road | 1,478,748 | 55.25% |
Non-City | Local Road | 212,875 | 7.95% |
Total | — | 2,676,583 | 100.00% |
Appendix A.4. Additional Experimental Results
- High-Proportion of Small Object:Figure A4a–c illustrates the distribution of object sizes in our dataset, revealing that many objects occupy only a small area of the image (i.e., low width and height). Object detection models often struggle with such small objects due to their limited pixel representation, which can negatively impact overall performance. When a significant portion of the dataset consists of these small objects, extracting meaningful features becomes even more challenging.
- Extreme Aspect Ratios of Road Facilities:Figure A4d shows the distribution of aspect ratios in our dataset. Object detection models often struggle with very tall and narrow or very wide and short objects, as these extremes occupy only a small portion of the image, making feature extraction and localization more difficult. Many detection frameworks (e.g., Faster R-CNN, YOLO) rely on predefined anchor boxes, which may not adequately cover such extreme shapes.
- Occlusion-Induced Detection Failures:Figure A5 shows two scenarios in which partial occlusion compromises object detection. In Figure A5a,b, a building partially obstructs certain objects, leading to incorrect bounding boxes or reduced IoU. In Figure A5c,d, tall trees obscure the target so severely that the model fails to detect it altogether. These examples underscore how occlusion—whether by buildings or vegetation—can result in detection errors or complete misses.
References
- Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Li, F.-F. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
- Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the Computer Vision—ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part V 13. Springer: Cham, Switzerland, 2014; pp. 740–755. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
- He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
- Kuznetsova, A.; Rom, H.; Alldrin, N.; Uijlings, J.; Krasin, I.; Pont-Tuset, J.; Kamali, S.; Popov, S.; Malloci, M.; Kolesnikov, A.; et al. The open images dataset v4: Unified image classification, object detection, and visual relationship detection at scale. Int. J. Comput. Vis. 2020, 128, 1956–1981. [Google Scholar] [CrossRef]
- Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The kitti dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
- Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The cityscapes dataset for semantic urban scene understanding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar]
- Neuhold, G.; Ollmann, T.; Rota Bulo, S.; Kontschieder, P. The mapillary vistas dataset for semantic understanding of street scenes. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4990–4999. [Google Scholar]
- Yu, F.; Xian, W.; Chen, Y.; Liu, F.; Liao, M.; Madhavan, V.; Darrell, T. Bdd100k: A diverse driving video database with scalable annotation tooling. arXiv 2018, arXiv:1805.04687. [Google Scholar]
- Rebuffi, S.A.; Bilen, H.; Vedaldi, A. Learning multiple visual domains with residual adapters. Adv. Neural Inf. Process. Syst. 2017, 30, 506–516. [Google Scholar]
- Zamir, A.R.; Sax, A.; Shen, W.; Guibas, L.J.; Malik, J.; Savarese, S. Taskonomy: Disentangling task transfer learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3712–3722. [Google Scholar]
- McCann, B.; Keskar, N.S.; Xiong, C.; Socher, R. The natural language decathlon: Multitask learning as question answering. arXiv 2018, arXiv:1806.08730. [Google Scholar]
- Everingham, M.; Van Gool, L.; Williams, C.K.; Winn, J.; Zisserman, A. The pascal visual object classes (voc) challenge. Int. J. Comput. Vis. 2010, 88, 303–338. [Google Scholar] [CrossRef]
- Acuna, D.; Ling, H.; Kar, A.; Fidler, S. Efficient interactive annotation of segmentation datasets with polygon-rnn++. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 859–868. [Google Scholar]
- Voigtlaender, P.; Krause, M.; Osep, A.; Luiten, J.; Sekar, B.B.G.; Geiger, A.; Leibe, B. Mots: Multi-object tracking and segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7942–7951. [Google Scholar]
- Milan, A. MOT16: A benchmark for multi-object tracking. arXiv 2016, arXiv:1603.00831. [Google Scholar]
- Choi, Y.; Kim, N.; Hwang, S.; Park, K.; Yoon, J.S.; An, K.; Kweon, I.S. KAIST multi-spectral day/night data set for autonomous and assisted driving. IEEE Trans. Intell. Transp. Syst. 2018, 19, 934–948. [Google Scholar] [CrossRef]
- Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The kitti vision benchmark suite. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
- Che, Z.; Li, G.; Li, T.; Jiang, B.; Shi, X.; Zhang, X.; Lu, Y.; Wu, G.; Liu, Y.; Ye, J. D2-city: A large-scale dashcam video dataset of diverse traffic scenarios. arXiv 2019, arXiv:1904.01975. [Google Scholar]
- Vinothkumar, S.; Dhanushya, S.; Guhan, S.; Krisvanth, P. Enhancing Road Infrastructure Maintenance Using Deep Learning Approach. In Proceedings of the International Conference on Intelligent Systems Design and Applications, Online, 11–13 December 2023; Springer: Cham, Switzerland, 2023; pp. 205–214. [Google Scholar]
- Silva, L.A.; Leithardt, V.R.Q.; Batista, V.F.L.; González, G.V.; Santana, J.F.D.P. Automated road damage detection using UAV images and deep learning techniques. IEEE Access 2023, 11, 62918–62931. [Google Scholar] [CrossRef]
- Angulo, A.; Vega-Fernández, J.A.; Aguilar-Lobo, L.M.; Natraj, S.; Ochoa-Ruiz, G. Road damage detection acquisition system based on deep neural networks for physical asset management. In Proceedings of the Advances in Soft Computing: 18th Mexican International Conference on Artificial Intelligence, MICAI 2019, Xalapa, Mexico, 27 October–2 November 2019; Proceedings 18. Springer: Cham, Switzerland, 2019; pp. 3–14. [Google Scholar]
- Sami, A.A.; Sakib, S.; Deb, K.; Sarker, I.H. Improved YOLOv5-based real-time road pavement damage detection in road infrastructure management. Algorithms 2023, 16, 452. [Google Scholar] [CrossRef]
- Kulambayev, B.; Gleb, B.; Katayev, N.; Menglibay, I.; Momynkulov, Z. Real-Time Road Damage Detection System on Deep Learning Based Image Analysis. Int. J. Adv. Comput. Sci. Appl. 2024, 15, 1051–1061. [Google Scholar] [CrossRef]
- Kim, H.; Kim, B.; Lu, W.; Li, L. No-code MLOps Platform for Data Annotation. In Proceedings of the 2023 IEEE International Conference on Memristive Computing and Applications (ICMCA), Jinan, China, 8–10 December 2023; pp. 1–6. [Google Scholar]
- Botezatu, A.P.; Burlacu, A.; Orhei, C. A review of deep learning advancements in road analysis for autonomous driving. Appl. Sci. 2024, 14, 4705. [Google Scholar] [CrossRef]
- Hwang, S.J.; Hong, S.W.; Yoon, J.S.; Park, H.; Kim, H.C. Deep Learning-based Pothole Detection System. J. Semicond. Disp. Technol. 2021, 20, 88–93. [Google Scholar]
Dataset | Ann. Images | Resolution | Classes of Road Infrastructure | Classes Total | Location | Instances | |||
---|---|---|---|---|---|---|---|---|---|
Rd. Safety | Rd. Mgmt | Traffic Mgmt | Bbox | Segment | |||||
BDD100K [10] | 100 k | 1 | 0 | 1 | 10 | 4 Cities | 1,841,435 | 189,053 | |
KAIST [18] | 8.9 k | 0 | 0 | 0 | 6 | 1 Met City | 308,913 | - | |
KITTI [19] | 15 k | 0 | 0 | 0 | 8 | 1 City | ∼200 k | - | |
D2-City [20] | 700 k | 0 | 0 | 0 | 12 | 5 Cities | ∼40 M | - | |
Cityscapes [8] | 25 k | 2 | 2 | 2 | 30 | 50 Cities | - | 65.4 k | |
Vistas [9] | 25 k | 4 | 14 | 8 | 66 | Global | - | ∼2 M | |
KRID (Ours) | 200 k | 16 | 8 | 8 | 34 | Nationwide | 2,031,633 | 644,950 |
Region | Non-City | City | ||||||||
---|---|---|---|---|---|---|---|---|---|---|
Highway | % | Nat. Road | % | Local Rd. | % | Nat. Road | % | Local Rd. | % | |
Gyeonggi-do | 2589.90 | 20% | 1793.57 | 8% | - | - | 1488.40 | 25% | - | - |
Gyeongsangbuk-do | 1910.31 | 15% | 4112.01 | 18% | - | - | 258.53 | 7% | - | - |
Gyeongsangnam-do | 1452.26 | 11% | 2669.52 | 12% | - | - | 335.26 | 9% | - | - |
Gangwon-do | 1156.60 | 9% | 3403.61 | 15% | - | - | 171.70 | 5% | - | - |
Chungcheongnam-do | 1139.77 | 9% | 2304.33 | 10% | 3064.07 | 53% | 420.05 | 9% | 132.52 | 46% |
Chungcheongbuk-do | 969.28 | 8% | 1608.34 | 7% | 2691.57 | 47% | 220.51 | 3% | 158.01 | 54% |
Jeollanam-do | 1041.50 | 8% | 3706.61 | 16% | - | - | 135.70 | 4% | - | - |
Jeollabuk-do | 1004.58 | 8% | 2554.93 | 11% | - | - | 282.68 | 8% | - | - |
Daegu | 343.41 | 3% | 44.90 | 0% | - | - | 170.65 | 5% | - | - |
Incheon | 314.90 | 2% | 45.17 | 0% | - | - | 116.45 | 3% | - | - |
Ulsan | 257.06 | 2% | 285.94 | 1% | - | - | 84.72 | 2% | - | - |
Daejeon | 208.47 | 2% | 59.68 | 0% | - | - | 115.51 | 3% | - | - |
Busan | 201.85 | 2% | 138.39 | 1% | - | - | 136.18 | 4% | - | - |
Gwangju | 91.91 | 1% | 62.19 | 0% | - | - | 147.02 | 4% | - | - |
Seoul | 93.38 | 1% | 2.98 | 0% | - | - | 335.14 | 9% | - | - |
Total (km) | 12,775.19 | 100% | 22,792.17 | 100% | 5755.64 | 100% | 4418.49 | 100% | 290.54 | 100% |
Percentage of total | 27.80% | 49.50% | 12.50% | 9.60% | 0.60% | |||||
Ratio of Non-City to City = | 89.80% | 10.20% |
No. | Field Name | Type | Required | Description |
---|---|---|---|---|
1 | license | array | N | Copyright |
2 | info[] | array | Y | General Information |
2-1 | info[].contributor | string | Y | Contributor |
2-2 | info[].date_created | string | Y | Creation Date |
2-3 | info[].description | string | Y | Data Description |
2-4 | info[].url | string | N | Image URL |
2-5 | info[].version | number | N | Processing Version |
2-6 | info[].year | number | N | Processing Year |
3 | images[] | array | Y | Image Information |
3-1 | images[].id | number | Y | Image ID |
3-2 | images[].file_name | string | Y | Image File Name |
3-3 | images[].width | number | Y | Image Width |
3-4 | images[].height | number | Y | Image Height |
3-5 | images[].latitude | number | N | Latitude |
3-6 | images[].longitude | number | N | Longitude |
3-7 | images[].day | string | Y | Shooting Date |
3-8 | images[].time | string | Y | Shooting Time |
3-9 | images[].camera | string | Y | Camera Model |
3-10 | images[].velocity | number | Y | Driving Speed |
3-11 | images[].direction | number | Y | Shooting Direction |
3-12 | images[].climate | string | Y | Weather at Shooting Time |
3-13 | images[].frame | number | Y | Frame Number |
4 | categories[] | array | Y | Object Information |
4-1 | categories[].id | number | Y | Object ID |
4-2 | categories[].name | enum | Y | Object Name |
4-3 | categories[].supercategory | string | N | Parent Category |
5 | annotations[] | array | Y | Annotation Information |
5-1 | annotations[].id | number | Y | Object ID |
5-2 | annotations[].image_id | number | Y | Image ID |
5-3 | annotations[].category_id | number | Y | Category ID |
5-4 | annotations[].bbox | array | Y | Bounding-Box Coordinates |
5-5 | annotations[].segmentation | array | Y | Segmentation Coordinates |
5-6 | annotations[].region1 | enum | Y | Province/City |
5-7 | annotations[].region2 | enum | N | Local District |
5-8 | annotations[].roadtype | enum | Y | Road Type |
5-9 | annotations[].regiontype | enum | Y | City/Non-City Type |
5-10 | annotations[].state | enum | Y | Facility Condition |
5-11 | annotations[].name | enum | Y | Facility Name |
5-12 | annotations[].subtype | string | N | Facility Subtype |
Facility Name | Type | Object Count | Ratio | Facility Category |
---|---|---|---|---|
CCTV | bbox | 22,476 | 0.84% | Traffic Management |
Utility pole | bbox | 293,999 | 10.98% | Other |
Seagull Mark | bbox | 32,726 | 1.22% | Road Safety |
Speed Bump | bbox | 4773 | 0.18% | Road Safety |
Traffic Signal | bbox | 65,591 | 2.45% | Traffic Management |
Diagonal Line | bbox | 74,291 | 2.78% | Road Safety |
Pillar | bbox | 611,618 | 22.85% | Other |
Emergency Contact Facility | bbox | 160 | 0.01% | Traffic Management |
Road Sign | bbox | 202,289 | 7.56% | Traffic Management |
Road Nameplate | bbox | 31,281 | 1.17% | Traffic Management |
Road Reflector | bbox | 7503 | 0.28% | Road Safety |
Road Guidepost | bbox | 240,859 | 9.00% | Traffic Management |
Variable Message Sign | bbox | 13,401 | 0.50% | Traffic Management |
Gaze-Directed Pole | bbox | 104,075 | 3.89% | Road Safety |
Delineator Sign | bbox | 159,330 | 5.95% | Road Safety |
Safety Sign | bbox | 23,775 | 0.89% | Traffic Management |
Obstacle Marker | bbox | 13,067 | 0.49% | Road Safety |
Lighting | bbox | 128,500 | 4.80% | Road Safety |
Raised Pavement Marker | bbox | 1919 | 0.07% | Road Safety |
Elevated Road | Polygon | 28,727 | 1.07% | Road Management |
Bridge | Polygon | 3139 | 0.12% | Road Management |
Rock-Fall Protection Wall | Polygon | 40,683 | 1.52% | Road Safety |
Rock-Fall Net | Polygon | 7814 | 0.29% | Road Safety |
Rock-Fall Fence | Polygon | 31,774 | 1.19% | Road Safety |
Guardrail | Polygon | 307,060 | 11.47% | Road Safety |
Eco-Engineering | Polygon | 289 | 0.01% | Road Safety |
Overpass Walkway | Polygon | 2585 | 0.10% | Road Management |
Interchange | Polygon | 1111 | 0.04% | Road Management |
Bus/Transit Stop | Polygon | 11,834 | 0.44% | Road Management |
Median Barrier | Polygon | 169,411 | 6.33% | Road Safety |
Underground Passage | Polygon | 1326 | 0.05% | Road Management |
Underpass | Polygon | 1602 | 0.06% | Road Management |
Shock Absorption | Polygon | 27,655 | 1.03% | Road Safety |
Tunnel | Polygon | 9940 | 0.37% | Road Management |
Total | - | 2,676,583 | 100.00% | - |
Images | Train | Validation | Test | Total |
---|---|---|---|---|
Number of Images | 159,426 | 20,287 | 20,287 | 200,000 |
Percentage | 79.70% | 10.10% | 10.10% | 100.00% |
Objects | Train | Validation | Test | Total |
---|---|---|---|---|
Count | 1,604,318 | 200,540 | 226,775 | 2,031,633 |
Ratio (%) | 79.00% | 9.90% | 11.20% | 100.00% |
Item | Specification |
---|---|
CPU | Ryzen9 7900 (12 cores, 24 threads) (AMD, CA, USA) |
Memory | 32 GB |
GPU | GeForce RTX 4080 (NVIDIA, CA, USA) |
Storage | 2 TB |
OS | Ubuntu 22.04.3 (64-bit) |
Development Language | Python 3.10.12 |
Training Algorithm | YOLOv4 |
Training Conditions | epoch: 15, learning_rate: 0.0013, optimizer: ADAM |
Dataset | Precision | Recall | mAP50 |
---|---|---|---|
Test Set | 0.78 | 0.73 | 0.705744 |
Validation Set | 0.78 | 0.73 | 0.7086 |
Objects | Train | Validation | Test | Total |
---|---|---|---|---|
Count | 514,316 | 64,290 | 66,344 | 644,950 |
Ratio (%) | 79.70% | 10.00% | 10.30% | 100.00% |
Item | Specification |
---|---|
CPU | Xeon(R) Silver 4210 CPU @ 2.20 GHz (Intel, CA, USA) |
Memory | 256 GB |
GPU | Tesla V100 (32 GB) (NVIDIA, CA, USA) |
Storage | 2 TB |
OS | Ubuntu 20.04.6 LTS |
Development Language | Python 3.10.12 |
Training Algorithm | YOLOv5 |
Training Conditions | epoch: 60, |
start_learning_rate: 0.01, optimizer: SGD |
Dataset | Precision | Recall | mAP50 |
---|---|---|---|
Test Set | 0.713 | 0.571 | 0.604 |
Validation Set | 0.711 | 0.573 | 0.594 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Kim, H.; Kim, E.; Ahn, S.; Kim, B.; Kim, S.J.; Sung, T.K.; Zhao, L.; Su, X.; Dong, G. KRID: A Large-Scale Nationwide Korean Road Infrastructure Dataset for Comprehensive Road Facility Recognition. Data 2025, 10, 36. https://doi.org/10.3390/data10030036
Kim H, Kim E, Ahn S, Kim B, Kim SJ, Sung TK, Zhao L, Su X, Dong G. KRID: A Large-Scale Nationwide Korean Road Infrastructure Dataset for Comprehensive Road Facility Recognition. Data. 2025; 10(3):36. https://doi.org/10.3390/data10030036
Chicago/Turabian StyleKim, Hyeongbok, Eunbi Kim, Sanghoon Ahn, Beomjin Kim, Sung Jin Kim, Tae Kyung Sung, Lingling Zhao, Xiaohong Su, and Gilmu Dong. 2025. "KRID: A Large-Scale Nationwide Korean Road Infrastructure Dataset for Comprehensive Road Facility Recognition" Data 10, no. 3: 36. https://doi.org/10.3390/data10030036
APA StyleKim, H., Kim, E., Ahn, S., Kim, B., Kim, S. J., Sung, T. K., Zhao, L., Su, X., & Dong, G. (2025). KRID: A Large-Scale Nationwide Korean Road Infrastructure Dataset for Comprehensive Road Facility Recognition. Data, 10(3), 36. https://doi.org/10.3390/data10030036