Next Article in Journal
TGN: A Temporal Graph Network for Physics Prediction
Previous Article in Journal
Transformer Fault Diagnosis Method Based on SCA-VMD and Improved GoogLeNet
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Development of Architectural Object Automatic Classification Technology for Point Cloud-Based Remodeling of Aging Buildings

School of Architecture, Kyungpook National University, Sangyeok-dong, Buk-gu, Daegu 41566, Republic of Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(2), 862; https://doi.org/10.3390/app14020862
Submission received: 16 November 2023 / Revised: 26 December 2023 / Accepted: 27 December 2023 / Published: 19 January 2024
(This article belongs to the Section Civil Engineering)

Abstract

:
In this study, we address the challenge of efficiently handling the maintenance and remodeling of buildings constructed post-1960s, lacking architectural drawings. The conventional approach involves manual measurements and data recording, followed by digital drawing creation. However, we leverage Fourth Industrial Revolution technologies to develop a deep learning-based automatic object classification system using point cloud data. We employ the FCAF3D network with multiscale cells, optimizing its configuration for classifying building components such as walls, floors, roofs, and other objects. While classifying walls, floors, and roofs using bounding boxes led to some boundary-related errors, the model performed well for objects with distinct shapes. Our approach emphasizes efficiency in the remodeling process rather than precise numerical calculations, reducing labor and improving architectural planning quality. While our dataset labeling strategy involved bounding boxes with limitations in numerical precision, future research could explore polygon-based labeling, minimizing loss of space and potentially yielding more meaningful results in classification. In summary, our technology aligns with the initial research objectives, and further investigations could enhance the methodology for even more accurate building object classification.

1. Introduction

Through the Industrial Revolution, mankind became capable of mass and automatic production through the rise of technological power and the development of machinery. Likewise, in the world of architecture, countless buildings are built. However, after World War I and World War II, many buildings were destroyed and lost, and people began to build residential spaces haphazardly. In particular, in the architectural industry, the invention of concrete led to the construction of structurally strong buildings, and such buildings have been produced in countless numbers since the 1950s. Now, over 60 years later, these buildings require remodeling, which is becoming more common through urban regeneration initiatives [1].
The conventional method for creating drawings before demolition relies heavily on manual labor. Workers use tools like rulers and laser measuring devices to measure dimensions and manually sketch architectural elements such as walls and columns. The data collected through this process is then used to create digital drawings, which, in turn, facilitate structural assessments and architectural planning to identify the components that need to be removed, contributing to waste reduction.
In the era of the Fourth Industrial Revolution, various smart technologies have emerged, finding applications across industries, including architecture [2].
These technologies encompass automated recognition and generation of drawings from data, as well as the use of deep learning for classifying building objects.
Smart devices like 3D scanners and drones, which employ laser-based point cloud data, have expanded the scope of work in the industry, allowing for tasks previously impossible with traditional 2D photography.
These technologies are particularly useful in indoor architecture for classifying and automatically arranging objects. The Fourth Industrial Revolution has the potential to replace labor-intensive, repetitive tasks, such as the manual measurement process, with the efficiency of machine and digital elements. This shift can free up time for more creative endeavors. Consequently, this research aims to develop a deep learning-based object classification technology using point cloud data to replace labor-intensive methods when remodeling aging buildings [2].

2. Background and Related Research

As mentioned, during the architectural measurement process, the step of creating architectural drawings separates the objects to be demolished from those to be preserved. To do this, a structural assessment phase is typically required. However, if we can automatically classify architectural objects before this step, it becomes easier to calculate the subsequent architectural waste and identify structural elements to be preserved. To achieve this, we aim to develop deep learning-based object classification technology, which requires the prior collection of training data.
In this research, we intend to classify architectural objects such as walls, columns, floors, and roofs, with an additional category for miscellaneous objects not included in actual buildings. To accomplish this, we plan to utilize point cloud public data from AI-Hub, provided by the Korea National Institute for Artificial Intelligence (AI-Hub), and real 3D scanning data of aging buildings. The methodology of this study is illustrated in Figure 1, as follows.
First, through an analysis of the architectural remodeling process, this study aims to determine at which stages our developed technology can be applied and identify the necessary elements. Subsequently, a preliminary analysis of prior research on point clouds is conducted. Then, the deep learning network to be used for training and classification in this study is analyzed and selected. Further modifications are made to adapt the chosen model to the specific needs of this research.
Following this, the collection of point cloud data is carried out using the aforementioned public data and actual measurement data. This data often contains unnecessary information when used for training, as it includes numerous data points, which significantly increase data volume compared to regular image data. Therefore, a data refinement process is applied to the point cloud data, followed by the training phase.
Finally, the trained model is used to verify the classification of architectural objects into their respective classes. This process is aimed at developing detailed construction techniques that can efficiently utilize the classified architectural objects in the later stages of the building remodeling process.

3. Analysis of Aging Building Remodeling and Object Recognition Technology

3.1. Guidelines for Aging Building Remodeling and Analysis of Related Research

Remodeling is defined as “the act of suppressing the aging of a building or partially expanding it for function enhancement, among other things”. In the context of extending the lifespan of a building during its lifecycle, remodeling is a crucial category in architecture, offering an efficient means to establish a sustainable architectural environment in the 21st century [3]. Currently, terms like remodeling and renovation are used interchangeably, but this text predominantly employs the term “remodeling” in a broad sense.
The remodeling process can be divided into planning, design, and construction phases. The planning phase is particularly important in the execution of remodeling projects, as it requires a more detailed analysis and process than new building construction plans to maintain and enhance various building performances [4].
Recent research directions in the field of aging building remodeling can be broadly categorized into three areas. First, there is research aimed at improving energy efficiency, with various studies exploring ways to reduce energy consumption in aging buildings and investigating energy-saving potential. The second area of research involves the study of recycled materials and building waste, focusing on sustainable remodeling materials and waste management practices, aiming to reduce environmental impact. Lastly, there is research on the introduction of renewable energy, such as solar and wind power systems, to enhance the sustainability of aging buildings, as seen in Figure 2.
Notably, there is a vibrant body of research in the planning phase of remodeling aimed at enhancing energy performance. These studies typically conduct phased analyses based on the building’s lifecycle and prioritize analyses depending on the extent of building reuse during remodeling.
Interest in aging building remodeling is also growing in South Korea, and guidelines have been established accordingly. The Ministry of Land, Infrastructure, and Transport has published the “Guidelines for Aging Building Remodeling”, which focus on safety, energy efficiency, and cultural heritage preservation. Additionally, the Korea Conformity Laboratories (KCL) provides technical guidelines for aging building remodeling, emphasizing energy efficiency and safety.
Internationally, several guidelines have been introduced. Notably, the International Union of Architects (UIA) offers international guidelines related to aging building remodeling, emphasizing cultural heritage preservation and sustainability. The U.S. Green Building Council (USGBC) provides LEED certification for improving the sustainability of existing buildings.
These guidelines are formulated to enhance building functionality, safety, and sustainability. To achieve this, architectural drawings and measurement data play a crucial role in the remodeling process.
Currently, approximately 31.3% of buildings in South Korea do not possess architectural drawings. This percentage is derived by dividing the number of buildings with registered drawings in the Road Address Building Data held by the National Geographic Information Institute by the total number of buildings.
During remodeling, architectural drawings are essential to understand the structure, layout, and facility locations of the existing building. In the absence of drawings, data collection must be done through measurements, particularly at the site, to precisely comprehend the current state of the building.
The measurement process for creating architectural drawings typically begins with preparing laser distance measurement devices, cameras, scanners, and safety equipment. Both the interior and exterior of the building are measured, capturing vital architectural elements like size, structure, walls, windows, doors, and stairs. Detailed measurements are taken for ceiling height, door width, window size, and the location of electrical and plumbing systems. All data, including photographs, is documented, enabling the generation of digital data. This data is then analyzed and processed in real-time to ensure accurate results.
Measurement techniques can be categorized into three methods: manual measurements, laser measurements, and photographic measurements. These methods vary in the tools they use, but they all involve extracting drawings based on the measured dimensions through on-site sketching and documentation [6]. In recent years, the development and advancement of Fourth Industrial Revolution technologies have led to increased efforts to utilize tools like 3D laser scanners, drones, and image-processing devices [7]. Figure 3 provides an overview of a proposed 3D object measurement tool equipped with a topographic LIDAR view.
These methods enable accurate and rapid measurement of a building’s structure and shape while capturing the external and internal aspects of the building at high resolutions. Therefore, in this study, we aim to develop technology that utilizes such devices and techniques to perform the measurement process in aging building remodeling more efficiently.

3.2. Analysis of Deep Learning-Based Point Cloud Object Recognition Technology

Point clouds and triangle meshes serve as robust representations, preserving intricate surface details essential for conveying comprehensive geometric information related to the structure of buildings and building components, as well as their interconnections [8]. The advent of deep learning has notably enhanced object recognition within this context. Deep learning leverages its capability to acquire prior knowledge of object features by performing extensive feature calculations on substantial datasets in advance [9].
The method for searching for objects within point cloud data is not significantly different from searching for objects in traditional images using networks [10,11]. In conventional images, objects are labeled with Bounding Boxes in x and y coordinates, and networks, primarily based on Fully Convolutional Neural Networks (FCNN), are used to locate objects in new images. The method for searching for objects within point cloud data is fundamentally similar [12].
Recent state-of-the-art methods for object recognition include point-cloud object recognition based on histograms of dual deviation angle features (HDDAF) [13] to detect unknown objects in point clouds by matching histograms of unknown objects with a model library. Shi et al. [14]. Developed an effective and efficient 3D object detector using a voxel-point geometry abstraction scheme for acceleration and the generation of proposals.
The previously mentioned network, FCNN, is one of the key networks that has rapidly advanced the development of deep learning technology. It is effectively used for object segmentation in high-resolution images. FCNN is an extension of traditional convolutional neural networks designed for image segmentation tasks [15]. This network has significant applications in semantic segmentation, object detection, which accurately predicts object positions and boundaries, and natural language processing, spanning fields such as roads, autonomous driving, medical image analysis, and more.
Semantic segmentation in point clouds is also quite similar to semantic segmentation in 2D images [16]. In traditional 2D images, labeling is done pixel by pixel for training, while in 3D, points are labeled similarly to pixels for training. Figure 4 illustrates the structure of FCNN. The foundational structure of FCNN involves augmenting the VGG 16-layer convolution network with a multilayer deconvolution network. This addition aims to produce a precise segmentation map for a given input proposal. Utilizing the feature representation derived from the convolution network, employ a sequence of unspooling, deconvolution, and rectification operations to construct a dense pixel-wise class prediction map [17].
Among the deep learning models used for processing point clouds, PointNet is currently one of the most prominent models in terms of performance. The generic object recognition algorithm is: Achieve high quality and accuracy while maintaining high efficiency [18].
PointNet efficiently addresses the structural disorder inherent in point cloud data by realizing effective end-to-end feature learning. PointNet avoids data transformation and handles displacement invariance of input points [19]. Additionally, PointNet has been successful in tasks such as segmentation and object classification, outperforming previous methods such as PointNetX.
PointNet is highly effective in performing tasks such as 3D object classification, segmentation, and recognition by extending traditional 2D image processing techniques into the 3D space [20]. PointNet takes 3D data represented as points or point clouds as input. This model learns the features of each point and integrates them to recognize or segment objects. One notable feature of PointNet is its ability to generate consistent results regardless of the input order.
PointNet, based on the principles shown in Figure 5, was first introduced by Qi et al. in 2017. Subsequently, an improved version called PointNet++ was developed [21]. While the initial version focused on processing single-point clouds, PointNet++ was designed to handle high-resolution point clouds through a hierarchical structure, resulting in enhanced performance in capturing spatial hierarchical information. These models are effectively utilized in the classification of 3D objects and the recognition of 3D objects and boundaries [20].
The lack of digital representations impedes the efficient operation and renovation of existing buildings. Digital information in the form of geometric information can be extracted from the shape of buildings and a high degree of surface detail in triangular meshes and point clouds.
In this study, we developed a technology that can efficiently be utilized in the remodeling process for numerous buildings without architectural drawings. Specifically, we employed point cloud data obtained through 3D scanning to classify architectural objects using the FCAF3D model, making it applicable for utilization.
The bottom-up approach refers to recognizing individual points or small parts first and then combining these small parts to recognize the entire object. It involves extracting features for small parts of 3D points and clustering or grouping points with similar features. By combining these grouped points, it recognizes and segments the entire object. While this approach performs well in finding details, it is challenging to recognize the overall structure, and it has a high computational complexity.
On the other hand, the top-down approach involves first recognizing the overall structure or large objects and then decomposing them into detailed parts for recognition. In the case of FCA3D, it classifies objects first and then proceeds with instance segmentation through learning. This method is effective in quickly recognizing large objects, but it may miss detailed details.
This research aims at the classification and segmentation of point cloud datasets to increase automation in the accuracy of the classification of architectural objects in the context of the renovation of aging buildings. The scope of architectural objects in this research encompasses point cloud-built objects scanned for renovation.

4. Construction Object Automatic Classification Dataset based on Point Cloud Data Construction

4.1. Building Object Segmentation Dataset Construction for Architectural Object Classification

The following Figure 6 represents the process of this study.
The majority of the training data used in this study was sourced from the public dataset “Indoor Space 3D Comprehensive Data” available on AI-Hub, as mentioned earlier. This dataset encompasses indoor space data from various types of buildings, including regular housing (apartments, villas, etc.), as well as special facilities such as hospitals, libraries, factories, amusement parks, and more. The indoor space data in this dataset was divided into Bounding Boxes, and the labeling was provided in ‘JSON’ format. Among the numerous data sets available, an example set of 3D point cloud data and ‘JSON’ files is illustrated in Figure 7.
Furthermore, actual scanned data was obtained from the “Bokhyeon 1-dong Refugee Village Documentation Project” conducted in the Bokhyeon 1-dong area of Buk-gu, Daegu, South Korea. This data includes 3D scanning data of aging buildings (houses) located within the project area. An example of this data is shown in Figure 8.
Based on this point cloud data, labeling was carried out for the categories of walls, columns, floors, and roofs using bounding boxes, as mentioned earlier. Other elements were labeled as miscellaneous categories. The labeled data includes the smallest and largest x, y, and z values for the coordinates of each object. When using this labeled data for training, it is important to consider that multiple objects may overlap and be present within the same point cloud. This can lead to the possibility of mixed or overlapping objects when modeling these points. Therefore, a preprocessing step similar to conventional image segmentation is necessary to prevent such situations. In other words, it was necessary to relabel the existing AI-Hub “Indoor Space 3D Comprehensive Data”, and an example of the newly labeled data is shown in Figure 9.
The data listed in Figure 8 represents individual points that have been labeled, and these points are organized into an “array” for labeling.

4.2. Building an Object Classification Network Establishment

For training, we based our network on ‘FCAF3D’ (https://github.com/SamsungLabs/fcaf3d, accessed on 7 July 2021) [23]. While there are other networks like PointNet and SSPC-Net for classifying 3D objects, the ‘SubRGBD’ dataset is widely used as an evaluation metric for the performance of 3D object classification networks in the fields of computer vision and networks. The term ‘SubRGBD’ is an abbreviation for “Subtractive RGB-D”, signifying RGB-D data processing and analysis techniques. RGB-D data typically refers to data types that combine color (RGB) and depth (D) information. This type of data provides depth information in 3D space alongside 2D images [24]. It denotes the methods used to extract desired information or eliminate unnecessary information from a given RGB-D dataset.
The FCAF3D network has been exhibiting the best performance on the ‘SubRGBD’ dataset, and it offers reasonable training and model speed. It operates using units referred to as ‘voxels’, which are concepts that correspond to pixels in 2D images [12]. Voxels are used as virtual minimum units in 3D datasets, and the training allows for adjusting voxel sizes. In this study, we used a base unit size of 1 cm. Figure 10 illustrates the definition of voxels in a visualized form.
Furthermore, Figure 11 represents the fundamental network architecture of FCAF3D.
As mentioned above, FCAF3D’s model is not significantly different from the models for semantic segmentation of 2D images. It learns the data through 3D convolution and pooling processes. In this study, we stacked 4 blocks for training, but for more precise training, you can increase the number of blocks. Conversely, when precision is not required, you can reduce the number of blocks for training. Increasing the number of blocks may slow down the training process, but it offers the advantage of detecting smaller objects.
The overall framework of the FCAF3D is outlined, featuring three-dimensional and sparse convolutions and transposed convolutions. This specific design facilitates the processing of the input point cloud in a single forward pass.

5. Automatic Classification Model for Architectural Objects in Renovating Aging Buildings

5.1. Segmentation Dataset Training

The training was conducted by varying the model’s block count to 4 and 5 or changing the voxel size. When the baseline was set with 4 blocks and the voxel sizes were changed to 1 cm, 1.5 cm, and 2 cm, the training results are shown in Figure 12a–c, respectively.
The Epoch denotes the fundamental unit iteratively undertaken during the learning process. Notably, with the current representation of point cloud data characterized by points and possessing a three-dimensional structure, a pivotal adjustment has been introduced. Specifically, the unit of Voxel for learning has been redefined, and the epoch has been formally designated.
It was observed that the performance decreased when the voxel size was increased to 2 cm. Furthermore, it was found that the 1 cm voxel size produced the most accurate results. The reason behind this is that reducing the voxel size reduces the overlap of cells. However, when considering the entire cell, the overlapping area increases. As a result, the accuracy of training is expected to improve accordingly, and the results confirmed this observation.
Figure 12d,e above show the results of training with the same voxel size while varying the number of blocks.
Changing the number of blocks did not significantly impact the network’s performance, but it was observed that the training speed was faster when using 4 blocks. This is also because, when looking at the structure of the voxels, the number of cases that are repeated during training increases, similar to what was observed when changing the voxel size.

5.2. Automatic Classification of Architectural Objects Based on Point Cloud Data

To classify architectural objects, it is necessary to establish a training dataset. While a larger dataset generally leads to higher accuracy in results, the abundance of irrelevant data is unnecessary. To address this, we utilized the “Comprehensive 3D Indoor Data” collected from the AIHub (aihub.or.kr), operated by the Korea Artificial Intelligence Society for the Promotion of Information and Communications, to gather raw 3D point cloud data. This dataset is expected to be valuable for applications such as 3D object recognition in indoor spaces, training AI robot systems, and rendering 3D perspectives for architecture and interior design.
The Comprehensive 3D Indoor Data comprises a total of 731 indoor spaces, encompassing various types such as typical residences, single-person dwellings (studios, one-bedroom apartments, etc.), schools, libraries, hospitals, public offices, offices, factories, farms, art galleries, amusement parks, cafes, restaurants, accommodations, and more. Within each indoor space, there are a total of 120,000 spatial elements. These spatial elements include basic spatial components, open-close spatial components, furniture elements, electronic device elements, facility safety elements, and others, as seen in Figure 13.
Basic spatial components consist of elements like floors, walls, ceilings, columns, beams, stairs, and roofs. Furniture elements encompass items such as beds, tables, desks, chairs, sofas, storage cabinets, and others.
The training data utilized the ‘Comprehensive 3D Indoor Data’ from AIHub for semantic segmentation labeling. The labeling classes included cupboard, floor, wall, column, beam, window, door, table, chair, sofa, bookshelf, board, and miscellaneous, totaling 13 classes, following the format of the FCAF3D dataset. The aim of this study is the classification and generation of architectural objects, to utilize the generated model in the remodeling process. Consequently, the classified point cloud should align with real-world objects to be effectively employed in practical applications, as seen in Figure 14.
Using the FCAF3D network and the training model employed in this study, we conducted classification verification for the three architectural objects among the previously mentioned five (walls, columns, floors, roofs, and others) using architectural data other than the point cloud data used for training. In the case of this experimental building, there were no columns and roofs, and their boundaries were ambiguous, so they were excluded from the classification target. The following Figure 15a–d represent the results classified into each architectural object.
Figure 13 represents images classified into walls, floors, and others (chairs, desks), respectively. In Figure 13, when classifying walls, you can see that there is space left above and below the yellow Bounding Box. While the roof and columns of the target building were not clearly distinguished, since the dataset was structured with five categories during data collection (walls, floors, roofs, columns, others), you can see that the upper part corresponding to the roof and the lower part corresponding to the floor were not included in the walls.
Figure 13 shows the result of classifying the floor. However, because the dataset consists of point cloud data, it was unable to classify the thickness or volume of the floor. When extracting point cloud data, points are obtained based on factors like the time it takes for the laser to reflect. In this process, the surface of the floor can be obtained through laser reflection. However, the underside of the floor, depending on its thickness, cannot be captured, resulting in the classification of only the floor’s surface.
Figure 13 are the results of classifying other objects, excluding architectural objects, specifically chairs and desks. The “other” category encompasses various objects. Similar to the situation when finding a single wall within the “wall” category of Figure 13 represents the recognition and classification of individual objects within the “other” category, namely chairs and desks.

6. Conclusions

In this study, an architectural object recognition technology was developed to efficiently assist in the remodeling process of countless buildings constructed without blueprints. Aging buildings often undergo remodeling, a process that currently relies heavily on manual labor and is inefficient. In recent times, various smart technologies have been developed and integrated into different industries during the era of the Fourth Industrial Revolution. Smart devices like 3D scanners and drones are widely used for surveying purposes. Therefore, this research aimed to develop deep learning-based automatic object classification and recognition technology using point cloud data and smart technology to replace the traditional, inefficient approach.
Although various deep learning models and networks exist, this study utilized the FCAF3D network, which is voxel-based. By adjusting the voxel size and the number of blocks, the optimal training network environment was determined to optimize performance. Specifically, the voxel size was set to 1 cm, and training was conducted with four blocks. Using this approach, objects within 3D scanning data of real buildings were classified into walls, floors, roofs, and other objects.
The results of classifying architectural objects, including walls, floors, and roofs, displayed some ambiguity at object boundaries due to labeling using Bounding Boxes. Overlapping areas caused errors in object boundary classification. However, the classification of “other” objects within the architectural category clearly distinguished them from other objects, even when the boundaries were adjacent. This model demonstrated that 3D scanning and point cloud data can be automatically classified as “other” objects, reducing the manual deletion of these objects during the alignment process.
Furthermore, the automatic classification of architectural objects allows for the automatic differentiation between elements to be removed and those to be retained during the remodeling process.
While the research aimed to minimize loss of space by utilizing the minimum and maximum values of x, y, and z during the labeling process, there are limitations in obtaining precise numerical values since Bounding Boxes were used. However, the primary goal of this research was to develop a technology that more efficiently distinguishes objects within buildings, reducing labor costs and improving time efficiency during the architectural remodeling process. Based on this perspective, it can be concluded that the research aligned with its initial objectives.
In the future, the research may transition from using Bounding Boxes for labeling to using polygon forms almost to eliminate loss of space. As a result, the numerical output of classification results could be effectively utilized. Through subsequent research that integrates these outcomes with Building Information Modeling (BIM), it is expected that the classified object information will be valuable for tasks such as estimating the amount of construction waste and identifying non-demolished architectural frameworks. Continuous research in this direction has the potential to significantly enhance the efficiency of the architectural remodeling process.

Author Contributions

T.K. conceived experiments, analyzed data, and wrote papers; H.G. investigated prior research; S.H. edited the thesis; and S.C. supervised the research. All authors have read and agreed to the published version of the manuscript.

Funding

This study was conducted with support from the Ministry of Land, Infrastructure, and Transport/Agency for Defense Development under project number RS-2021-KA163269. Additionally, this project (results) represents this research outcomes of the third-stage Industry-Academia Cooperation Leading University Cultivation Project (LINC 3.0), supported by the Ministry of Education and the National Research Foundation of Korea. And This Paper was supported by “National Fire Agency” R&D program (20016433).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kim, K.S. A study on the Phase step of Applying Devices of Remodeling process. Korean Soc. Basic Des. Art 2003, 4, 272–279. [Google Scholar]
  2. Simeone, D. BIM and Behavioural Simulation for existing buildings re-use design. TEMA 2018, 4, 59–69. [Google Scholar] [CrossRef]
  3. Min, J.S.; Park, H.K. A Study on the Remodeling Process of Building. In Proceedings of the Autumn Annual Conference of AIK 2001, Los Angeles, CA, USA, 27 October 2001; Volume 212, pp. 127–130. [Google Scholar]
  4. Kwon, W.; Chun, J.Y. Constitution of Work Process for the Remodeling Construction Project in Planning Phase. Korean J. Constr. Eng. Manag. 2006, 7, 165–174. [Google Scholar]
  5. Hasik, V.; Escott, E.; Bates, R.; Carlisle, S.; Faircloth, B.; Bilec, M.M. Comparative whole-building life cycle assessment of renovation and new construction. Build. Environ. 2019, 161, 106218. [Google Scholar] [CrossRef]
  6. Kwon, S.C.; Kim, M.Y.; Kim, T.Y. The Choice of Measurement Techniques in Survey and Recording of Historic Buildings in Modern Ages of Korea. In Proceedings of the Autumn Annual Conference of AIK, Gwangju, Republic of Korea, 24 October 2008; Volume 28, pp. 431–434. [Google Scholar]
  7. Ryu, J.W.; Byun, N.H. Analysis study on patent for Scan-to-BIM related technology. J. Korea Acad. Ind. Coop. Soc. 2020, 21, 107–114. [Google Scholar]
  8. Czerniawski, T.; Leite, F. Automated digital modeling of existing buildings: A review of visual object recognition methods. Autom. Constr. 2020, 113, 103131. [Google Scholar] [CrossRef]
  9. Wu, Z.; Jiang, Y.; Wang, J.; Pu, J.; Xue, X. Exploring Inter-feature and Inter-class Relationships with Deep Neural Networks for Video Classification. In Proceedings of the 22nd ACM International Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014. [Google Scholar] [CrossRef]
  10. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907v4. [Google Scholar]
  11. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  12. Zhou, Y.; Tuzel, O. VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection. arXiv 2018, arXiv:1711.06396v1. [Google Scholar]
  13. Shi, C.; Wang, C.; Liu, X.; Sun, S.; Xi, G.; Ding, Y. Point cloud object recognition method via histograms of dual deviation angle feature. Int. J. Remote Sens. 2023, 44, 3031–3058. [Google Scholar] [CrossRef]
  14. Shi, G.; Wang, K.; Li, R.; Ma, C. Real-Time Point Cloud Object Detection via Voxel-Point Geometry Abstraction. IEEE Trans. Intell. Transp. Syst. 2023, 24, 5971–5982. [Google Scholar] [CrossRef]
  15. Sultana, F.; Sufian, A.; Dutta, P. Evolution of Image Segmentation using Deep Convolutional Neural Network: A Survey. Knowl. Based Syst. 2020, 201–202, 106062. [Google Scholar] [CrossRef]
  16. Feng, M.; Zhang, L.; Lin, X.; Gilani, S.Z.; Mian, A. Point attention network for semantic segmentation of 3D point clouds. Pattern Recognit. 2020, 107, 107446. [Google Scholar] [CrossRef]
  17. Ghosh, A.; Sufian, A.; Sultana, F.; Chakrabarti, A.; De, D. Fundamental Concepts of Convolutional Neural Network. In Intelligent Systems Reference Library; Springer: Berlin/Heidelberg, Germany, 2019; pp. 519–567. [Google Scholar] [CrossRef]
  18. Zhi, S.; Liu, Y.; Li, X.; Guo, Y. Toward real-time 3D object recognition: A lightweight volumetric CNN framework using multitask learning. Comput. Graph. 2018, 71, 199–207. [Google Scholar] [CrossRef]
  19. Fei, B.; Yang, W.; Chen, W.M.; Li, Z.; Li, Y.; Ma, T.; Hu, X.; Ma, L. Comprehensive Review of Deep Learning-Based 3D Point Cloud Completion Processing and Analysis. IEEE Trans. Intell. Transp. Syst. 2022, 23, 22862–22883. [Google Scholar] [CrossRef]
  20. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet: Deep Learning on Point Sets for 3D Classification and Segmentation. arXiv 2017, arXiv:1711.06396v1. [Google Scholar]
  21. Qi, C.R.; Su, H.; Mo, K.; Guibas, L.J. PointNet++: Deep Hierarchical Feature Learning on Point Sets in a Metric Space. arXiv 2017, arXiv:1706.02413v1. [Google Scholar]
  22. Liu, L.; Ouyang, W.; Wang, X.; Fieguth, P.; Chen, J.; Liu, X.; Pietikäinen, M. Deep Learning for Generic Object Detection: A Survey. Int. J. Comput. Vis. 2020, 128, 261–318. [Google Scholar] [CrossRef]
  23. Rukhovich, D.; Vorontsova, A.; Konushin, A. FCAF3D: Fully Convolutional Anchor-Free 3D Object Detection. arXiv 2021, arXiv:2112.00322v2. [Google Scholar]
  24. Stjepandić, J.; Sommer, M. Object Recognition Methods in a Built Environment. In DigiTwin: An Approach for Production Process Optimization in a Built Environment; Springer Series in Advanced Manufacturing; Springer: Berlin/Heidelberg, Germany, 2021. [Google Scholar] [CrossRef]
Figure 1. Flow of study.
Figure 1. Flow of study.
Applsci 14 00862 g001
Figure 2. Renovation LCA stage and boundary diagram, Hasik et al. [5].
Figure 2. Renovation LCA stage and boundary diagram, Hasik et al. [5].
Applsci 14 00862 g002
Figure 3. Overview of the proposed 3D object detector from the Bird’s Eye View (BEV) of the LIDAR Point Cloud.
Figure 3. Overview of the proposed 3D object detector from the Bird’s Eye View (BEV) of the LIDAR Point Cloud.
Applsci 14 00862 g003
Figure 4. Fully Convolutional Neural Network (FCNN). Adapted with permission from Ref. [15]. 2020, Sultana et al.
Figure 4. Fully Convolutional Neural Network (FCNN). Adapted with permission from Ref. [15]. 2020, Sultana et al.
Applsci 14 00862 g004
Figure 5. Application of PointNet. Adapted with permission from Ref. [22].
Figure 5. Application of PointNet. Adapted with permission from Ref. [22].
Applsci 14 00862 g005
Figure 6. Process of research.
Figure 6. Process of research.
Applsci 14 00862 g006
Figure 7. ‘JSON’ File and Labeling by Bounding box.
Figure 7. ‘JSON’ File and Labeling by Bounding box.
Applsci 14 00862 g007
Figure 8. An Example of an aging building’s 3D Point cloud data.
Figure 8. An Example of an aging building’s 3D Point cloud data.
Applsci 14 00862 g008
Figure 9. An Example of Labeling individual Point data by forming them into an array.
Figure 9. An Example of Labeling individual Point data by forming them into an array.
Applsci 14 00862 g009
Figure 10. Definition of Multi-Cell; (ac) illustrates the definition of voxels in a visualized form.
Figure 10. Definition of Multi-Cell; (ac) illustrates the definition of voxels in a visualized form.
Applsci 14 00862 g010
Figure 11. Structure of the FCAF3D Network.
Figure 11. Structure of the FCAF3D Network.
Applsci 14 00862 g011
Figure 12. (a) Result with a Multi-Cell size of 1 cm; (b) Result with a Multi-Cell size of 1.5 cm; (c) Result with a Multi-Cell size of 2 cm; (d) Result with 4 blocks; (e) Result with 5 blocks.
Figure 12. (a) Result with a Multi-Cell size of 1 cm; (b) Result with a Multi-Cell size of 1.5 cm; (c) Result with a Multi-Cell size of 2 cm; (d) Result with 4 blocks; (e) Result with 5 blocks.
Applsci 14 00862 g012
Figure 13. Raw Point Cloud Data.
Figure 13. Raw Point Cloud Data.
Applsci 14 00862 g013
Figure 14. Training Point Clouds Data (Labeling Data).
Figure 14. Training Point Clouds Data (Labeling Data).
Applsci 14 00862 g014
Figure 15. (a) Classification example of Walls; (b) Classification example of Floors; (c) Classification of chairs; (d) Classification of desks.
Figure 15. (a) Classification example of Walls; (b) Classification example of Floors; (c) Classification of chairs; (d) Classification of desks.
Applsci 14 00862 g015
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, T.; Gu, H.; Hong, S.; Choo, S. Development of Architectural Object Automatic Classification Technology for Point Cloud-Based Remodeling of Aging Buildings. Appl. Sci. 2024, 14, 862. https://doi.org/10.3390/app14020862

AMA Style

Kim T, Gu H, Hong S, Choo S. Development of Architectural Object Automatic Classification Technology for Point Cloud-Based Remodeling of Aging Buildings. Applied Sciences. 2024; 14(2):862. https://doi.org/10.3390/app14020862

Chicago/Turabian Style

Kim, Taehoon, Hyeongmo Gu, Soonmin Hong, and Seungyeon Choo. 2024. "Development of Architectural Object Automatic Classification Technology for Point Cloud-Based Remodeling of Aging Buildings" Applied Sciences 14, no. 2: 862. https://doi.org/10.3390/app14020862

APA Style

Kim, T., Gu, H., Hong, S., & Choo, S. (2024). Development of Architectural Object Automatic Classification Technology for Point Cloud-Based Remodeling of Aging Buildings. Applied Sciences, 14(2), 862. https://doi.org/10.3390/app14020862

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop