Conference PaperPDF Available

Automated Wall Detection in 2D CAD Drawings to Create Digital 3D Models

Authors:

Abstract and Figures

Generating digital 3D buildings models from scratch is time-consuming and labor-intensive. In this paper, we present an automated detection process leveraging computer vision and the information available in 2D drawings to reduce 3D modeling time. The recognition system is limited to walls and has two parts: (1) Image classification on walls by ResNet-50 model, (2) Object Detection on walls by YOLOv3 model. The system accepts new 2D drawings and outputs parameters of recognized walls. The parameters are input into Dynamo for 3D model reconstruction. We anticipate these types of systems, which rely on 2D drawings as recognition priors, will be pivotal to the industry’s transition from 2D to 3D information modalities.
Content may be subject to copyright.
39th International Symposium on Automation and Robotics in Construction (ISARC 2022)
Automated Wall Detection in 2D CAD Drawings
to Create Digital 3D Models
Chialing Weia, Mohit Guptaaand Thomas Czerniawskia
aSchool of Sustainable Engineering and the Built Environment, Arizona State University, USA
E-mail: cwei32@asu.edu,mgupta70@asu.edu,Thomas.Czerniawski@asu.edu
Abstract –
Generating digital 3D buildings models from
scratch is time consuming and labor intensive. In
this paper, we present an automated detection
process leveraging computer vision and the
information available in 2D drawings to reduce 3D
modeling time. The recognition system is limited to
walls and has two parts: (1) Image classification on
walls by ResNet-50 model, (2) Object Detection on
walls by YOLOv3 model. The system accepts new 2D
drawings and outputs parameters of recognized
walls. The parameters are input into Dynamo for 3D
model reconstruction. We anticipate these types of
systems, which rely on 2D drawings as recognition
priors, will be pivotal to the industry’s transition
from 2D to 3D information modalities.
Keywords –
Deep Learning; Image Classification; Object
Detection; 2D Drawings; 3D Building Models
1 Introduction
Building Information Modeling (BIM) plays an
important role during the entire life cycle of a building.
First, Cloud hosted BIMs enhance communication by
ensuring all views into the model are synchronized.
Secondly, achieving better visualization allows clients
to have a good understanding at each stage reducing the
possibility of design changes in the future. Thirdly,
performing clash detection tasks on BIM models can be
cost effective and improve safety. Lastly, environmental
analysis and simulation on BIM models accomplish
sustainable and AI-based architectural design [1].
For projects during preconstruction phase, it is quite
beneficial if 3D modeling process is efficient since it
helps push construction start date forward by enhancing
the collaboration among entities early on. Other projects
are 3D reconstruction for as-built buildings which can
assist in reconstructing historical buildings and better
visualization through digital representation. In detail,
doing structural analysis on 3D models of historical
buildings can reach better maintenance. Moreover,
taking this paper application as an example,
constructing campus 3D representation can let online
students have experience of looking around campus
through virtual tours and even have interaction, such as
reading posters on bulletin boards. To speed up the 3D
modeling process with high accuracy, computer vision
is a promising approach.
Computer vision is a prevailing implementation of
artificial intelligence. It can achieve pattern recognition,
object detection, image classification, and instance
segmentation tasks on images, such as 2D drawings.
This paper is experimenting with computer vision
techniques in the automatic 3D model generation
process to solve time-consuming manual modeling
conditions. The scale and distribution of the dataset,
quality of labeling and model choosing would all affect
the model performance.
In our study, we proposed an object detection model
to localize the vertical and horizontal walls on 2D
drawings by computer vision technique. Some
researchers were using semantic segmentation models
to achieve this process [2, 3, 4, 8, 10], however, to make
the process more efficient, we chose an object detection
model which can reduce much labeling time and still
achieve good performance. After we expand this work
to multi-classes detection on 2D drawings, the
performance metrics would include a confusion matrix
and mean Average Precision (mAP).
2 Related Work
We focus on reviewing related work on how deep
learning assists the process of 3D model reconstruction
from 2D drawings, the benefits and applications of later
usage.
2.1 Deep Learning Approach on 2D Drawings
There are lots of researchers doing recognition and
segmentation of components on 2D drawings by deep
learning approaches. Xiao et al. [2] cropped the original
39th International Symposium on Automation and Robotics in Construction (ISARC 2022)
2D drawings into smaller image dimensions for feeding
into a neural network model. They manually did
pixel-labeling on 300 2D drawings and implemented a
transfer learning from ResNet-152. This model is
pretrained on the ImageNet dataset and then executed
recognition and localization on five architectural
components: wall, window, door, column, stairs. Liu et
al. [3] transformed rasterized images to vector-graphics
representation since vector images can make further 3D
reconstruction models having better visualization, being
easier to manipulate and do computational analysis.
They first did deep representation learning through
Convolutional Neural Network (CNN) to convert a
raster image to a set of junctions and pixel-wise
semantics and then assembled junctions to lines and
boxes by setting constraints through integer
programming with a straightforward post-processing to
achieve vectorization. Dodge et al. [4] utilized Fully
convolutional networks (FCN) to do wall segmentation
after trying different pixel strides. They conducted
Faster R-CNN and Optical Character Recognition
(OCR) to estimate room sizes. Mishra et al. [5] did
furniture and architectural components detection on
floorplans by Cascade Mask R-CNN network with CNN
and Deformable Convolutional Networks (DCN)
separately.
The recognition and segmentation tasks are not
limited to architectural floor plans and can be more
generalized to different types of buildings. Zhao et al.
[6] detected structural components and grid reference by
YOLO model. The structural elements include columns,
horizontal beams, vertical beams and sloped beams.
Kalervo et al. [7] presents a large-scale floor plan
dataset, CubiCasa5K, which is carefully annotated by
applying the Quality Assessment process and including
5000 samples with over 80 object classes.
In this paper, we adopt ResNet-50 and YOLOv3 in
our methodology for wall recognition and segmentation
tasks since both algorithms have their prevalence in
pretrained models and can be easily deployed by
industries’ practitioners. Moreover, implementing the
ResNet-50 model for classification tasks can be trained
easily without increasing the training error with a large
number of layers and this algorithm could help vanish
gradient problems during the backpropagation process
[8]. The YOLOv3 model for object detection has high
speed and comprehends generalized object
representation [9].
2.2 3D Model Generation and Application
Kalervo et al. [7] mentioned that the application of
3D models includes 3D real estate virtual tours and
AR/VR technology. Jang et al. [8] created CityGML
(City Geography Markup Language) and IndoorGML
(Indoor Geography Markup Language) 3D data models
by extracting indoor spatial information from floorplan
images. These 3D models support a structure for 3D
geospatial data integration, storage and exchange.
Kippers et al. [9] claimed that 3D models can support
better decision making, data analysis and scenario
simulation. They integrated CityJSON and floor plan
images to reconstruct the 3D model. Some researchers
indicated that CityJSON is easier to use than CityGML.
Seo et al. [10] stated the application of the 3D model
reconstruction process at each stage. Architectural
component recognition can contribute to evacuation
paths generation and evacuation distances calculation.
Analysis of building energy ratings with window area
ratios could be executed by automatically calculating
the window and wall areas. Moreover, Using Generative
Adversarial Networks (GAN) can generate much more
virtual drawing images to integrate into AI-based
architectural design in the future.
For indoor furniture fitting, Dodge et al. [4]
performed Optical Character Recognition (OCR) to
extract text information from 2D drawings which can
measure each room size and compute the pixel density
for fitting interior components.
3 Methodology
This section presents data generation process,
training, inference pipelines and 3D model
reconstruction process in Revit.
3.1 Data Creation
In this study, Arizona State University Tempe
campus 2D Drawings were assembled by the University
Facilities Management department. We utilized 29
sheets of size 3400 x 2200 pixels as our original dataset.
We then used the LabelMe annotation tool [11] to
manually label walls’ location. We drew a rectangle as a
bounding box surrounding each wall so that we can get
the location of each wall which includes the coordinates
of the upper-left point, width and height of each
bounding box.
To feed into the CNN architecture, we randomly
sampled each sheet with 100 crops of 256 x 256 pixels
images, and computed the labeled information to the
relative position for the small crop images. In the
meantime, we programmed the small cropped images to
two folders depending on whether or not the image
contained the wall for further training.
Ultimately, we assembled all labeled information
into a csv file and got 1250 non-wall images and 1650
wall-contained images individually.
39th International Symposium on Automation and Robotics in Construction (ISARC 2022)
3.2 Model Training
Figure 1. The training pipeline from labeling to Models’
Generation
3.2.1 Wall Classification
The first step of doing model training is determining
programming languages, deep learning libraries,
programming environment and referred neural network
models. We used Python with Tensorflow and Keras
libraries on Google Colaboratory which is an online
cloud-based Jupyter notebook environment. We also
referred to the YOLOv3 model architecture from
PyLessons [12] website which posts deep learning
applications and tutorials for customized data. We chose
their architecture and did further modification since the
descriptions and explanations are clear, code is precise
and the results they have shown are promising.
To classify wall objects, we utilized the ResNet-50
model to train a binary classification task. The training
dataset was collected during the preprocessing step and
a 70/30 ratio to split training and test dataset was
determined. We utilized all training set as our input for
the training pipeline shown in Figure 1, and test set for
inference pipeline shown in Figure 3, separately.
We first programmed the ImageNet pretrained
model for transfer learning. The setting of the pretrained
model includes average pooling, 256 x 256 x 3 pixels
input shape, 2 classes classification, and all layers inside
the pretrained model are not trainable, which could
reduce time complexity during the training process.
We flattened the output of the pretrained model into
one dimension and added a fully connected layer with
512 nodes and a ReLU activation function. Finally,
using a sigmoid function to determine the probability of
wall existence.
During the training process, we set a 32 batch size,
100 epochs, and Adam optimizer with a 0.001 learning
rate. The result shows that the training and test accuracy
converged to 98% and 94%, respectively, as shown in
Figure 2.
Figure 2. ResNet-50 classification training and test
accuracy versus number of Epochs
3.2.2 Wall Detection
This section describes the wall detection task using
the wall-contained images as our training set and
YOLOv3 model for training. For this supervised
learning training, we parsed the images and loaded the
labeled information csv file with image filename and
corresponding bounding box of wall information. We
used an 80/20 ratio to split the training and test dataset,
set the batch size as 4, Adam optimizer and 30 epochs
in total. For learning rate, we used two warmup epochs
first and then decayed linearly from 1e-4 to 1e-6.
39th International Symposium on Automation and Robotics in Construction (ISARC 2022)
3.3 Model Inference
Figure 3. The inference pipeline for 3D wall
reconstruction in Revit
To do model inference on new 2D drawings, we first
cropped each drawing into 256 x 256 pixels size in a
sliding windows sequence since we need to match the
input size of deep learning algorithms that we set during
the training process.
We feed each crop into our ResNet-50 model to
classify if the crop contains walls or not. The inferred
non-wall images stay the same since there are no walls
to detect the location, however, we parsed the inferred
wall images into our YOLOv3 model to infer bounding
box information and probability of walls. The inferred
rectangle bounding box surrounding walls and
probability would be shown on each crop. Eventually,
we can combine each crop to the original sheet size in a
sequence after the inference process. The overall
inference process is shown as Figure 3.
3.4 3D Model Reconstruction in Revit
After the model inference process described in
section 3.3, we can get the prediction value of the wall
location which represents the bounding boxes
information. We extract and calculate the end points of
each predicted wall location by their predicted bounding
boxes. These parameters are saved in a csv file with
height information on the elevation plan and then
imported to Dynamo BIM. In the Dynamo environment,
we inject the csv file into wall generation nodes and
connect to the wall family type. Ultimately, we
reconstruct the walls in Revit through Dynamo with our
csv file and the family type we specify. The overall
procedure is shown in Figure 4.
39th International Symposium on Automation and Robotics in Construction (ISARC 2022)
4 Results and Discussion
The results are shown in Figure 5 which are
zooming a portion of the new 2D drawing. The red
rectangles indicate the bounding box and the number on
the upper-left corner above the rectangles is the neural
network’s confidence probability of the wall.
Figure 4. The procedure of creating a 3D model from extraction wall location through Dynamo and CSV file
39th International Symposium on Automation and Robotics in Construction (ISARC 2022)
Figure 5. Original small portion of floorplan
(upper-left), bounding boxes (upper-right) and floorplan
with bounding boxes (lower)
We raised three failure cases on our inferred
drawings as shown in Figure 6. First, the rectangle
surrounding “MATCH LINE” on the drawing causes the
CNN model to possibly be misinterpreted as a wall
appearance. Secondly, we discovered that “curved
walls” and “diagonal walls” are not suitable for object
detection tasks during manual labeling process since the
bounding boxes are axis aligned which could not tightly
delineate the curved and diagonal lines. Therefore, we
did not label the curved and diagonal walls during the
training process so that our model could not recognize
them on the inferred dataset.
Future work will include, first, we can remove
legend, title and match line symbol before doing the
inference process so that we can avoid detecting the
rectangle surrounding the match line symbol as walls.
Secondly, we will make a list of all failed inference
cases and explore the potential improvement. Fourthly,
we would like to explore different methods to do
“diagonal walls'' and “curved walls'' detection since the
original labeled method would easily make a bounding
box of walls covering other objects. The most promising
method we would attempt first is an instance
segmentation algorithm. Lastly, expanding the
automation process to more architectural components,
doors and windows, would be a critical contribution for
this research and our strategy would be doing multiclass
classification and detection, creating precise interface
connection among all components.
5 Conclusion
BIM has become an essential methodology in the
AEC industry in the past few decades. To facilitate BIM
applications, we hope that we could make the BIM
model generation effective and precise.
In the paper, we demonstrated that using computer
vision techniques to detect the location of walls on the
2D drawings so that we could automatically generate
3D models later on. We achieve good results on the
inferred dataset and plan to expand the scale to other
components inference and experimental instance
segmentation for non-straight walls.
Deep learning related research is very active and
advanced, lots of machine learning models are being
optimized and well-performed on many projects.
Therefore, using those techniques on 2D drawings to 3D
model generation could be of practical use in the AEC
industry in the near future. These automated systems
could reduce budget, improve safety, enhance
communication among entities and promote more
advanced implementation for researchers and expertise.
Figure 6. Failure cases: “MATCH LINE” (left), “diagonal wall” (middle) and “curved wall” (right)
39th International Symposium on Automation and Robotics in Construction (ISARC 2022)
References
[1] Michael T. Benefits of Building
Information Modeling. On-line:
https://www.ny-engineers.com/blog/bene
fits-of-building-information-modeling,
Accessed: 24/01/2022
[2] Xiao Y., Chen S., Ikeda Y. and Hotta K.
Automatic recognition and segmentation
of architectural elements from 2D
drawings by convolutional neural
network. In Proceedings of the 25th
International Conference of the
Association for Computer-Aided
Architectural Design Research in Asia
(CAADRIA), pages 843–852, Hong
Kong, 2020.
[3] Liu C., Wu J., Kohli P. and Furukawa Y.
Raster-to-vector: Revisiting floorplan
transformation. In Proceedings of the
IEEE International Conference on
Computer Vision, pages 2195–2203,
Venice, Italy, 2017.
[4] Dodge S., Xu J. and Stenger B. Parsing
floor plan images. In Proceedings of the
15th IAPR International Conference on
Machine Vision Applications (MVA),
pages 358–361, Nagoya, Japan, 2017.
[5] Mishra S., Hashmi K., Pagani A.,
Liwicki M., Stricker D and Afzal M.
Towards robust object detection in floor
plan images: A data augmentation
approach. Applied Sciences.
11(23):11174, 2021.
[6] Zhao Y., Deng X. and Lai H. A Deep
Learning-Based Method to Detect
Components from Scanned Structural
Drawings for Reconstructing 3D Models.
Applied Sciences. 10(6):2066, 2020.
[7] Kalervo A., Ylioinas J., Häikiö M.,
Karhu A. and Kannala J. CubiCasa5k: A
dataset and an improved multi-task
model for floorplan image analysis. In
Proceedings of the Scandinavian
Conference on Image Analysis, pages
28–40, 2019.
[8] Jang H., Yu K. and Yang J. Indoor
Reconstruction from Floorplan images
with a Deep Learning Approach. ISPRS
International Journal of
Geo-Information. 9(2), 2020.
[9] Kippers R., Koeva M., Van Keulen M.
and Oude Elberink S. Automatic 3d
Building Model Generation Using Deep
Learning Methods Based on Cityjson
and 2d Floor Plans. In Proceedings of
International Archives of the
Photogrammetry, Remote Sensing and
Spatial Information Sciences, pages
49–54, 2021.
[10] Seo J., Park H. and Choo S. Inference of
drawing elements and space usage on
architectural drawings using semantic
segmentation. Applied Sciences.
10(20):7347, 2020.
[11] MIT, Computer Science and Artificial
Intelligence Laboratory. LabelMe.
On-line:
http://labelme.csail.mit.edu/Release3.0/,
Accessed: 04/11/2021.
[12] Rokas B. PyLessons. On-line:
https://pylessons.com/, Accessed:
22/11/2021.
... Learning-based approaches: Learning-based approaches have been gaining popularity in the field of entity recognition in scanned drawings and consist of the use of deep learning for training a network to identify building components in technical drawings. Different types of networks have been used throughout the literature, including Graph Neural Networks (GNN) [18,39], Generative Adversarial Networks (GAN) [39,55], Convolutional Neural Networks (CNN) [56][57][58], Global Convolutional Networks (GCN) [59], Fully Convolutional Networks (FCN) [60], Faster Region-based Convolutional Neural Networks (Faster R-CNN) [25], Cascade Mask R-CNN [61,62] and ResNet-50 [63][64][65]. These networks rely on datasets containing large quantities of floor plans to train the network to produce reliable results. ...
... Semantic richness of the reviewed literature[7,[14][15][16][17][18][19][20][21][23][24][25]28,37,[44][45][46][47][48][49][50][51][52]54,[61][62][63][64][65][70][71][72][73]. ...
Article
Full-text available
Among the methods of 3D reconstruction, the automatic generation of 3D models from building documentation is one of the most accessible and inexpensive. For 30 years, researchers have proposed multiple methods to automatically generate 3D models from 2D drawings. This study compiles this research and discusses the different methods used to generate 3D models from 2D drawings. It offers a critical review of these methods, focusing on the coverage and completeness of the reconstruction process. This review allows us to identify the research gaps in the literature, and opportunities for improvement are identified for future research.
... Instead of commercial 3D modeling software, such as Revit 2022, s to use open-source modeling software, such as Blender 3.6, due to its rel and flexibility. To explore the capabilities of interoperability for Blende tection output from our previous work is utilized [24] in an attempt to model. The neural network output is aggregated in a CSV file, as shown the coordinate system transformation and merging of adjacent wall insta are conducted. ...
... Instead of commercial 3D modeling software, such as Revit 2022, some users prefer to use open-source modeling software, such as Blender 3.6, due to its relative accessibility and flexibility. To explore the capabilities of interoperability for Blender 3.6, the wall detection output from our previous work is utilized [24] in an attempt to create a 3D wall model. The neural network output is aggregated in a CSV file, as shown in Figure 41, and the coordinate system transformation and merging of adjacent wall instances post-process are conducted. ...
Article
Full-text available
Building owners are working on converting their legacy documentation 2D floor plans into digital 3D representations, but the manual process is labor-intensive and time-consuming. In this paper, deep learning is leveraged to automate the process. This automation requires interoperability between artificial neural networks and prevailing 3D modeling software. The system processes 2D floor plans and outputs parameters of recognized walls, single doors, double doors, and columns. The parameters include the start point and end point of the wall and the center point of the door and column. These parameters are input into Revit 2022 through the Revit API 2022 after post-processing. The dimensional parameter integration affordances of object detection and instance segmentation are studied and compared using Faster R-CNN and Mask R-CNN models. Instance segmentation was found to require more time for data labeling but was more capable of informing the modeling of irregularly shaped objects. The mean Average Precision (mAP) of object detection and instance segmentation are 71.7% and 69.3%, respectively. Apart from single doors, the average precision for other categories falls within the range of 74% to 96%. The results provide software developers with guidance on choosing between object detection and instance segmentation strategies for processing legacy building documents. These types of systems are anticipated to be pivotal to the industry’s transition from 2D to 3D information modalities and advise practitioners to carefully choose suitable models and consider the recommendations provided in this study to mitigate potential failure cases.
... Fan et al. (2021) collected 15,000 large-scale 2D CAD drawings of architectural floor plans, aiming to detect walls and symbols and annotated 30 different objects from various datasets and used different detection approaches. Similarly, Wei et al. (2022) employed a method to detect walls in 2D floor plans, using 29 sheets of specific size and training a model with 1,250 non-wall and 1,650 wall-contained images. However, curved and inclined walls were not well detected. ...
Conference Paper
Full-text available
This research presents a framework for the creation of georeferenced three-dimensional as-is models of railway signals from 2D railway signalling plans. Signal plans are stored as image-based information, where analysis and transformation into 3D require object-oriented data representation. Therefore, extraction into structured information is essential. Machine Learning techniques have been applied to recognize different types of signals within the plans and to extract the affiliated data to them, such as signal type and geographical position. With the help of those information, dummy meshes, including the extracted alphanumerical data, are placed at the global geolocation and expressed within the open format Industry Foundation Classes (IFC). The result can serve as a digitised planning base for future and/or ongoing railway projects.
... Overall, a larger number of publications exist on the reconstruction of high-rise building models from drawings. Wei et al. (2022) proposes a pipeline for detecting and reconstructing walls from floor plans. Firstly, the drawing is divided into patches, and the ResNet model is employed to identify patches containing walls. ...
Conference Paper
Full-text available
Digital methods such as Building Information Modeling (BIM) can be leveraged, to improve the efficiency of maintenance planning of bridges. However, this requires digital building models, which are rarely available. Consequently, these models must be created retrospectively, which is time-consuming when done manually. Naturally, there is a great interest in the industry to automate the process of retro-digitization. This paper contributes to these efforts by proposing a multistage pipeline to automatically extract the gradient of a bridge from pixel-based construction drawings using deep learning. The bridge gradient, a key element of the structure’s axis, is critical for describing the elevation profile and axis slope. This information is implicitly contained in the longitudinal view of bridge drawings as gradient symbols. To extract this information, the wellestablished object detection model YOLOv5 is employed to locate the gradient symbols inside the drawings. Subsequently, EasyOCR and heuristic rules are applied to extract the relevant gradient parameters associated with each detected symbol. The extracted parameters are then exported in a machine-interpretable format to facilitate seamless integration into other applications. The results show a promising 98% accuracy in symbol detection and an overall accuracy of 70%. Consequently, the pipeline represents a significant advance in automating the retro-digitization process for existing bridges by reducing the time and effort required.
... We divided 75 indoor scan data into 60 training set, 8 validation set and 7 test set. As our previous work (Wei et al. 2022), each projected data is rescaled to 2200 x 3400 pixels and we performed 100 times random cropping on each one. The size of small crops is 800 x 800 pixels for feeding into the neural network model as shown in Figure 6. ...
Conference Paper
Full-text available
We present an automated scan-to-BIM pipeline that simplifies the 3D building object recognition problem into a 2D recognition problem. We used the Habitat Matterport 3D Dataset (HM3D) for training wall detection model. The weakly supervised learning is conducted since we used the noisy depth-projected annotation. We isolated individual building levels and projected the points to 2D along the Z-axis (up/down). The architectural components recognition system detects walls within the plan view projection of the indoor point cloud. We compare the performance metric of validation on noisy annotation with human-labeled annotation and analyze the wall inference results from visualization. We assume the human-labeled annotation as ground truth and noisy annotation is prediction to calculate the average precision. The average precision values are compared to the neural network performance. We anticipate this experiment can provide a feasible weak supervision method for simplifying 3D digital model creation from scan data.
Article
Full-text available
Object detection is one of the most critical tasks in the field of Computer vision. This task comprises identifying and localizing an object in the image. Architectural floor plans represent the layout of buildings and apartments. The floor plans consist of walls, windows, stairs, and other furniture objects. While recognizing floor plan objects is straightforward for humans, automatically processing floor plans and recognizing objects is challenging. In this work, we investigate the performance of the recently introduced Cascade Mask R-CNN network to solve object detection in floor plan images. Furthermore, we experimentally establish that deformable convolution works better than conventional convolutions in the proposed framework. Prior datasets for object detection in floor plan images are either publicly unavailable or contain few samples. We introduce SFPI, a novel synthetic floor plan dataset consisting of 10,000 images to address this issue. Our proposed method conveniently exceeds the previous state-of-the-art results on the SESYD dataset with an mAP of 98.1%. Moreover, it sets impressive baseline results on our novel SFPI dataset with an mAP of 99.8%. We believe that introducing the modern dataset enables the researcher to enhance the research in this domain.
Article
Full-text available
Artificial intelligence presents an optimized alternative by performing problem-solving knowledge and problem-solving processes under specific conditions. This makes it possible to creatively examine various design alternatives under conditions that satisfy the functional requirements of the building. In this study, in order to develop architectural design automation technology using artificial intelligence, the characteristics of an architectural drawings, that is, the architectural elements and the composition of spaces expressed in the drawings, were learned, recognized, and inferred through deep learning. The biggest problem in applying deep learning in the field of architectural design is that the amount of publicly disclosed data is absolutely insufficient and that the publicly disclosed data also haves a wide variety of forms. Using the technology proposed in this study, it is possible to quickly and easily create labeling images of drawings, so it is expected that a large amount of data sets that can be used for deep learning for the automatic recommendation of architectural design or automatic 3D modeling can be obtained. This will be the basis for architectural design technology using artificial intelligence in the future, as it can propose an architectural plan that meets specific circumstances or requirements.
Article
Full-text available
Among various building information model (BIM) reconstruction methods for existing building, image-based method can identify building components from scanned as-built drawings and has won great attention due to its lower cost, less professional operators and better reconstruction performance. However, this kind of method will cost a great deal of time to design and extract features. Moreover, the manually extracted features have poor robustness and contain less non-geometric information. In order to solve this problem, this paper proposes a deep learning-based method to detect building components from scanned 2D drawings. Taking structural drawings as an example, in this article, 1500 images of structural drawings were firstly collected and preprocessed to guarantee the quality of data. After that, the neural network model—You Only Look Once (YOLO) was trained, verified and tested. In addition, a series of metrics were utilized to evaluate the performance of recognition. The results of test experiments show that the components in structural drawings (e.g., grid reference, column and beam) can be successfully detected, while the average detection accuracy of the whole image is over 80% and the average detection time for each image is 0.71 s. The experimental results demonstrate that the proposed method is robust and timesaving, which provides a good basis for the reconstruction of BIM from 2D drawings.
Article
Full-text available
Although interest in indoor space modeling is increasing, the quantity of indoor spatial data available is currently very scarce compared to its demand. Many studies have been carried out to acquire indoor spatial information from floorplan images because they are relatively cheap and easy to access. However, existing studies do not take international standards and usability into consideration, they consider only 2D geometry. This study aims to generate basic data that can be converted to indoor spatial information using IndoorGML (Indoor Geography Markup Language) thick wall model or the CityGML (City Geography Markup Language) level of detail 2 by creating vector-formed data while preserving wall thickness. To achieve this, recent Convolutional Neural Networks are used on floorplan images to detect wall and door pixels. Additionally, centerline and corner detection algorithms were applied to convert wall and door images into vector data. In this manner, we obtained high-quality raster segmentation results and reliable vector data with node-edge structure and thickness attributes that enabled the structures of vertical and horizontal wall segments and diagonal walls to be determined with precision. Some of the vector results were converted into CityGML and IndoorGML form and visualized, demonstrating the validity of our work.
Conference Paper
Full-text available
Article
Full-text available
Building Information Modeling (BIM) has recently attained widespread attention in the Architectural, Engineering and Construction (AEC) industry. BIM represents the development and use of computer-generated n-dimensional (n-D) models to simulate the planning, design, construction and operation of a facility. It helps architects, engineers and constructors to visualize what is to be built in simulated environment and to identify potential design, construction or operational problems. In this paper, the benefits and possible risks of BIM and future challenges for the construction industry are discussed. First presented is the main concept of BIM with its advantages and possible applications in construction. Then the role of BIM in the construction industry and academia is discussed based on the results of three recent questionnaire surveys. After that, a case study of Hilton Aquarium project in Atlanta is presented to quantitatively illustrate the cost and time savings realized by developing and using a building information model. It is followed by data from 10 construction projects to determine the net BIM savings and BIM return on investment. At the end, BIM risks and future challenges for the construction industry are discussed.
CubiCasa5k: A dataset and an improved multi-task model for floorplan image analysis
  • A Kalervo
  • J Ylioinas
  • M Häikiö
  • A Karhu
  • J Kannala
Kalervo A., Ylioinas J., Häikiö M., Karhu A. and Kannala J. CubiCasa5k: A dataset and an improved multi-task model for floorplan image analysis. In Proceedings of the Scandinavian Conference on Image Analysis, pages 28-40, 2019.
Automatic 3d Building Model Generation Using Deep Learning Methods Based on Cityjson and 2d Floor Plans
  • R Kippers
  • M Koeva
  • M Van Keulen
  • Oude Elberink
Kippers R., Koeva M., Van Keulen M. and Oude Elberink S. Automatic 3d Building Model Generation Using Deep Learning Methods Based on Cityjson and 2d Floor Plans. In Proceedings of International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, pages 49-54, 2021.