Computer Vision-Bases Construction Progress Monitoring
Computer Vision-Bases Construction Progress Monitoring
Automation in Construction
journal homepage: www.elsevier.com/locate/autcon
Review
A R T I C L E I N F O A B S T R A C T
Keywords: Automating the process of construction progress monitoring through computer vision can enable effective
Progress monitoring control of projects. Systematic classification of available methods and technologies is necessary to structure this
Computer vision complex, multi-stage process. Using the PRISMA framework, relevant studies in the area were identified. The
Automated construction
various concepts, tools, technologies, and algorithms reported by these studies were iteratively categorised,
Data acquisition
developing an integrated process framework for Computer-Vision-Based Construction Progress Monitoring (CV-
3D reconstruction
As-built modelling CPM). This framework comprises: data acquisition and 3D-reconstruction, as-built modelling, and progress
Point cloud assessment. Each stage is discussed in detail, positioning key studies, and concurrently comparing the methods
Scan to BIM used therein. The four levels of progress monitoring are defined and found to strongly influence all stages of the
Literature review framework. The need for benchmarking CV-CPM pipelines and components are discussed, and potential research
Digital Twin questions within each stage are identified. The relevance of CV-CPM to support emerging areas such as Digital
Twin is also discussed.
1. Introduction computing power required for processing. However, with the increasing
ubiquity of both devices and high-performance computing, it has
Monitoring the progress of construction projects is crucial, as it become feasible to implement vision-based technologies in applications
provides vital inputs for managers to make timely and informed de that automate construction processes. Computer vision enables com
cisions. Improper progress monitoring leads to a loss of control of the puters to derive numeric information from digital images, videos, depth
project, resulting in time and cost overruns. Traditional progress- images and 3D point clouds, process the information, and take action.
monitoring methods require manual data entry, which proves to be Sub-domains of computer vision relevant to this paper include scene
tedious, time-consuming, and prone to human error [1]. Therefore, its reconstruction, 3D pose estimation, motion estimation, object detection,
effectiveness can be improved through automation. object recognition and labelling, learning, and 3D scene modelling.
Technologies investigated for their applicability in automated Vision-based technologies have been studied on a wide range of
progress monitoring in construction include RFID, barcodes and QR construction applications, such as workforce tracking, resource tracking,
codes, laser scanning, photogrammetry, videogrammetry, range imag condition assessment, quality inspections, safety management, and
ing (RGB–D), web-based CCTV, and structural sensing [2]. Among automated layout generation [3]. Several studies on vision-based con
these, the impact of vision-based technologies was initially limited due struction progress monitoring have also been reported. However, the
to the sophistication of the data-acquisition devices and the high processes for effective progress monitoring are diverse, intricate, and
Abbreviations: AEC, Architecture, Engineering and Construction; AI, Artificial Intelligence; AR, Augmented Reality; BIM, Building Information Modelling; BoQ,
Bill of Quantities; CAD, Computer-Aided Design; CCTV, Closed-Circuit Television; CMVS, Clustering Multi-View Stereo; CNN, Convolutional Neural Network; CV-
CPM, Computer Vision-based Construction Progress Monitoring; DT, Digital Twin; DL, Deep Learning; GPS, Global Positioning System; GPU, Graphical Processing
Unit; GUI, Graphical User Interface; HSI, Hue-Saturation-Intensity; ICP, Iterative Closest Point; IFC, Industry Foundation Class; IMU, Inertial Measurement Unit; IoT,
Internet of Things; LiDAR, Light Detection and Ranging; LoD, Level of Detail; LPM, Level of Progress Monitoring; MEP, Mechanical, Electrical, and Plumbing; ML,
Machine Learning; MR, Mixed Reality; MVS, Multi-View Stereo; PCA, Principal Component Analysis; PMVS, Patch-based Multi-View Stereo; PSR, Poisson’s Surface
Reconstruction; QR, Quick Response; RANSAC, Random Sample Consensus; RFID, Radio Frequency Identification; RGBD, Red Green Blue Depth; SfM, Structure-
from-Motion; SGM, Semi-Global Matching; SLAM, Simultaneous Localisation and Mapping; SMS, Short Message Service; TLS, Terrestrial Laser Scanning; UAV,
Unmanned Aerial Vehicle; UGV, Unmanned Ground Vehicle; VR, Virtual Reality; XR, Extended Reality.
* Corresponding author at: BTCM Division, Department of Civil Engineering, IIT Madras, India.
E-mail addresses: [email protected] (V.K. Reja), [email protected] (K. Varghese), [email protected] (Q.P. Ha).
https://doi.org/10.1016/j.autcon.2022.104245
Received 8 November 2021; Received in revised form 1 March 2022; Accepted 31 March 2022
Available online 8 April 2022
0926-5805/© 2022 Elsevier B.V. All rights reserved.
V.K. Reja et al. Automation in Construction 138 (2022) 104245
complex. In addition, there are many concepts and technologies that can The organisation of this paper is as follows. Section 2 characterises
be applied to automate each stage of the process. An integrated frame and classifies the key studies in this area, based on the three macro-
work to characterise and categorise the current and future studies in this stages of the CV-CPM process. Section 3 expands the three macro-
area will enable systematic investigation and documentation in this stages into a detailed process framework, and for each process in the
domain. pipeline, the relevant concepts, technologies, and tools used are
Review papers in this domain have focused on specific aspects of the compared and discussed. Section 4 discusses a strategy for a systematic
progress monitoring process. Omar et al. have presented a review of data approach in this domain and identifies the areas for further work. The
acquisition technologies [2], while Ma et al. [4] have comprehensively conclusions of the study are presented in Section 5.
studied 3D reconstruction techniques. Ekanayake et al. [5] and Patel
et al. [6] have recently presented a bibliometric analysis of the literature 2. Overview of vision-based monitoring of construction progress
and highlighted specific challenges to be addressed. However, there are
several stages in the process, and a broad study of both literature and This section presents the macro-level stages proposed for vision-
practice has resulted in identifying the following key stages: (i) data based progress monitoring in construction and references the key
acquisition and 3D reconstruction, (ii) generation of as-built models, progress monitoring processes within these stages. Each of these pro
and (iii) monitoring progress. Currently, there is no integrated frame cesses is also referred to as a pipeline. Fig. 1 shows the six macro-level
work that addresses the details of the processes of Computer Vision- processes that constitute vision-based progress monitoring.
Based Construction Progress Monitoring (CV-CPM), from data acquisi These processes are grouped into three stages, as follows:
tion to progress estimation. Therefore, to address this gap, the following
objectives are identified: i. Acquisition of as-built data and conversion into a point cloud
using 3D reconstruction techniques.
1. Develop an integrated framework that captures the process re ii. Generation of the as-built model from the point cloud (with
quirements of vision-based construction progress monitoring and reference to the as-planned model, if available)
enables the characterisation and categorisation of current and future iii. Comparison of the as-built and as-planned models to assess
work in the area. progress by visualisation and/or by quantifying the work done.
2. Utilise the framework to position and compare the various concepts
and tools adopted by published research studies. All studies on vision-based progress monitoring focus on one or more
3. Identify areas and strategies for future work in the area. of these stages. Table 1, a key output of this study, was developed
through a detailed review of the published works on vision-based con
As a first step towards these objectives, relevant articles were struction progress monitoring. The articles were identified through the
selected through the broader steps defined in the PRISMA framework. systematic process described in the introduction. The categories shown
These include identification, screening, eligibility, and inclusion. An in Table 1 were refined iteratively based on objectives, solutions and
extensive keyword search in the Scopus and Web of Science (WoS) concepts addressed by the reviewed papers.
reputable scientific databases was conducted. The keywords used were The studies referred to in Table 1 are arranged based on the category
“progress monitoring in construction,” “construction progress moni of progress monitoring, as indicated in the last column. The first column
toring,” “progress monitoring,” “automated progress monitoring,” contains the reference to the study, and the subsequent columns contain
“computer vision in construction progress monitoring,” and “vision- the categories and sub-categories based on the macro-level model. The
based progress monitoring in construction.” One hundred eighty-two cells in Table 1 identify the technology, concept, and algorithm explored
articles were retrieved from the keyword search. Primary screening of in each category and sub-category. It can be observed in Table 1 that for
the articles was carried out by reading their abstracts. In order to capture each category, there is a wide range of alternatives for all process
relevant developments in this rapidly evolving area, only key articles components, construction elements, and the key foci that the studies
appearing after 2014 were defined as eligible for meta-analysis. Other have explored, as discussed briefly in the following paragraphs.
articles which address specific parts of the framework were also referred Data Acquisition and 3D Reconstruction: This section of the table
to and positioned appropriately in the manuscript to support the documents the data acquisition technologies, type of mounting, and
framework. For an article to qualify as a key article for review, the usage environment (indoor/outdoor) that the studies use. Depending on
following two criteria were set: the input type, the data undergoes 3D reconstruction using various
commercial software, as shown in Table 1. The workflow from data
• “The article should present an applied computer vision-based prog acquisition to 3D reconstruction is discussed in detail in Subsection 3.1
ress monitoring pipeline by demonstrating it with a case study or of this paper.
experimental results.” As-built Modelling: As shown in Table 1, there are varied output
• “Pipelines addressing both indoor and outdoor progress monitoring, data types that are generated by the pipelines. The generation of these
as well as pipelines ending at different stages of a type of progress as-built models requires specific as-built modelling techniques, which
estimation, were considered (i.e., only visualisation as well as visu are classified and discussed in detail in Subsection 3.2.
alisation with quantification).” Progress Monitoring: Table 1 also indicates the level of progress
monitoring (LPM) carried out, judging by the types of platforms used to
Twenty-four key articles were selected and included for meta-
analysis to formulate the framework. Among other articles referred for
specific tools and techniques, fifty-four articles are from the civil engi
neering domain, and eight are from the computer science domain.
The key articles were systematically reviewed, analysed and cat
egorised based on the process they addressed and the concepts they
used, following the methodology defined by [7]. The framework evolved
as more studies were characterised, categorised, and positioned within
it, and the application contexts of concepts and technologies used were
studied and compared. After the reviewed works were positioned in the
framework, the processes, concepts, and technologies that need to be
further developed/explored were identified. Fig. 1. Macro-level conceptualisation of vision-based progress monitoring.
2
V.K. Reja et al.
Table 1
Review of pipelines for vision-based progress monitoring.
Data Acquisition Progress Monitoring
3D As-Built Identified Operational
Papers Technology Method Indoor/ Visualisation Comparison Quantification Schedule Key Focus LPM
Reconstruction Output Elements State
Outdoor Updating
Immersive VR
Laser Scanning Extended Target (iVR)–based
and Reality XR (VR Cuboidal visualisation of
[45] Photogrammetry Manual Indoor SLAM 3D Mesh + AR) Visualisation N N Objects N progress
Building
CNN-based object
Virtual Segmented Interactive Object Elements
[13] RGBD Camera Manual Both N N Y detection and use of
photogrammetry Images VRE Detection (Walls, floor,
gaming engine
beams etc.,)
Computer vision-
based element
Image and
Segmented detection using
Commercial Point cloud-
[57] Digital Images UAV Outdoor Image with 3D Viewer N N Columns Y images, camera
Software based Object
Point Cloud poses, as-planned-
Detection
BIM, precedence
relationships
Gaming engine
utility for
Laser Scanning or Commercial Commercial
[44] Manual Both Point Cloud VRE N N Not specified N visualisation of as-
Photogrammetry Software Software
built and as-planned
data L-2
Uses Geometry and Visualisation
Photogrammetry/ Construction
[38] UAV Both SfM & MVS Point Cloud 3D Viewer Thresholding N N Y Appearance based and
LS Materials
reasoning comparison
3
MATLAB-based
MATLAB Columns, element cropping
Object in Image beams, block and quantity
[59] Videography Fixed Both NA Image Image viewer Processing Y N masonry N estimation L-3
Hybrid data Quantification
Laser Scanning Mesh collection using
Manual Overlapping Structural
[60] and UAV Both Model/3D 3D Viewer Y N N drone-based
Modelling & BoQ Components
Photogrammetry BIM photogrammetry
and Laser Scanning
V.K. Reja et al.
Data Acquisition Progress Monitoring
3D As-Built Identified Operational
Papers Technology Method Indoor/ Visualisation Comparison Quantification Schedule Key Focus LPM
Reconstruction Output Elements State
Outdoor Updating
Object detection,
segmentation and
Object multiple objects
Detected & tracking from images
Segmented Object to detect precast
[61] Video Camera Fixed Outdoor NA Images 3D Viewer Detection Y N Precast Walls N walls.
big regular-
Commercial Mixed Rays Mixed Reality–based
[12] RGBD Camera Manual Indoor 3D Mesh Y N shaped N
Software Reality Thresholding progress monitoring
objects
SfM & MVS Enhanced method for
[31] Photogrammetry Manual Both with Coded BIM 3D Viewer BoQ N N Walls N 3D reconstruction
Targets using coded targets
Extraction of columns
Overlapping
Commercial Rectangular using geometric and
[26] Laser Scanning Manual Both Point Cloud 3D Viewer Distance Y N N
Software Columns relationship-based
Threshold
reasoning
Percentage of
[48] NA NA Both NA Point Cloud 3D Viewer Thresholding Y N Walls N completion of in-situ
cast concrete walls L-3
Uses UAS images, Quantification
VSfM/ Building low-cost
[10] Videogrammetry UAS Outdoor 3D BIM 3D Viewer Thresholding Y N N
4
Photogrammetry-
based monitoring
Enclosed with automated
L-4
Commercial Volume notification system
Schedule
[47] Photogrammetry Manual Outdoor Software Point Cloud NA Calculation Y Y Columns N and schedule update
Update with
Commercial Continuous
notifications
Commercial Project monitoring using
[14] RGBD Camera Manual Both 4D BIM 3D Viewer N Y Not specified N
Software Management helmet-mounted
Software scanners
V.K. Reja et al. Automation in Construction 138 (2022) 104245
Fig. 2. Integrated framework for computer vision-based construction progress monitoring (CV-CPM).
5
V.K. Reja et al. Automation in Construction 138 (2022) 104245
implement progress estimation. This study classifies these into four A systematic review can be found in [2] on technologies used for
levels. These are: (L-1) only visualising the as-built model, (L-2) visu progress tracking in construction. The study finds technology selection a
alisation incorporating a comparison with the as-planned model, (L-3) “multi-faceted issue” which depends on “required degree of accuracy,
quantification of progress, and (L-4) quantification with schedule up project size, level of automation, and the ultimate purpose of progress
date and warning notifications. tracking”.
For progress visualisation, different immersive and non-immersive Fig. 3 shows a dual matrix illustrating the data acquisition technol
environments have been used. For quantification, numerous ap ogy and method selection criteria. The technologies and methods are
proaches have been used based on the availability of an as-planned listed on the concentric rings in their respective matrix, where colours
model. The details are discussed in Subsection 3.3. and icons are added for easier comprehension. The key factors that
In addition to the three process components—data acquisition and govern their selection have been identified and are positioned along the
3D reconstruction, as-built modelling, and progress monitoring—T outer ring. Eight key factors for the sensor mounting method include
able 1 also characterises the pipelines on two additional parameters. statutory clearance, cost of mounting equipment, range of operation,
First, the types of elements being worked on, and second, whether the preferred use case, operator training, navigation, manoeuvring speed,
structure’s current operational state can be recognised (e.g., a wall’s and accessibility. In addition, twelve key factors for data acquisition
stage can be brickwork- completed, plastered, or painted). technology have been identified. They include the level of automation
As inferred from Table 1, numerous combinations of tools, concepts, for data capture, real-time data availability, range of equipment oper
and algorithms need to be evaluated to arrive at the appropriate ation, spatial resolution, spatial accuracy, adequate lightning require
approach for a specific progress monitoring need. Though all the com ment, user training requirement, time for data capture, preparations for
binations may not be technically viable, a significantly large set of op data capture, computation cost for processing, equipment portability,
tions needs to be experimentally explored and documented to ensure and equipment cost.
systematic progress in this domain. Therefore, an integrated framework This matrix is constructed based on the review and classification of
encompassing all the processes will support the systematic study. literature. The relative comparison indicator for technology versus
In the next section, the three macro-level steps are expanded to sensor mounting method is indicated along the radial direction. The low,
formulate the detailed framework wherein the concepts/processes/ medium, and high ratings are shown using green, yellow, and red colour
technologies mentioned in these studies are characterised, positioned, codes, respectively. The proposed rating is a preliminary step towards
and discussed. assisting in the selection of a combination. As the proposed ratings are
based on the authors’ perspective, experimental studies are required to
3. Integrated framework for computer vision-based construction
progress monitoring (CV-CPM) Table 2
Summary of stage-wise specific contributions.
The integrated framework for Computer Vision-based Construction
Stage Published review papers Detailed contribution of
Progress Monitoring is shown in Fig. 2. This framework was derived this paper
after evaluating the existing literature studies in detail using the macro- Reference Their contribution
level framework lens described in the previous section and the meth Data [2] Categorised, listed, • Identified and
odology defined in [7]. This section explores the micro-level details of Acquisition and compared various proposed factors and
data acquisition ratings that can be
each of the three main components. In addition to the basic concepts and
technologies and used to guide
references to the work done in this area, Table 2 summarises the key studies in detail. selection of sensors-
contributions of this paper for the micro-level stages of the framework mount combinations
with reference to other review papers. for data acquisition.
The following three sub-sections are on the various processes, algo 3D Reconstruction [4]
Presented and reconstruction using • Comparison SLAM
rithms, methods, and technologies that could be applied to the relevant
overall SfM. Categorised, and SfM technologies
subsections of the framework. pipeline for listed, and compared for construction
3D techniques for SfM progress monitoring.
3.1. Data acquisition and 3D reconstruction technologies and
studied in detail.
As-Built [32] Presented an overview • Defined types of as-
This section addresses the data acquisition (Fig. 2 (a)) and 3D Modelling of the heuristic-based built models based on
reconstruction (Fig. 2 (b)) components of the framework. The data modelling process and existing work and
acquisition depends on the acquisition technology as well as the sensor classification of proposed guidelines
mounting method selected. There are various steps involved in 3D categories. for selecting as-built
model type for prog
reconstruction algorithms. The decision as to which technologies to use ress monitoring based
is a multi-faceted issue. Defining these based on the characteristics of a on project
construction project and other factors is the primary contribution of this requirements.
section. • Detailed review of as-
built modelling ap
proaches and ratio
3.1.1. Data acquisition: technologies and methods nale for using a
Data acquisition depends on selecting the technology and the sensor hybrid approach.
mounting method (Fig. 2 (a)). As seen in Table 1, several combinations Progress NA NA • Categorised progress
of sensing technologies and sensor mounting methods have been Visualisation monitoring into four
Progress Quantification levels and associates
explored. Digital cameras [8,9], video cameras [10], laser scanners [11], these levels as the
and range imaging (or RGB-D cameras) [12–14] are the vision-based requirements driver
technologies used for collecting data for progress monitoring. They for deciding on
generate inputs to the framework in the form of image frames or point options in the
upstream stages of
clouds. The sensor mounting method for data acquisition can be in the
process.
form of fixed devices, handheld devices, robotic systems mounted on NA NA
unmanned ground vehicles (UGV) [15], unmanned aerial vehicles
(UAV) [10], or a combination of these systems [16].
6
V.K. Reja et al. Automation in Construction 138 (2022) 104245
Fig. 3. Dual matrix for sensor mounting method and data acquisition technology selection.
make them more objective. and stereo images, and video frames using SfM for photogrammetry and
its corresponding algorithms for 3D reconstruction. In our paper, only
3.1.2. 3D reconstruction the basic concept of SfM is presented, and a detailed comparison with
Following the image-based input from the previous step, the next SLAM is made to facilitate decision making for 3D reconstruction in
step is to generate a point cloud model from a series of algorithms for 3D construction.
reconstruction (Fig. 2 (b)). As observed from Table 1, most SfM (Visual SFM [18] or Open SfM [19]) is a technique consisting of
photogrammetry-based progress monitoring pipelines use commercially a combination of algorithms for photogrammetric 3D reconstruction
available software to generate a 3D point cloud from optical camera from numerous image frames (Fig. 2 (b). It is an offline approach for
images or depth images. estimating the scene’s camera motion information. The process is to
Optical camera images do not contain depth information. SfM match all the images to each other, find the correspondences, and then
(Structure-from-Motion) and SLAM (Simultaneous Localisation and delete mismatched images to obtain relative camera positions and the
Mapping) are the two conceptual approaches to add depth information structure without any prior geometric or semantic information. The
for sparse 3D reconstruction. concept of SfM is based on stereoscopic photogrammetry, as stated in
Depth images are captured using RGB-D cameras, for example, [20]. A detailed review of the 3D reconstruction by SfM and its algo
Microsoft Kinect. They have depth information (XYZ coordinates) along rithm can be obtained in [4,21].
with colour information (RGB values) on a per-pixel basis. This infor SLAM [22] is a more general framework, in real-time, where a
mation is used for mapping and performing dense 3D reconstruction mapping system starts motion from an unknown location in an unknown
using intrinsic and extrinsic camera parameters [17]. The output from a environment, and during movement, simultaneously keeps track of its
laser scanner is a 3D point cloud, and hence does not require 3D position with respect to the surrounding environment, building an in
reconstruction. cremental map. SLAM may depend on sensors like cameras, LiDAR, GPS,
The work in [4] focuses on generating point clouds from monocular IMU, etc. Modern visual SLAM has gradually developed into a multi-
7
V.K. Reja et al. Automation in Construction 138 (2022) 104245
8
V.K. Reja et al. Automation in Construction 138 (2022) 104245
views for multi-view stereo (CMVS), patch-based multi-view stereo outlier removal, and down-sampling, along with various techniques and
(PMVS), and semi-global matching (SGM). The open-source algorithms algorithms used in the corresponding literature. The table also shows
which, perform dense 3D reconstruction include VisualSfM [18], studies that have addressed fundamental aspects of these sub-processes.
OpenSfM [19], and Bundler SfM [25]. Registration involves merging multiple partial point clouds obtained
A dense 3D reconstruction step is required if the reconstructed point into a single file or registering an as-built point cloud over an as-planned
cloud is sparse, typically in the case of reconstruction from monocular or BIM model for progress monitoring. There are two stages of registration,
stereoscopic images or video frames. The reconstruction output from coarse and fine.
depth images is already dense enough to be used directly and does not Coarse registration aligns a set of point clouds using correspondences
require dense 3D reconstruction. between them. In contrast, the fine registration algorithm further
Only a handful of CV-CPM pipelines have stated the method used for matches multiple point clouds by estimating a rigid transform and
absolute scale recovery and dense 3D reconstruction; others have not minimising the distance between the corresponding matched points. The
explicitly mentioned it in the paper. The choices available for absolute selection of these approaches depends on the type of data acquisition
scale recovery and dense 3D reconstruction add additional variables to technology used and location capture, as registration depends on
our framework. Once a dense point cloud is generated, the next step is to localising the data. While the manual and marker-based methods require
pre-process it through registration, noise filtering, outlier removal and iterative human efforts, the sensor and feature-based approaches are
down sampling. These sub-processes and model generation are dis reported to be more automated. The Iterative Closest Point (ICP) algo
cussed in the next section. rithm is the most used in CV-CPM pipelines for fine registration. It
searches for an optimum solution, and the convergence speed is directly
proportional to the accuracy of coarse registration. For image-based
3.2. Model generation methods, image registration over the BIM model is made using
georeferencing.
In this section, first, the characteristics of the as-planned model are Next, noise and outliers are unwanted points and are removed to
discussed. All the steps in pre-processing and the process of generating improve the point cloud quality prior to further processing. Finally, due
the as-built model are then discussed in detail (Fig. 2 (d)). The existing to the merging of point clouds, the overlapped regions become much
techniques for as-built model generation are categorised into heuristic- denser, affecting processing efficiency. Hence down-sampling is per
based and learning-based methods. The definition of these categories formed to make the point cloud uniform. Relevant works which include
and the mapping of existing pipelines to these categories/sub-categories these steps are shown in Table 3.
is the primary contribution of this section. This section also presents the Although pre-processing of data has been presented in some progress
trade-off between the levels of information the various as-built outputs monitoring pipelines, there is no detailed exploration of the available
provide for progress estimation and the computational complexity techniques. There is a need for a comprehensive study and further
required to generate them. research to customise pre-processing methods based on the input pipe
line’s requirement. There is also a need to explore the other techniques
3.2.1. As-planned model mentioned in Table 3 for better results using the CV-CPM framework.
As-planned models may or may not be available for a particular fa
cility or project. Generally, they are prepared in the design phase of the 3.2.2.2. Types of as-built models. Existing pipelines have used several
project. These can be 2D/3D CAD or 3D/4D BIM models. The approach combinations of as-built and as-planned model comparisons for progress
to progress monitoring depends on the availability and type of the as- estimation (Table 1). Based on the algorithm and method used, the as-
planned model, as it facilitates comparison with the as-built model. built output can be a pre-processed point cloud [26], voxel model
[27], mesh model [12,28], surface model [29,30], or IFC/BIM models
3.2.2. As-built model [11,14,31] (Fig. 2 (d)).
The following sub-sections include the details for the various sub- The level of progress monitoring proposed for a project plays a vital
processes and options for as-built modelling (Fig. 2 (d)). role in deciding the type of as-built model to be generated. Table 4
compares computational requirements and quality of output for the
3.2.2.1. Point cloud pre-processing. Point cloud pre-processing is a different levels of progress monitoring with the corresponding types of
crucial step towards as-built modelling to improve the point cloud as-built models. Here, green represents a performance improvement,
quality. The level of automation in pre-processing has not been and red indicates a decrement in performance.
adequate, and manual intervention is still used to get the required It can be seen from Table 4 that for L-1 or L-2 based monitoring,
quality. Table 3 outlines the critical sub-processes of point cloud pre- voxel-based, mesh-based, or point cloud models are recommended to be
processing, which broadly consist of registration, noise filtering,
Table 4
Relative computational requirements and output accuracy for the level of progress monitoring, based on the type of as-built models.
9
V.K. Reja et al. Automation in Construction 138 (2022) 104245
[89]
quality is low, it is adequate for visualisation. While visual output
x
quality is better for IFC-based or surface models, it is not recommended
[88]
for progress visualisation due to very high computing requirements.
x
For L-3 or L-4 based monitoring, while it will take higher computa
[34]
tion to generate an IFC or surface-based model initially, the computation
x
required for progress quantification will be substantially lower with
[87] higher accuracy; hence it is recommended. On the other hand, the mesh
x
x
model, voxel model, and point clouds are easy to generate but will take
[86]
x
x
x
x
x
proaches in detail.
[11]
methods resulting in IFC classes (Fig. 2 (d)). Only two existing studies
x
recognition in detail.
x
[33]
A. Heuristics-based approach
[39]
x
x
x
Learning – Appearance x
Learning – Geometry
elements.
Heuristic Based
Learning-Based
10
V.K. Reja et al. Automation in Construction 138 (2022) 104245
Table 6
Classification of applied heuristics to obtain specific elements.
Heuristics Classification Elements and Applied Heuristics
rules. Apart from a few complex elements, most elements can be computational challenges in automatic registration as the acquired
identified by applying these rules. point cloud data is voluminous.
3. Relationship-based constraints: The construction sequence is of
utmost importance, as it can accurately define the objects’ de It can be summarised that heuristics-based techniques usually
pendencies from schedules, enabling better element detection for require a significant amount of prior information. Generation of this
progress monitoring. These time-based relationships can be helpful information requires domain knowledge about construction and the
while there is limited visibility of the scene or occlusions. Space- type of elements being recognised. A recommended solution is to define
based constraints have been used in the literature using ontological a set of metaheuristics that generally applies to any construction dataset
relations [34] and shape grammar [35], but they are challenging to as a subset of all the heuristics; using them will make element identifi
formulate and apply in practice. cation viable with less detailed information requirements.
4. Model-based constraints: Model-based element retrieval uses Though using some heuristics makes element identification easier, a
reverse engineering to recognise the elements. The point cloud is pure heuristic-based approach for CV-CPM can be unreliable and chal
registered over the as-planned BIM model, and the elements are lenging to be implemented.
identified based on overlap occupancy. Although the method results
in high accuracy, its use is limited because it requires a detailed as- B. Learning-based approach
planned model. It has specific point cloud quality and reference
BIM model requirements [36]. The approach also faces Learning-based techniques (Fig. 2 (d), Table 5) are used on point
11
V.K. Reja et al. Automation in Construction 138 (2022) 104245
clouds for 3D shape classification, object detection, and segmentation. heuristics-based approaches may work for general geometric compo
They are based on training features using a neural network and nents, they fail to perform when the construction scene becomes com
computed weights to predict object or material classes. Two types of plex and do not follow conventional rules, relationships, and geometries.
learning approaches have been used: Learning-based approaches are more adaptive, as they can generalise
on-site data and are not based on structured domain knowledge.
1. Geometric Learning extracts geometrical features using descriptors Therefore, to utilise the advantages of both the approaches and
and feeds neural networks on statistics like distance, area, and an overcome their individual shortcomings, there is a need to explore the
gles. The geometric features are used to identify the element’s hybrid approach, i.e., using heuristics in conjunction with learning
physical presence. algorithms.
2. Appearance Learning extracts features like HSI (Hue-Saturation-
Intensity) colour values, material reflectance, and surface roughness 3.3. Progress monitoring
as features. The appearance-based features describe its operational
state. Progress monitoring requires a visual or quantitative comparison
between the as-planned and as-built models. Four progress monitoring
As observed from Table 5, only a few studies have detected elements levels identified as a part of this study are L-1, L-2, L-3, and L-4, as
for as-built modelling using learning-based techniques in the past. Most discussed in Section 2. The following sub-section discusses the key
of these methods detect geometric features and do not recognise the technologies and methods used for implementing these levels.
elements’ operational state. A few studies recognised the operational
state by using image processing techniques [37,38] and learning 3.3.1. Progress visualisation and comparison (L-1 and L-2)
appearance-based features [39]. L-1 based monitoring is currently the predominant method used at
For progress monitoring applications, learning-based approaches construction sites and only requires visualising the 3D as-built model
have been predominantly used on 2D image data, and there are very few built from spatial data obtained from the site. As an extension to L-1, L-2
studies involving 3D point cloud learning. Recent advances in machine based monitoring requires a visual comparison between as-planned and
learning and deep neural networks have motivated researchers to as-built status at a given time. This is also used on several projects and
explore this area. However, their application in point cloud-based explored by numerous research studies.
element recognition for construction has been limited because of chal These visualisations or comparisons can be made using selected en
lenges such as: vironments, as shown in Fig. 2 (e). The literature review in Table 1
shows that most pipelines have used non-immersive environments such
1. 3D point cloud data are unstructured, large, and with varying point as a 3D viewer or web-based viewer. A few pipelines have used
densities, making 3D feature detection computing-intensive and immersive Extended Reality (XR) environments, which include
time-consuming. Augmented Reality (AR) [21,28], Virtual Reality (VR) [13,44], and
2. Scan data are generally noisy and have occluded components and Mixed Reality (MR) [12,45].
cluttered scenes. Also, the data are often susceptible to changes in While a few studies used traffic light-based coding to show progress
environmental factors such as rain, fog, dust, or mist. [21], the use of comparative and interactive UI has also been investi
3. Learning models do not perform well when there is a significant gated [28]. Interactive UIs are gaining popularity, as they can pass in
difference between the training and test data. formation among stakeholders and assign tasks or comments for the
4. Supervised learning requires extensive annotated training and workforce by adding features like annotating PDF notes into the envi
testing datasets for various categories of incomplete building ele ronments. Advancements in gaming engines can be used to represent
ments: currently, the data is limited. point clouds and BIM models by importing and overlaying in a single VR
environment utilising Unity3D [13,44].
ScanNet [40], S3DIS [41], and ModelNET-40 [42] are 3D point cloud Based on the review of literature and practice, this study finds that
datasets consisting of different classes of building elements that are the capability of visualisation environments is underutilised for progress
currently used in studies. However, these data sets represent completed monitoring applications. With advancements in computing skills, rapid
buildings; for construction progress monitoring studies, data sets are development in immersive technologies, and required bandwidths
required to be acquired from buildings under construction. For material supported by the 5G standard, these platforms can provide the required
recognition, one such example of an image-based dataset is the Con information in real-time [46]. Focused research is required to generate
struction Material Library (CML) created in [9,37], which consist of interactive UI and integrate AR and VR environments to give the user the
more than 3000 images categorised into 20 construction material best of both worlds and utilise gaming engines to take progress visual
classes. isation to the next level.
There are two approaches to generate the data. The first approach is
by generating realistic, rendered elements-based graphic models to 3.3.2. Progress quantification, schedule updating, and notifications (L-3
create synthetic data [43]. The second approach is to capture and pro and L-4)
cess real-world data with the high computing power of state-of-the-art The next level of progress monitoring, L-3, involves quantifying the
data acquisition devices. completed elements. Therefore, in continuation to Table 4, Fig. 5 shows
If the challenges mentioned earlier are solved, the learning-based a trade-off for selecting the quantity estimation method corresponding
techniques open doors to autonomous element detection and identifi to the type of model generated. Here, on the left-hand scale, the green-
cation for robust progress monitoring. to-red colour variation indicates the increase in complexity from low to
high. Similarly, the green-to-red colour variation indicates the decrease
Hybrid approaches in the ease of progress quantification from high to low on the right-hand
scale.
As discussed, both the heuristics-based approach and the learning- While the IFC/BIM models can produce direct BoQs from IFC-based
based approach face a set of challenges. While the former requires element geometries [31], other models require comparison with the as-
extensive hard coding of the domain knowledge, which is complex and planned model. For surface models and mesh models, object detection
challenging to define for all situations faced in construction, the latter (based on models overlapping) [13], detection by ray thresholding [12],
approach requires extensive training datasets, higher computational or enclosed volumetric comparison [47] can be made for estimating
resources, and knowledge of advanced computing techniques. Though quantities of completed elements.
12
V.K. Reja et al. Automation in Construction 138 (2022) 104245
Fig. 5. Trade-off for selecting Quantity Estimation method corresponding to the type of as-built model for progress quantification.
For comparison using voxel occupancy, the as-planned model is or discrepancies in the as-built model from the as-planned model is
voxelised and overlapped with as-built point clouds, after which essential for in-time decision making and can improve the project’s
thresholding is done to detect the occupied voxels [8]. Although it works performance.
well for smaller models, it may be computationally costly for larger In addition to providing updated information on progress status,
models; hence, it is not a preferred option. For direct usage of the point features to suggest interventions and control strategies and forecast
cloud, the most straightforward and widely-used method is Scan-vs-BIM, resulting outcomes will automate project controls even further.
which uses thresholding-based element identification. This method re Conventional control decisions taken by the project managers are
quires an as-planned 3D BIM and has its shortcomings [48], but the as- based on the know-how of issues specific to the project. Automated
built model generation step can be skipped, saving computation. Com control suggestions can be generated using cognitive computing on the
mercial project management software are also used for detecting de progress status and the data from sources such as the organisation’s
viations between as-planned and as-built models that can result in the Enterprise Resource Portal (ERP). Such an approach can recognise pat
quantification of contractual scheduling metrics [14,44]. terns in activities that suffer delays and suggest interventions to bring
The selection of a model for progress monitoring depends on the the project back on track. There have been selected deployments using
project’s characteristics (type, budget, duration), the required level of AI-based simulation engines for addressing complications inherent in
progress monitoring, and other factors. A detailed study of these factors scheduling and project coordination [53]. Such deployments can also
is required to converge onto an appropriate pipeline for progress evaluate alternatives and provide real-time suggestions. This is an
monitoring based on the project-specific use case. emerging area with significant potential, and further exploration is
For L-4 progress monitoring, updating schedules and generating re required.
ports/notifications are included with progress quantification. This is This section presented a micro-analysis of the stages of the assembled
challenging; therefore, as shown in Table 1, only a few pipelines have CV-CPM framework. Various gaps were identified while discussing each
explored this area. Though schedule updating has not been fully auto section in detail. The following section discusses the application of the
mated yet in practice, conceptual schemes have been proposed [49]. CV-CPM framework to enable broader evaluation of various pipelines/
Some schemes have utilised the concept of precedence relationship components to develop benchmarks. It also summarises specific chal
graphs and articulation points, where all objects must be finished before lenges identified within each stage.
the element linked to the articulation point can be built [50]. Few
studies have compared as-planned and as-built BIM for schedule 4. Discussion
updating [31,51]. As the as-built IFC/BIM model is updated, parametric
information has also been utilised to generate the updated schedules As presented and discussed in the previous section, numerous com
automatically [52]. binations of input-process-output options are possible at each stage of
The primary challenge in schedule updating is that schedule LoD is the CV-CPM framework. For an integrated pipeline across all stages, the
different from BIM LoD, resulting in mismatch and incompatibility. combinations of options are immense. Given the pace at which tech
Secondly, it requires an as-planned 4D model for smooth implementa nology is developing, the number of options will continue to increase.
tion, that is not always available for a construction project. Other However, only a limited combination of options will be feasible, and this
challenges include deciding the frequency of updating 4D-BIM. paper has proposed guidelines to structure these combinations.
Although a few research papers on CV-CPM have mentioned the A systematic approach is essential to explore and document the
concept of schedule updating in their pipelines, they have not specified performance of specific approaches/algorithms and integrated CV-CPM
the details. Therefore, there is a necessity to conduct and integrate pipelines. This section discusses requirements for benchmarking to
focused studies for tackling these specific challenges. enable a systematic study. The section also summarises potential
Likewise, only a few studies have explored an automatic notification research questions within each stage of CV-CPM and discusses its rele
system that can trigger notifications while monitoring and controlling a vance to support the emerging area of Digital Twin.
construction site [47]. These can be obtained via SMS or e-mail, con
taining warnings, reports, or graphical outputs based on the settings and
LoD required. The frequency and content of notifications should be 4.1. Benchmarking
selected to give construction managers broad insights about areas to
target for meeting schedules. As discussed earlier, providing managers Benchmarking evaluates pipelines, algorithms, devices, and tools
with real-time information/alerts about any deviation in the schedules using standard datasets and test methods. Fig. 6 outlines the broad re
quirements to benchmark pipelines and components of CV-CPM, and
13
V.K. Reja et al. Automation in Construction 138 (2022) 104245
these are discussed below: quantitative metrics, qualitative measures will also be required to
evaluate subjective outcomes. For example, for benchmarking visual
1. Datasets isation environments, qualitative factors like ease of use, skills required
for navigation, immersive features for comparison, and training re
Though existing datasets contain relevant information for building quirements can be compared from a user feedback-based evaluation.
elements [40,41], these represent completed elements and not the Correspondingly, for benchmarking the algorithms, the evaluation
schedule-based construction progress of the elements as required for metrics can be in the form of accuracy, processing time, computational
progress monitoring. Hence, there is a requirement for creating an open- cost etc., which are quantitative.
source dataset that contains the data on elements as construction pro The integrated CV-CPM framework and the guidelines proposed for
gresses. As shown in Fig. 6, these can be synthetically created or ac each stage is expected to assist in developing a strategy and a roadmap
quired from the real world. The dataset type will change based on the for benchmarking. As a wide range of studies is required, the technology
particular stage under investigation. For example, for 3D reconstruction, roadmap needs to be developed collaboratively by the research com
the type of dataset required for benchmarking reconstruction algorithms munity. A starting point for these studies can be to investigate the
will be in the form of raw images, videos, and depth images. subjective ratings proposed in this study and quantify these ratings
through controlled experimental testbeds. The CV-CPM framework can
2. Testbeds help in structuring and prioritising the areas to be explored in devel
oping a roadmap so that they form a standard reference for bench
While existing studies in the area have contributed significantly to marking studies.
knowledge about the applicability of specific algorithms and tools used,
the experimental set-ups and procedures used for these investigations 4.2. Future research directions
vary. As a result, repeatability, and comparison of results across research
groups are not feasible. Hence, for each CV-CPM stage, there is a need to This section highlights potential research questions within the three
design standard testbeds that define the experimental setup, procedures, stages of the CV-CPM framework. Several of these questions can be
and measurement parameters. As shown in Fig. 6, the testbed can be addressed through benchmarking studies, while others require explo
either computational for benchmarking software (algorithms) or phys ration of new concepts.
ical for benchmarking the hardware (devices/sensors). For example, a Stage 1: Data acquisition and 3D reconstruction
testbed for data acquisition can have sensors capturing data with various
physical constraints, like the speed of capture, distance from an object, • How can we objectively assess the factors affecting data acquisition
lightning condition, resolution settings etc., that can be varied to technology and sensor mounting method shown in Fig. 3?
perform multiple experiments. Further, the testbeds should be flexible to • What are the appropriate sensing technology and mounting method
accommodate the requirements for collaborative explorations in the for data acquisition based on project types and project characteristics
area. (ex., Linear, underground, elevated etc.)?
• Based on factors identified in Fig. 4, How can we decide on using SfM
3. Component of CV-CPM and SLAM for a specific project?
• Vision-based data is usually enormous, and storage requirements are
As shown in Fig. 6, for CV-CPM, the components can be bench high. How sparse can data be without impacting the output?
marked independently or as pipelines integrating multiple components.
Each component can be evaluated for its independent performance. Stage 2: As-built modelling
Additionally, the pipelines can also be benchmarked for their effec
tiveness for overall progress monitoring by varying specific components • How can the data be effectively pre-processed to meet the point
within the pipeline. This will standardise the approach to sequence and cloud quality requirements for an as-built model generation?
investigate various input-process-output combinations and enable a • How can a hybrid of heuristics and learning-based approaches be
systematic comparison of results with other studies. deployed, such that it exploits the advantages and overcomes the
shortcomings of each?
4. Evaluation Metrics • How can cognitive computing be successfully implemented to
recognise and measure complex geometries at reduced computa
There are several factors to be considered for evaluation. While most tional cost accurately?
benchmarking studies within this framework require the definition of
14
V.K. Reja et al. Automation in Construction 138 (2022) 104245
Stage 3: Progress Monitoring papers, including recent papers in this area. Hence, it can be inferred
that it is holistic and robust.
• How can we objectively compare the computational requirements It was found that the four levels of progress monitoring identified in
and output accuracy for the level of progress monitoring based on the this study strongly influence the technology selection for the pipeline at
type of as-built models as shown in Table 4? each stage. For implementing progress monitoring on a construction
• How can virtual and physical environments be seamlessly connected project, it is recommended that the pipeline and components required
to improve visualisation? are selected based on the chosen level.
• How can we select an appropriate as-built model and progress Several areas for further research were discussed in the previous
quantification technique based on project characteristics and the section. Among these areas, advancements in the as-built modelling
trade-off mentioned in Fig. 5? stage are required to facilitate the quantification of progress. To enable
• How can we generate automatic schedules update and control sug this, exploring a hybrid approach that combines learning with heuristics
gestions by applying cognitive computing techniques to the progress is recommended.
data produced by CV-CPM? The need and requirements for benchmarking and future research
directions derived from the CV-CPM framework have been presented in
In addition to the specific areas identified above, the CV-CPM the Discussion section. Addressing these requirements and following a
framework is also relevant to support emerging areas of research and well-developed roadmap in this area is essential to move CV-CPM
development, such as Digital Twin. research from laboratory studies to field applications.
Digital Twin technologies enable the built facility’s virtual repre
sentation to mirror and predict the state/behaviour of the physical fa Declaration of Competing Interest
cility. Key technologies for Digital Twin implementation include:
The authors confirm that all authors agreed with the current version
1. continuous streaming and processing of sensing data, and there is no conflict of interest among all authors with respect to this
2. threaded models and simulations for current/future state/behaviour paper.
of the facility,
3. formulating/implementing interventions to limit future state/ Acknowledgements
behaviour within specified limits.
The authors would like to thank the journal editor and reviewers for
Studies have envisioned the utility of Digital Twin in all phases of a their detailed constructive comments, which helped improve the paper.
construction project lifecycle. However, details on twinning an under- Additionally, the support, the first author received from the ‘Prime
construction facility are limited [54]. Commercial platforms such as Minister Research Fellowship (PMRF)’ Scheme from the ‘Ministry of
Autodesk Tandem [55] and Bentley iTwin [56] are available for creating Education, Government of India’ is gratefully acknowledged.
Digital Twin solutions. However, technological features available in
these platforms are currently in the early stages and focus on the oper References
ations and maintenance phase of the built asset.
A key issue in creating a Digital Twin for construction progress [1] J. Teizer, Status quo and open challenges in vision-based sensing and tracking of
temporary resources on infrastructure construction sites, Adv. Eng. Inform. 29
monitoring is that the geometry of the facility keeps expanding; hence (2015) 225–238, https://doi.org/10.1016/j.aei.2015.03.006.
the corresponding sensor positioning is also dynamic. In this context, [2] T. Omar, M.L. Nehdi, Data acquisition technologies for construction progress
computer vision-based sensing would be the most appropriate input to tracking, Autom. Constr. 70 (2016) 143–155, https://doi.org/10.1016/j.
autcon.2016.06.016.
capture geometrical attributes directly and rapidly. The proposed CV-
[3] S. Paneru, I. Jeelani, Computer vision applications in construction: current state,
CPM framework identifies the essential technologies to embed prog opportunities & challenges, Autom. Constr. 132 (2021), 103940, https://doi.org/
ress monitoring capabilities in a Digital Twin. 10.1016/j.autcon.2021.103940.
[4] Z. Ma, S. Liu, A review of 3D reconstruction techniques in civil engineering and
Pipelines that can stream and process the progress data accurately
their applications, Adv. Eng. Inform. 37 (2018) 163–174, https://doi.org/10.1016/
and rapidly to mirror the progress in the virtual model will support j.aei.2018.05.005.
Digital Twinning for progress monitoring. For this, the process within [5] B. Ekanayake, J.K.-W. Wong, A.A.F. Fini, P. Smith, Computer vision-based interior
and across the stages of the CV-CPM framework needs to be optimised construction progress monitoring: a literature review and future research
directions, Autom. Constr. 127 (2021), 103705, https://doi.org/10.1016/j.
for real-time updates. autcon.2021.103705.
In addition to mirroring the progress, a Digital Twin can forecast [6] T. Patel, B.H.W. Guo, Y. Zou, A scientometric review of construction progress
construction progress as well as simulate and evaluate control measures monitoring studies, Eng. Constr. Archit. Manag. (2021), https://doi.org/10.1108/
ECAM-10-2020-0799.
to bring the project back on track. As the usage of autonomous con [7] B. Omair, A. Alturki, An improved method for taxonomy development in
struction equipment increases, the control measures can be transmitted information systems, Int. J. Adv. Comput. Sci. Appl. 11 (2020) 535–540, https://
from the Digital to the physical world to be implemented by the doi.org/10.14569/IJACSA.2020.0110470.
[8] M. Golparvar-Fard, F. Peña-Mora, S. Savarese, Automated progress monitoring
equipment. using unordered daily construction photographs and IFC-based building
information models, J. Comput. Civ. Eng. 29 (2015) 04014025, https://doi.org/
5. Conclusion 10.1061/(ASCE)CP.1943-5487.0000205.
[9] A. Dimitrov, M. Golparvar-Fard, Vision-based material recognition for automated
monitoring of construction progress and generating building information modeling
CV-CPM has the potential of creating an immense impact by from unordered site image collections, Adv. Eng. Inform. 28 (2014) 37–49, https://
providing real-time, accurate, reliable information to construction doi.org/10.1016/j.aei.2013.11.002.
[10] J.R. Bognot, C.G. Candido, A.C. Blanco, J.R.Y. Montelibano, Building construction
managers. Though a significant amount of work has been done in the last progress monitoring using unmanned aerial system (UAS), low-cost
decade, specific challenges remain due to the construction industry’s photogrammetry, and geographic information system (GIS), ISPRS Ann.
dynamic nature and complexities at sites. Some of these challenges have Photogramm. Remote Sens. Spat. Inform. Sci. IV–2 (2018) 41–47, https://doi.org/
10.5194/isprs-annals-IV-2-41-2018.
been addressed; however, significant gaps need to be filled to make
[11] F. Bosché, M. Ahmed, Y. Turkan, C.T. Haas, R. Haas, The value of integrating scan-
pipelines accurate and automated to meet rising user expectations of to-BIM and scan-vs-BIM techniques for construction monitoring using laser
real-time feedback. scanning and BIM: the case of cylindrical MEP components, Autom. Constr. 49
This study has comprehensively reviewed individual pipelines and (2015) 201–213, https://doi.org/10.1016/j.autcon.2014.05.014.
[12] M. Kopsida, I. Brilakis, Real-time volume-to-plane comparison for mixed
formulated an integrated CV-CPM framework, as shown in Table 1 and reality–based progress monitoring, J. Comput. Civ. Eng. 34 (2020) 04020016,
Fig. 2, respectively. The proposed framework positions all the reviewed https://doi.org/10.1061/(ASCE)CP.1943-5487.0000896.
15
V.K. Reja et al. Automation in Construction 138 (2022) 104245
[13] F. Pour Rahimian, S. Seyedzadeh, S. Oliver, S. Rodriguez, N. Dawood, On-demand [39] H. Hamledari, B. McCabe, S. Davari, Automated computer vision-based detection
monitoring of construction projects through a game-like hybrid application of BIM of components of under-construction indoor partitions, Autom. Constr. 74 (2017)
and machine learning, Autom. Constr. 110 (2020), 103012, https://doi.org/ 78–94, https://doi.org/10.1016/j.autcon.2016.11.009.
10.1016/j.autcon.2019.103012. [40] A. Dai, A.X. Chang, M. Savva, M. Halber, T. Funkhouser, M. Niessner, ScanNet:
[14] Z. Pučko, N. Šuman, D. Rebolj, Automated continuous construction progress Richly-annotated 3D reconstructions of indoor scenes, in: Conference on Computer
monitoring using multiple workplace real time 3D scans, Adv. Eng. Inform. 38 Vision and Pattern Recognition (CVPR), IEEE, 2017, pp. 2432–2443, https://doi.
(2018) 27–40, https://doi.org/10.1016/j.aei.2018.06.001. org/10.1109/CVPR.2017.261.
[15] A. Adán, B. Quintana, S.A. Prieto, F. Bosché, An autonomous robotic platform for [41] I. Armeni, O. Sener, A.R. Zamir, H. Jiang, I. Brilakis, M. Fischer, S. Savarese, 3D
automatic extraction of detailed semantic models of buildings, Autom. Constr. 109 semantic parsing of large-scale indoor spaces, in: Conference on Computer Vision
(2020), 102963, https://doi.org/10.1016/j.autcon.2019.102963. and Pattern Recognition (CVPR), IEEE, 2016, pp. 1534–1543, https://doi.org/
[16] K. Asadi, A. Kalkunte Suresh, A. Ender, S. Gotad, S. Maniyar, S. Anand, 10.1109/CVPR.2016.170.
M. Noghabaei, K. Han, E. Lobaton, T. Wu, An integrated UGV-UAV system for [42] Zhirong Wu, S. Song, A. Khosla, Fisher Yu, Linguang Zhang, Xiaoou Tang, J. Xiao,
construction site data collection, Autom. Constr. 112 (2020), 103068, https://doi. 3D ShapeNets: A deep representation for volumetric shapes, in: Conference on
org/10.1016/j.autcon.2019.103068. Computer Vision and Pattern Recognition (CVPR), IEEE, 2015, pp. 1912–1920,
[17] Q. Wang, Y. Tan, Z. Mei, Computational methods of acquisition and processing of https://doi.org/10.1109/CVPR.2015.7298801.
3D point cloud data for construction applications, Arch. Comput. Method Eng. 27 [43] P. Bhadaniya, V.K. Reja, K. Varghese, Mixed reality-based dataset generationfor
(2020) 479–499, https://doi.org/10.1007/s11831-019-09320-4. learning-based scan-to-BIM, in: Pattern Recognition - Lecture Notes in Computer
[18] C. Wu, VisualSFM: A Visual Structure from Motion System. http://ccwu.me/vsfm/, Science (LNCS), Springer International Publishing, Cham, 2021, pp. 389–403,
2011 (accessed July 19, 2020). https://doi.org/10.1007/978-3-030-68787-8.
[19] Mapillary, OpenSfM: Open Source Structure from Motion Pipeline. https://github. [44] S. Vincke, R. de Lima Hernandez, M. Bassier, M. Vergauwen, Immersive
com/mapillary/OpenSfM, 2018 (accessed July 19, 2020). visualisation of construction site point cloud data, meshes and BIM models in a VR
[20] Y. Furukawa, B. Curless, S.M. Seitz, R. Szeliski, Reconstructing building interiors environment using a gaming engine, Int. Arch. Photogramm. Remote. Sens. Spat.
from images, in: 12th International Conference on Computer Vision, IEEE, 2009, Inf. Sci. XLII-5 (W2) (2019) 77–83, https://doi.org/10.5194/isprs-archives-XLII-5-
pp. 80–87, https://doi.org/10.1109/ICCV.2009.5459145. W2-77-2019.
[21] M. Golparvar-Fard, F. Peña-Mora, S. Savarese, Integrated sequential as-built and [45] A. Khairadeen Ali, O.J. Lee, D. Lee, C. Park, Remote indoor construction progress
as-planned representation with D4AR tools in support of decision-making tasks in monitoring using extended reality, Sustainability 13 (2021) 2290, https://doi.org/
the AEC/FM industry, J. Constr. Eng. Manag. 137 (2011) 1099–1116, https://doi. 10.3390/su13042290.
org/10.1061/(ASCE)CO.1943-7862.0000371. [46] V.K. Reja, K. Varghese, Impact of 5G technology on IoT applications in
[22] T. Bailey, H. Durrant-Whyte, Simultaneous localization and mapping (SLAM): part construction project management, in: 36th International Symposium on
II, IEEE Robot. Automat. Mag. 13 (2006) 108–117, https://doi.org/10.1109/ Automation and Robotics in Construction (ISARC 2019), Banff, Canada, 2019,
MRA.2006.1678144. pp. 209–217, https://doi.org/10.22260/ISARC2019/0029.
[23] B. Triggs, P.F. McLauchlan, R.I. Hartley, A.W. Fitzgibbon, Bundle adjustment — a [47] H. Omar, L. Mahdjoubi, G. Kheder, Towards an automated photogrammetry-based
modern synthesis, in: International Workshop on Vision Algorithms, Greece, 2000, approach for monitoring and controlling construction site activities, Comput. Ind.
pp. 298–372, https://doi.org/10.1007/3–540-44480-7_21. 98 (2018) 172–182, https://doi.org/10.1016/j.compind.2018.03.012.
[24] C.-Z. Sun, B. Zhang, J.-K. Wang, C.-S. Zhang, A review of visual SLAM based on [48] M. Bassier, S. Vincke, L. Mattheuwsen, R. de Lima Hernandez, J. Derdaele,
unmanned systems, in: 2nd International Conference on Artificial Intelligence and M. Vergauwen, Percentage of completion of in-situ cast concrete walls using point
Education (ICAIE), IEEE, 2021, pp. 226–234, https://doi.org/10.1109/ cloud data and BIM, Int. Arch. Photogramm. Remote. Sens. Spat. Inf. Sci. XLII-5
ICAIE53562.2021.00055. (W2) (2019) 21–28, https://doi.org/10.5194/isprs-archives-XLII-5-W2-21-2019.
[25] I. Mitsugami, Bundler: structure from motion for unordered image collections, [49] H. Son, C. Kim, Y. Kwon Cho, Automated schedule updates using as-built data and
J. Inst. Image Inform. Telev. Eng. 65 (2011) 479–482, https://doi.org/10.3169/ a 4D building information model, J. Manag. Eng. 33 (2017) 04017012, https://doi.
itej.65.479. org/10.1061/(ASCE)ME.1943-5479.0000528.
[26] R. Maalek, D.D. Lichti, J.Y. Ruwanpura, Automatic recognition of common [50] A. Braun, A. Borrmann, S. Tuttas, U. Stilla, Big data in smart building operation, in:
structural elements from point clouds for automated progress monitoring and eWork and eBusiness in Architecture, Engineering and Construction, CRC Press,
dimensional quality control in reinforced concrete construction, Remote Sens. 11 2014, pp. 765–804, https://doi.org/10.1201/b17396-20.
(2019) 1102, https://doi.org/10.3390/rs11091102. [51] H. Hamledari, B. McCabe, S. Davari, A. Shahi, Automated schedule and progress
[27] P. Hübner, M. Weinmann, S. Wursthorn, Voxel-based indoor reconstruction from updating of IFC-based 4D BIMs, J. Comput. Civ. Eng. 31 (2017) 04017012, https://
HoloLens triangle meshes, ISPRS Ann. Photogramm. Remote Sens. Spat. Inform. doi.org/10.1061/(ASCE)CP.1943-5487.0000660.
Sci. V-4–2020 (2020) 79–86, https://doi.org/10.5194/isprs-annals-V-4-2020-79- [52] H. Kim, K. Anderson, S. Lee, J. Hildreth, Generating construction schedules
2020. through automatic data extraction using open BIM (building information
[28] S. Zollmann, C. Hoppe, S. Kluckner, C. Poglitsch, H. Bischof, G. Reitmayr, modeling) technology, Autom. Constr. 35 (2013) 285–295, https://doi.org/
Augmented reality for construction site monitoring and documentation, Proc. IEEE 10.1016/j.autcon.2013.05.020.
102 (2014) 137–154, https://doi.org/10.1109/JPROC.2013.2294314. [53] ALICE Technologies, AI for Construction Management. https://blog.alicetech
[29] H. Macher, T. Landes, P. Grussenmeyer, From point clouds to building information nologies.com/whitepapers/ai-for-construction-management-whitepaper, 2021
models: 3D semi-automatic reconstruction of indoors of existing buildings, Appl. (accessed January 10, 2022).
Sci. 7 (2017) 1030, https://doi.org/10.3390/app7101030. [54] C. Boje, A. Guerriero, S. Kubicki, Y. Rezgui, Towards a semantic construction
[30] X. Xiong, A. Adan, B. Akinci, D. Huber, Automatic creation of semantically rich 3D digital twin: directions for future research, Autom. Constr. 114 (2020), 103179,
building models from laser scanner data, Autom. Constr. 31 (2013) 325–337, https://doi.org/10.1016/j.autcon.2020.103179.
https://doi.org/10.1016/j.autcon.2012.10.006. [55] Autodesk Tandem, Autodesk. https://intandem.autodesk.com/, 2021 (accessed
[31] H. Mahami, F. Nasirzadeh, A. Hosseininaveh Ahmadabadian, S. Nahavandi, January 10, 2022).
Automated Progress controlling and monitoring using daily site images and [56] Bentley, An Open Platform for Infrastructure Digital Twins: How Bentley Systems’
building information modelling, Buildings 9 (2019) 70, https://doi.org/10.3390/ iTwin Platform is Enabling a Digital Twin Ecosystem for the World’s Infrastructure,
buildings9030070. A Frost Sullivan White Pap. https://prod-bentleycdn.azureedge.net/-/media/files/
[32] V. Pătrăucean, I. Armeni, M. Nahangi, J. Yeung, I. Brilakis, C. Haas, State of documents/white-papers/wp_frost_sullivan_open_platform_for_infrastructure_di
research in automatic as-built modelling, Adv. Eng. Inform. 29 (2015) 162–171, gital_twins.pdf?la=pt-br&modified=00010101000000, 2021.
https://doi.org/10.1016/j.aei.2015.01.001. [57] A. Braun, S. Tuttas, A. Borrmann, U. Stilla, Improving progress monitoring by
[33] S. Zeng, J. Chen, Y.K. Cho, User exemplar-based building element retrieval from fusing point clouds, semantic data and computer vision, Autom. Constr. 116
raw point clouds using deep point-level features, Autom. Constr. 114 (2020), (2020), 103210, https://doi.org/10.1016/j.autcon.2020.103210.
103159, https://doi.org/10.1016/j.autcon.2020.103159. [58] K. Karsch, M. Golparvar-Fard, D. Forsyth, ConstructAide: analyzing and visualizing
[34] K.K. Han, D. Cline, M. Golparvar-Fard, Formalized knowledge of construction construction sites through photographs and building models, ACM Trans. Graph.
sequencing for visual monitoring of work-in-progress via incomplete point clouds 33 (2014) 1–11, https://doi.org/10.1145/2661229.2661256.
and low-LoD 4D BIMs, Adv. Eng. Inform. 29 (2015) 889–901, https://doi.org/ [59] F. Arif, W.A. Khan, Smart progress monitoring framework for building construction
10.1016/j.aei.2015.10.006. elements using videography–MATLAB–BIM integration, international, J. Civ. Eng.
[35] H. Tran, K. Khoshelham, A. Kealy, L. Díaz-Vilariño, Shape grammar approach to 3D 19 (2021) 717–732, https://doi.org/10.1007/s40999-021-00601-3.
modeling of indoor environments using point clouds, J. Comput. Civ. Eng. 33 [60] S. Kim, S. Kim, D.-E. Lee, Sustainable application of hybrid point cloud and BIM
(2019) 04018055, https://doi.org/10.1061/(ASCE)CP.1943-5487.0000800. method for tracking construction progress, Sustainability 12 (2020) 4106, https://
[36] D. Rebolj, Z. Pučko, N.Č. Babič, M. Bizjak, D. Mongus, Point cloud quality doi.org/10.3390/su12104106.
requirements for scan-vs-BIM based automated construction progress monitoring, [61] Z. Wang, Q. Zhang, B. Yang, T. Wu, K. Lei, B. Zhang, T. Fang, Vision-based
Autom. Constr. 84 (2017) 323–334, https://doi.org/10.1016/j. framework for automatic Progress monitoring of precast walls by using
autcon.2017.09.021. surveillance videos during the construction phase, J. Comput. Civ. Eng. 35 (2021)
[37] K.K. Han, M. Golparvar-Fard, Appearance-based material classification for 04020056, https://doi.org/10.1061/(ASCE)CP.1943-5487.0000933.
monitoring of operation-level construction progress using 4D BIM and site [62] A. Pushkar, M. Senthilvel, K. Varghese, Automated progress monitoring of masonry
photologs, Autom. Constr. 53 (2015) 44–57, https://doi.org/10.1016/j. activity using photogrammetric point cloud, in: 35th International Symposium on
autcon.2015.02.007. Automation and Robotics in Construction (ISARC 2018), Berlin, Germany, 2018,
[38] K. Han, J. Degol, M. Golparvar-Fard, Geometry- and appearance-based reasoning https://doi.org/10.22260/ISARC2018/0125.
of construction progress monitoring, J. Constr. Eng. Manag. 144 (2018) 04017110, [63] K.K. Han, M. Golparvar-Fard, Automated monitoring of operation-level
https://doi.org/10.1061/(ASCE)CO.1943-7862.0001428. construction progress using 4D BIM and daily site photologs, in: Construction
16
V.K. Reja et al. Automation in Construction 138 (2022) 104245
Research Congress 2014, American Society of Civil Engineers, Reston, VA, 2014, Symposium on Automation and Robotics in Construction (ISARC 2021), Dubai,
pp. 1033–1042, https://doi.org/10.1061/9780784413517.106. UAE, 2021, pp. 349–356, https://doi.org/10.22260/ISARC2021/0049.
[64] A. Rashidi, I. Brilakis, P. Vela, Generating absolute-scale point cloud data of built [78] A. Adán, B. Quintana, S.A. Prieto, F. Bosché, Scan-to-BIM for ‘secondary’ building
infrastructure scenes using a monocular camera setting, J. Comput. Civ. Eng. 29 components, Adv. Eng. Inform. 37 (2018) 119–138, https://doi.org/10.1016/j.
(2015) 04014089, https://doi.org/10.1061/(ASCE)CP.1943-5487.0000414. aei.2018.05.001.
[65] M. Bueno, H. González-Jorge, J. Martínez-Sánchez, H. Lorenzo, Automatic point [79] S. Nikoohemat, A.A. Diakité, S. Zlatanova, G. Vosselman, Indoor 3D reconstruction
cloud coarse registration using geometric keypoint descriptors for indoor scenes, from point clouds for optimal routing in complex buildings to support disaster
Autom. Constr. 81 (2017) 134–148, https://doi.org/10.1016/j. management, Autom. Constr. 113 (2020), 103109, https://doi.org/10.1016/j.
autcon.2017.06.016. autcon.2020.103109.
[66] M. Bueno, F. Bosché, H. González-Jorge, J. Martínez-Sánchez, P. Arias, 4-plane [80] S. Ochmann, R. Vock, R. Klein, Automatic reconstruction of fully volumetric 3D
congruent sets for automatic registration of as-is 3D point clouds with 3D BIM building models from oriented point clouds, ISPRS J. Photogramm. Remote Sens.
models, Autom. Constr. 89 (2018) 120–134, https://doi.org/10.1016/j. 151 (2019) 251–262, https://doi.org/10.1016/j.isprsjprs.2019.03.017.
autcon.2018.01.014. [81] C. Wang, Y.K. Cho, C. Kim, Automatic BIM component extraction from point clouds
[67] L. Lei, Y. Zhou, H. Luo, P.E.D. Love, A CNN-based 3D patch registration approach of existing buildings for sustainability applications, Autom. Constr. 56 (2015)
for integrating sequential models in support of progress monitoring, Adv. Eng. 1–13, https://doi.org/10.1016/j.autcon.2015.04.001.
Inform. 41 (2019), 100923, https://doi.org/10.1016/j.aei.2019.100923. [82] S. Pu, G. Vosselman, Knowledge based reconstruction of building models from
[68] C. Kim, H. Son, C. Kim, Fully automated registration of 3D data to a 3D CAD model terrestrial laser scanning data, ISPRS J. Photogramm. Remote Sens. 64 (2009)
for project progress monitoring, Autom. Constr. 35 (2013) 587–594, https://doi. 575–584, https://doi.org/10.1016/j.isprsjprs.2009.04.001.
org/10.1016/j.autcon.2013.01.005. [83] H. Tran, K. Khoshelham, Procedural reconstruction of 3D indoor models from Lidar
[69] C.C. Kim, C.C. Kim, H. Son, C.C. Kim, Automated construction progress data using reversible jump Markov chain Monte Carlo, Remote Sens. 12 (2020)
measurement using a 4D building information model and 3D data, Autom. Constr. 838, https://doi.org/10.3390/rs12050838.
31 (2013) 75–82, https://doi.org/10.1016/j.autcon.2012.11.041. [84] S. Franz, R. Irmler, U. Rüppel, Real-time collaborative reconstruction of digital
[70] S. Roh, Z. Aziz, F. Peña-Mora, An object-based 3D walk-through model for interior building models with mobile devices, Adv. Eng. Inform. 38 (2018) 569–580,
construction progress monitoring, Autom. Constr. 20 (2011) 66–75, https://doi. https://doi.org/10.1016/j.aei.2018.08.012.
org/10.1016/j.autcon.2010.07.003. [85] L. Barazzetti, Parametric as-built model generation of complex shapes from point
[71] J. Lee, H. Son, C. Kim, C. Kim, Skeleton-based 3D reconstruction of as-built clouds, Adv. Eng. Inform. 30 (2016) 298–311, https://doi.org/10.1016/j.
pipelines from laser-scan data, Autom. Constr. 35 (2013) 199–207, https://doi. aei.2016.03.005.
org/10.1016/j.autcon.2013.05.009. [86] L. Díaz-Vilariño, K. Khoshelham, J. Martínez-Sánchez, P. Arias, 3D modeling of
[72] Y.-F. Liu, S. Cho, B.F. Spencer, J.-S. Fan, Concrete crack assessment using digital building indoor spaces and closed doors from imagery and point clouds, Sensors 15
image processing and 3D scene reconstruction, J. Comput. Civ. Eng. 30 (2016) (2015) 3491–3512, https://doi.org/10.3390/s150203491.
04014124, https://doi.org/10.1061/(ASCE)CP.1943-5487.0000446. [87] A. Arnaud, J. Christophe, M. Gouiffes, M. Ammi, 3D reconstruction of indoor
[73] H. Son, C. Kim, C. Kim, Fully automated as-built 3D pipeline extraction method building environments with new generation of tablets, in: Proceedings of the 22nd
from laser-scanned data based on curvature computation, J. Comput. Civ. Eng. 29 ACM Conference on Virtual Reality Software and Technology, ACM, New York, NY,
(2015) 1–9, https://doi.org/10.1061/(ASCE)CP.1943-5487.0000401. USA, 2016, pp. 187–190, https://doi.org/10.1145/2993369.2993403.
[74] Y. Li, W. Li, S. Tang, W. Darwish, Y. Hu, W. Chen, Automatic indoor as-built [88] X. Liu, Y. Zhang, X. Ling, Y. Wan, L. Liu, Q. Li, TopoLAP: topology recovery for
building information models generation by using low-cost RGB-D sensors, Sensors building reconstruction by deducing the relationships between linear and planar
20 (2020) 293, https://doi.org/10.3390/s20010293. primitives, Remote Sens. 11 (2019) 1372, https://doi.org/10.3390/rs11111372.
[75] J. Chen, Z. Kira, Y.K. Cho, Deep learning approach to point cloud scene [89] J. Chen, Y. Fang, Y.K. Cho, Unsupervised recognition of volumetric structural
understanding for automated scan to 3D reconstruction, J. Comput. Civ. Eng. 33 components from building point clouds, in: Computing in Civil Engineering 2017,
(2019) 04019027, https://doi.org/10.1061/(ASCE)CP.1943-5487.0000842. American Society of Civil Engineers, Reston, VA, 2017, pp. 34–42, https://doi.org/
[76] D. Iwaszczuk, Z. Koppanyi, N.A. Gard, B. Zha, C. Toth, A. Yilmaz, Semantic 10.1061/9780784480823.005.
labeling of structural elements in buildings by fusing RGB and depth images in an [90] M. Previtali, L. Díaz-Vilariño, M. Scaioni, Towards automatic reconstruction of
encoder-decoder CNN framework, Int. Arch. Photogramm. Remote. Sens. Spat. Inf. indoor scenes from incomplete point clouds: door and window detection and
Sci. XLII–1 (2018) 225–232, https://doi.org/10.5194/isprs-archives-XLII-1-225- regularization, Int. Arch. Photogramm. Remote Sens. Spat. Inform. Sci. 42 (2018)
2018. 573–578, https://doi.org/10.5194/isprs-archives-XLII-4-507-2018.
[77] V.K. Reja, P. Bhadaniya, K. Varghese, Q.P. Ha, Vision-based progress monitoring of [91] B. Quintana, S.A. Prieto, A. Adán, F. Bosché, Door detection in 3D coloured point
building structures using point-intensity approach, in: Proc. 38th International clouds of indoor environments, Autom. Constr. 85 (2018) 146–166, https://doi.
org/10.1016/j.autcon.2017.10.016.
17