scispace - formally typeset
Search or ask a question
Journal ArticleDOI

Automatic Creation of Semantically Rich 3D Building Models from Laser Scanner Data

01 May 2013-Automation in Construction (Elsevier)-Vol. 31, pp 325-337
TL;DR: A method to automatically convert the raw 3D point data from a laser scanner positioned at multiple locations throughout a facility into a compact, semantically rich information model that is capable of identifying and modeling the main visible structural components of an indoor environment despite the presence of significant clutter and occlusion.
About: This article is published in Automation in Construction.The article was published on 2013-05-01 and is currently open access. It has received 576 citations till now. The article focuses on the topics: Laser scanning & Point cloud.

Summary (2 min read)

1. INTRODUCTION

  • In the Architecture, Engineering, and Construction (AEC) industry, semantically rich 3D models are increasingly used throughout a building’s lifecycle, from the design phase, through construction, and into the facility management phase.
  • Currently, as-is BIMs are created through a manual process, typically using data from laser scanners as input [1].
  • Applied to all the rooms in a building, their method will automatically produce a compact, semantically rich, 3D model which, while not strictly a BIM in the traditional sense, contains the geometric and identity information that substantially makes up the BIM.
  • The authors algorithm addresses the challenges of clutter and occlusion by explicitly reasoning about them throughout the process.
  • These two phases are described in more detail in the next two sections.

2. CONTEXT-BASED MODELING

  • Distinguishing between different types of objects and between clutter and non-clutter can be difficult or impossible if those objects are seen only in isolation.
  • The more easily recognized structures can provide the scaffolding that enables the recognition of other, more challenging instances.
  • This phase of the algorithm consists of four steps .
  • Patches are found using a region growing algorithm to connect nearby points that have similar surface normals and that are welldescribed by a planar model.
  • The result of this process is a compact model of the walls, floor, and ceiling of a room, with each patch labeled according to its type.

3. DETAILED SURFACE MODELING

  • The surfaces produced by the first phase represent idealized surfaces that are perfectly planar and unoccluded and with no openings.
  • By tracing a line from the origin of the laser scanner to a detected measurement, a pixel is classified as empty if the ray passes through the plane of the patch, as occupied if the ray stops approximately at the plane, and occluded if the ray stops before reaching the plane.
  • Detecting openings in unoccluded surfaces can be achieved by analyzing the data density and classifying low density areas as openings.
  • The authors use a number of features as input to the SVM, including opening width, height, and distance from the boundaries of the patch (sides, and top, and bottom).
  • This step is not strictly necessary, but it improves the visualization of the results .

4. EXPERIMENTAL RESULTS

  • The authors conducted experiments using data from a building that was manually modeled by a professional laser scanning service provider.
  • The patch detection resulted in 389 patches in the interior of the building.
  • The highest confusion was between walls and clutter.
  • The authors experiments suggest that the context aspect of their algorithm improves recognition performance by about 6% and that the most useful contextual features are coplanarity and orthogonality.
  • Figure 5 shows some example results, and Figure 6 shows one floor of the entire reconstructed model.

Did you find this useful? Give us your feedback

Figures (6)
Citations
More filters
Journal ArticleDOI
TL;DR: Results show scarce BIM implementation in existing buildings yet, due to challenges of (1) high modeling/conversion effort from captured building data into semantic BIM objects, (2) updating of information in BIM and (3) handling of uncertain data, objects and relations in B IM occurring inexisting buildings.

1,499 citations

Proceedings ArticleDOI
27 Jun 2016
TL;DR: This paper argues that identification of structural elements in indoor spaces is essentially a detection problem, rather than segmentation which is commonly used, and proposes a method for semantic parsing the 3D point cloud of an entire building using a hierarchical approach.
Abstract: In this paper, we propose a method for semantic parsing the 3D point cloud of an entire building using a hierarchical approach: first, the raw data is parsed into semantically meaningful spaces (e.g. rooms, etc) that are aligned into a canonical reference coordinate system. Second, the spaces are parsed into their structural and building elements (e.g. walls, columns, etc). Performing these with a strong notation of global 3D space is the backbone of our method. The alignment in the first step injects strong 3D priors from the canonical coordinate system into the second step for discovering elements. This allows diverse challenging scenarios as man-made indoor spaces often show recurrent geometric patterns while the appearance features can change drastically. We also argue that identification of structural elements in indoor spaces is essentially a detection problem, rather than segmentation which is commonly used. We evaluated our method on a new dataset of several buildings with a covered area of over 6, 000m2 and over 215 million points, demonstrating robust results readily useful for practical applications.

1,320 citations


Cites background from "Automatic Creation of Semantically ..."

  • ...[40] also employs similar planar surface assumption in order to estimate the semantics of a single room using contextual information....

    [...]

Journal ArticleDOI
TL;DR: An overview of recent advances in computer vision techniques as they apply to the problem of civil infrastructure condition assessment and some of the key challenges that persist toward the goal of automated vision-based civil infrastructure and monitoring are presented.

500 citations


Cites methods from "Automatic Creation of Semantically ..."

  • ...[154] investigated an automated method for converting dense 3D pointcloud data from a room to a semantically rich 3Dmodel represented by planar walls, floors, ceilings, and rectangular openings (a process referred to as Scan-to-Building Information Modeling (BIM))....

    [...]

Journal ArticleDOI
TL;DR: The research presented in this paper combines the Hough transform and “Scan-vs-BIM” systems in a unified approach for more robust automated comparison of as-built and as-planned cylindrical MEP works, thereby providing the basis for automated earned value tracking, automated percent-built-as-planned measures, and assistance for the delivery of as -built BIM models from as-designed ones.

358 citations

Journal ArticleDOI
TL;DR: A scientometric review of global BIM research in 2005–2016, through co-author analysis, co-word analysis and co-citation analysis provides researchers and practitioners with an in-depth understanding of the status quo and trend of the BIMResearch in the world.

277 citations

References
More filters
Journal ArticleDOI
TL;DR: The conclusion is that for almost any real-world generalization problem one should use some version of stacked generalization to minimize the generalization error rate.

5,834 citations

Proceedings ArticleDOI
01 Jul 2000
TL;DR: A novel algorithm for digital inpainting of still images that attempts to replicate the basic techniques used by professional restorators, and does not require the user to specify where the novel information comes from.
Abstract: Inpainting, the technique of modifying an image in an undetectable form, is as ancient as art itself. The goals and applications of inpainting are numerous, from the restoration of damaged paintings and photographs to the removal/replacement of selected objects. In this paper, we introduce a novel algorithm for digital inpainting of still images that attempts to replicate the basic techniques used by professional restorators. After the user selects the regions to be restored, the algorithm automatically fills-in these regions with information surrounding them. The fill-in is done in such a way that isophote lines arriving at the regions' boundaries are completed inside. In contrast with previous approaches, the technique here introduced does not require the user to specify where the novel information comes from. This is automatically done (and in a fast way), thereby allowing to simultaneously fill-in numerous regions containing completely different structures and surrounding backgrounds. In addition, no limitations are imposed on the topology of the region to be inpainted. Applications of this technique include the restoration of old photographs and damaged film; removal of superimposed text like dates, subtitles, or publicity; and the removal of entire objects from the image like microphones or wires in special effects.

3,830 citations

Journal ArticleDOI
01 Jul 2006
TL;DR: This work presents a system for interactively browsing and exploring large unstructured collections of photographs of a scene using a novel 3D interface that consists of an image-based modeling front end that automatically computes the viewpoint of each photograph and a sparse 3D model of the scene and image to model correspondences.
Abstract: We present a system for interactively browsing and exploring large unstructured collections of photographs of a scene using a novel 3D interface. Our system consists of an image-based modeling front end that automatically computes the viewpoint of each photograph as well as a sparse 3D model of the scene and image to model correspondences. Our photo explorer uses image-based rendering techniques to smoothly transition between photographs, while also enabling full 3D navigation and exploration of the set of images and world geometry, along with auxiliary information such as overhead maps. Our system also makes it easy to construct photo tours of scenic or historic locations, and to annotate image details, which are automatically transferred to other relevant images. We demonstrate our system on several large personal photo collections as well as images gathered from Internet photo sharing sites.

3,398 citations


"Automatic Creation of Semantically ..." refers methods in this paper

  • ...Approaches based on (Snavely et al., 2006) have been shown to be useful for visualizing construction progress by overlaying design models with imagery (Golparvar-Fard et al....

    [...]

  • ...More recently, multi-view image-matching techniques have been employed to perform largescale structure-from-motion on large image collections (Snavely et al., 2006; Agarwal et al., 2009)....

    [...]

Proceedings ArticleDOI
01 Aug 1996
TL;DR: This work presents a new approach for modeling and rendering existing architectural scenes from a sparse set of still photographs, which combines both geometry-based and imagebased techniques, and presents view-dependent texture mapping, a method of compositing multiple views of a scene that better simulates geometric detail on basic models.
Abstract: We present a new approach for modeling and rendering existing architectural scenes from a sparse set of still photographs. Our modeling approach, which combines both geometry-based and imagebased techniques, has two components. The first component is a photogrammetricmodeling method which facilitates the recovery of the basic geometry of the photographed scene. Our photogrammetric modeling approach is effective, convenient, and robust because it exploits the constraints that are characteristic of architectural scenes. The second component is a model-based stereo algorithm, which recovers how the real scene deviates from the basic model. By making use of the model, our stereo technique robustly recovers accurate depth from widely-spaced image pairs. Consequently, our approach can model large architectural environments with far fewer photographs than current image-based modeling approaches. For producing renderings, we present view-dependent texture mapping, a method of compositing multiple views of a scene that better simulates geometric detail on basic models. Our approach can be used to recover models for use in either geometry-based or image-based rendering systems. We present results that demonstrate our approach’s ability to create realistic renderings of architectural scenes from viewpoints far from the original photographs. CR Descriptors: I.2.10 [Artificial Intelligence]: Vision and Scene Understanding Modeling and recovery of physical attributes; I.3.7 [Computer Graphics]: Three-Dimensional Graphics and Realism Color, shading, shadowing, and texture I.4.8 [Image Processing]: Scene Analysis Stereo; J.6 [Computer-Aided Engineering]: Computer-aided design (CAD).

2,159 citations


"Automatic Creation of Semantically ..." refers methods in this paper

  • ...Early work by Debevec used images to semi-automatically model building exteriors (Debevec et al., 1996)....

    [...]

Book
01 Jan 2003

1,814 citations


"Automatic Creation of Semantically ..." refers background in this paper

  • ...We find the optimal value for w using stochastic gradient ascent (Bertsekas et al. (2003))....

    [...]

Frequently Asked Questions (11)
Q1. What are the contributions in "Automatic creation of semantically rich 3d building models from laser scanner data" ?

This paper presents a method to automatically convert the raw 3D point data from a laser scanner positioned at multiple locations throughout a building into a compact, semantically rich model. Then, the authors perform a detailed analysis of the recognized surfaces to locate windows and doorways. The authors evaluated the method on a large, highly cluttered data set of a building with forty separate rooms yielding promising results. 

Their experiments suggest that the context aspect of their algorithm improves recognition performance by about 6% and that the most useful contextual features are coplanarity and orthogonality. 

In the first phase, planar patches are extracted from the point cloud and a context-based machine learning algorithm is used to label the patches as wall, ceiling, floor, or clutter. 

The detailed surface modeling phase of the algorithm operates on each planar patch produced by the contextbased modeling process, detecting the occluded regions and regions within openings in the surface. 

A learning algorithm is used to encode the characteristics of opening shape and location, which allows the algorithm to infer the shape of an opening even when it is partially occluded. 

The authors are currently working on completing the points-to-BIM pipeline by implementing an automated method to convert the surface-based representation produced by their algorithm into a volumetric representation that is commonly used for BIMs. 

Detecting openings in unoccluded surfaces can be achieved by analyzing the data density and classifying low density areas as openings. 

The classifier uses local features computed on each patch in isolation as well as features describing the relationship between each patch and its nearest neighbors. 

These models, which are generally known as building information models (BIMs), are used for many purposes, including planning and visualization during the design phase, detection of mistakes made during construction, and simulation and space planning during the management phase. 

The result of this process is a compact model of the walls, floor, and ceiling of a room, with each patch labeled according to its type. 

Building modeling algorithms are frequently demonstrated on simple examples like hallways that are devoid of furniture or other objects that would obscure the surfaces to be modeled.