Next Article in Journal
Quantitative Infrared Thermography to Evaluate the Humidification of Lightweight Concrete
Previous Article in Journal
Diagnosis of Occlusal Caries with Dynamic Slicing of 3D Optical Coherence Tomography Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comparison of Different Remote Sensing Methods for 3D Modeling of Small Rock Outcrops

1
Department of Forest Management and Applied Geoinformatics, Faculty of Forestry and Wood Technology, Mendel University in Brno, 613 00 Brno, Czech Republic
2
Department of Geology and Pedology, Faculty of Forestry and Wood Technology, Mendel University in Brno, 613 00 Brno, Czech Republic
*
Author to whom correspondence should be addressed.
Submission received: 7 February 2020 / Revised: 11 March 2020 / Accepted: 14 March 2020 / Published: 17 March 2020
(This article belongs to the Section Optical Sensors)

Abstract

:
This paper reviews the use of modern 3D image-based and Light Detection and Ranging (LiDAR) methods of surface reconstruction techniques for high fidelity surveys of small rock outcrops to highlight their potential within structural geology and landscape protection. LiDAR and Structure from Motion (SfM) software provide useful opportunities for rock outcrops mapping and 3D model creation. The accuracy of these surface reconstructions is crucial for quantitative structural analysis. However, these technologies require either a costly data acquisition device (Terrestrial LiDAR) or specialized image processing software (SfM). Recent developments in augmented reality and smartphone technologies, such as increased processing capacity and higher resolution of cameras, may offer a simple and inexpensive alternative for 3D surface reconstruction. Therefore, the aim of the paper is to show the possibilities of using smartphone applications for model creation and to determine their accuracy for rock outcrop mapping.

Graphical Abstract

1. Introduction

Light Detection and Ranging (LiDAR) and Structure from Motion (SfM) software provide large amounts of digital data from which virtual outcrops can be created. The accuracy of these surface reconstructions is crucial for quantitative structural analysis [1]. The use of photogrammetry and laser scanning significantly increased the possibilities of documentation of geological structures and archaeological monuments.
In [2] the Structure from Motion approach and the Dense Image Matching (DIM) or Multi View Stereo (MVS) algorithms are characterized by five main steps, which require, as their input data, only the acquired images:
  • Feature extraction and identification of the matching points between images through different algorithms, such as the Scale Invariant Feature Transform (SIFT) algorithm [3], or its modified version [4];
  • Outliers filtering, exploiting the Random Sample Consensus (RANSAC) algorithm [5], and a robust estimation of the acquisition geometry through linear models that impose geometric constraints;
  • Estimate of the 3D coordinates of the object, internal and external orientation and the relative statistical information concerning the accuracy of the calculation, through the Bundle-Adjustment technique [6];
  • Dense surface reconstruction of the point cloud identified for the orientation, exploiting several algorithms, including the Exact, Smooth, Height-field and Fast methods, for the object surface description and its main discontinuities [7];
  • Dense point cloud interpolation and texturing in order to obtain a photorealistic visualization.
The SfM method offers an economical and simple solution for many industries, including geological sciences. Close-range terrestrial photogrammetry and computer vision approach is suitable to obtain high-resolution spatial data suitable for modelling meso and micro-scale landforms [8]. SfM produces models in which a high-resolution photograph is fully integrated with the underlying 3-D surface and is therefore suitable for rock slopes mapping [9]. This method is particularly useful at steep dam abutments as an alternative to rope access and can provide necessary accuracy for detailed analyses [10]. In [11] the authors recommend wherever possible the use of remote sensing derived 3D models in combination with high resolution photographs of the rock cuts. In tunnels, SfM is used to characterize underground rocks [12], to monitor deformations, displacements and to analyze tunnel stability [13,14]. Furthermore, the photogrammetric SfM method has its place in archeology [15]. Stones and rocks can be also part of cultural heritage and the method of 3D modelling can improve the archeological research. In [16] carved stones modelling by SfM was tested to obtain photo-realistic 3D models of the stones.
The Terrestrial Laser Scanning (TLS) provides a high-density point cloud, up to thousands of points per square meter, depending on the setting parameters. During a short period of time, large numbers of high-quality data can be obtained to characterize the object of interest. TLS has a wide range of applications, especially in architecture and construction [17], but it can also be exploited in geosciences. Examples include archeology or cultural heritage documentation [18,19,20,21,22,23], or speleology [24,25]. In recent years, TLS has increasingly been used in geomorphological studies to capture more detailed features of given objects and phenomena, such as riverbeds and sediments [26], landslides [27,28,29] or erosion [30,31]. TLS data also allows to derive and illustrate the orientation and slope of rock hillsides [32], respectively destructive phenomena such as falling rocks and avalanches [33,34], debris currents (debris flows) [35] or the boulders instability [36].
With the rapid development of mobile devices, especially smartphones, the computing capacity and data collection through digital photography have been growing. The resolution of the latest smartphone cameras is commonly from 48 to 108 MPix and increasing camera sensor size allows image capturing even in the dark conditions. Three-dimensional point cloud creation from SfM is included in number of photogrammetric applications based on image processing in smartphones [2,37]. Images and processed 3D models are thus useful for mapping of smaller objects such as rock formations, roads or individual tree trunks in forest stands. One of SfM app, commonly used for scientific research, is SCANN3D [2,37]. SCANN3D locally performs the 3D reconstruction, not on the cloud, in quite a rapid way, working entirely offline. The main benefit of the application is the real-time guidance during the image capturing phase, tracking multiple points in the viewpoints, which go from red to green when the camera has been moved sufficiently, ensuring a good image overlap. However, this application does not lock the focus and exposure after acquiring the first image, and this aspect can negatively affect the 3D reconstruction [2].
The Google Tango is technology, that enables to measure and create the models of objects directly using a smartphone. It uses infrared cameras (the RGB-D sensor) to measure image depth. The technology combined three basic principles: image depth estimation based on RGB-D sensor, motion mapping based on inertial navigation (accelerometers and gyroscopes) and so-called SLAM technology [38,39], which can refine the position based on recognized objects in space. The technology has been successfully tested for forest inventory [40,41]. However, Tango was not fully utilized by the market, as the sensor had a high battery consumption and was redundant for the common user. Since 2017, Google has stopped supporting the Tango project and has started offering ARCore technology, which partially replaces it.
ARCore is Google platform for augmented reality experiences building. Using different Application Programming Interfaces (API), ARCore enables a smartphone to sense its environment, understand the world and interact with information. Some of the APIs are available across Android and iOS to enable shared ARCore experiences. ARCore uses three key capabilities to integrate virtual content with the real world as seen through phone’s camera [42]:
  • Motion tracking allows the phone to understand and track its position relative to the world;
  • Environmental understanding allows the phone to detect the size and location of all type of surfaces: horizontal, vertical and angled like the ground, coffee table or walls;
  • Light estimation allows the phone to estimate the environment current lighting conditions.
ARCore’s motion tracking relies on both visual information from the camera and inertial measurements from the device’s inertial measurement unit (IMU). ARCore uses a process called concurrent odometry and mapping, or COM, to understand where the phone is relative to the world around it. ARCore detects visually distinct features in the captured camera image called feature points and uses these points to compute its change in location. The visual information is combined with inertial measurements from the device’s IMU to estimate the pose (position and orientation) of the camera relative to the world over time [43].
Although majority of new mobile devices support this technology, there are not too many applications on the market that allow direct mapping of larger objects. Available applications focus only on smaller objects scanning, such as statues or furniture or building interiors, except 3D Scanner for the ARCore app. ARCore’s 3D Scanner uses code from the Tango project library, but it replaces pixel depth calculation with ArCore technology. The 3D Scanner for ARCore uses a sequence of images, which is connected using the SfM algorithm [44] and emulating key points from ARCore. Thanks to the emulation, any mobile phone whose camera supports ARCore technology can be used to scan an object. After the scan is completed, the models are filled using the Poisson reconstruction and texturing using the Tango library. The application is primarily intended for scanning flats and interiors. Prerequisite for good quality results is to dispose of an object with quite complicated surface texture. The main advantage of the application is the reconstruction of objects in real time with very short modeling time depending on the mobile phone performance. Newer application versions also support Time-of-Flight cameras (ToF), which measure the depth of each pixel based on the time of radiation and radiation reception, and primarily serve for accurate focusing [45].
The possibility of reducing the costs of the survey with low-cost equipment, such as digital and amateur cameras, smartphones, tablets and action cams, is one of the main advantages of the photogrammetric technique, which also allows for a high repeatability of the survey [2]. Therefore, the main aim of this research is to compare different remote sensing methods for 3D modeling of small rock outcrops on the example of two boulders and to answer the question whether modern technologies of 3D modeling using modern smartphones can offer a simpler and cheaper method for 3D reconstructions of small rock outcrops and other similar objects in other applications, such as archeology. The results of the study should answer whether smartphone applications can be an alternative to conventional data collection methods such as TLS and close-range photogrammetry with sufficient accuracy without the need of expensive instruments (TLS) or professional SfM software.

2. Materials and Methods

2.1. Study Case

Rock formations in the vicinity of Přečkov are one of the most visited locations in the Třebíčsko Nature Park (Figure 1). They have the character of boulders that are not firmly anchored in the geological bedrock, as it makes them less resistant to natural and anthropic dynamic processes, resulting in rather significant morphological changes. Within a few decades, they are often perceptibly shifted, damaged or destroyed. Three-Dimensional Scanning Technology is a unique method in this case of rock outcrops that greatly supports the observation of relatively rapid changes in their macro and micro shapes, stability and changes in the overall distribution. This monitoring is important for the necessary inventory of this area and for the geodiversity protection and preservation in this part of the Třebíč Nature Park as a valuable natural asset, which has been primarily endangered by anthropic activity. In the past, the rock outcrops were used as an accessible building material and for the colostrum production by burning a pitched pine wood. To do this, it was necessary to modify them by hollowing out the bowl and then to keep the fire and grooves out of the drain. At present, rock outcrops are a very popular place for bouldering—rock climbing without protection at low altitudes, as the surface microstructure is damaged by this activity [46].
In order to compare remote sensing methods, it was necessary to select suitable and easily accessible rock outcrops of a size that would be easy to capture using a smartphone (Figure 2). The selected boulder No. 1 lies on a small hill and is surrounded by pine forest (Figure 3). It´s size is approximately 5 × 4 × 3 m and occupies an area of about 20 m2. The selected boulder No. 2 is located at the edge of the ridge above the watercourse at the edge of the forest clearing (Figure 3). It has more complex shape with a size of approximately 6 × 5 × 4 m and occupies an area of about 25 m2.

2.2. Data Acqusition

Rock formations mapping was performed on 5th December 2018. Technologies of TLS by the FARO FOCUS 3D scanner, imaging using the integrated Xiaomi MI 8 smartphone camera for further SfM processing, 3D Scanner for ARCore and SCANN3D apps were used to create the models. For the purpose of subsequent model referencing, 3 control points near the object were laid out and measured in local coordinate system by total station Trimble M3 for each rock outcrop so that they were visible on the models. Wooden stakes of 5 × 5 cm cross-section were used as control points.

2.2.1. Terrestrial Laser Scanning

The Faro Focus 3D static panoramic scanner was used for the measurement (Table 1). Before that process, basic scanning parameters were set. The resolution 1/8 was chosen, which equals a point spacing of 24 mm at distance of 10 m from the scanner with scanning speed of 122,000 points per second. This point density is sufficient to create an accurate 3D model. Scanning was performed from 6 stations at a distance less than 10 m. The total of 6 spherical spheres with diameter of 20 cm were placed around the rock outcrops for subsequent alignment of individual scans. Individual scans were processed in Faro Scene version 5.3 software by Faro Technologies. The following process consisted in combining individual scans into point cloud based on spherical spheres (separately for each object). The mean registration error of the object Boulder No. 1 was 0.0018 m with a deviation of 0.0015 m, the second object (Boulder No. 2) mean error was 0.0011 m with a deviation of 0.0012 m. No point cloud coloring by images was used to speed up the scanning process. The final model was exported in LAS format.

2.2.2. Close-Range Photogrammetry with SfM Processing in AGISOFT PhotoScan Professional

Both objects were captured by smartphone Xiaomi Mi 8 (Table 2) in the automatic camera mode by gradually bypassing in a circle so that the images overlapped about 70% for successful processing into a model. Altogether, 73 images were taken for the boulder No. 1 and 111 images for the boulder No. 2 (Figure 4). Depending on the incoming light, smartphone camera automatically changed some parameters during data acquisition (ISO about 100, shutter from 1/100 to 1/2000). Images were downloaded from the smartphone and uploaded to AGISOFT PhotoScan software. During processing, autocalibration of cameras and highest quality of alignment and dense cloud generation were used. To determine the scale and position in the coordinate system, it was necessary to mark the control points in the images in advance and enter their coordinates. Subsequently, the model was stored in LAS format for further processing.

2.2.3. 3D Scanner for ARCore

The 3D Scanner for ARCore is an app for Android smartphones that uses the same code base for 3D scanning as the app for Tango. The main difference in these two apps is in providing depth data. The Tango version uses an active depth sensor and ARCore uses passive depth sensing [44]. The mapping was performed using the Xiaomi MI8 with a dual camera by bypassing boulders around with continuous data acquisition. The app uses an automatic regime of camera with no options of settings. The advantage of app is the real-time pre-modeling and the visibility of the scanned surface on the phone display. This allows you to monitor scanned locations and fill in gaps in the model. Both created 3D models were exported to OBJ format for further processing.

2.2.4. SCANN3D App

SCANN3D was developed by SmartMobileVision company and is available only for Android. SCANN3D deploys patent pending photogrammetry technology to enable true 3D model capture and reconstruction for smartphones and tablets. The device becomes a standalone tool to turn images into 3D models—all images are processed by and on it. The resulting 3D models can be stored, shared, and edited by third party applications, and can be used in augmented or virtual reality applications [47]. SCANN3D performs the photogrammetry locally, not in the cloud. SCANN3D supports different formats for further processing (e.g., STL, PLY, OBJ). The mapping was again performed using the Xiaomi MI8 with a dual camera by bypassing rock outcrops around with continuous data acquisition and final models were exported to OBJ format. The application uses the automatic camera mode without the ability to set image properties.2.3. Data Processing and Model Comparison
Exported models from all four methods were processed and compared in the Cloud Compare software and then in ArcGIS for Desktop 10.6.1. In the Cloud Compare all models were edited and aligned to the local coordinate system based on measured control points. Coordinates of control points were assigned to positions of wooden stakes identified in the models. Surface area and volume were calculated, cross-sectioning in 3 axes, as well as comparing the distance between surfaces were performed. It was also necessary to remove terrain points from all models manually so that the resulting model contained only objects of rock outcrops.
In the case of TLS data, created point cloud already had its own scale and altitude determined by the built-in barometric sensor (altimeter). This model can be used directly to calculate the size and volume of an object. Since TLS is a proven method and provides very accurate results, the model created from TLS was used as a reference model for comparison with the other methods (Figure 5).
During the image processing in the AGISOFT software, the coordinates of the control points were entered, so the resulting model had its scale again and was immediately usable for further calculations (Figure 6). The pictures show that the dense clouds on the top of the models have gaps that are caused by insufficient photo overlap.
The model created in 3DScanner for ArCore is also usable almost without modification, as it has its own scale due to the smartphone accelerometer and gyroscope information. It was necessary to rotate the model manually, because the coordinate system axes were interchanged (Figure 7).
The SCANN3D model has no scale and was significantly smaller than other models when it was opened in Cloud Compare. For this reason, it was aligned to the local coordinate system based on the control points coordinates (Figure 8). Figure 5 shows model errors that are caused by overexposure of the first model; in the second model, the shaded areas were underexposed as the model could not be calculated in some places.
The Compute 2.5 Volume tool from Cloud Compare software was used to calculate the volume, surface area and maximum height of rock outcrops. This tool calculates an interpolated raster model with an average point height above the reference plane. To compute the volume Cloud Compare sums the contribution of each cell. This contribution is simply the volume of the elementary parallelepiped corresponding to the cell footprint multiplied by the difference in heights (dV = grid step * grid step * difference of height). Only the cells which have a valid height value for both the ‘ground’ and the ‘ceil’ are used for the global volume estimation. Empty cells or cells for which only one height value is defined are simply ignored [48]. Resolution of 1 cm was chosen for the calculation. To create slices of all models for comparison, it was necessary to align the models so that the slices correspond positively to each other. For this purpose, the models were firstly manually aligned based on 3 identical points and then accurately registered using ICP registration (Interactive Closest Point Registration) [49]. The models were not scaled during registration to maintain their original size and shape. After ICP registration, the Cross Section tool was used to create 3 slices for each model with the largest circumference in the X, Y, and Z axes (Figure 9). These slices were then exported to shapefile format and their perimeter and area were calculated in ArcGIS software. Furthermore, model differences were calculated based on the mutual distance of individual points compared to the reference (TLS) model using the Cloud/Cloud Distance tool.
Models were compared using several methods. At first, the calculated volumes and surfaces were compared with the reference models (TLS) and the deviation values were expressed in units and in % compared to the TLS model. In the next step, perimeters and areas of cross sections in 3 axes were compared.

3. Results

3.1. Boulder No. 1

In the case of boulder No. 1, the Sfm processing method in AGISOFT PhotoScan achieved the best results as it is professional photogrammetry software and good quality results were expected. The smallest deviations from the TLS model were achieved both when comparing the surface, volume and height of the object (−0.3%, 0.6% and 2.7%) as well as when comparing the perimeter and area of the sections (Table 3, Table 4, Table 5 and Table 6). Comparison of the distance between points and the reference model shows the best results in the case of the 3DScanner for ARCore application with an RMSE of 0.045 m (Table 7). However, in all cases the deviations are relatively small, only the model from SCANN3D achieved lower accuracy (Figure 10).

3.2. Boulder No. 2

In the case of boulder No. 2, the results are not so obvious. The 3DScanner for ARCore application, which achieves deviations of up to 5%, is the best based on volume and surface model comparisons. Other methods are less accurate due to imaging conditions (excessive sun exposure of the southern part and shade in the north) as the images could not be optimally aligned (Table 8, Table 9, Table 10 and Table 11). Taking the cross sections and the distance of points from the reference model in comparison, SCANN3D, which most accurately captures the shape of the object, works best (Table 12). However, the upper and northern part of the boulder is missing and therefore it is not the best to assess the surface and volume (Figure 11).

3.3. Overall Results

The overall results show that all methods have very small deviations from TLS in percentage units in all observed parameters. Comparison of generated models from different methods with the TLS reference model shows that almost all methods achieve a volume difference of up to 6 m3 (10% respectively), with the biggest error in SCANN3D (15 m3 and 24% respectively). In the case of the surface, a difference of up to 2 m2 (6%) was achieved, again with the greatest error in SCANN3D application (almost 6 m2 and 19%). At maximum height, errors are even less than 20 cm (7%). Comparison of cross sections circumference resulted in errors up to 2 m and 10%, in case of cross section area similarly about 2 m2 and 10%. A deviation of up to 13 cm was found in all models when comparing the distance between surfaces.

4. Discussion and Conclusions

The results of comparison of different remote sensing methods for 3D modeling of small rock outcrops are not entirely clear, and it is not possible to state which mapping method is optimal. Nevertheless, it can be stated that all methods are usable for the purpose of creating rock outcrop models. These surprisingly good results in all image-based methods used are mainly due to the texture of the rocks, as the high texture ensures finding of a high number of tie points. The accuracy of a 3D model depends on the amount of feature points extracted from the captured set of images. If a surface being captured is texture rich, the quality of the produced data would be higher in comparison with poorly textured surfaces. The problem of feature extraction from a texture-less surface is mentioned in many previous studies [50,51,52]. The limiting factors are mainly light conditions during imaging because all methods are passive and are based on the visible part of the electromagnetic spectrum. Good quality model processing is also dependent on the overlay of images, which puts demands during the manual imaging process. In the case of ready-to-use apps like SCANN3D and 3D Scanner for ARCore is process driven by app itself. Compared to TLS, they do not require expensive instruments and achieve very similar results. However, when processing photos in SCANN3D App and AGISOFT PhotoScan, the disadvantage is the necessity to measure at least 3 control points to scale and position in the coordinate system. The 3DScanner for ARCore is the only application that works completely independently, allowing to create a scale model that is directly applicable for further use. Another advantage is the creation of the model in real time, when it is possible to check calculated model during the scanning and fill in the places with the gaps. Ready-to-use applications have limitations primarily in terms of the size of the subject. In the case of rock formations with detailed texture, it is possible to create models of different sized objects, especially depending on accessibility, as these applications require shorter distance to capture feature points. Based on the results presented here, smartphone apps should be restricted to objects with maximum height of 5 m, but the particular constraint is due to the object’s accessibility rather than its size.
Although there is number of 3D scanning applications for smartphones on the market, there are not many studies dealing with the accuracy of the created models of large objects. Similar studies include e.g., [50], where authors compared cloud processing of images taken with different types of smartphones for the purposes of archeological documentation of buildings with the conclusion that the models created are accurate enough, but there are errors at the edges of the model up to 50 cm. Another study [51] focused on reliability of smartphones that incorporate 3D depth sensors for 3D reconstruction of cultural heritage objects. The authors used Google Tango technology and compared results with photogrammetry. The results are very similar as [53], showing the larger deviations on the edges of the model where no perfect match was achieved. The model produced by photogrammetry was of higher quality than the Tango-based one. In [54] the authors concluded that it is possible to obtain high-quality results from numerous images of building captured by smartphone cameras using appropriate software solutions. Very similar results were gained in [55], comparison of riverbank models from TLS and close-range photogrammetry resulting in RMSE from 0,02–0,05 m, depending on the device used (digital single-lens camera x smartphone). In [2] authors tried to create model of the interior of cave by SCANN3D app with conclusion, that due to the higher number of images and to the particularly poor light conditions, the application was able to reconstruct the models only partially, with a quality that was not excellent in terms of the number of points and completeness of the model, even when setting the highest quality level.
Due to the rapid development, in near future, it will be possible to count on further improvements in the 3D data collection using smartphones. In many cases, new smartphones for 2020 already have another (mostly third or fourth) camera for determining the depth of images (ToF cameras—Time of Flight), which should further refine the possibility of creating 3D models according to manufacturer´s information. In [56] the authors tested the ToF camera system with different objects to check the surface of reconstruction and accuracy evaluation. The results have demonstrated that the proposed technique is feasible for dense 3D measurement applications. In [57] the authors developed a machine vision system consisting of RGB and ToF camera to estimate the size of apples in tree canopies. With this system, they were able to not only measure the size of apple fruits but even to calculate the local coordinates of each fruit. For example, the 3D Scanner for ARCore app also allows the use of ToF camera in the new version. The disadvantage is the lower resolution of ToF cameras. However, phones for the year 2020 can have up to 5 Mpix ToF cameras and can improve the data acquisition. As a result, smartphones can replace expensive instruments such as Terrestrial Laser Scanners as well as photogrammetric cameras and desktop software for photogrammetric processing.

5. Conclusions

In this work, we compared different remote sensing methods for 3D modeling of small rock outcrops on the example of two boulders. Four different methods were used for 3D reconstruction: TLS, desktop image processing with SfM and two smartphone apps SCANN3D and 3D Scanner for ARCore. The model size was compared with the TLS reference model. All methods have proved the possibility of using for this purpose 3D modeling. The 3D scanner for ARCore seems to be the best and easiest method because it does not require additional scale and georeferencing of the model with sufficient accuracy. The results confirmed that smartphones can offer cheaper and simpler method of 3D reconstruction than expensive instruments such as Terrestrial Laser Scanners as well as photogrammetric cameras and desktop software for photogrammetric processing. Future research should address the usage of ToF cameras in smartphone for faster and more accurate 3D modeling.

Author Contributions

A.B. and T.M. initiated the study; T.M. conducted the GIS and statistical analysis and drafted the manuscript; M.C. carried out terrestrial laser scanning and processed the data; T.M., A.B. carried out photogrammetry and TLS data and their processing; Z.P. carried out and reviewed the statistical analysis; M.B. and A.B. revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by Internal Grant Agency of Faculty of Forestry and Wood Technology, Mendel University in Brno by a project LDF_VP_2018033 “Comparison of modern non-contact methods for forest inventory”.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cawood, A.J.; Bond, C.E.; Howell, J.A.; Butler, R.W.H.; Totake, Y. LiDAR, UAV or compass-clinometer? Accuracy, coverage and the effects on structural models. J. Struct. Geol. 2017, 98, 67–82. [Google Scholar] [CrossRef]
  2. Dabove, P.; Grasso, N.; Piras, M. Smartphone-Based Photogrammetry for the 3D Modeling of a Geomorphological Structure. Appl. Sci. 2019, 9, 3884. [Google Scholar] [CrossRef] [Green Version]
  3. Lowe, D.G. Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
  4. Farenzena, A.M.; Fusiello, A.; Gherardi, R. Structure-and-Motion Pipeline on a Hierarchical Cluster Tree. In Proceedings of the IEEE International Workshop on 3-D Digital Imaging and Modeling, Kyoto, Japan, 27 September–4 October 2009. [Google Scholar]
  5. Fischler, M.A.; Bolles, R.C. Random Sample Consensus: A Paradigm for Model Fitting with Applications to Image Analysis and Automated Cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  6. Triggs, B.; McLauchlan, P.F.; Hartley, R.I.; Fitzgibbon, A.W. Bundle adjustment—A modern synthesis. In International Workshop on Vision Algorithms; Springer: Berlin/Heidelberg, Germany, 1999; pp. 298–372. [Google Scholar]
  7. Furukawa, Y.; Ponce, J. Accurate, dense, and robust multiview stereopsis. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 1362–1376. [Google Scholar] [CrossRef]
  8. Westoby, M.; Brasington, J.; Glasser, N.; Hambrey, M.; Reynolds, J. ‘Structure-from-Motion’ photogrammetry: A low-cost, effective tool for geoscience applications. Geomorphology 2012, 179, 300–314. [Google Scholar] [CrossRef] [Green Version]
  9. Haneberg, W.C. Using close range terrestrial digital photogrammetry for 3-D rock slope modeling and discontinuity mapping in the United States. Bull. Eng. Geol. Environ. 2008, 67, 457–469. [Google Scholar] [CrossRef]
  10. Krosley, L.K.; Shaffner, P.T.; Oerter, E.; Ortiz, T. Digital ground-based photogrammetry for measuring discontinuity orientations in steep rock exposures. In Proceedings of the 41st U.S. Symposium of Rock Mechanics, Golden, Colorado, 17–21 June 2006; pp. 1–13. [Google Scholar]
  11. Sturznegger, M.; Stead, D. Close-range terrestrial digital photogrammetry and terrestrial laser scanning for discontinuity characterization on rock cuts. Eng. Geol. 2009, 106, 163–182. [Google Scholar] [CrossRef]
  12. Senent, S.; Jimenez, R. Structure from Motion photogrammetry to characterize underground rock masses: Experiences from two real tunnels. Tunn. Undergr. Space Technol. 2019, 83, 262–273. [Google Scholar] [CrossRef]
  13. Roncella, R.; Umili, G.; Forlani, G. A novel image acquisition and processing procedure for fast tunnel DSM production. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2012, 297–302. [Google Scholar] [CrossRef] [Green Version]
  14. Chaiyasarn, K.; Kim, T.; Viola, F.; Cipolla, R.; Soga, K. Distortion-free image mosaicing for tunnel inspection based on robust cylindrical surface estimation through structure from motion. J. Comput. Civ. Eng. 2016, 30, 04015045. [Google Scholar] [CrossRef]
  15. Lerma, J.L.; Muir, C. Evaluating the 3D documentation of an early Christian upright stone with carvings from Scotland with multiples images. J. Archaeol. Sci. 2004, 46, 311–318. [Google Scholar] [CrossRef]
  16. Monna, F.; Esin, Y.; Magail, J.; Granjon, L.; Navarro, N.; Wilczek, J.; Chateau, C. Documenting carved stones by 3D modelling—Example of Mongolian deer stones. J. Cult. Herit. 2018, 34, 116–128. [Google Scholar] [CrossRef]
  17. Abdul Shukor, S.A.; Wong, R.; Rushforth, E.; Basah, S.N.; Zakaria, A. 3D terrestrial laser scanner for managing existing building. J. Teknol. 2015, 76, 133–139. [Google Scholar] [CrossRef] [Green Version]
  18. Ergincan, F.; Çabuk, A.; Avdan, U.; Tün, M. Advanced technologies for archaeological documentation: Patara case. Sci. Res. Essays 2005, 5, 2615–2629. [Google Scholar]
  19. Torres, J.A.; Hernandez-Lopez, D.; Gonzalez-Aguilera, D.; Hidalgo, M.A.M. A hybrid measurement approach for archaeological site modelling and monitoring: The case study of Mas D’Is, Penàguila. J. Archaeol. Sci. 2014, 50, 475–483. [Google Scholar] [CrossRef]
  20. Castagnetti, C.; Bertacchini, E.; Capra, A.; Dubbini, M. Terrestrial Laser Scanning for Preserving Cultural Heritage: Analysis of Geometric Anomalies for Ancient Structures. In Proceedings of the FIG Working Week 2012 on Knowing to Manage the Territory, Protect the Environment, Evaluate the Cultural Heritage, Rome, Italy, 6–10 May 2012. [Google Scholar]
  21. Conde, B.; Ramos, L.F.; Oliveira, D.V.; Riveiro, B.; Solla, M. Structural assessment of masonry arch bridges by combination of non-destructive testing techniques and three-dimensional numerical modelling: Application to Vilanova bridge. Eng. Struct. 2017, 148, 621–638. [Google Scholar] [CrossRef]
  22. Sánchez-Aparicio, L.J.; Del Pozo, S.; Ramos, L.F.; Arce, A.; Fernandes, F.M. Heritage site preservation with combined radiometric and geometric analysis of TLS data. Autom. Constr. 2018, 85, 24–39. [Google Scholar] [CrossRef]
  23. Di Filippo, A.; Sánchez-Aparicio, L.J.; Barba, S.; Martín-Jiménez, J.A.; Mora, R.; González Aguilera, D. Use of a wearable mobile laser system in seamless indoor 3D mapping of a complex historical site. Remote Sens. 2018, 10, 1897. [Google Scholar] [CrossRef] [Green Version]
  24. Gallay, M.; Kaňuk, J.; Hochmuth, Z.; Meneely, J.D.; Hofierka, J.; Sedlák, V. Large-scale and high-resolution 3-D cave mapping by terrestrial laser scanning: A case study of the Domica Cave, Slovakia. Int. J. Speleol. 2015, 44, 277–291. [Google Scholar] [CrossRef] [Green Version]
  25. Fabbri, S.; Sauro, F.; Santagata, T.; Rossi, G.; De Waele, J. High-resolution 3-D mapping using terrestrial laser scanning as a tool for geomorphological and speleogenetical studies in caves: An example from the Lessini mountains (North Italy). Geomorphology 2017, 280, 16–29. [Google Scholar] [CrossRef]
  26. Brasington, J.; Vericat, D.; Rychkov, I. Modeling river bed morphology, roughness, and surface sedimentology using high resolution terrestrial laser scanning. Water Resour. Res. 2012, 48, 11519. [Google Scholar] [CrossRef] [Green Version]
  27. Jaboyedoff, M.; Oppikofer, T.; Abellan, A.; Derron, M.H.; Alexandre, L.; Metzger, R.; Pedrazzini, A. Use of LIDAR in landslide investigations: A review. Nat. Hazards 2015, 61, 5–28. [Google Scholar] [CrossRef] [Green Version]
  28. Wang, G.; Joyce, J.; Phillips, D.; Shrestha, R.; Carter, W. Delineating and defining the boundaries of an active landslide in the rainforest of Puerto Rico using a combination of airborne and terrestrial LIDAR data. Landslides 2013, 10, 503–513. [Google Scholar] [CrossRef]
  29. Barbarella, M.; Fiani, M. Landslide monitoring using terrestrial laser scanner: Georeferencing and canopy filtering issues in a case study. ISPRS Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2012, 39, 157–162. [Google Scholar] [CrossRef] [Green Version]
  30. Schürch, P.; Densmore, A.L.; Rosser, N.J.; McArdell, B.W. Dynamic controls on erosion and deposition on debris-flow fans. Geology 2011, 39, 827–830. [Google Scholar]
  31. Day, S.; Gran, K.; Belmont, P.; Wawrzyniec, T. Measuring bluff erosion part 2: Pairing aerial photographs and terrestrial laser scanning to create a watershed scale sediment budget. Earth Surf. Process. Landf. 2013, 38, 1068–1082. [Google Scholar] [CrossRef]
  32. Abellan, A.; Oppikofer, T.; Jaboyedoff, M.; Rosser, N.J.; Lim, M.; Lato, M.J. Terrestrial laser scanning of rock slope instabilities. Earth Surf. Process. Landf. 2013, 39, 89–97. [Google Scholar] [CrossRef]
  33. D’Amato, J.; Guerin, A.; Hantz, D.; Rossetti, J.-P.; Jaboyedoff, M. Investigating Rock Fall Frequency and Failure Configurations Using Terrestrial Laser Scanner. In Engineering Geology for Society and Territory; Springer: Cham, Switzerland, 2015; pp. 1919–1923. [Google Scholar]
  34. Kasperski, J.; Delacourt, C.; Allemand, P.; Potherat, P.; Jaud, M.; Varrel, E. Application of a Terrestrial Laser Scanner (TLS) to the Study of the Séchilienne Landslide (Isère, France). Remote Sens. 2010, 2, 2785–2802. [Google Scholar] [CrossRef] [Green Version]
  35. Bremer, M.; Sass, O. Combining airborne and terrestrial laser scanning for quantifying erosion and deposition by a debris flow event. Geomorphology 2012, 138, 49–60. [Google Scholar] [CrossRef]
  36. Abellan, A.; Calvet, J.; Vilaplana, J.; Blanchard, J. Detection and spatial prediction of rockfalls by means of terrestrial laser scanner monitoring. Geomorphology 2010, 119, 162–171. [Google Scholar] [CrossRef]
  37. Muratov, O.; Slynko, Y.; Chernov, V.; Lyubimtseva, M.; Shamsuarov, A.; Bucha, V. 3DCapture: 3D Reconstruction for a Smartphone. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Las Vegas, NV, USA, 26 June–1 July 2016. [Google Scholar] [CrossRef]
  38. Durrant-Whyte, H.; Bailey, T. Simultaneous localization and mapping: Part I. IEEE Robot. Autom. Mag. 2006, 13, 99–110. [Google Scholar] [CrossRef] [Green Version]
  39. Engelhard, N.; Endres, F.; Hess, J.; Sturm, J.; Burgard, W. Real-time 3d visual slam with a hand-held rgb-d camera. In Proceedings of the RGB-D Workshop on 3D Perception in Robotics, Vasteras, Sweden, 8 April 2011. [Google Scholar]
  40. Hyyppä, J.; Virtanen, J.-P.; Jaakkola, A.; Yu, X.; Hyyppä, H.; Liang, X. Feasibility of Google Tango and kinect for crowdsourcing forestry information. Forests 2017, 9, 6. [Google Scholar] [CrossRef] [Green Version]
  41. Tomaštík, J.; Saloň, Š.; Tunák, D.; Chudý, F.; Kardoš, M. Tango in forests—An initial experience of the use of the new Google technology in connection with forest inventory tasks. Comput. Electron. Agric. 2017, 141, 109–117. [Google Scholar] [CrossRef]
  42. ARCore Overview. 2019. Available online: https://developers.google.com/ar/discover/ (accessed on 15 January 2020).
  43. ARCore Fundamental Concepts. 2019. Available online: https://developers.google.com/ar/discover/concepts (accessed on 15 January 2020).
  44. 3D Scanner for ARCore. 2019. Available online: https://github.com/lvonasek/tango/wiki/3D-Scanner-for-ARcore (accessed on 15 January 2020).
  45. Velten, A.; Willwacher, T.; Gupta, O.; Veeraraghavan, A.; Bawendi, M.; Raskar, R. Recovering Three-Dimensional Shape around a Corner Using Ultrafast Time-of-Flight Imaging. Nat. Commun. 2012, 3, 745. [Google Scholar] [CrossRef] [Green Version]
  46. Turista. 2019. Available online: http://turista.707.cz/kolomaz-uvod.html (accessed on 15 January 2020).
  47. SCANN3D. 2019. Available online: https://play.google.com/store/apps/details?id=com.smartmobilevision.scann3d&hl=en_US (accessed on 15 January 2020).
  48. Compute 2.5D Volume. 2019. Available online: https://www.cloudcompare.org/doc/wiki/index.php?title=2.5D_Volume (accessed on 15 January 2020).
  49. Besl, P.J.; McKay, N. A method for registration of 3-d shapes. IEEE Trans. Pattern Anal. Mach. Intell. 1992, 14, 239–256. [Google Scholar] [CrossRef]
  50. Apuzzo, N. Measurement and modeling of human faces from multi images. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2002, 34, 241–246. [Google Scholar]
  51. Wong, S.S.; Chan, K.L. 3D object model reconstruction from image sequence based on photometric consistency in volume space. Pattern Anal. Appl. 2010, 13, 437–450. [Google Scholar] [CrossRef]
  52. Sansoni, G.; Trebeschi, M.; Docchio, F. Fast 3D profilometer based upon the projection of a single fringe pattern and absolute calibration. Meas. Sci. Technol. 2006, 17, 1757. [Google Scholar] [CrossRef]
  53. Bakuła, K.; Flasiński, A. Capabilities of a smartphone for georeferenced 3dmodel creation: An evaluation. In International Multidisciplinary Scientific GeoConference Surveying Geology and Mining Ecology Management; SGEM: Vienna, Austria, 2014; Volume 3, pp. 85–92. [Google Scholar]
  54. Yilmazturk, F.; Gurbak, A.E. Geometric Evaluation of Mobile-Phone Camera Images for 3D Information. Int. J. Opt. 2019, 2019, 8561380. [Google Scholar] [CrossRef]
  55. Micheletti, N.; Chandler, J.H.; Lane, S.N. Investigating the geomorphological potential of freely available and accessible Structure-from-Motion photogrammetry using a smartphone. Earth Surf. Process. Landf. 2014, 40, 473–486. [Google Scholar] [CrossRef] [Green Version]
  56. Ye, Q.; Liu, H.; Yao, Y.; Gui, P. Study on Depth Measurement and 3D Reconstruction Accuracy Based on Portable RGB-D Sensor. J. Tongji Univ. 2019, 47, 870–877. [Google Scholar]
  57. Gongal, A.; Karkee, M.; Amatya, S. Apple fruit size estimation using a 3D machine vision system. Inf. Process. Agric. 2018, 5, 498–503. [Google Scholar] [CrossRef]
Figure 1. Location of Přečkov within the Czech Republic.
Figure 1. Location of Přečkov within the Czech Republic.
Sensors 20 01663 g001
Figure 2. Localization of boulders.
Figure 2. Localization of boulders.
Sensors 20 01663 g002
Figure 3. Boulder No. 1 (left) and boulder No. 2 (right) with visible wooden control points.
Figure 3. Boulder No. 1 (left) and boulder No. 2 (right) with visible wooden control points.
Sensors 20 01663 g003
Figure 4. The camera’s selected geometry configuration for data acquisition (Boulder No.1—left, boulder No.2—right).
Figure 4. The camera’s selected geometry configuration for data acquisition (Boulder No.1—left, boulder No.2—right).
Sensors 20 01663 g004
Figure 5. Reference models gained from TLS (Boulder No.1—left, boulder No.2—right).
Figure 5. Reference models gained from TLS (Boulder No.1—left, boulder No.2—right).
Sensors 20 01663 g005
Figure 6. Models by AGISOFT PhotoScan (Boulder No.1—left, boulder No.2—right).
Figure 6. Models by AGISOFT PhotoScan (Boulder No.1—left, boulder No.2—right).
Sensors 20 01663 g006
Figure 7. Models created by 3DScanner for ArCore (Boulder No.1—left, boulder No.2—right).
Figure 7. Models created by 3DScanner for ArCore (Boulder No.1—left, boulder No.2—right).
Sensors 20 01663 g007
Figure 8. Models from SCANN3D App (Boulder No.1—left, boulder No.2—right).
Figure 8. Models from SCANN3D App (Boulder No.1—left, boulder No.2—right).
Sensors 20 01663 g008
Figure 9. Boulders with marked cross sections in X, Y and Z axes (Boulder No.1—left, boulder No.2—right).
Figure 9. Boulders with marked cross sections in X, Y and Z axes (Boulder No.1—left, boulder No.2—right).
Sensors 20 01663 g009
Figure 10. Shapes in cross sections boulder No. 1 (a) X axis, (b) Y axis, (c) Z axis.
Figure 10. Shapes in cross sections boulder No. 1 (a) X axis, (b) Y axis, (c) Z axis.
Sensors 20 01663 g010
Figure 11. Shapes in cross sections boulder No. 2 (a) X axis, (b) Y axis, (c) Z axis.
Figure 11. Shapes in cross sections boulder No. 2 (a) X axis, (b) Y axis, (c) Z axis.
Sensors 20 01663 g011
Table 1. Faro Focus 3D specifications.
Table 1. Faro Focus 3D specifications.
Faro Focus 3D Scanner
Range:0.6–20 m at normal incidence on >10% matte reflective surface, 0.6–120 m indoor or outdoor with low ambient light and normal incidence to a 90% reflective surface
Measurement speed:122,000 / 244,000 / 488,000 / 976,000 points/sec
Ranging error:±2 mm
Table 2. Xiaomi Mi2 camera specifications.
Table 2. Xiaomi Mi2 camera specifications.
Dual CameraStandard LensTelephoto Lens
Resolution12 Mpix12 Mpix
Sensor size1/2.55″1/3.4″
Aperturef/1.8f/2.4
Pixel size1.40 µm1.00 µm
Table 3. Boulder No. 1: volume, surface area and height and particular model differences.
Table 3. Boulder No. 1: volume, surface area and height and particular model differences.
Volume (m3)Surface (m2)Height (m)Vol.Diff. (m3)Sur.Diff. (m2)H.Diff. (m)Vol.Diff. (%)Sur.Diff (%)Height.Diff (%)
TLS45.33721.9062.875refrefrefrefrefref
AGI45.03321.9772.7970.304−0.0710.0780.671−0.3242.713
3DS46.53620.4643.018−1.1991.1442−0.143−2.6456.583−4.974
S3D49.75223.5463.067−4.415−1.640−0.192−9.738−7.487−6.678
Table 4. Boulder No. 1: circumference and the section area.
Table 4. Boulder No. 1: circumference and the section area.
Circumference (m)Area (m2)
XYZXYZ
TLS15.63618.25212.91214.44520.6808.431
AGI15.93718.26912.96413.84920.0908.191
3DS15.71218.06513.27614.35920.0728.032
S3D15.40517.43712.51813.70119.0967.819
Table 5. Boulder No. 1: circumference and the section area—model differences.
Table 5. Boulder No. 1: circumference and the section area—model differences.
Circumference Diff. (m)Area Diff. (m2)
XYZXYZ
TLSrefrefrefrefrefref
AGI−0.301−0.017−0.0520.5970.5900.240
3DS−0.0750.118−0.3640.0860.6080.400
S3D0.2320.8150.3940.7441.5840.612
Table 6. Boulder No. 1: circumference and the section area—model differences (in percents).
Table 6. Boulder No. 1: circumference and the section area—model differences (in percents).
Circumference Diff. (%)Area Diff. (%)
XYZXYZ
TLSrefrefrefrefrefref
AGI−1.922−0.001−0.4044.1302.8542.845
3DS−0.4820.010−2.8190.5952.9414.739
S3D1.4820.0453.0555.1517.6627.262
Table 7. Boulder No. 1: point distances
Table 7. Boulder No. 1: point distances
Mean (m)Stand. Dev. (m)RMSE (m)
TLSrefrefRef
AGI0.0280.0460.054
3DS0.0340.0290.045
S3D0.0680.0570.089
Table 8. Boulder No. 2: volume, surface area and height and particular model differences.
Table 8. Boulder No. 2: volume, surface area and height and particular model differences.
Volume (m3)Surface (m2)Height (m)Vol.Diff. (m3)Sur.Diff. (m2)H.Diff. (m)Vol.Diff. (%)Sur.Diff (%)Height.Diff (%)
TLS64.06829.0184.076refrefrefrefrefref
AGI58.05727.2823.9076.0111.7360.1669.3825.9824.076
3DS66.68127.6944.178−2.6131.324−0.105−4.0784.563−2.578
S3D48.72023.6343.86015.3485.3840.21323.95618.5545.230
Table 9. Boulder No. 2: circumference and the section area
Table 9. Boulder No. 2: circumference and the section area
Circumference (m)Area (m2)
XYZXYZ
TLS17.88613.97013.44515.95523.89513.027
AGI18.64714.99914.93316.42326.29514.700
3DS18.80514.65715.07216.95126.29914.206
S3D17.80813.78212.59815.63022.10312.504
Table 10. Boulder No. 2: circumference and the section area—particular model differences.
Table 10. Boulder No. 2: circumference and the section area—particular model differences.
Circumference Diff. (m)Area Diff. (m2)
XYZXYZ
TLSrefrefrefrefrefref
AGI−0.757−1.029−1.483−0.473−2.395−1.670
3DS−0.915−0.687−1.622−1.001−2.399−1.176
S3D0.0820.1880.8520.3201.7970.526
Table 11. Boulder No. 2: circumference and the section area—particular model differences (in percent).
Table 11. Boulder No. 2: circumference and the section area—particular model differences (in percent).
Circumference Diff. (%)Area Diff. (%)
XYZXYZ
TLSrefrefrefrefrefref
AGI−4.232−7.365−11.025−2.968−10.019−0.128
3DS−5.113−4.916−12.056−6.273−10.039−0.090
S3D0.4581.3446.3312.0047.5190.040
Table 12. Boulder No. 2: point distances.
Table 12. Boulder No. 2: point distances.
Mean (m)Stand. Dev. (m)RMSE (m)
TLSrefrefRef
AGI0.0970.0850.129
3DS0.0810.0610.101
S3D0.0590.0470.075

Share and Cite

MDPI and ACS Style

Mikita, T.; Balková, M.; Bajer, A.; Cibulka, M.; Patočka, Z. Comparison of Different Remote Sensing Methods for 3D Modeling of Small Rock Outcrops. Sensors 2020, 20, 1663. https://0-doi-org.brum.beds.ac.uk/10.3390/s20061663

AMA Style

Mikita T, Balková M, Bajer A, Cibulka M, Patočka Z. Comparison of Different Remote Sensing Methods for 3D Modeling of Small Rock Outcrops. Sensors. 2020; 20(6):1663. https://0-doi-org.brum.beds.ac.uk/10.3390/s20061663

Chicago/Turabian Style

Mikita, Tomáš, Marie Balková, Aleš Bajer, Miloš Cibulka, and Zdeněk Patočka. 2020. "Comparison of Different Remote Sensing Methods for 3D Modeling of Small Rock Outcrops" Sensors 20, no. 6: 1663. https://0-doi-org.brum.beds.ac.uk/10.3390/s20061663

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop