1 Introduction

With a large number of metro tunnels being constructed and operated in China, the fast and effective tunnel deformation measuring method becomes imperative as tunnel’s deformation condition is important for evaluating its performance. Most of the metro tunnels are constructed using the shield method. Shield tunnels’ deformation is generally induced by cyclic train loads, foundation deformation and adjacent construction activities, not only impacting tunnels’ durability but also reducing their safety factors (Wang et al. 2009) [1]. In soft soil regions such as Hangzhou city and Shanghai city, tunnels’ deformation is usually large. The maximum deformation observed in Shanghai metro was 148 mm (2.4%D) (Zhang et al. 2014) [2].

Traditional deformation measuring techniques mainly include Bassett convergence instruments and total stations. The convergence instruments, which measure tunnels’ deformation through the change of the measuring lines’ length, are typical contacting measurements. Setting an enormous number of measuring lines is time-consuming, costly and tedious (Bassett et al. 1999) [3]. Measurement of total station is discrete and inflexible even though it implements precise and non-contact measurements (Hope et al. 2008; Yang et al. 2005; Yang et al. 2006) [4,5,6]. These two methods cannot satisfy the increasing measuring requirements any longer due to the shortcomings aforementioned.

Some new measuring methods have appeared in recent years such as light detection and ranging (LiDAR), which is able to obtain the complete data of cross sections. Han et al. (2013) [7] used Minimum-Distance Projection (MDP) algorithm to establish point correspondences to measure tunnels’ deformation. Some researchers established 3D models of tunnels using terrestrial laser scanning to measure tunnels’ geometry deformation and settlement (Xie et al. 2017; Argüelles et al. 2013) [8, 9], but for most subway companies it’s still too expensive.

The close range photogrammetry technology advances rapidly as digital single-lens reflex cameras mature. Study on photogrammetry using non-metric cameras has become a research hotspot during the past years. The shapes, dimensions and locations of the vertical cross-sections of tunnels can be derived from the photographs based on the control data provided by two targets’ locations on the plates (Ethrog et al. 1982) [10]. Zhang et al. (2001) [11] used a CCD area array camera to measure areas of irregular plane objects. Yang et al. (2001) [12] proposed a novel algorithm to improve CCD cameras’ measuring precision. Lee et al. (2006) [13] found that the displacement and strain data obtained with close range photogrammetric technique in a model tunnel showed a remarkable agreement with the physical data. Hu (2008) [14] used close range photogrammetry to measure a mountain tunnel’s deformation, but the precision was unsatisfactory. Chen et al. (2014) [15] proposed using a mobile metal bracket to offer control points, which improved the practicability and maneuverability of DLT method in tunnel engineering.

This article tries to create a novel method for measuring shield tunnel cross sections based on laser orientation, laser range-finding and close range photogrammetry. Based on a novel way of setting control points and target points by using laser, this article studies on how to apply DLT method in tunnel monitoring. A specially-designed device was used in the field experiment. The results proved the practicability and accuracy of the proposed method.

2 Close Range Photogrammetry

2.1 Imaging Principle

The pinhole imaging is the fundamental theory of close range photogrammetry (see Fig. 1).

Fig. 1.
figure 1

The theory of pinhole imaging

Based on the similar triangles principle, we have:

$$ \frac{h}{f} = \frac{{h^{'} }}{{S_{2} - f}} $$
(1)
$$ \frac{h}{{S_{1} }} = \frac{{h^{'} }}{{S_{2} }} $$
(2)

Where \( h \) stands for the object height, \( h^{'} \) stands for the image height, \( f \) represents the focal length, \( \text{S}_{1} \) represents the distance between the object plane and the optical center and \( {\text{S}}_{2} \) represents the distance between the imaging plane and the optical center.

Rearranging the two equations, we can get:

$$ \frac{1}{f} = \frac{1}{{S_{1} }} + \frac{1}{{S_{2} }} $$
(3)

As \( {\text{S}}_{1} \) is usually far greater than \( {\text{S}}_{2} \) in engineering projects, it is easy to get:

$$ h^{'} = \frac{fh}{{S_{1} }} $$
(4)

2.2 General Coordinate Systems

Four general coordinate systems are used in DLT method: world coordinate system, camera coordinate system, image coordinate system and pixel coordinate system.

World coordinates are absolute coordinates that objectively describe the three-dimensional space, signified with (X w , Y w , Z w ). Camera coordinates are signified with (X c , Y c , Z c ) and the origin point of them is located at the lens’ optical center. Z c represents the lens’ main optical axis. Image coordinates and pixel coordinates are signified with (x, y) and (u, v) separately. The main difference between them lies in the measurement unit. The former one is measured in millimeter or centimeter whereas the later one is measured in pixel.

Figure 2 shows the conversion relations of the four general systems. It should be pointed out that the conversion between camera coordinates and image coordinates is based on Eq. (4).

Fig. 2.
figure 2

General coordinate systems

The transformation equation between pixel coordinates and world coordinates is:

$$ Z_{C} \left( {\begin{array}{*{20}c} u \\ v \\ 1 \\ \end{array} } \right) = \left( {\begin{array}{*{20}c} {\frac{1}{{d_{x} }}} & 0 & {u_{0} } \\ 0 & {\frac{1}{{d_{y} }}} & {v_{0} } \\ 0 & 0 & 1 \\ \end{array} } \right)\left( {\begin{array}{*{20}c} f & 0 \\ 0 & f \\ 0 & 0 \\ \end{array} \begin{array}{*{20}c} 0 & 0 \\ 0 & 0 \\ 1 & 0 \\ \end{array} } \right)\left( {\begin{array}{*{20}c} R & T \\ {0^{T} } & 1 \\ \end{array} } \right)\left( {\begin{array}{*{20}c} {\begin{array}{*{20}c} {X_{W} } \\ {Y_{W} } \\ {Z_{W} } \\ \end{array} } \\ 1 \\ \end{array} } \right) $$
(5)

Where R represents a certain 3 \( \times \) 3 rotation matrix and T represents a 3 \( \times \) 1 translation matrix. The parameters d x and d y are the physical dimensions of one pixel in x-axis direction and y-axis direction in the image plane respectively.

2.3 Direct Linear Translation Method

Among the commonly used parsing methods for close range photogrammetry, DLT is especially suitable for non-metric photography. It does not involve initial approximations for the unknown parameters of inner and outer orientation of the camera (Abdel-Aziz et al. 2015) [16]. DLT contains a direct linear transformation from pixel coordinates into world coordinates with Eq. (6) which is based on Eq. (5).

$$ A \cdot m = B $$
(6)
$$ A = \left[ {\begin{array}{*{20}c} {\begin{array}{*{20}c} {\begin{array}{*{20}c} {\begin{array}{*{20}c} {{\text{X}}_{{{\text{w}}1}} } \\ 0 \\ \end{array} } & {\begin{array}{*{20}c} {{\text{Y}}_{{{\text{w}}1}} } \\ 0 \\ \end{array} } \\ {\begin{array}{*{20}c} {{\text{X}}_{{{\text{w}}2}} } \\ 0 \\ \end{array} } & {\begin{array}{*{20}c} {{\text{Y}}_{{{\text{w}}2}} } \\ 0 \\ \end{array} } \\ \end{array} } & {\begin{array}{*{20}c} {\begin{array}{*{20}c} {{\text{Z}}_{{{\text{w}}1}} } \\ 0 \\ \end{array} } & {\begin{array}{*{20}c} 1 \\ 0 \\ \end{array} } \\ {\begin{array}{*{20}c} {{\text{Z}}_{{{\text{w}}2}} } \\ 0 \\ \end{array} } & {\begin{array}{*{20}c} 1 \\ 0 \\ \end{array} } \\ \end{array} } \\ \end{array} } & {\begin{array}{*{20}c} {\begin{array}{*{20}c} 0 & 0 \\ {{\text{X}}_{{{\text{w}}1}} } & {{\text{Y}}_{{{\text{w}}1}} } \\ \end{array} } & {\begin{array}{*{20}c} 0 & 0 \\ {{\text{Z}}_{{{\text{w}}1}} } & 1 \\ \end{array} } \\ {\begin{array}{*{20}c} 0 & 0 \\ {{\text{X}}_{{{\text{w}}2}} } & {{\text{Y}}_{{{\text{w}}2}} } \\ \end{array} } & {\begin{array}{*{20}c} 0 & 0 \\ {{\text{Z}}_{{{\text{w}}2}} } & 1 \\ \end{array} } \\ \end{array} } & {\begin{array}{*{20}c} {\begin{array}{*{20}c} { - {\text{u}}_{1} {\text{X}}_{{{\text{w}}1}} } \\ { - {\text{v}}_{1} {\text{X}}_{{{\text{w}}1}} } \\ \end{array} } & {\begin{array}{*{20}c} { - {\text{u}}_{1} {\text{Y}}_{{{\text{w}}1}} } \\ { - {\text{v}}_{1} {\text{Y}}_{{{\text{w}}1}} } \\ \end{array} } & {\begin{array}{*{20}c} { - {\text{u}}_{1} {\text{Z}}_{{{\text{w}}1}} } \\ { - {\text{v}}_{1} {\text{Z}}_{{{\text{w}}1}} } \\ \end{array} } \\ {\begin{array}{*{20}c} { - {\text{u}}_{2} {\text{X}}_{{{\text{w}}2}} } \\ { - {\text{v}}_{2} {\text{X}}_{{{\text{w}}2}} } \\ \end{array} } & {\begin{array}{*{20}c} { - {\text{u}}_{2} {\text{Y}}_{{{\text{w}}2}} } \\ { - {\text{v}}_{2} {\text{Y}}_{{{\text{w}}2}} } \\ \end{array} } & {\begin{array}{*{20}c} { - {\text{u}}_{2} {\text{Z}}_{{{\text{w}}2}} } \\ { - {\text{v}}_{2} {\text{Z}}_{{{\text{w}}2}} } \\ \end{array} } \\ \end{array} } \\ \vdots & \vdots & \vdots \\ {\begin{array}{*{20}c} {\begin{array}{*{20}c} {{\text{X}}_{\text{wn}} } & {{\text{Y}}_{\text{wn}} } \\ 0 & 0 \\ \end{array} } & {\begin{array}{*{20}c} {{\text{Z}}_{\text{wn}} } & 1 \\ 0 & 0 \\ \end{array} } \\ \end{array} } & {\begin{array}{*{20}c} {\begin{array}{*{20}c} 0 & 0 \\ {{\text{X}}_{\text{wn}} } & {{\text{Y}}_{\text{wn}} } \\ \end{array} } & {\begin{array}{*{20}c} 0 & 0 \\ {{\text{Z}}_{\text{wn}} } & 1 \\ \end{array} } \\ \end{array} } & {\begin{array}{*{20}c} { - {\text{u}}_{\text{n}} {\text{X}}_{\text{wn}} } & { - {\text{u}}_{\text{n}} {\text{Y}}_{\text{wn}} } & { - {\text{u}}_{\text{n}} {\text{Z}}_{\text{wn}} } \\ { - {\text{v}}_{\text{n}} {\text{X}}_{\text{wn}} } & { - {\text{v}}_{\text{n}} {\text{Y}}_{\text{wn}} } & { - {\text{v}}_{\text{n}} {\text{Z}}_{\text{wn}} } \\ \end{array} } \\ \end{array} } \right]_{{2{\text{n}} \times 11}} $$
(7)
$$ B = \left[ {\begin{array}{*{20}c} {\begin{array}{*{20}c} {u_{1} } & {v_{1} } \\ \end{array} } & {\begin{array}{*{20}c} {\begin{array}{*{20}c} {u_{2} } & {v_{2} } \\ \end{array} } & \cdots \\ \end{array} } & {\begin{array}{*{20}c} {u_{n} } & {v_{n} } \\ \end{array} } \\ \end{array} } \right]_{2n \times 1}^{T} $$
(8)

Where m represents the coefficient matrix with Eq. (9). Interior and exterior orientation elements are implicit in the m matrix.

$$ m = \left[ {\begin{array}{*{20}c} {\begin{array}{*{20}c} {m_{1} } & {m_{2} } \\ \end{array} } & {\begin{array}{*{20}c} {\begin{array}{*{20}c} {m_{3} } & {m_{4} } \\ \end{array} } & \cdots \\ \end{array} } & {\begin{array}{*{20}c} {m_{10} } & {m_{11} } \\ \end{array} } \\ \end{array} } \right]_{1 \times 11}^{T} $$
(9)

In general, pixel coordinates in a photo are easily obtained. Then Eq. (6) could be used to calculate the world coordinates of any target point if the m matrix is calculated. To solve the m matrix, several control points are required, of which the world coordinates are known. Three-dimensional DLT method requires at least six control points as there are 11 unknown Parameters in the m matrix. In tunnel engineering, these points are normally set on the inner surface of the linings (Ma et al. 2006) [17]. To improve the measuring precision, they are generally more than six. Actually, it’s difficult to set many control points in a shield tunnel, which has restricted the application of DLT method in tunnel projects.

The calculating process of DLT method is divided into four steps: (1) identifying and positioning control points; (2) computing the m matrix (camera calibration); (3) extracting pixel coordinates of target points; (4) calculating world coordinates of target points.

Chen’s research is an example of applying DLT method in tunnel engineering (Chen et al. 2014) [15]. A metal bracket was placed in the tunnel with predefined target points on the linings when taking a photo. The bracket established a reference coordinate system in photos. The m matrix could be calculated through these points’ coordinates. However, solving any target point needed at least two photos captured from different vision angles because only two equations can be built from one photo for each target point which contains three unknowns: (X w , Y w , Z w ).

2.4 Laser-Based Calibration Method

This article proposes a novel calibration method based on laser range-finding, which aims to achieve higher precision at lower cost. A laser source is fixed on a self-made inspection device (MTI-200A) to create a bright outline on the lining surface. On the same device, a laser range-finder is placed on a 360° protractor, which is coplanar with the outline plane (see Fig. 3).

Fig. 3.
figure 3

The related devices

The distances and their corresponding angles of control points are recorded by rotating the laser range-finder (see Fig. 4). The recorded data will be used to calculate world coordinates of control points with Eqs. (10) and (11).

Fig. 4.
figure 4

A control point on the outline

$$ X_{wi} = L_{i} \times sin\theta_{i} $$
(10)
$$ Y_{wi} = L_{i} \times cos\theta_{i} $$
(11)

Where \( L_{i} \) stands for the distance and \( \theta_{i} \) stands for the corresponding angle.

The parameter \( {\text{Z}}_{w} \) in 3D DLT equations is assumed as zero on the hypothesis that the outline and control points are located in the same plane. Then the matrix \( A \) in Eq. (7) transforms into 2D form as shown in Eq. (12).

$$ A = \left[ {\begin{array}{*{20}c} {\begin{array}{*{20}c} {X_{w1} } & {Y_{w1} } & 1 \\ 0 & 0 & 0 \\ \end{array} } & {\begin{array}{*{20}c} 0 & 0 & 0 \\ {X_{w1} } & {Y_{w1} } & 1 \\ \end{array} } & {\begin{array}{*{20}c} { - u_{1} X_{w1} } & { - u_{1} Y_{w1} } \\ { - v_{1} X_{w1} } & { - v_{1} Y_{w1} } \\ \end{array} } \\ {\begin{array}{*{20}c} \vdots & \vdots & \vdots \\ \end{array} } & {\begin{array}{*{20}c} \vdots & \vdots & \vdots \\ \end{array} } & {\begin{array}{*{20}c} \vdots & \vdots \\ \end{array} } \\ {\begin{array}{*{20}c} {X_{wn} } & {Y_{wn} } & 1 \\ 0 & 0 & 0 \\ \end{array} } & {\begin{array}{*{20}c} 0 & 0 & 0 \\ {X_{wn} } & {Y_{wn} } & 1 \\ \end{array} } & {\begin{array}{*{20}c} { - u_{n} X_{wn} } & { - u_{n} Y_{wn} } \\ { - v_{n} X_{wn} } & { - v_{n} Y_{wn} } \\ \end{array} } \\ \end{array} } \right]_{2n \times 8} $$
(12)

Solving the m parameters of 2D DLT requires at least 4 control points due to the unknown number of m has decreased from 11 to 8. To achieve higher measuring precision, more control points are needed. Least squares method as shown in Eq. (13) is used to process the data if there are extra control points.

$$ m = (A^{T} A)^{ - 1} A^{T} B $$
(13)

The most significant improvement of this method is that solving each target point needs only one photo captured from any vision angle.

The measuring system frame is shown in Fig. 5. The relevant algorithms were programed with Mathematica.

Fig. 5.
figure 5

The measuring system frame

3 Field Test and Verification

3.1 Preparation

The main advantages of non-metric digital cameras are listed as follows: (1) The interior and exterior orientation elements of them can be calculated accurately; (2) They do not need negatives like film cameras do; (3) They storage data in digital form and that data can be easily taken use of by a computer; (4) They are fairly cheaper than metric cameras. (5) Specially-designed programs are able to complete the calculating process automatically and correctly.

Nikon D7000 digital camera and Tokina AT-X PRO SD 11-16 F2.8 IF DX IINikon lens were chosen as the test equipment as shown in Fig. 6. Full frame cameras were not considered as they had not been in widespread use. The image sensor of D7000 has sixteen million pixels, which is enough for the experiments. A wide-angle lens was required and Tokina 11–16 met the requirements perfectly. The 11 mm focus was preferred because it allowed the experimenter to approach closer to the outline than the 16 mm focus.

Fig. 6.
figure 6

The D7000 camera

The only problem lied in the noticeable barrel image distortion at the 11 mm focus as shown in Fig. 7, released by the famous image quantity reference website DXOMARK [18]. Distortion means that the original straight lines are distorted into curved lines which will undoubtedly lead to some errors in measurements. To avoid that, the Adobe software Lightroom was used to correct the distortion. Figure 8 shows the original image of a gridding board shot by D7000 and Fig. 9 shows the modified image using the Lightroom software. Although the distortion in the image edges is still detectable, the maximum deviation of straight lines has decreased from 75 pixels to 15 pixels. Furthermore, the outline range seldom ever appears on the image edges as Fig. 10 shows, in fact.

Fig. 7.
figure 7

The image distortion

Fig. 8.
figure 8

The original image

Fig. 9.
figure 9

The corrected image

Fig. 10.
figure 10

The outline range

In terms of correcting distortion, lots of papers talked about different correcting algorithms. For example, Jin et al. (2011) [19] used BP neural network model to correct image distortion.

To demonstrate the feasibility of DLT method, four randomly selected points E, F, G, H in Fig. 9 were chosen as control points to solve the relative coordinates of other points. Each little square was 17.5 mm in width and 24.8 mm in height. The maximum error appeared at the point S which was 1.1315 mm. The pixel distances between target points and the image center were defined as PD. The percentage errors were defined as absolute errors divided by measuring distances. Then the curves of PE in x-axis and y-axis to PD were obtained as shown in Fig. 11. It is clearly indicated that errors in the image edges are larger than those in the image center.

Fig. 11.
figure 11

The curve of percentage errors to pixel distances

3.2 Field Experiment

A series of tests were carried out in the subway tunnel of Shanghai Metro Line 1 using the specially-designed device. Specific experiment steps were as follows: (1) The laser range-finder and the protractor were installed on the tunnel inspection vehicle. (2) The laser source was powered and it formed a bright red circular outline on the inner surface of the lining. (3) The laser range-finder was rotated with a 10-degree interval. (4) The real-time distances and their corresponding angles were recorded. (5) During the test, a total station was used to get the precise coordinates of these recorded points.

The model of the inspection device was MTI-200A (see Fig. 10). The laser range-finder’s model was HZH-80 with an accuracy of 1 mm. The total station’s model was SOKKIA CX-102. Its accuracy was (3 + 2 ppm*D) mm without reflection prism.

3.3 Test Results

Pixel coordinates of tens of thousands of target points on the outline were extracted automatically on the basis of color differences using Mathematica software. In total, about seventeen thousand pixel points were extracted of each outline.

Twenty uniformly-distributed control points were taken into consideration. Two fitting outlines are shown in Fig. 12. The red one was fitted based on the absolute coordinates measured by the total station while the blue one was fitted according to the data obtained by the proposed method. The horizontal diameter of the red outline was 5539 mm whereas that of the blue one was 5533 mm, meaning that the error was below 6 mm.

Fig. 12.
figure 12

Two fitting outlines (Color figure online)

According to the design document as Fig. 13 shows, the designed dimension of the tunnel’s inner diameter was 5500 mm. The maximum magnitude of horizontal convergence was around 39 mm, which was in accordance with the monitoring data on site.

Fig. 13.
figure 13

The geometry of shield tunnel’s linings

Compared with the total station, the precision of the proposed method was acceptable and the method was more flexible and more effective. The specially-designed mobile device was not only low-cost but also handy. With the device, the DLT method was potential to be applied in tunnel deformation monitoring. The abundant measurement results would provide basing data for stress analysis and performance analysis.

3.4 Discussion

As Fig. 12 shows, there are still some deviations between the red outline and the blue one. Error sources in close range photogrammetry were discussed thoroughly by Song et al. (2010) [20]. In this test, there were various factors that led to errors. They were analyzed as follows: (1) The image distortion still had a little influence on the results. (2) The laser range-finder was rotated manually, leading to inaccuracy in angles. (3) The measuring equipment had systemic errors. (4) The sensor’s pixels were not abundant enough.

More control points will lead to higher measuring precision intuitively. However, it means a higher cost meanwhile. The optimal option should satisfy the precision requirements at minimum costs. An important question is determining that how many control points are optimal.

Assuming that eight control points are optimal, then eight laser range-finders are installed uniformly on a protractor. It is fixed on the inspection device in practical applications. When the device moves along the tracks, the only thing necessary is to photograph the outline continually. Each photo records a certain outline and hundreds of outlines are recorded. Photoelectric encoder is used to record the real-time distances the device has moved. Once the outlines and their corresponding locations are combined correctly, the complete 3D shape of a tunnel will be gained, which is extremely helpful for engineers to evaluate the tunnel’s performance.

4 Conclusions

A novel method for measuring tunnel cross sections was put forward. It was based on laser orientation module, laser range-finding module and the theory of close range photogrammetry. A series of tests in the subway tunnel of Shanghai Metro Line 1 were carried out to validate the method being low-cost and precise. Measurement results were satisfied with twenty uniformly-distributed control points for the tunnel around 5.5 m in diameter. The maximum error between the calculated outline and the actual outline was just 6 mm.