1. Field of the Invention
The present invention relates to a surveying system. In particular, it relates to a surveying instrument or system, including a total station, an electronic tacheometer, or the like, that is able to measure an angle and a distance relating to a measurement point.
2. Description of the Related Art
Generally, a surveying operation requires at least two persons. Namely, one holds a target, such as a prism or the like, at a measurement point, while the other person operates a surveying instrument. In the surveying operation, in order to obtain measurement data for a plurality of measurement points, the person with the target should move to each of the measurement points, and the other person should sight the surveying instrument to the target, each time the target moves to a measurement point. Therefore, the conventional surveying operation is quite cumbersome and requires at least two people and a lot of time, so it is inefficient and expensive. Recently, a non-prism-type surveying instrument, which requires no target, has been developed. With the non-prism-type surveying instrument, the person who carries the target can be excluded and the surveying can be carried out by a single person who operates the surveying instrument, so that the efficiency of the surveying operation is improved. However, even when using the non-prism-type surveying instrument, the operator should still carry out sighting operations for each of the measurement points. This becomes extremely cumbersome and requires time particularly when the number of measurement points is large, such as when estimating the volume of a basin or hollow ground to be buried.
Therefore, an object of the present invention is to improve efficiency of the surveying operation. More particularly, the invention aims to cooperate measurement information obtained by a surveying instrument and image information of a surveying area which is obtained by a camera, simply and efficiently, thereby improving efficiency.
According to the present invention, a surveying system is provided. The system comprises a surveying device, a collinear-line calculating processor, a sighting direction control processor, and an object-point searching processor.
The collinear-line calculating processor obtains a collinear line which provides a collinear condition for an arbitrarily designated point on a schematic image of which the positional relation to the surveying device is known. The sighting direction control processor controls the sighting direction of the surveying device to move along the collinear line. The object-point searching processor searches for an object point that can be determined as a point on the collinear line by measuring the sighting direction with the surveying device, while carrying out the sighting direction control procedure. Further, the position obtained by the object point searching processor coincides with an object point that corresponds to the designated point on the schematic image.
Further according to the present invention, an automatic surveying system is provided that comprises a position relation calculating processor, a correspondence establishing processor, an input device, and a sighting direction control processor.
The position relation calculating processor calculates a positional relation between a coordinate system to which measurement information of a measurement point refers and a schematic image of a surveying field, which includes the measurement point. The correspondence establishing processor establishes correspondence between measurement information of the measurement point and position information of a point corresponding to the measurement point on the schematic image with respect to the above positional relation. The input device enables a designation of a measurement point on the schematic image. The sighting direction control processor controls a sighting direction of a surveying device in accordance with the position of the designated measurement point on the schematic image, which is designated by using the input device.
The objects and advantages of the present invention will be better understood from the following description, with reference to the accompanying drawings in which:
The present invention is described below with reference to the embodiments shown in the drawings.
A surveying instrument 10 of the present embodiment may be a total station or the like that comprises a sighting telescope 17. The sighting telescope 17 has a horizontal axis Lh for measuring an altitude θp, and a vertical axis Lp for measuring a horizontal angle θh. Namely, the sighting telescope 17 is vertically rotatable about the horizontal axis Lh and horizontally rotatable about the vertical axis Lp. The horizontal axis Lh and the vertical axis Lp intersect at point OS (referred to as the sighting origin OS in the following) at a right angle. The optical axis LNO (or collimation axis) of the sighting telescope 17 passes through the sighting origin OS. The optical axis LNO is bifurcated by a half-mirror 18, so that one of the bifurcated optical axes reaches the eyepiece lens and the other reaches the distance measurement component 11.
The distance measurement component 11 detects an oblique distance to a measurement point (which is sighted) by using a phase modulation measurement method, a pulse radar method, or the like, while an angle measurement/control component 12 detects a vertical angle θp and a horizontal angle θh. Further, the sighting telescope 17 can be rotated about each of the horizontal and vertical axes Lh and Lp by using an actuator (not shown), such as a stepping motor. This motor driven rotation of the sighting telescope 17 is controlled by the angle measurement/control component 12.
The distance measurement component 11 and the angle measurement/control component 12 are connected to a system control circuit 13, whereby they are controlled by signals from the system control circuit 13. For example, the distance measurement component 11 detects a distance in accordance with signals from the system control circuit 13, and outputs the detected data or measurement data to the system control circuit 13. On the other hand, the angle measurement/control component 12 continuously detects angles at a regular timing and outputs the detected data or measurement data to the system control circuit 13 when it is required. Further, the angle measurement/control component 12 controls the rotational movement of the sighting telescope 17 about the horizontal and vertical axes Lh and Lp by driving the stepping motor. The detected data, such as an oblique distance, horizontal angle, and vertical angle, are processed in the system control circuit 13. The system control circuit 13 is also connected to a switch group 14, a display 15 (ex. LCD), an interface circuit 16, and so on.
A computer 40, such as a notebook sized personal computer, may be connected to the interface circuit 16 through an interface cable. Namely, the surveying instrument 10 can transmit measurement data to the computer 40 via the interface cable. Further, the surveying instrument 10 can be controlled by control signals from the computer 40. For example, a sighting operation of the sighting telescope 17, to an arbitrary direction (horizontal angle θh and vertical angle θp), may be controlled by the computer 40. Note that, the interface circuit 16 is also available for other peripheral devices, such as a data collector, and the like, which are not shown.
The computer 40 is generally comprised of a CPU 41, a recording medium 42, input devices including a mouse 43 and a keyboard 44, a display (image indicating device) 45, and an interface circuit 46. Examples of the recording medium 42 are a hard disk, DVD, MO, and IC card. Further, the display 45 includes an LCD, CRT, and so on. As described above, the interface circuit 46 is connected to the interface circuit 16 of the surveying instrument 10 via the interface. Further, the interface circuit 46 can be connected to a digital still camera 20. Namely, an image captured by the digital still camera 20 is transmitted to the computer 40 as digital image data and stored in the recording medium 42 via the interface circuit 46, for example.
With reference to
In Step 101, an operator takes a picture or image around the surveying area with the digital still camera (DSC) 20. Further, the obtained digital image data are transmitted to the computer 40 via the interface cable, and then stored in the recording medium 42. Note that, a plurality of measurement points, which are to be measured, is included within the above captured digital image (referred to as a schematic image) taken by the digital still camera 20.
In Step 102, the schematic image is indicated on the display 45 of the computer 40, for example. At this time a plurality of points (or pixels) on the schematic image are selected by the operator by using a pointing device such as a mouse 43. Thereby, substantial points in the substantial or object space, which correspond to the selected pixels, are designated as control points Pi (i=1, 2, . . . , n), where a control point is a measurement point that is arbitrarily selected for calculating the orientation of a schematic image. Further, the positions of image points Pi′ that correspond to each of the control points Pi are derived as two-dimensional image coordinates (xpi′,ypi′). Note that, the image coordinate system is a two-dimensional coordinate system of which the origin is at the upper left corner of the image with the y′-axis being positive in the downward direction. Further, the number “n” of control points which are three dimensionally arranged, may be more than 11.
In Step S103, an oblique distance and the angles (such as vertical and horizontal angle) of each control point Pi, which are appointed in Step S102, are measured by an operator by using the survey instrument 10. Measured values are then fed to the interface circuit 46 of the computer 40 via the interface circuit 16. At the CPU 41, three-dimensional coordinates (Xpi,Ypi,Zpi) for each of the control points Pi are calculated in a predetermined surveying coordinate system. At this time, the correspondence between the surveying coordinates (Xpi,Ypi,Zpi) for each of the control points Pi and the image coordinates (xpi′,ypi′) for image points Pi′ is determined. Note that, the origin of the surveying coordinate system may be taken at the sighting origin OS of the surveying instrument 10, and the absolute coordinates, including latitude and longitude or any type of positioning format defined by the relevant surveying authority, may be adapted to the above survey coordinates. Further, the surveying coordinates calculation may be carried out in the surveying instrument 10, and later the calculated data may be sent to the computer 40.
As will be explained in detail later, exterior orientation parameters for the digital still camera 20, which represent the position and the inclination of the camera 20 at the moment when the schematic image was taken, are calculated in Step S104, in accordance with the correspondence between the surveying coordinates and the image coordinates for each of the control points Pi, by means of space resection. Further, inner orientation parameters for compensating for aberrations from the collinear condition due to lens distortion or displacement of the principal point from the image center, may be calculated. Namely, the position or the surveying coordinates (XO,YO,ZO) of the origin OC of the three-dimensional camera coordinate system, which is fixed in the digital still camera 20, and the rotational angle (ω,φ,κ) about the x-axis, y-axis, and z-axis of the camera coordinate system, at the time, are derived as exterior orientation parameters. Further, the inner orientation parameters (f: distance from the center of projection for the lens or the principal distance; D2,D4,D6: second, fourth, and sixth order components of the distortion; N1,N2: unsymmetrical components of distortion; XC,YC: displacement of the principal point from the center of the image) are obtained. Thereby, the perspective projection relationship between the image coordinates and the surveying coordinates is established. Note that, when the inner orientation parameters are definite for the above (f,D2,D4,D6,N1,N2,XC,YC), the number of control points required for calculating the exterior and inner orientation parameters is at least seven. Among these control points, at least three are required to calculate the exterior orientation parameters (XO,YO,ZO,ω,φ,κ) Note that, in the present embodiment, eleven (or more) control points are required to calculate the exterior and inner orientations.
Note that, the camera coordinate system is a left-handed coordinate system of which the origin OC is located at the center of the lens or the center of the projection, and in which the y-axis and z-axis are parallel to each of the s′-axis and t′-axis of the screen coordinate system. Further, the x-axis of the camera coordinate system is normal to the imaging surface and is oriented to the side opposite to the imaging surface from the center of the projection. Namely, any point on the imaging surface is represented by the camera coordinates (−f,y,z). Here, the screen coordinate system is a two-dimensional coordinate system of which the origin is defined at the principal point with each of the s′-axis and t′-axis arranged in parallel with each of the x′-axis and y′-axis, that is, in parallel with each of the horizontal and vertical lines of the imaging device 21 (see
As described above, the single photograph orientation process of the present embodiment ends. Note that, the image capturing operation in Step S101 can be excluded and an image taken previously may be used as a schematic image, in place. Further, although the single photograph orientation process in the present embodiment is described with reference to the flowchart of
With reference to
In Step 201, appropriate initial values (XGO,YGO,ZGO,ωG,φG,κG) and (fG,D2G,D4G,D6G,N1G,N2G,XCG,YCG) are given as approximate values for the exterior orientation parameters (XO,YO,ZO,ω,φ,κ), representing the position and the inclination of the camera, and the inner orientation parameters (f,D2,D4,D6,N1,N2,XC,YC). Then in Step S202, approximate image coordinates (xpGi′,ypGi′) of the image point Pi′ (i=1,2, . . . ,11), which correspond to each of the eleven control points Pi, are calculated from the surveying coordinates (Xpi,Ypi,Zpi) of the respective control points Pi, by using the given exterior orientation parameters (XGO,YGO,ZGO,ωG,φG,κG) and inner orientation parameters (fG,D2G,D4G,D6G,N1G,N2G,XCG,YCG).
Namely, the coordinates (xpi,ypi,zpi) of the control points Pi (i=1,2,3) in the camera coordinate system are derived from coordinates (Xpi,Ypi,Zpi) of the surveying coordinate system by Eq. (1), thereby, approximate camera coordinates (xpGi,ypGi,zpGi) of the control points Pi are obtained by substituting the approximate exterior orientation parameters (XGO,YGO,ZGO,ωG,φG,κG) and the surveying coordinates (xpi,ypi,zpi) of the control points Pi into Eq. (1).
where {Tjk} represents a rotational matrix, and each of the elements Tjk are described in the following forms.
T11=cos φ·cos κ
T12=cos ω·sin κ+sin ω·sin φ·cos κ
T13=sin ω·sin κ−cos ω·sin φ·cos κ
T21=−cos φ·sin κ
T22=cos ω·cos κ−sin ω·sin φ·sin κ
T23=sin ω·cos κ+cos ω·sin φ·sin κ
T31=sin φ
T32=−sin ω·cos φ
T33=cos ω·cos φ
The screen coordinates (spi′,tpi′) of the image point Pi′ corresponding to the control points Pi, which have not being yet been compensated by the inner orientation parameters, are derived from the collinear condition (wherein a control point, the center of projection, and the corresponding image point are aligned on the same line). Therefore, the uncompensated screen coordinates (spi′,tpi′) are calculated by substituting the exterior orientation parameters (XO,YO,ZO,ω,φ,κ) and the camera coordinates (xpi,ypi,zpi) of control points Pi into the following Eq. (2).
Although, the uncompensated screen coordinates (spi′,tpi′) are affected by distortion, the effect is compensated for by substituting the screen coordinates (spi′,tpi′) of each image point Pi′ and the approximate inner orientation parameters (fG,D2G,D4G,D6G,N1G,N2G,XCG,YCG) into the following Eq. (3). Namely, compensated approximate screen coordinates (scpGi′,tcpGi′) are obtained.
R2=spi′2+tpi′2
scpi′=spi′(1+D2R2+D4R4+D6R6)+(R2+2spi′2)N1+2spi′tpi′N2+XC
tcpi′=tpi′(1+D2R2+D4R4+D6R6)+2spi′tpi′N1+(R2+2tpi′2)N2+YC (3)
Further, approximate image coordinates (xpGi′,ypGi′) of the image points Pi′ can be derived by substituting the compensated approximate screen coordinates (scpGi′,tcpGi′) into the following Eq. (4).
xpi′=scpi′/(−Px)+W/2
ypi′=tcpi′/Py+H/2 (4)
where Px and Py are the pixel pitches of the CCD or the image in the horizontal and vertical directions respectively, and W and H are the numbers of pixels in the CCD or the image, which are aligned in the horizontal and vertical directions, respectively.
In Step S203, a merit function Φ is calculated in order to determine whether the approximately given exterior orientation parameters (XGO,YGO,ZGO,ωG,φG,κG) and inner orientation parameters (fG,D2G,D4G,D6G,N1G,N2G,XCG,YCG) are appropriate. For example, the merit function Φ is determined by the following Eq. (5).
Namely, in the present embodiment, the merit function Φ corresponds to the total amount of squared distances between the image coordinates (xpi′,ypi′) of image points Pi′ corresponding to the control points Pi (which are selected in the schematic image), and the approximate image coordinates (xpGi′,ypGi′) of the image points Pi′ calculated from the surveying coordinates (Xpi,Ypi,Zpi) of the control points Pi, the approximately given exterior orientation parameters (XGO,YGO,ZGO,ωG,φG,κG) and inner orientation parameters (fG,D2G,D4G,D6G,N1G,N2G,XCG,YCG).
Then, whether the merit function is smaller than a predetermined value is determined, in Step S204. That is, whether the approximate image coordinates (xpGi′,ypGi′) of the image points Pi′, which are obtained from the approximately given exterior orientation parameters (XGO,YGO,ZGO,ωG,φG,κG) and inner orientation parameters (fG,D2G,D4G,D6G,N1G,N2G,XCG,YCG), are sufficiently close to the image coordinates (xpi′,ypi′) of the image points Pi′ corresponding to the control points Pi, which are designated on the schematic image, is determined. When the value Φ is smaller than the predetermined value, this process is terminated and the values of the presently given exterior orientation parameters (XGO,YGO,ZGO,ωG,φG,κG) and inner orientation parameters (fG,D2G,D4G,D6G,N1G,N2G,XCG,YCG) are determined as the exterior and inner orientation parameters that represent the exterior orientation parameters (which indicate position and the inclination of the camera) and inner orientation parameters when the schematic image was captured.
On the other hand, when the value of Φ is determined to be larger or equal to the predetermined value, in Step S204, then in step S205, compensation values (δX,δY,δZ,δω,δφ,δκ,δf,δD2,δD4,δD6,δN1,δN2,δXC,δYC) for the approximately given exterior orientation parameters (XGO,YGO,ZGO,ωG,φG,κG) and inner orientation parameters (fG,D2G,D4G,D6G,N1G,N2G,XCG,YCG) are calculated by using the least squares method, for example. Namely, (scpi′,tcpi′) of Eq. (3) are substituted for (spi′,tpi′) of Eq. (2), which represents the collinearity condition. Eq. (2) is then subjected to Taylor's expansion at the approximate exterior orientation parameters (XGO,YGO,ZGO,ωG,φG,κG) and inner orientation parameters (fG,D2G,D4G,D6G,N1G,N2G,XCG,YCG) and the higher order terms are neglected so as to linearize the formula. Accordingly, the normal equations, for which the compensation values (δX,δY,δZ,δω,δφ,δκ) and (δf,δD2,δD4,δD6,δN1,δN2,δXC,δYC) are regarded as unknown values, are derived in order to obtain the appropriate compensation values (δX,δY,δZ,δω,δφ,δκ) and (δf,δD2,δD4,δD6,δN1,δN2,δXC,δYC).
In Step S206, the approximate exterior orientation parameters (XGO,YGO,ZGO,ωG,φG,κG) and inner orientation parameters (fG,D2G,D4G,D6G,N1G,N2G,XCG,YCG) are altered by the compensation values (δX,δY,δZ,δω,δφ,δκ) and (δf,δD2,δD4,δD6,δN1,δN2,δXC,δYC) which are obtained in Step S205. That is, each of the values (XGO,YGO,ZGO,ωG,φG,κG) and (fG,D2G,D4G,D6G,N1G,N2G,XCG,YCG) is replaced by the values (XGO+δX,YGO+δY,ZGO+δZ,ωG+δω,φG+δφ,κG+δκ) and (fG+δf,D2G+δD2,D4G+δD4,D6G+δD6,N1G+δN1,N2G+δN2,XCG+δXC,YCG+δYC) to thereby renew or compensate the position, inclination, and inner orientation parameters of the camera. Then the process returns to Step S202, so that Steps S202 through S206 are repeated until the value Φ is determined to be smaller than the predetermined value in Step S204.
With reference to
When the surveying instrument 10 is sighted to the point R2 on the collinear line LNC, a position which is measured is the object point Q2 which is outside the surveying area “A”. The image coordinates (xq2′,yq2′) on the schematic image, which corresponds to the object point Q2, is calculated by substituting the exterior orientation parameters (XO,YO,ZO,ω,φ,κ) and the inner orientation parameters (f,D2,D4,D6,N1,N2,XC,XC) of which the values are obtained by the single photograph orientation process, and the surveying coordinates (Xq2,Yq2,Zq2) of the object point Q2, into Eq. (1) through Eq. (4). However, the image coordinates (xq2′,yq2′) which correspond to the measured object point Q2 are different from the image coordinates (xq6′,yq6′) of the designated image point Q6′ since the object point Q2 is not equal to the object point Q6 which corresponds to the designated image point Q6′.
On the other hand, the point R6 on the collinear line LNC coincides with the object point Q6 where the collinear line LNC intersects with a subject within the surveying area “A”. Namely, image coordinates calculated from surveying coordinates that are obtained when the surveying instrument 10 is sighted to the point R6, coincide with the image coordinates (xq6′,yq6′) of the designated image point Q6′. Consequently, automatic measurement for an object point that corresponds to an arbitrarily designated pixel on the schematic image can be achieved by carrying out surveying along a collinear line corresponding to the arbitrarily designated pixel (a pixel within the surveying area “A” in the present embodiment), and searching for an object point of which the image coordinates obtained from its surveying coordinates substantially coincide with the image coordinates of the arbitrarily designated pixel.
With reference to
In Step S302, “1” is set for a pixel number “i”, so that Step S303 through Step S306 are repeated until “i” reaches “N”. In Step S303, surveying along the collinear line that corresponds to the image coordinates (xi′,yi′) of a pixel number “i” is automatically carried out. Namely, the position of an object point that corresponds to the pixel with the pixel number “i” is searched and measured by comparison between the image coordinates corresponding to measured surveying coordinates and the image coordinates (xi′,yi′) of the pixel with the pixel number “i” (object point searching process). Note that, the detail of the object point searching process will be explained latter.
In Step S304, the measurement data of the image coordinates (xi′,yi′) obtained by the object point searching process of Step S303 is recorded in the recording medium 42. In Step S305, whether the pixel number “i” is larger than the total pixel number “N” is determined. Namely, whether the object point searching process has been carried out for all pixels inside the surveying area “A” is determined. When it is determined that it has not, i.e. i=N, the number “i” is incremented by one in Step S306 and the process returns to Step S303. On the other hand, when it is determined that i=N, this automatic surveying operation ends.
In Step S403, the surveying coordinates (XRj,YRj,ZRj) of the hypothetical sighting points Rj are calculated from the unit collinear vector (rix,riy,riz), which is obtained in Step S401, and the distance “L”, as follows.
XRj=XO+rix×L
YRj=YO+riy×L
ZRj=ZO+riz×L (6)
Further, the horizontal angle θh and the altitude θp for sighing the sighting telescope 17 to the hypothetical sighting point Rj are obtained by Eq. (7) and Eq. (8).
Note that, as described in
In Step S404, the measurement or surveying is carried out for the direction which is sighted in Step S403. According to the above measurement, the surveying coordinates (XQi,YQi,ZQi) for the object point Qi is calculated from the substantial measurement data for the object point Qi which corresponds to the hypothetical sighting point Ri. The image coordinates (xQi′,yQi′) which correspond to the surveying coordinates (XQi,YQi,ZQi) of the object point Qi are then calculated from the exterior orientation parameters (XO,YO,ZO,ω,φ,κ) and the inner orientation parameters (f,D2,D4,D6,N1,N2,XC,XC), in Step S405.
In Step S406, whether the image coordinates (xQi,yQi) of the object point Qi coincide with the image coordinates (xi′,yi′) of the collinear line LNC is determined. For example, when the distance between the image coordinates (xQi′,yQi′) and the image coordinates (xi′,yi′) is within a predetermined value, it is determined that the coordinates (xQi′,yQi′) of the object point Qi coincide with the image coordinates (xi′,yi′) of the collinear line LNC. In this situation, the surveying coordinates (XQi,YQi,ZQi) of the object point Qi, which are calculated in Step S407, are defined as the surveying coordinates of the pixel (xi′,yi′) that corresponds to the pixel number “i”. Accordingly, the object point searching process for the pixel number “i” ends. On the other hand, when it is determined that the above two sets of coordinates do not coincide, the distance “L” is replaced by “L+ΔL” in Step S408, and the processes from Step S403 are repeated. Note that, “ΔL” is an increment for the distance “L” and defined by regarding the required precision and the surveying time.
Next, with reference to
In the present embodiment, a unit collinear vector is calculated by means of the least square method. Namely, in the unit collinear vector calculating process of the present embodiment, an arbitrary unit vector (rGix,rGiy,rGiz) is given as an approximate unit collinear vector for a pixel (xi′,yi′), in the first step. Therefore in Step S501, the image coordinates (xGi′,yGi′) for the surveying coordinates (XO+rGix,YO+rGiy,ZO+rGiz) that correspond to the end of the unit vector (rGix,rGiy,rGiz) which extends from the origin OC: (XO,YO,ZO) of the camera coordinates system, are calculated using Eq. (1) to Eq. (4) with respect to the exterior orientation parameters (XO,YO,ZO,ω,φ,κ) and the inner orientation parameters (f,D2,D4,D6,N1,N2,XC,XC).
In Step S502, a merit function Ψ=(xi′−xGi′)2+(yi′−yGi′)2 is calculated for the image coordinates (xGi′,yGi′) obtained in Step S501. Whether the merit function Ψ is smaller than a predetermined value is determined in Step S503. When it is determined that the merit function Ψ is smaller than the predetermined value, the unit vector (rGix,rGiy,rGiz) is defined as the unit collinear vector (rix,riy,riz) of the pixel that corresponds to the pixel number “i”, in Step S504, so that this unit collinear vector calculating process for the pixel number “i” ends.
On the other hand, when it is determined in Step S503 that the merit function Ψ is not smaller than the predetermined value, compensation values (δrix,δriy,δriz) for the approximate unit vector (rGix,rGiy,rGiz) are calculated by means of the least square method, in Step S505. Further, the approximate unit vector is replaced by a new unit vector (rGix+δrix,rGiy+δriy,rGiz+δriz). The process then returns to Step S501, so that Step S501 to Step S505 are repeated until the merit function Ψ is determined to be smaller than the predetermined value, in Step S503.
Note that, the volume of the surveying area is calculated from the measurement data of the surveying area, which is obtained from the above automatic surveying operation. Further, the area, the circumferential length, or the like, which relates to the surveying area can also be calculated by surveying only the circumference of the designated surveying area.
As described above, according to the present embodiment, a measurement point which is designated on a schematic image can be automatically sighted by a surveying instrument, so that the measurement point can be automatically measured. Further, although it is not depicted, when using a surveying instrument with a CCD provided inside, the sighting telescope can be more precisely sighted to a point by designating the point on an image captured by the CCD, which has a higher magnification than the schematic image, after rough sighting to the point. Further, according to the present embodiment, the surveying operation within a surveying area can be automatically carried out when an operator merely defines the surveying area on the schematic image, so that efficiency of the surveying operation can be greatly improved.
Note that, when using a lens system having a long focal length, such as a telephoto lens, distortion which is a part of the inner orientation parameters is sufficiently small, so that on some occasions it can be neglected. Namely, the inner orientation parameters (D2,D4,D6,N1,N2) can be neglected, so that the remaining inner orientation parameters are reduced to (f,XC,YC). In this case, only five control points Pi which are three-dimensionally arranged are required for obtaining the inner orientation parameters (f,XC,YC). Further, when the displacement of the principal point from the image center, the unsymmetrical components of the distortion, and the fourth and sixth distortion components are negligible, the remaining inner orientation parameters are reduced to (f,D2), so that the number of control points required for obtaining the inner parameters is reduced to four. As described above, since the number of unknown inner orientation parameters is reduced, the number of control points is also reduced, so that labor and time for the surveying can be reduced.
In the present embodiment, the digital still camera 20 is arbitrarily disposed with respect to the surveying instrument 10, however, the digital still camera 20 may be disposed at the position which is optically equivalent to the sighting telescope 17 of the surveying instrument 10, by using a camera mounting accessory equipped on the surveying instrument 10. In this case, the number of unknown exterior orientation parameters can be reduced, so that the number of control points can be reduced. Further, when a schematic image is captured with the digital still camera 20 placed at the position which is optically equivalent to the sighting telescope 17 of the surveying instrument 10 (when the center of projection for an imaging optical system of an imaging device is placed at a position that is optically equivalent to the sighting origin of a sighting telescope), the horizontal angle and the vertical angle can be directly calculated from image coordinates, so that the process is facilitated.
In the following, an alternate embodiment when the digital still camera 20 is positioned at a point where it is optically equivalent to the sighting telescope 17 is explained with reference to
xpn′=f·tan(θhn−θhO)/Px+W/2,
ypn′=f·tan(θpn−θpO)/Py+W/2 (9)
where θhO and θpO are initial sighting directions when the schematic image is captured (note that, when distortion compensation is necessary, the inner orientation parameters are also used). Therefore, when a sighting position (xpn′,ypn′) is designated on the schematic image, the horizontal angle θhn and the altitude θpn of the measurement point Pn can be calculated as the inverse of Eq. (9) that is:
θhn=tan−1(Px(xpn′−W/2)/f)+θhO
θpn=tan−1(Py(ypn′−H/2)/f)+θpO. (10)
Therefore, the surveying instrument can be sighted in the direction represented by Eq. (10).
In Step S601, correspondence between the schematic image and the surveying coordinates is establish based on the single photograph orientation process described in
According to the present alternate embodiment, an operator can visually recognize whether the sighting is finished on the schematic image, so that the efficiency of the sighting is improved.
Further, another alternate embodiment is described, wherein a surveying operation is carried out for a sighting point which has three-dimensional position information, such as when a measurement point to be sighted is given three-dimensional positional information as planning values, such as a staking point, or when sighting a measurement point has already been measured in a previous surveying. When three-dimensional information of a measurement point Pn is given, the surveying instrument 10 can be automatically sighted by calculating the horizontal angle θhn and vertical angle θpn by substituting the coordinates (Xn,Yn,Zn) of the measurement point Pn into Ri (XRi,YRi,ZRi) of Eq. (7) and Eq. (8).
In Step S701, correspondence between the schematic image and the surveying coordinates is establish according to the single photograph orientation process described in
According to the above alternate embodiment, a similar effect as that in the previous alternate embodiment can be obtained. Thus, the completion of the sighting can be visually confirmed on the schematic image and efficiency of the surveying operation is improved. Note that, in Step S703, a mark indicating a selected measurement point may be altered when the measurement point is designated. In the case of surveying a staking point and when the planning coordinate system for the staking point is not equivalent to the surveying coordinate system, the coordinate values of the staking point in the planning coordinate system are transformed to the surveying coordinate system. On the other hand, the inverse transformation can also be carried out.
In the present embodiments, the surveying instrument is oriented to the sighting point Ri on the collinear line LNC that corresponds to an arbitrary designated pixel, and the image coordinates (xQi′,yQi′) on the schematic image are calculated from the surveying coordinates (XQi,YQi,ZQi) measured in the above sighting. Further, whether the image coordinates (xQi′,yQi′) coincide with the image coordinates (xi′,yi′) of the arbitrarily designated pixel is determined. However, this may be determined by whether the surveying coordinates (XQi,YQi,ZQi) coincide with the surveying coordinates (XRi,YRi,ZRi) of the sighting point Ri.
Note that, although in the present embodiments, the control points are arbitrarily designated on the schematic image by using a pointing device, it is also possible to capture an image of a reference scale of which the dimensions are known, or reference marks at arbitrary positions within the schematic image, and to calculate the exterior orientation parameters by regarding them as the control points. In this case, the points on the reference scale or the reference marks may be designated on the schematic image by a pointing device or the like. Further, when the reference scale or the reference marks are applied the control points on the schematic image may be automatically detected by means of image processing.
In the present embodiments, a computer connected to a surveying instrument is used, however, the function of the computer may be integrated into the surveying instrument or a digital still camera. Further, although in the present embodiments, a digital image is obtained by a digital still camera, any type of image capturing device, such as a digital video camera and the like, can be used as long as it can finally produce a digital image.
Although the embodiments of the present invention have been described herein with reference to the accompanying drawings, obviously many modifications and changes may be made by those skilled in this art without departing from the scope of the invention.
The present disclosure relates to subject matter contained in Japanese Patent Application No. 2002-190599 (filed on Jun. 28, 2002) which is expressly incorporated herein, by reference, in its entirety.
Number | Date | Country | Kind |
---|---|---|---|
P2002-190599 | Jun 2002 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
4682218 | Kaneko | Jul 1987 | A |
6034722 | Viney et al. | Mar 2000 | A |
6137569 | Sasaki et al. | Oct 2000 | A |
6144761 | Kaneko et al. | Nov 2000 | A |
6304669 | Kaneko et al. | Oct 2001 | B1 |
6487517 | Sakai et al. | Nov 2002 | B1 |
6563574 | Ohtomo et al. | May 2003 | B1 |
6600511 | Kaneko et al. | Jul 2003 | B1 |
6618497 | Nakayama | Sep 2003 | B1 |
6618498 | Nakayama | Sep 2003 | B1 |
20030048355 | Shimoyama et al. | Mar 2003 | A1 |
Number | Date | Country |
---|---|---|
11325883 | Nov 1999 | JP |
11337336 | Dec 1999 | JP |
3261359 | Dec 2001 | JP |
Number | Date | Country | |
---|---|---|---|
20040004706 A1 | Jan 2004 | US |