The present invention relates to an image display apparatus and an image display method, and more particularly, to an apparatus and method for synthesizing images taken by a plurality of cameras into one image and displaying the image.
JP 09-114979 A proposes a camera system in which a subject is taken by a plurality of cameras located at different positions from each other and a plurality of the images taken by these cameras are used to generate a virtual image viewed from a viewpoint different from the viewpoints of the respective cameras. The three-dimensional position coordinates of the subject in the image taken by each of the cameras is obtained, and an image viewed from an arbitrary viewpoint is reconstructed based on these coordinates.
However, with respect to the system disclosed in JP 09-114979 A, the images taken by a plurality of cameras each having fixed positions are used to generate virtual images from different viewpoints. Accordingly, for example, in a case of parking a vehicle into a parking space, if the virtual image such as an overhead image and so on that enable a comfortable driving operation is generated, a plurality of cameras each having fixed positions in the parking space side must be installed. This causes such problems that space for installing the plurality of cameras is required, that costs are increased by installing a plurality of cameras, and that it takes time to install a plurality of cameras.
Recently, with the aim of improving the safety and operability of driving, a system has been used in which some cameras for taking the images of the rear part and the side part of a vehicle are mounted on the vehicle and the images taken by the cameras are displayed on a monitor located at the driver's seat at the time of, for example, backing the vehicle up.
By utilizing the aforementioned cameras that are mounted on a vehicle and installing only one fixed camera installed in the parking space side so as to be able to generate an overhead image, the driving operation at the time of parking can be easily and cheaply made more comfortable. However, since the movement of the vehicle changes the relative positional relationship between the cameras mounted on the vehicle and the fixed camera installed in the parking space side, it is difficult for the system disclosed in JP 09-114979 A to generate overhead images or the like.
The present invention has been made to solve the above-mentioned problem in the conventional art, and therefore has an object of providing an image display method that can display an image viewed from a given viewpoint even if a plurality of cameras having a variable relative positional relationship between each other are used.
The present image display apparatus comprises: a plurality of cameras for which the relative positional relationship changes; a relative position calculation unit for calculating the relative positional relationship of the plurality of cameras; an image composition unit for creating an image viewed from a given viewpoint by synthesizing images taken by the plurality of cameras based on the relative positional relationship calculated by the relative position calculation unit; and a monitor for displaying the image created by the image composition unit.
The present image display method comprises: calculating a relative positional relationship of a plurality of cameras, for which the relative positional relationship changes; creating an image viewed from a given viewpoint by synthesizing images taken by the plurality of cameras based on the calculated relative positional relationship; and displaying the created image viewed from the given viewpoint.
According to the present invention, because the relative positional relationship of the plurality of cameras is calculated by the relative position calculation unit and the images taken by the plurality of cameras are synthesized by the image composition unit based on the relative positional relationship, an image viewed from a given viewpoint can be displayed even if a plurality of cameras having a variable relative positional relationship between each other are used.
Hereinafter, Embodiments according to the present invention are described with reference to the accompanying drawings.
The moving object U is connected to a relative position calculation unit 1 that calculates a relative positional relationship of the moving camera B with respect to the fixed camera A. The relative position calculation unit 1, the fixed camera A, and the moving camera B are connected to an image composition unit 2. Furthermore, the image composition unit 2 is connected to a monitor 3.
Furthermore, the moving object U includes a sensor (not shown) for obtaining positional information such as its own current position, orientation and so on, and a detection signal from the sensor is input to the relative position calculation unit 1.
Next, referring to the flow chart of
Furthermore, in Step S3, the relative positional relationship of the moving camera B with respect to the fixed camera A is calculated by the relative position calculation unit 1. At this point, because the detection signal from the sensor for obtaining the positional information of the moving object U has been input to the relative position calculation unit 1, the relative position calculation unit 1 can recognize the current position and orientation of the moving object U based on the input detection signal. Then, by comparing the current position and orientation of the moving object U with the given place at which the fixed camera A is installed, the relative position calculation unit 1 calculates the relative positional relationship of the moving camera B with respect to the fixed camera A.
After the relative positional relationship between the fixed camera A and the moving camera B is calculated as described above, in Step S4, the image composition unit 2 synthesizes the image taken by the fixed camera A and the image taken by the moving camera B into one image based on the relative positional relationship, thereby creating an overhead image which is a composite image viewed downward.
Next, in Step S5, the overhead image is displayed on the monitor 3.
It should be noted that, as an example of synthesizing the image taken by the fixed camera A and the image taken by the moving camera B and creating the overhead image, there can be mentioned JP 03-099952 A.
Furthermore, as the sensor for obtaining the positional information of the moving object U, any sensor can be employed as long as it can recognize the current position and orientation of the moving object U. Examples of the sensor include; a speed sensor or a distance amount sensor and a yaw-rate sensor; a GPS sensor and a yaw-rate sensor; and two GPS sensors mounted at different positions of the moving object U.
The relative position calculation unit 1, the image composition unit 2 and the monitor 3 can be mounted on the moving object U. In this case, the fixed camera A and the image composition unit 2 may be connected by wire or by wireless. When the fixed camera A and the image composition unit 2 are connected by wireless, communication means has to be provided with both the fixed camera A and the image composition unit 2 respectively.
Furthermore, the relative position calculation unit 1, the image composition unit 2 and the monitor 3 may be installed at a given place instead of being mounted on the moving object U. In this case, the sensor for obtaining the positional information of the moving object U is connected to the relative position calculation unit 1 by wire or by wireless, and also, the fixed camera A and the moving camera B are connected to the image composition unit 2 by wire or by wireless.
Further, only the relative position calculation unit 1 may be mounted on the moving object U while the image composition unit 2 and the monitor 3 are installed at a given place. Otherwise, the relative position calculation unit 1 and the image composition unit 2 may be mounted on the moving object U while only the monitor 3 is installed at a given place.
As illustrated in
It is assumed that the mark M as the fixed target is fixed at a given place having a given positional relationship with respect to the fixed camera A and that the given positional relationship of the mark M with respect to the fixed camera A is previously recognized. For the mark M, as illustrated in
As described above, as long as the positional relationship of the mark M with respect to the fixed camera A is previously recognized, the mark M may be placed outside the field of view of the fixed camera A.
Similarly, in Embodiment 2, a composite image such as an overhead image and so on is displayed on the monitor 3 in accordance with Steps S1 to S5 of the flow chart illustrated in
At first, in Step S6, the image processing means 4 of the relative position calculation unit 1 extracts the five feature points C1 to C5 of the mark M from the image of the mark M taken by the moving camera B, and then detects and obtains two-dimensional coordinates of each of the feature points C1 to C5 on the image.
Next, in Step S7, based on the two-dimensional coordinates of the respective feature points C1 to C5 detected by the image processing means 4, the positional parameter calculating means 5 calculates positional parameters including six parameters which are three-dimensional coordinates (x, y, z), a tilt angle (angle of depression), a pan angle (angle of direction), and a swing angle (angle of rotation).
Here, the calculation method for the positional parameters by the positional parameter calculating means 5 is described.
First, a point on the ground, which moves in a downward direction perpendicular to the road surface from a given point on the moving object U, is set as an origin O, a road surface coordinate system in which an x-axis and a y-axis are set in the horizontal direction and a z-axis is set in the vertical direction is assumed, and an image coordinate system in which an X-axis and a Y-axis are set on the image taken by the moving camera B is assumed.
Coordinate values Xm and Ym (m=1 to 5) of the feature points C1 to C5 of the mark M in the image coordinate system are expressed, with the six positional parameters of the respective feature points C1 to C5 of the mark M in the road surface coordinate system, that is, coordinate values xm, ym, and zm, angle parameters Kn (n=1 to 3) of the tilt angle (angle of depression), the pan angle (angle of direction), and the swing angle (angle of rotation), by using functions F and G as follows:
Xm=F(xm,ym,zm,Kn)+DXm
Ym=G(xm,ym,zm,Kn)+DYm,
where DXm is the deviation between the X coordinate of each of the feature points C1 to C5 calculated by means of the functions F and G and the coordinate value Xm of each of the feature points C1 to C5 detected by the image processing means 4 and where DYm is the deviation between the Y coordinate of each of the feature points C1 to C5 calculated by means of the functions F and G and the coordinate value Ym of each of the feature points C1 to C5 detected by the image processing means 4.
Specifically, by expressing the X-coordinate and the Y-coordinate of each of the five feature points C1 to C5, a total of ten relational expressions are created with respect to the six positional parameters (xm, ym, zm, Kn).
Here, positional parameters (xm, ym, zm, Kn) which minimize the sum of squares S of the deviations DXm and DYm, which is expressed as follows:
S=Σ(DXm2+DYm2),
are obtained. In other words, an optimization problem for minimizing S is solved. Well-known optimization methods such as a simplex method, steepest descent method, Newton method, quasi-Newton method and so on may be used.
Because the positional parameters are determined by creating more relational expressions than the number (six) of the positional parameters (xm, ym, zm, Kn) to be calculated, it is possible to obtain the positional parameters (xm, ym, zm, Kn) with high accuracy.
In Embodiment 2 of the present invention, ten relational expressions are created by the five feature points C1 to C5 with respect to the six positional parameters (xm, ym, zm, Kn). However, the number of relational expressions merely has to be more than the number of the positional parameters (xm, ym, zm, Kn) to be calculated. Accordingly, if six relational expressions are created by at least three feature points, the six positional parameters (xm, ym, zm, Kn) can be calculated.
By using the positional parameters of the moving camera B calculated in the aforementioned manner, in Step S8, the relative position identifying means 6 identifies the relative positional relationship of the moving camera B with respect to the fixed camera A. Specifically, based on the positional parameters calculated by the positional parameter calculating means 5, the relative positional relationship between the moving camera B and the mark M is identified. Furthermore, the relative positional relationship between the moving camera B and the fixed camera A is identified because the given positional relationship of the mark M with respect to the fixed camera A is previously recognized.
The relative positional relationship between the fixed camera A and the moving camera B, which has been calculated by the relative position calculation unit 1 in the aforementioned manner, is transmitted to the image composition unit 2. The image composition unit 2 combines, based on the relative positional relationship, the image taken by the fixed camera A and the image taken by the moving camera B into one image to create an overhead image, which is then displayed on the monitor 3.
In Embodiment 2 of the present invention, the positional parameters including the six parameters which are the three-dimensional coordinates (x, y, z) of the moving camera B with reference to the mark M, the tilt angle (angle of depression), the pan angle (angle of direction), and the swing angle (angle of rotation) are calculated. Accordingly, even if there is a step or a tilt between the floor surface on which the mark M is located and the road surface on which the moving object U is currently located, the relative positional relationship between the mark M and the moving camera B, and further, the relative positional relationship between the fixed camera A and the moving camera B are accurately identified, thereby enabling creation of an overhead image with high accuracy.
However, when there is no tilt between the floor surface on which the mark M is located and the road surface on which the moving object U is currently located, by calculating the positional parameters including at least four parameters which are the three-dimensional coordinates (x, y, z) of the moving camera B with reference to the mark M and the pan angle (angle of direction), the relative positional relationship between the mark M and the moving camera B can be identified. In this case, four positional parameters can be obtained by creating four relational expressions by means of two-dimensional coordinates of at least two feature points of the mark M. However, it is desirable that two-dimensional coordinates of more feature points are used to calculate the four positional parameters for higher accuracy by a least-squares method or the like.
Furthermore, in a case where the mark M and the moving object U are on the same plane and there is no step or tilt between the floor surface on which the mark M is located and the road surface on which the moving object U is currently located, by calculating the positional parameters including at least three parameters the two-dimensional coordinates (x, y) of the moving camera B with reference to the mark M and the pan angle (angle of direction), the relative positional relationship between the mark M and the moving camera B can be identified. Similarly, in this case, three positional parameters can be obtained by creating four relational expressions by means of two-dimensional coordinates of at least two feature points of the mark M. However, it is desirable that two-dimensional coordinates of more feature points are used to calculate the three positional parameters for higher accuracy by a least-squares method or the like.
A parking-space-side apparatus 11 is provided within a parking space S or in the vicinity thereof, and a vehicle-side apparatus 12 is mounted on the vehicle to be parked in the parking space S.
The parking-space-side apparatus 11 includes the fixed camera A located at the back of the parking space S to take an image of the vicinity of an entrance of the parking space S, and the fixed camera A is connected to a communication unit 14 via an encoder 13. The encoder 13 is for compressing an image taken by the fixed camera A into a format suitable for wireless transmission. The communication unit 14 is mainly for transmitting image data compressed by the encoder 13 to the vehicle-side apparatus 12. A control unit 15 is connected to the fixed camera A and the communication unit 14. Furthermore, the parking-space-side apparatus 11 includes the mark M fixedly provided on the floor surface in the vicinity of the entrance of the parking space S.
It is assumed that internal parameters (focal length, strain constant, and the like) and external parameters (relative position, angle, and the like, with respect to the parking space S) of the fixed camera A are previously known. Similarly, it is assumed that the relative position of the mark M with respect to the fixed camera A is known.
On the other hand, the vehicle-side apparatus 12 includes the moving camera B provided at the rear of the vehicle to take an image behind the vehicle. At the time of backing into the parking space S, an image of the mark M in the parking space S is taken by the moving camera B. The moving camera B is connected to the relative position calculation unit 1. Furthermore, the vehicle-side apparatus 12 includes a communication unit 16 for communicating with the communication unit 14 of the parking-space-side apparatus 11, and the communication unit 16 is connected to a decoder 17. The decoder 17 is for decoding the compressed image data received by the communication unit 16 from the parking-space-side apparatus 11.
The decoder 17 is connected to the image composition unit 2. An image selection unit 18 is connected to both the image composition unit 2 and the moving camera B. The image selection unit 18 is connected to the monitor 3 located at the driver's seat of the vehicle.
Furthermore, a control unit 19 is connected to the moving camera B, the relative position calculation unit 1, the communication unit 16, and the image selection unit 18.
It is assumed that internal parameters (focal length, strain constant, and the like) of the moving camera B and the relative position and angle of the moving camera B with respect to the vehicle are previously known.
As illustrated in
Next, with reference to the flow chart of
First, in Step S11, in a state where the vehicle is located in the vicinity of the parking space S so that the mark M is within the field of view of the moving camera B, the control unit 19 of the vehicle-side apparatus 12 operates the moving camera B to take an image of the mark M.
The image taken by the moving camera B is input to the relative position calculation unit 1. In the subsequent Step S12, the relative position calculation unit 1 calculates the relative positional relationship between the moving camera B and the fixed camera A according to the flow chart illustrated in
Accordingly, based on the relative positional relationship calculated by the relative position calculation unit 1, the control unit 19 calculates a relative distance L of the moving camera B with respect to the fixed camera A. In Step S13, the relative distance L is compared with a given value Lth, and when the relative distance L is larger than the given value Lth, the control unit 19 causes the image selection unit 18 to select the image data from the moving camera B in Step S14. Consequently, as illustrated in
On the other hand, when the relative distance L of the moving camera B with respect to the fixed camera A is less than or equal to the given value Lth, the control unit 19 transmits a request signal for the image data from the communication unit 16 to the parking-space-side apparatus 11 in Step S15.
In the parking-space-side apparatus 11, when the communication unit 14 receives the request signal for the image data from the vehicle-side apparatus 12, the control unit 15 operates the fixed camera A, thereby taking an image of the vicinity of the entrance of the parking space S in Step S16. Then, after the image data taken by the fixed camera A is compressed by the encoder 13 into a format suitable for wireless transmission, the compressed image data is transmitted from the communication unit 14 to the vehicle-side apparatus 12.
In Step 17, after the image data from the parking-space-side apparatus 11 is received by the communication unit 16 of the vehicle-side apparatus 12, the image data is decoded by the decoder 17, and then transmitted to the image composition unit 2.
In Step S18, the image composition unit 2 synthesizes the image taken by the fixed camera A and the image taken by the moving camera B into one image to thereby create an overhead image based on the relative positional relationship between the fixed camera A and the moving camera B.
In Step S19, because the relative distance L of the moving camera B with respect to the fixed camera A is less than or equal to the given value Lth, the control unit 19 causes the image selection unit 18 to select the image data from the image composition unit 2. Consequently, as illustrated in
It should be noted that the given value Lth used in Step S13 may be set as follows. For example, after assuming a garage in which the both sides and the rear side of the parking space S are divided by walls or the like, the given value Lth is set with reference to a relative distance between the moving camera B and the fixed camera A when a part of a vehicle V during backward parking begins to enter the field of view of the fixed camera A.
If the given value Lth is set in the aforementioned manner, the image of the area behind the vehicle taken by the moving camera B is displayed on the screen of the monitor 3 as illustrated in
As illustrated in
Next, with reference to the flow chart of
At first, in Step S21, in a state where the mark M is within the field of view of the first moving camera B1, an image is taken by the first moving camera B1. In the subsequent Step S22, the first calculation unit 22 of the relative position calculation unit 21 extracts five feature points C1 to C5 of the mark M from the image of the mark M taken by the first moving camera B1, and then detects and obtains two-dimensional coordinates of the respective feature points C1 to C5 of the image. Afterwards, in Step S23, based on the two-dimensional coordinates of the feature points C1 to C5, positional parameters including six parameters which are the three-dimensional coordinates (x, y, z) of the first moving camera B1 with reference to the mark M, the tilt angle (angle of depression), pan angle (angle of direction), and swing angle (angle of rotation) are calculated.
Next, in Step S24, in a state where the mark M is within the field of view of the second moving camera B2, an image is taken by the second moving camera B2. In the subsequent Step S25, the second calculation unit 23 of the relative position calculation unit 21 extracts the five feature points C1 to C5 of the mark M from the image of the mark M taken by the second moving camera B2, and then detects and obtains two-dimensional coordinates of the respective feature points C1 to C5 of the image. Afterwards, in Step S26, based on the two-dimensional coordinates of the feature points C1 to C5, positional parameters including six parameters which are the three-dimensional coordinates (x, y, z) of the second moving camera B2 with reference to the mark M, the tilt angle (angle of depression), the pan angle (angle of direction), and the swing angle (angle of rotation) are calculated.
Furthermore, in Step S27, the third calculation unit 24 of the relative position calculation unit 21 identifies the relative positional relationship between the first moving camera B1 and the second moving camera B2 based on the positional parameters of the first moving camera B1 and the second moving camera B2 calculated by the first calculation unit 22 and the second calculation unit 23 respectively.
The relative positional relationship between the first moving camera B1 and the second moving camera B2 calculated by the relative position calculation unit 21 is transmitted to the image composition unit 2 in the aforementioned manner, and in Step S28, the image composition unit 2 synthesizes the image taken by the first moving camera B1 and the image taken by the second moving camera B2 into one image based on the relative positional relationship, thereby creating an overhead image. In Step S29, the overhead image is displayed on the monitor 3.
It is assumed that the relative positional relationship between the mark M1 and the mark M2 is previously known. For example, the relative positional relationship between the mark M1 and the mark M2 may be stored in the relative position calculation unit 21. Alternatively, the marks M1 and M2 may be configured to record the respective positional information by means of a barcode or the like so that the positional information can be detected from the image taken by each camera. Still alternatively, the positional information of each of the marks M1 and M2 may be output from a wireless transmitter located in the vicinity of each mark, thereby allowing the positional information to be detected by receiving at the first moving object U1 and the second moving object U2.
In the case where the first moving camera B1 takes an image of the mark M1 and the second moving camera B2 takes an image of the mark M2, the first calculation unit 22 of the relative position calculation unit 21 calculates the positional parameters of the first moving camera B1 with reference to the mark M1 based on the image of the mark M1 taken by the first moving camera B1 and the second calculation unit 23 of the relative position calculation unit 21 calculates the positional parameters of the second moving camera B2 with reference to the mark M2 based on the image of the mark M2 taken by the second moving camera B2.
Here, because the relative positional relationship between the mark M1 and the mark M2 is previously recognized, the third calculation unit 24 of the relative position calculation unit 21 can identify the relative positional relationship between the first moving camera B1 and the second moving camera B2 based on the respective positional parameters calculated by the first calculation unit 22 and the second calculation unit 23 and the relative positional relationship between the marks M1 and M2.
As a result, the image composition unit 2 can synthesize the image taken by the first moving camera B1 and the image taken by the second moving camera B2 into one image based on the relative positional relationship, thereby creating an overhead image which is then displayed on the monitor 3.
In a case where the mark photographed by each of the moving cameras B1 and B2 is not specified as one of the marks M1 and M2, or in a case where both the marks M1 and M2 are photographed at the same time, in order to determine whether the photographed mark is the mark M1 or the mark M2, it is desirable that the marks M1 and M2 be different from each other in shape, color, or the like.
In each of Embodiments described above, the composite image created by the image composition unit 2 is not limited to the overhead image, and hence the image composition unit 2 can create an image viewed from an arbitrary virtual viewpoint P different from the viewpoint of each camera. Furthermore, the composite image is not necessarily limited to the image viewed from the virtual viewpoint P different from the viewpoint of each camera, and hence the image composition unit 2 can create an image viewed from the viewpoint of any one of a plurality of cameras. For example, in the aforementioned Embodiment 3 of the present invention, the image composition unit 2 may convert the image taken by the fixed camera A in the parking space side into an image viewed from the viewpoint of the moving camera B mounted on the vehicle, and synthesize the converted image with the image taken by the moving camera B, whereby it is possible to obtain a composite image that is added with a see-through area of such a blind spot within the parking space S as illustrated in
In the above-mentioned Embodiments of the present invention, the images taken by one fixed camera A and one moving camera B, or the images taken by two moving cameras B1 and B2 are synthesized. However, it is also possible to create an image viewed from the arbitrary virtual viewpoint P by synthesizing images taken by more than three cameras of which the relative positional relationships change.
Furthermore, the moving objects U, U1, and U2 are not limited to moving equipment such as a vehicle and so on but, for example, may be a person walking with a camera.
The mark used in Embodiments 2 to 5 of the present invention has a square-like shape obtained by making four isosceles right triangles abut to one another, but is not limited to such marks, as various kinds of marks may be used. However, it is desirable that the mark have a specific shape, color, or the like that makes it easier to identify against other shapes existing in the natural world, whereby the presence of the mark is easily recognized through image detection by the image processing means 4. Moreover, it is desirable that the feature points within the mark are easily detected.
Furthermore, it is desirable that the mark is sufficiently large for the relative positional relationship between the camera and the mark to be calculated with high accuracy based on the two-dimensional coordinates of the detected feature points and also that the mark is placed at a place where the mark cam be easily detected by the camera. Specifically, the mark may be placed in the vicinity of where the camera is located, such as a floor surface, a wall surface, a ceiling surface and so on.
Number | Date | Country | Kind |
---|---|---|---|
2007-272654 | Oct 2007 | JP | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/JP2008/068152 | 10/6/2008 | WO | 00 | 3/2/2010 |