This invention relates to a photographic apparatus, device and method for taking separate textural and silhouette images of objects for use in generating three-dimensional models of such objects.
Recently, technologies generating a three-dimensional model of an object from a plurality of two-dimensional images taken from a plurality of positions and/or orientations have been developed. In prior art devices, it is known to determine the silhouettes of a number of photographed images of the device and to use those silhouettes to generate a three-dimensional model of the object, the model consisting of a number of polygons. Photographed images are used to generate textures for application to each polygon of the three-dimensional images to generate the final model of the object.
For example, U.S. Pat. No. 6,317,139 discloses a method and apparatus in which two-dimensional images are taken from a number of positions using a video camera and binary silhouettes are determined for each image. Individual silhouettes are projected onto a volume and the projected volumes are combined to generate a single combined volume representing the shape of the three-dimensional object. The accuracy of the projection is dependent on the number of images taken and the positions from which the images are taken. A 3-D isosurface is determined from the combined volume. A set of polygons approximating to the 3-D isosurface is generated and texture images are used to fill the polygons with textural data to generate the final 3-D image.
Technologies to display such three-dimensional models using an internet browser and to manipulate the view of model, for example by rotating the model, have also been developed. Using these technologies to make three-dimensional models observable through the internet browser, it becomes possible for electronic commerce (E-Commerce) customers to observe merchandise as three-dimensional objects. Thus, it is expected that such three-dimensional object modelling technologies will greatly contribute to the advancement of E-Commerce businesses.
The methods known from the prior art generally require the kind of specialist equipment normally found in a photographic studio for photographing the objects. For example, the lighting conditions and backgrounds for each of the photographs must be arranged so as to be able to take both textural and silhouette images effectively. This is significantly complicated by the need to take images from a number of positions and orientations. Furthermore, the accuracy of the final three-dimensional model is dependent to a large extent on the positions from which the two-dimensional images are taken. Thus, the photographer must be sufficiently skilled in generating three-dimensional images to determine where the two-dimensional images should be taken from.
One prior art method for generating silhouettes of objects is the choroma-key technique. Using the choroma-key technique, it is necessary to prepare a single coloured background of a colour that is not significantly used in an object to be photographed. To accomplish this, the Japanese-Laid Open Patent 2001-148021 introduces a technology in which the colour and/or brightness of the background and the stand on which the object is supported in front of that background can be varied. Before taking images of the object, the colour and/or the brightness of the background and the stand must be manually chosen depending on the colour and brightness of the object to enable the object to be easily extracted from the photographed image, which includes the background, the table and the object. The extraction is performed by comparing pixels in the images of the object and finding places where the difference between adjacent pixels exceeds a predetermined threshold: these positions are the edges of the object. JP 2001-148021 also discloses displaying the texture pattern on the stand and the background in the case of a reflective object.
Another known means for extracting the silhouette of an object is to take an image of the background of the object and the table (an “object-setting surface”) without the object and then take an image including all of the background, the object-setting surface and the object. After these photographs have been taken, a difference between images can be determined to obtain a silhouette.
When silhouette information is obtained by using the techniques described above, a user has to manually change the background and object-setting surface or has to set and remove the object manually. Therefore, it is impossible to complete all photographs without a manual operation by the user. Accordingly, using these techniques, it is not possible to provide a three-dimensional modelling system that is fully automatic and in which the user operation is relatively straightforward.
By not only photographing lateral images but also top and bottom images of the object, it is possible to create a three-dimensional model observable from all orientations. However, in order to take photographic images from below the object, it is necessary for an operator to invert the object so that the bottom of the object can be photographed. Inverting the object also causes another adjustment between the image of the bottom and other pre-photographed images to be necessary, since the picture shooting distance for photographing the bottom is normally slightly different from the one for photographing other images. An operator must then adjust the size and orientation of these images by using computer programs. This adds to both the workload and operating skill required of the operator and also adds to the time it takes to create a three-dimensional model of an object.
There are a number of problems associated with the methods described above. For example, the known three-dimensional modelling systems generally required skilled operators, both in terms of the photographic skills required and the judgement of where to take images from in order to obtain a good three-dimensional image. Preparing all the required silhouettes is crucial to the quality of the final model. Getting the lighting right to remove shadows, setting up the camera, and getting the background and stand conditions correct results in a process that is both skilful and laborious.
In the known methods mentioned above, it takes a long time to take a small number of images. As a result, the number of images that can realistically be taken is often limited owing to the time available. This reduces the quality of the final three-dimensional images and adds the importance of the skill and experience of the operators. Consider, for example, a cylinder. Taking six images, from six different locations, will result in the cylinder being modelled as a six-sided object; taking twenty images will give a more accurate twenty-sided model. More complicated shapes lead to further problems for the operator to consider.
There are also a number of problems associated with the table on which an object to be modelled is placed. In some prior art devices, the table is supported by a frame. The table may be a turntable and may include a turning mechanism as part of the frame. A spindle may be included to assist with the rotation of the turntable. All such support means and turning mechanisms are visible in the image taken by the image capturing device which can cause complications to the processing of the image.
In order to create an entire visual hull, it is necessary to capture silhouette image from various longitudinal and latitudinal orientations in respect of the object. A backlight unit needs to be placed behind the target object with respect to image capturing device so that image frame of the camera covers the entire target object with certain amount of border to detect entire silhouette. A problem with such a system is that the mechanical requirements of the system tend to lead to the system being large.
The present invention seeks to overcome at least some of the problems identified above.
The present invention provides a photographic apparatus for taking images of an object for use in generating a three dimensional model of the object, the photographic apparatus comprising an object placing unit for placing the object, an image capturing unit for capturing images of the object for use in generating the three dimensional model, and an illumination unit, the image capturing unit and the illumination unit being connectedly moveable relative to the object placing unit such that, in use, the object may be placed by the placing unit both in the field of view of the image capturing unit and in a position where the illumination unit is capable of providing illumination for the image capturing device to take silhouette images of the object.
Providing an illumination unit that is connectedly moveable with the image capturing unit leads to a number of advantages. As the image capturing unit is moved to a number of different positions, the same illumination unit can be used. This leads to consistency in the lighting provided. Further, the system is more flexible than a system in which a multiplicity of fixed illumination units are provided since the image capturing unit and illumination unit can be positioned in any desired position.
The image capturing unit and the illumination unit may be arranged to be rotatably moved about an axis of rotation such that, whatever angle the image capturing unit is at relative to the object, the object is positioned between the image capturing device and the illumination unit.
The placing unit may include a transparent table (such as a glass table) on which the object is placed, with said axis of rotation being closely located above the table. Moreover, the placing unit may include a rotatable turntable to enable the image capturing unit to be used to take images of the object at two or more different orientations. In one embodiment of the invention, the turntable is settable to sixteen turntable positions.
Preferably, the image capturing unit can be used to take both silhouette and textural images of the object, wherein the illumination unit provides different illumination when textural images are taken from when the silhouette images are taken. For example, a backlight unit may be provided for taking silhouette images and a front light unit may be provided for taking textural images.
The image capturing unit may take two or more silhouette images of the object at different orientations in a first period and two or more textural images of the object at different orientations in a second period, the first and second periods being non-overlapping. This may be preferred to taking one silhouette image and then one textural image if the lighting needs to be changed between taking the silhouette and textural images.
A second illumination unit may be provided attached to said image capturing unit for providing illumination for the image capturing device to take textural images.
The image capturing unit may include an image capturing device and an optical device with the optical device deflecting an optical axis extending from the object to the image capturing device. A relative angle of the optical device and the image capturing device may be adjustable in order to move an image of the object towards the centre of an optical view of the image capturing device. Further, the relative angle the optical device and the image capturing device may be dependent on the angle of the image capturing device relative to the object and/or on the size of said object. In particular, the amount of adjustment may be at its greatest when the angle of the image capturing device relative to the object is at its smallest. In one embodiment of the invention, the optical device is a tiltable mirror.
In embodiments where the object placing unit is a turntable, the table may be supported by one or more support wheel arrangements and driven by a driving wheel arrangement. The support and/or driving arrangements are preferably arranged so that they are not in the area of the target object with some redundant border area within the entire field of view of the image capturing unit.
In a preferred arrangement of the invention, the illumination unit is mounted between a right illumination arm and a left illumination arm and the image capturing unit is mounted between a right camera arm and a left camera arm, with the right illumination arm and said right camera arm meeting at a right arm joint, and the left illumination arm and the left camera arm meeting at a left arm joint. The apparatus may further comprise an arm drive, wherein said arm drive is arranged to rotate said illumination and camera arms so as to rotate said illumination unit and said image capturing unit about an axis of rotation. The arm drive may be driven by a stepping motor.
The placing unit may include a transparent table on which the object is placed, and, in use, the image capturing unit may be settable to at least four angles relative to the table. In one form of the invention, the image capturing unit is settable to angles of 80, 45 and 10 degrees above the horizontal and 70 degrees below the horizontal. Clearly other suitable angles may be chosen instead of, or in addition to, those listed above. The image capturing unit may take a relatively large number of images of said object when said image capturing unit is at a lower angle relative to the table and a relatively small number of images of said object when said image capturing unit is at a greater angle relative to the table as it generally considered that this arrangement produces the most accurate three dimensional models with the smallest number of images taken.
Furthermore, in the use of the invention, an exposure parameter of said image capturing unit may be set such that the resulting image is underexposed when said image capturing unit is capturing silhouette data. This increases the contrast between light and dark and makes the determination of silhouettes easier.
The present invention also provides a method of generating a three dimensional model of an object, the method comprising the steps of:
The image capturing unit and the illumination unit may be rotatably mounted about an axis of rotation such that, whatever angle the image capturing unit is at relative to the object, the object is positioned between the image capturing unit and the illumination unit. The centre of rotation may be closely located above a table on which the object is placed.
The placing unit may include a turntable and may be rotatable to enable the image capturing unit to be used to take images of the object at two or more different orientations. In one form of the invention, the turntable is settable to sixteen different positions.
The image capturing unit preferably includes an image capturing device and an optical device with the optical device deflecting an optical axis extending from the object to the image capturing device in said step of taking a plurality of silhouette images. The optical device may be a mirror.
The image capturing device and the optical device are relatively tiltable in order to move an image of the object towards the centre of an optical view of the image capturing device in said step of taking a plurality of silhouette images. The magnitude and direction of the tilt may be dependent on the angle of the image capturing device relative to the object and/or on the size of said object. For example, the tilt angle may be at its greatest when the angle of the image capturing device relative to the object is at its smallest
The method may also include a step of performing a calibration subroutine to generate calibration data prior to the step of placing said object, wherein said calibration subroutine comprises the steps of:
Preferably, the images of the calibration mat are taken from every orientation at which said silhouette images are to be taken of an object to be modelled.
The method may include the step of taking a plurality of textural images of the object to be modelled from different orientations. In one form of the invention, a period for said step of taking the silhouette images and a period for said step of taking the textural are non-overlapping.
Another illumination unit attached to said image capturing unit may provided to provide illumination for the image capturing device to take the textural images of the object. In one form of the invention, a backlight unit provides the illumination for silhouette images and a front light unit provides the illumination for textural images.
The present invention also provides a photographic apparatus for taking images of an object for use in generating a three dimensional model of the object, the photographic apparatus comprising an object placing unit for placing the object and an image capturing unit for capturing images of the object for use in generating the three dimensional model, the image capturing unit including an image capturing device and an optical device to deflect an optical axis extending from the object to the image capturing device, the apparatus being arranged such that, in use, the image capturing unit is arranged to be rotatably moved about an axis of rotation such that, whatever angle the image capturing unit is at relative to the object the object may be placed by the object placing unit in the field of view of the image capturing device.
The relative angle of the optical device and the image capturing device may be adjustable in order to move an image of the object towards the centre of an optical view of the image capturing device. The relative angle the optical device and the image capturing device may be dependent on the angle of the image capturing device relative to the object and/or on the size of said object. In one form of the invention, the optical device is a tiltable mirror.
The object placing unit may include a table on which the object is placed, and the angle of deflecting the optical axis may be greater when the angle of image capturing device relative to the turntable is smaller.
In one form of the invention, an illumination unit is provided that is connectedly moveable with the image capturing unit relative to the object placing unit such that, in use, the object may be placed by the placing unit both in the field of view of the image capturing unit and in a position where the illumination unit is capable of providing illumination for the image capturing device to take silhouette images of the object. That illumination device may be a backlight unit. A front light unit may also be provided in addition to the backlight unit.
The placing unit may include a rotatable turntable to enable the image capturing unit to be used to take images of the object at two or more different orientations. In one form of the invention, the turntable is settable to sixteen different positions.
The present invention also provides a system for generating three dimensional models of an object, the system comprising any of the apparatuses described above, the system further comprising control means for obtaining image data and means for generating a three dimensional model from said images. The control means may include a graphical user interface, a display for displaying information for an operator, and input means to enable an operator to communicate with the system.
The present invention also provides a method of generating a three dimensional model of an object, the method comprising the steps of:
By way of example only, embodiments of the present invention will now be described with reference to the accompanying drawings, of which:
The enclosure 2 includes a door 10 that is shown in an open position in
Photographic apparatus 12 includes a glass turntable 14 on which an object to be photographed can be placed. The glass turntable is rotatable about a central vertical axis 16 to enable an object placed on the turntable 14 to be photographed from many angles. A camera unit 18 is provided to take photographic images of an object on the turntable 14. The camera unit 18 comprises a camera 20, a zoom lens 22 and a mirror 23 with a tilting mechanical stage 23a. The zoom position of the zoom lens 22 is electrically controllable. Detailed description of suitable controlling mechanisms for such a zoom lens are omitted from the present description since they do not relate directly to the present invention and suitable implementations are well known to persons skilled in the art.
A front fluorescent light unit 24 is provided on the camera unit 18 and a diffusion panel 25 is provided in front of the front fluorescent light unit to diffuse the light from front fluorescent light unit 24, to reduce glare from the light unit, for example. The front fluorescent light unit 24 is used to provide appropriate lighting to enable the camera 20 to take photographs of an object placed on the turntable 14 for the generation of textural data for use by the three dimensional modelling software. The camera unit 18 is mounted on a central camera arm 26. Central camera arm 26 extends from a left camera arm 28 to a right camera arm 30.
A backlight unit 32 comprising rear fluorescent light tubes 34 having a diffusion panel 35 in front of the rear fluorescent light tubes is provided. The backlight unit 32 is positioned such that an object placed on the turntable 14 is located between the backlight unit 32 and the camera unit 18. The backlight unit 32 is illuminated when the camera unit 18 is being used to capture a silhouette image of an object placed on the turntable 14.
The backlight unit 32 is mounted between a right backlight arm 36 and a left backlight arm 38. The right backlight arm 36 is connected to the right camera arm 30 by a right arm joint 40. The left backlight arm 38 is connected to the left camera arm 28 by a left arm joint 42.
Polarising filters (not shown) having dimensions similar to the diffusion panels 25 and 35 may be provided for the front fluorescent light unit 24, the backlight unit 32 and for the image capturing camera 20. By setting the polarisation angle of the polarisation filters approximately at right angles to one another, the effect of reflected light from the diffusion panels 25 and 35 can be reduced.
It should be noted that whilst the front light unit 24 and backlight unit 32 are described herein as fluorescent light units, other types of lights, such as flashlights and/or tungsten lights may be used.
A frame 44 is provided to support the elements that support the turntable 14 (described further below). Further, a right arm pillar 46 extends from the support frame 44 to the right camera arm 30 to support the right camera arm and the right backlight arm 36. In a similar manner, a left arm pillar 48 extends from the support frame 44 to the left camera arm 28 to support the left camera arm and the left backlight arm 38.
The turntable support frame 44 includes a drive wheel arrangement indicated generally by the reference numeral 50, a first support wheel arrangement indicated generally by the reference numeral 52a, a second support wheel arrangement indicated generally by the reference numeral 52b and a third support wheel arrangement indicated generally by the reference numeral 52c, which is shown in
The first support wheel arrangement 52a is best shown in
The first support wheel arrangement 52a includes a support 55, a lower roller 58 and a side roller 60. As shown in
The second and third support wheel arrangement 52b and 52c support the glass turntable 14 in a similar manner.
The drive wheel arrangement 50 is shown in
Clearly, the drive wheel arrangement 50 shown in
A photodetector device 76 is also shown in
As shown in
As noted above, the camera arms and the backlight arms are held at fixed positions with respect to one another, but those arms can be rotated together relative to the glass turntable. The arms are driven by arm drive 80. When driven by arm drive 80, the camera unit 18 moves in an arc indicated in
Left arm pillar 48 is penetrated by a spindle 82 carrying a gear 84. Gear 84 (and hence spindle 82) is driven by a toothed belt 86. Belt 86 extends around roller 84 and a smaller gear 88. Gear 88 is driven by stepping motor 90. Stepping motor 90 drives roller 88, which, via toothed belt 86 and gear 84, drives spindle 82. The stepping motor 90 is secured with a plate 90a which is attached to the left pillar 48 by three screws 90b, 90c, and 90d. Spindle 82 is connected to the camera arm 28 and operates to rotate both the camera arm 28 and the backlight arm 38 (as well as right camera arm 30 and right backlight arm 36). The axis of rotation 77 passes through the centre of the spindle 82.
Of course, the arm drive 80 shown in
As described above, the glass turntable 14 is supported by support wheel arrangements 52a, 52b and 52c and is driven by drive wheel arrangement 50. The support wheel arrangements 52a, 52b and 52c support the glass turntable on three sides.
There is no support on a fourth side that is in the field of view of the camera.
As discussed above, the left and right camera arms 28 and 30, and the left and right backlight arms 38 and 36, are connected together and can be rotated relative to the turntable 14 by arm drive 80. FIGS. 9 to 12 show the camera and backlight arms in a number of different positions relative to the turntable. None of the support wheel arrangements 52a, 52b and 52c or drive wheel arrangement 50 are within the optical view of the camera 20 (with some boundary area) of either the small object 92 or the large object 94. Further, no mechanical gear or belt is used to rotate the glass turntable, neither is a spindle used. As a result, there are no obstacles in the area of target object with some boundary area within the optical field of view of the digital camera 20. Clearly, this is advantageous because any such obstacles would be visible in the images taken by the camera 20.
In
In the use of the photographic apparatus 12 to capture a plurality of images of an object, different images can be taken at different elevations. For example, views can be taken at raised positions relative to the turntable (as in
As shown in FIGS. 9 to 12, the camera unit 18 and the backlight unit 32 rotate relative to turntable 14 on which an object to be modelled can be placed. Thus, the same backlight unit 32 is used for all positions of the camera 20. This ensures uniformity in the distance from the camera 20 to the backlight unit 32 and also ensures uniformity in the brightness and hence in the image generated. The use of a single movable backlight unit is preferable to the use of multiple fixed backlight units for a number of reasons. For example, with fixed backlight units there is the potential for backlight units to be in the background of a captured image. Also, the use of multiple backlight units increases the size and cost of the photographic apparatus. The use of a single camera and backlight unit increases the flexibility of the system since the camera and backlight can be positioned at any angle relative to the turntable. This is simply not possible if fixed devices are used. In the preferred embodiment described below, there are four possible elevation angles of camera arm, namely +80, +45, +10, and −70 degrees. However, the single camera with backlight can be positioned at any desired angle depending, for example, on the target object and the geometric accuracy required for the 3D object model.
One other advantage of utilising a single camera and backlight is that it is possible to have a larger and more flexible range of elevation angles. In the particular embodiment described, the camera arm is rotated between +80 degrees through −70 degrees with respect to the glass turntable 14. The range of possible elevation angles of camera position may be much reduced in the case of one or more fixed backlight units since part of a fixed backlight unit placed for a higher camera angle might be in the way of a camera position at a lower angle, or vice versa.
FIGS. 13 to 16 show simplified views of the photographic apparatus 12 having the camera arm orientated at different angles relative to the turntable. Of the elements of the photographic apparatus 12, only the turntable 14 and the digital camera 20, zoom lens 22 and mirror 23 of the camera unit 18 are shown in FIGS. 13 to 16. Also shown in FIGS. 13 to 16 are the small object 92 and the extremities of the corresponding optical vertical view 96a (which is related to the horizontal optical view 96 described with reference to
In each of the examples of FIGS. 13 to 16, the optical axis 100 is deflected by the mirror 23. This feature is discussed further below.
In
The optical axis 100 of the photographic apparatus passes through a fixed point relative to the glass turntable 14. That point is the axis of rotation 77 of the arc 82 (marked with a cross in FIGS. 13 to 16). In the examples of FIGS. 13 to 16, the point 77 is near the top of the object 92. The vertical optical view 96a extends a small distance either side of the point 77 so that the object 92 is within the optical view of the digital camera 20 with some boundary space on either side of the object 92. In the example of
Clearly, if an object moves outside the optical view of the camera, then it is not possible to obtain the required image data for that view of the object. Thus, if the object moves in the optical view vertically, then that optical view must be made sufficiently large (i.e. wide) to ensure that the object does not move outside of the optical view. The effect of this is that the optical view is significantly larger than the target object itself. If the object could be prevented from moving within the optical view, the optical view could be made smaller (i.e. narrower) and hence the image of the object could be captured with a higher resolution, leading to an increased quality of three-dimensional image, without risking the object moving out of the optical view.
FIGS. 21 to 24 show similar views of the photographic apparatus 12 as FIGS. 14 to 16, except that the small object 92 of FIGS. 13 to 16 is replaced with the large object 94 shown in
In a similar way to the example of
The problem of objects moving within the optical view of the digital camera 20 can be reduced in a simple manner by tilting the mirror 23 when the camera arm is orientated close to the horizontal.
FIGS. 17 to 20 are identical to FIGS. 13 to 16 respectively with the exception that the angle of the mirror 23 (and hence the position of the optical view relative to the object 92) in
Refer to
Refer now to
The mirror is not titled in the examples of
FIGS. 25 to 28 are identical to FIGS. 21 to 24 respectively with the exception that the angle of the mirror (and hence the position of the optical view relative to the object 94) in
Refer to
Refer now to
The mirror is not titled in the examples of
The amount of tilting required depends on the orientation of the camera unit 18 with respect to the glass turntable 14 since the problem is reduced when that angle increases. The direction of tilting required depends on the size of the object, since small objects will tend to appear low down in the optical view of the camera and large objects will tend to appear higher in the optical view of the camera.
Appropriate angles of tilting of the mirror have been determined by experimentation with the photographic apparatus. In the use of the photographic device 2 described in detail below, the camera arm is positioned at one of 80, 45, 10 and −70 degrees relative to the turntable 14. As discussed above, when the camera arm approaches the vertical, the object remains in the centre of the optical view and no tilting of the mirror is required. Hence, with the camera arm at either 80 or −70 degrees relative to the turntable, no mirror tilting is necessary. When the camera moves closer to the horizontal, small objects tend to move lower in the optical view and large objects tend to move higher in the optical view. The effect becomes more pronounced the closer the camera arm is to the horizontal. With a small object, a clockwise (negative) rotation of the mirror is required. Angles of 3 degrees and 5 degrees have been found to be effective when the camera arm is at 45 and 10 degrees to the horizontal respectively. With a large object, an anticlockwise (positive) rotation of the mirror is required. Angles of 8 and 10 degrees have been found to be effective when the camera arm is at 45 and 10 degrees to the horizontal respectively. These values are shown in the table of
Instead of having a mirror described above, an optical glass prism can be employed to deflect optical axis, which might have flatter reflection surface than a mirror.
Instead of tilting the mirror to move the object closer to the centre of the optical view of the camera, the position of the camera could itself be adjusted. This approach is relatively straightforward, however, repeating mirror adjustments accurately is easier than repeating camera adjustments accurately as the mechanical mass and weight of the camera 20 and zoom lens 22 is considerably greater than that of mirror 23 and the associated supporting mechanism. As will be seen later, it is important that successive images are taken from the same locations in order to provide accurate three dimensional models; accordingly, tilting the mirror is generally preferred to moving the camera itself.
A number of images of the calibration mat are taken by the digital camera 20 during a calibration process. The images are processed to detect the calibration dots 108 on the calibration mat 106 in the captured image. The detected calibration dots are analysed to determine a central position of the calibration mat 106 for creating supposed three-dimensional coordinates. In accordance with the supposed three-dimensional coordinates, a position, an orientation and a focal length of the digital camera 20 can be obtained from the image of the calibration dots 38 by using perspective information. Further details of the calibration process, and how the calibration data obtained is used in the generation of three-dimensional objects of models are given below.
Different mats may be provided in order to calibrate the photographic apparatus for different sizes of object. For example, the large mat of
The computer system 110 includes a central processing unit (CPU) 112 that is used to execute an application program. Normally, the application program is stored in a ROM or a hard disk within the computer system 110 as object code. That program is read from storage and written into memory within the CPU 112 at system launch for execution by the computer system 110. Detailed descriptions of data flow, control flow and memory construction are omitted from the present description since they do not relate directly to the present invention and suitable implementations are well known to persons skilled in the art.
A video monitor 114 is connected to the computer system 110. A video signal to be displayed by the video monitor 114 is output from a video board 116 to which the monitor 114 is connected. The video board 116 is driven by a video driver 118 consisting of a set of software programs. A keyboard 120 and mouse 122 are provided to enable an operator of the system to manually input data. Such input data are interpreted by a keyboard and mouse interface 124 to which the keyboard 120 and mouse 122 are connected. Of course, other data input and output devices could be used in addition to, or instead of, the video monitor 114, keyboard 120 and mouse 122 in order to enable the operator to communicate with the computer system 110.
The digital camera 20 and zoom lens 22 are connected to the computer system 110 by a Universal Serial Bus (USB) port and HUB interface 126. A USB device manager 128 manages USB port 126 (and any other USB ports under its control). The digital camera 20 and zoom lens 22 are controlled by a USB driver 130. Control functions, including image capturing, exposure control, and zoom positioning are controlled by the computer system 110.
An interface box 132, external to the computer system 110, controls communications between STM drivers 134, 136 and 138, photodetector monitor 140, lighting control unit 142 and the computer system 110.
STM driver 134 drives and controls a stepping motor 144 used to tilt the mechanical tilting stage 23a of the mirror 23 as described above. STM driver 136 drives and controls the stepping motor 90 used to drive the arm drive 80. STM driver 138 drives and controls the stepping motor 68 used to drive the drive wheel arrangement 50. STM drivers 134, 136 and 138 control steeping motors 144, 90 and 68 respectively in accordance with outputs from digital-to-analogue converters (DACs) 146, 148 and 150 respectively. DACs 146, 148 and 150 each convert digital data received from the computer system 110 into analogue signals for use by the STM drivers 134, 136 and 138 respectively.
Photodetector monitor 140 detects an output from photodetector device 76 indicating positions of one or more marks 152 composed of evaporated aluminium thin films or thin material located on a circumference of the turntable 14. The analogue output of the photodetector monitor 140 is converted into digital data by analogue-to-digital converter (ADC) 154 for use by the computer system 110.
The lighting control unit 142 has a register that controls front fluorescent light unit 24 and backlight unit 32. This register is a 2-bit register, the first bit (control signal #F-FL) controlling front fluorescent light unit 24, the second bit (#B-FL) controlling backlight unit 32. These control signals are created in accordance with the application program of computer system 110.
The computer system 110 and interface box 132 communicate via serial interface 156 under the control of communication serial port driver (COM port driver) 158. Digital data for use by STM drivers 134, 136 and 138 are sent from CPU 112 to those STM drivers via the serial interface 156 and the appropriate DACs 146, 148 and 150. Data from photodetector monitor 140 is passed to the CPU via ADC 154 and serial interface 156.
A hard disk unit 160 stores data 162 of texture images and silhouette images. A three-dimensional object model creating program is stored in a ROM or a hard disk within the computer system 110 as an object code and is represented in the block diagram by 3D Object Modelling Engine 164. The program is read out from storage and written into a memory within the CPU 112 when the system is launched. The code is executed from the CPU 112. The application program and the model creating program communicate through a communication (COM) interface. A program for displaying a graphical user interface (GUI) for the application is stored in the CPU 112 and is represented by the GUI block 166.
Flowcharts describing the operation of the system of
The operation of the system of
The initialisation subroutine then terminates and the program returns to the main program of
As mentioned above, in order to generate three dimensional object models, both image data of the object concerned and calibration data of a suitable calibration mat are required. As is known in the art, the calibration subroutine is used to determine a central position of the calibration mat for creating supposed three-dimensional co-ordinates. In accordance with the supposed three-dimensional coordinates, a position, an orientation and a focal length of the digital camera 20 can be obtained from the image of the calibration dots 38 by using perspective information. This information is essential to the three-dimensional modelling algorithm described in more detail below.
The camera calibration subroutine begins with the operator being asked at step #301 whether the camera is to be calibrated for the purposes of taking images of a large object (in which case the flowchart moves to step #302) or a small object (in which case the flowchart moves to step #305). It is possible for a system operator to choose the size option by using a scale specifically prepared for the system. For example, a scale having two indications of size: one representing the maximum size of large option and the other representing the maximum size of small option may be provided. The detailed explanation of the scale is omitted in this embodiment. Examples of large sized objects are training shoes or a toy doll with a height of the order of a few tens of centimetres. Examples of small sized objects include a wrist watches or a miniature car toy with a size of the order of a few centimetres, for example. The number of size options available to the operator and the maximum object size for each size option can be predetermined depending on the pixel resolution of the camera 20, target image resolution and quality of the 3D model to be created based on the image data by the photographing apparatus 12. Detailed explanations of the configuration to determine each size option is omitted in this embodiment here as they are not directly related to the objective of the present invention.
The step #301 is implemented by displaying a size-selecting window on the display of the video monitor 114. This is controlled by the GUI 166 via the CPU 112 and the video board 116. The operator utilises the keyboard 120 and/or the mouse 122 to select the size of the object by referring the displayed page. As noted above, since the GUI 166 and the selecting page itself are not important to describe this invention, detailed descriptions thereof are omitted.
If the user indicates that the object is large, then a software variable “obj_size” is set to be large in step #302 and the subroutine moves to step #303. The user is then instructed on the video monitor 114, via the GUI 166, to place the large calibration mat shown in
If the user indicates that the object is small, then a software variable “obj_size” is set to be small in step #305 and the process moves to step #306. The user is then instructed on the video monitor 114, via the GUI 166, to place the small calibration mat shown in
Regardless of the object size, the system waits at step #307a until the system indicates that the required zoom lens position has been set and that it is ready for calibration. When it is ready for calibration, the front fluorescent lights 24 is turned on (step #308). In detail, in accordance with the application program, the CPU 112 instructs the writing of a flag “1” in the front fluorescent light control bit of the register in the lighting control unit 142 of the interface box 132 though the serial interface 156 under control of the COM port Driver 158. Accordingly an output of the port of front lights #F-FL switches to a predetermined level representing “1” and, as shown in the table of
The next step, as indicated at step #309 is to initialise the turntable 14. This is achieved by calling the turntable initialisation subroutine of
The turntable initialisation routine begins at step #401, the STM driver 138 issues instructions to the stepping motor 68 such that the driving arrangement 50 drives the turntable 14 is driven in a clockwise direction. The instructions to the STM driver 138 are received from the CPU 112 via the serial interface 156 and DAC 150.
The turntable 14 is driven in a clockwise direction until, at step #402, an initial mark 152 is located by the photodetector device 76. The output of photodetector device 76 is monitored by the CPU 112 by means of the photodetector monitor 140, ADC 154 and the serial interface 156. When it is detected that the mark 152 is aligned with the photodetector 76, the STM driver 138 instructs the stepping motor 68 to stop rotating the turntable 14 (step #403).
At this point, the turntable is in a known position, with the mark 152 aligned with the photodetector 76 and the turntable initialisation subroutine is terminated.
The camera calibration subroutine continues at step #310 where the CPU 112 transmits the instruction “exp_param_set; #0” to the digital camera 20 to set the imaging parameters for the camera. From
The process moves to step #311 where a loop variable C is set at 0, 1, 2 or 3. Initially, C is set at zero.
In step #312, the variable “camera_arm_set” is made equal to the value of loop variable C. As stated above, this value is initially 0. The camera_arm_set variable determines the elevation angle of the camera arms 28 and 30 relative to the turntable 14. An initial value of C=0 sets the camera arm angle to 80 degrees (i.e. nearly vertical, as in
Thus the camera arm angle is initially set at 80 degrees (since C=0). This is achieved by CPU 112 issuing instructions to STM driver 136 to driver stepping motor 90 until the camera angle is set at 80 degrees by arm drive 80. The above correspondence between the variable C and the elevation angle of the camera arms are listed in the table of
With the camera arm angle set, the process moves to step #313, which calls a mirror tilt subroutine, as shown in
The mirror tilt subroutine sets a software variable “mirror_tilt set” depending on the value of the variable C and the variable “obj_size” (step #601). The variable mirror_tilt set is set to one of #0, #1, #2, #3 or #4. As shown in
If the variable C has a value 1 and the object size is small, then the variable mirror_tilt set is #1 (step #602). If the variable C has a value 1 and the object size is large, then the variable mirror_tilt set is #3 (step #603). If the variable C has a value 2 and the object size is small, then the variable mirror_tilt set is #2 (step #604). If the variable C has a value 2 and the object size is large, then the variable mirror_tilt set is #4 (step #605). Otherwise, the variable mirror_tilt set is #0 (step #606).
With the mirror tilt angle set, the mirror tilting subroutine is terminated.
The camera calibration subroutine returns to step #314, where a loop variable N is set as one of sixteen integers from “0” to “15”. Initially the variable N is set as “0”. At step #315, a software variable STOP is set to 16.
By this point, the front fluorescent light is on, the exposure parameters for the digital camera 20 have been set (initially to #0) and the camera arm and mirror tilting positions have been set (initially to 80 degrees and 0 degrees respectively) At this point, the digital camera 20 is instructed to take an image (step #316).
Image data obtained by the digital camera 20 is transferred to the hard disc unit 160 through the USB ports HUB interfaces 126 after compressing the image data in conformity with well-known JEPG compression scheme. Clearly, data compression is not essential but transmission time and data storage space requirements make such compression schemes attractive. Also, data compression schemes other than the JPEG scheme can be used. At step #317, the image data captured by the camera 20 is stored as file “img_cam_#C#N”, thus the first file, with C=0 and N=0 will be “img_cam—0—0.jpg”. Since the image is mirrored by the mirror 23, the original image data captured by digital camera 20 is flipped digitally in order to restore the proper orientation.
The image taken is of the calibration mat placed on the turntable by the user at either step #303 (large object) or step #306 (small object). At step #318, the CPU detects the calibration pattern of the calibration mat in the captured image data, namely, “img_cam_#C#N.jpg”.
The JEPG compressed image data obtained in the step #316 and stored in the step #317 is processed and developed and the CPU 112 detects the calibration dots 108 on a calibration mat 106 (or 106′ in the case of a small object) in the captured image in accordance with the application program and three-dimensional object model creating program 164 in a step #318. The CPU 112 processes and analyses the detected calibration dots and determines a central position of the calibration mat 106 for creating supposed three-dimensional coordinates. As described above, in accordance with the supposed three-dimensional coordinates, camera calibration parameters consisting of camera position, an orientation and a focal length of the digital camera are obtained from the image of the calibration dots 38 by using perspective information. Detailed methods or processes for obtaining the central position of the calibration mat 106, the supposed three-dimensional coordinates, the camera calibration parameters such as the position and the orientation of the digital camera, and the focal length were disclosed in several former patent applications, for example, a Japanese Laid-Open Patents numbered 00-96374, a Japanese Laid-Open Patents numbered 98-170914 and a UK patent application numbered 0012812.4 These methods or processes, and others which are known to the persons skilled in the art, can be adopted for the step #318. Therefore, in this specification, detailed descriptions of such concrete methods and processes have been omitted.
Once the calibration pattern detection process is complete (step #319), the calibration data can be stored (step #320) as filed “cal_large_#C#N” for a large object or “cal_small_#C#N” for small object. Thus the initial calibration data for a small object would be stored as file “cal_small—0—0”.
With the first calibration data point stored, the turntable is incremented at step #321 by calling a turntable rotation subroutine, as shown in
The first step of the turntable rotation subroutine (step #501) is to restore a value P. The value P represents the number of pulses of the stepping motor 68 required to turn the turntable 14 through one complete revolution. This value may be determined when the photographic apparatus 12 is first assembled, or may be set as a design requirement of the photographic apparatus.
The variable P is divided by the variable STOP (set at 16 at step #315) to obtain a value DR step (step #502). The value DR step is a measure of the number of pulses of the stepping motor 68 required to turn the turntable 14 through 1/16of a complete revolution.
At step #503, the CPU 122 instructed STM driver 138 to drive the stepping motor 68 for DR steps, thereby moving rotating the turntable 14 through 1/16 of a complete revolution.
Once the turntable has been rotated, the value N is incremented (step #504) and the CPU 122 determines whether or not the value N has reached 15 (step #505). If the value N has not reached 15, the turntable rotation subroutine terminates and the program returns to the main algorithm at step #322. A further check regarding whether N=15 is made at step #322, before the steps #316 to #321 are repeated with the new value N.
Thus with the camera in the first position (80 degrees above the horizontal), calibration data is taken for the glass turntable at 16 different positions, the calibration data stored (for a large object) being cal_large—0—0, cal_large—0—1 . . . cal_large—0—15.
Once the data cal_large—0—15 has been stored, the value N is incremented at step #504 and becomes 15. The turntable rotation subroutine no longer terminates at step #505 and proceeds to #506. The turntable is driven clockwise at step #506 and continues until the photodetector indicates that the mark 152 is once again aligned with the photodetector. Thus, in a similar manner as in the turntable calibration subroutine, the turntable rotation subroutine terminates at step #508 with the mark 152 aligned with the photodetector 76 so that the turntable 14 is once again in its original position.
The camera calibration routine returns to step #322, which is answered positively and the program proceeds to step #323 where the value C is incremented.
At this stage, 16 calibration readings have been taken with the camera arm at position #0 (80 degrees). With C incremented to #1, the camera arm is moved to position #1 (45 degrees). The steps #312 to #322 are repeated so that 15 calibration readings are taken with the camera arm at position #1. At step #323, C is incremented to #2 and the process is repeated with the camera arm at position #2 (10 degrees) and is repeated once more with the camera arm at position #3 (−70 degrees).
When the program returns to step #323 with C set at #3, the following calibration data (assuming a large object) has been stored:
cal_large—0—0, cal_large—0—1 . . . cal_large—0—15
cal_large—1—0, cal_large—1—1 . . . cal_large—1—15
cal_large—2—0, cal_large—2—1 . . . cal_large—2—15
cal_large—3—0, cal_large—3—1 . . . cal_large—3—15
The front fluorescent light is switched off at step #324 and at step #325 the operator is asked whether the camera calibration should be performed for the other size of object (i.e. if the calibration has been performed for a large object, does it also need to be done for a small object?). If so, the camera calibration algorithm is repeated for the other size of device. If not, the camera calibration subroutine terminates.
As discussed above, the zoom position of zoom lens 22 is set depending on the size of the object to be modelled. Each zoom position of the zoom lens 22 must therefore be calibrated. The objective of providing different zoom options for different objects is to provide the appropriate pixel resolution for each object, regardless of its size. However, if the operator is allowed to set the zoom position to any possible position, each of these must be calibrated. This takes time and requires a large amount of data to be stored.
By allowing only a small number of object sizes (in this case two) and setting the zoom position according to the size of object indicated by the operator, the number of calibration steps is reduced and, as a result, both the time taken to perform the calibration steps and the data storage requirements are reduced. Furthermore, the operator is only required to indicate the size of the object being modelled. There is no need for the operator to directly set the zoom position of the zoom lens 22. This contributes to the overall aim of reducing the skill required of the operator.
At this stage, all calibration data has been obtained. To this point, all the operator of the photographic apparatus 12 has needed to do is to indicate via GUI 166 that calibration is required (step #102), indicate the object size (step #301), place the appropriate calibration mat on the turntable (step #303 and/or step #306) and remove the calibration mat at the end of the calibration process. The remaining steps of the calibration procedure are entirely automatic. Specifically, the operator does not need to position the camera arm, the backlight unit or the turntable to their required positions for each calibration image.
Once the camera calibration subroutine (step #103) has been completed, modelling can begin. The operator is asked at step #104 whether modelling should start. If yes, the process proceeds to step #105, if not the step #104 is repeated until the operator is ready to begin modelling. The step #104 is implemented by displaying a window on the display of the video monitor 114 asking the operator to indicate whether or not modelling should begin. This is controlled by the GUI 166 via the CPU 112 and the video board 116. The operator utilises the keyboard 120 and/or the mouse 122 to indicate whether or not modelling should begin. As noted above, since the GUI 166 and the selecting page itself are not important to describe this invention, detailed descriptions thereof are omitted.
The modelling begins with the operator being asked at step #105 whether the object to be modelled is large (in which case the routine moves to step #106) or small (in which case the routine moves to step #108). As explained in step #301, the target object size is assessed by the operator by referring to a scale prepared specifically for the photographing apparatus. The step #105 is implemented by displaying a size-selecting window on the display of the video monitor 114. This is controlled by the GUI 166 via the CPU 112 and the video board 116. The operator utilises the keyboard 120 and/or the mouse 122 to select the size of the object by referring the displayed page. Again, detailed descriptions of these elements are omitted.
If the user indicates that the object is large, then a software variable “obj_size” is set to be large in step #106. A software parameter zoom_pos_set is then set as #0 at step #107. From the table of
If the user indicates that the object is small, then a software variable “obj_size” is set to be small in step #108 The software variable zoom_pos_set is set as #5 indicating, as shown in
Once the zoom lens has been set to the appropriate setting, the operator is instructed on the video monitor 114, via the GUI 166, to put the object to be modelled on the turntable, as close to the centre of the turntable as possible (step #110). Once the user has indicated that this is complete, for example by using the keyboard 120 or mouse 122, the process proceeds to step #111, which calls the turntable initialisation subroutine. The process for initialising the turntable is described above with reference to
With the turntable initialised, the process proceeds to step #112 where a loop variable C is set at 0, 1, 2 or 3. Initially, C is set at zero.
As described above with reference to the camera calibration process, the variable “camera_arm_set” is made equal to the value of the loop variable C. An initial value of C=0 sets the camera arm angle to 80 degrees. Values of C of 1, 2 and 3 respectively set the camera arm angle at 45 degrees, 10 degrees and −70 degrees.
With the camera angle set according to the value of the loop variable C (step #113), the process moves on to step #114, which calls the mirror tilting subroutine. As described above, the mirror tilt subroutine sets a software variable “mirror_tilt set” depending on the value of the variable C and the variable obj_size set at steps #105, #106 and #108. The mirror tilt subroutine terminates with the mirror tilt angle set at one of 0, −3, −5, +8 or +10 degrees, as described above with reference to
The process proceeds to step #115 at which point the front fluorescent light unit 24 is turned on. In accordance with the application program, the CPU 112 instructs the writing of a flag “1” in the front fluorescent light control bit of the register in the lighting control unit 142 of the interface box 132 though the serial interface 156 under control of the COM port Driver 158. Accordingly an output of the port of front lights #F-FL switches to a predetermined level representing “1” and, as shown in the table of
At step #116, the CPU 112 transmits the instruction “exp_param_set; #0” to set the imaging parameters for the digital camera 20 and zoom lens 22. From
At step #117, a loop variable N is set as one of sixteen integers from “0” to “15”. Initially the variable N is set as “0”. At step #118, a software variable STOP is set to 16.
By this point, the front fluorescent light is on, the exposure parameters of the digital camera 20 and zoom lens 22 have been set (initially to #0) and the camera arm and mirror tilting positions have been set (initially to 80 degrees and 0 degrees respectively). At this point, the digital camera 20 is instructed to take an image (step #119).
Image data obtained by the digital camera 20 is transferred to the hard disc unit 160 through the USB ports HUB interfaces 126 after compressing the image data, advantageously in conformity with well-known JEPG compression scheme as discussed above. At step #120, the image data captured by the camera 20 is stored as file “img_cam_#C#N”, thus the first file, with C=0 and N=0 will be “img_cam—0—0.jpg”. As noted above, since the image is mirrored by mirror 23, the original image data captured by digital camera 20 may be flipped digitally in order to restore the proper orientation.
With the first image stored, the turntable is incremented at step #121 by calling a turntable rotation subroutine, as shown in
Thus, with the camera in the first position (80 degrees above the horizontal), image data is taken for the glass turntable at 16 different positions, the image data stored being img_cam—0—0.jpg, img_cam—0—1.jpg . . . img_cam—0—15.jpg. Of course, as a result of the calibration process, the positions at which each of these images is taken is known to the 3D object modelling engine 164.
The front fluorescent light is then turned off at step #123 and the back fluorescent light turned on at step #124. The loop variable N is reset to 0 (step #125) and the variable STOP set to 16 (step #126).
At step #127, the CPU 112 transmits the instruction “exp_param_set; #1” to set the imaging parameters for the digital camera 20 and zoom lens 22. From
By this point, the backlight unit 32 is on, the exposure parameters have been set (initially to #1) and the camera arm and mirror tilting positions have been set (initially to 80 degrees and 0 degrees respectively). At this point, the digital camera 20 is instructed to take an image (step #128). Since the backlight unit 32 is on, the image taken will be a silhouette of the object on the turntable 14.
The imaging parameters may be varied for a number of reasons. For example, the imaging parameters when silhouette data is being captured may be such that the images are underexposed since this increases the contrast between the object and the background, thereby making it easier to determine the edges of the silhouette.
Silhouette data obtained by the digital camera 20 is transferred to the hard disc unit 160 through the USB ports HUB interfaces 126 after compressing the image data in conformity with well-known JEPG compression scheme. At step #129, the silhouette data captured by the camera 20 is stored as file “sil_cam_#C#N”, thus the first file, with C=0 and N=0 will be “sil_cam—0—0.jpg”. As noted above, since the image is mirrored by mirror 23, the original image data captured by digital camera 20 may be flipped digitally in order to restore the proper orientation.
With the first silhouette data point stored, the turntable is incremented at step #130 by calling a turntable rotation subroutine, as shown in
Thus, with the camera in the first position (80 degrees above the horizontal), silhouette data is taken for the glass turntable at 16 different positions, the silhouette data stored being sil_cam—0—0.jpg, sil_cam—0—1.jpg . . . sil_cam—0—15.jpg.
The backlight unit 32 is then turned off at step #132.
At this stage, both image and silhouette data has been captured with the camera in the first position (80 degrees above the horizontal). At step #133, C is incremented to #1 and the steps #113 to 132 are repeated, first with C=1, then C=2, and finally with C=3.
When the program returns to step #133 with C set at #3, the following data has been stored:
img_cam—0—0.jpg, img_cam—0—1.jpg . . . img_cam—0—15.jpg
sil_cam—0—0.jpg, sil_cam—0—1.jpg . . . sil_cam—0—15.jpg
img_cam—1—0.jpg, img_cam—1—1.jpg . . . img_cam—1—15.jpg
sil_cam—1—0.jpg, sil_cam—1—1.jpg . . . sil_cam—1—15.jpg
img_cam—2—0.jpg, img_cam—2—1.jpg . . . img_cam—2—15.jpg
sil_cam—2—0.jpg, sil_cam—2—1.jpg . . . sil_cam—2—15.jpg
img_cam—3—0.jpg, img_cam—3—1.jpg . . . img_cam—3—15.jpg
sil_cam—3—0.jpg, sil_cam—3—1.jpg . . . sil_cam—3—15.jpg
The above description assumes that sixteen images are taken at every orientation of the camera. Of course, more or fewer images can be taken at any orientation. Moreover, a different number of images may be taken at different orientations. This may be of use since it is considered that in order to obtain the optimum three dimensional image with the smallest number of images, more images are required at lower angles than at higher angles relative to the turntable. Accordingly, in one embodiment of the invention, sixteen images are taken with the camera at 10 degrees to the horizontal, 8 images are taken with the camera at 45 degrees to the horizontal, and 4 images are taken with the camera at 80 degrees and at −70 degrees to the horizontal. Also it is not always necessary to take texture images and silhouette images at the same turntable and camera elevation positions. It is considered that having more silhouette image data creates a more accurate geometry shape. However, it is not always necessary to take as many texture images in order to provide the required quality of texture image.
The routine proceeds to step #134 (
In the step #134, before creating three-dimensional geometry, camera calibration parameters including photographing positions, orientations, and focal lengths for each of digital cameras and each photographing are calibrated by using the obtained camera information, including the position, the orientation and the focal length of the digital cameras stored in the hard disc unit 160 as files named “cal_large_#C_#N” or “cal_small_#C_#N” in the step #320.
In a step #135, in accordance with the three-dimensional object model creating program 164, the CPU 112 creates texture data to be put on surfaces of each polygon created in the step #134 by using texture images stored in the hard disc unit 160. The additional of textural data to the polygon surfaces completes the three dimensional model. Detailed methods or processes for obtaining the three-dimensional texture data for each polygon by using two-dimensional texture images taken from different positions and orientations are known in the art and are not described in detail here.
In a step #136, the resultant three-dimensional object having geometry on which texture images have been put is displayed on the display of the video monitor 114. As known, such resultant three-dimensional model can be rotated, magnified or like by the user, using the keyboard 120 or the mouse 122.
In a step #137, all information of such a resultant three-dimensional object including three-dimensional geometry information and texture information to be put on the geometry is stored in the hard disc unit 160 as a VRML file, for example. Clearly, any other suitable file format could be used. The process is completed after the step #137.
As noted above, the operator was only required to perform a few simple tasks in order to obtain the required calibration data. Similarly, the operator is only required to perform a few simple operations in order to obtain the image data. The operator must indicate, via GUI 166, the object size (step #105), place the object on the turntable (step #110) and remove the object at the end of the process. Furthermore, the operator is not required to have any photographic experience.
Accordingly, a user with no previous experience of using the photographic apparatus 2 can be trained to make three dimensional models of objects relatively quickly.
It is noted here that although the titling angle of the mirror 23 depends on camera elevation angle and/or on the size of target object in the above embodiment, it is also possible for the object height to be assessed by the operator, and for the height information to be inputted by the operator, with the titling angle being determined on the basis of the information provided by the operator.
It is also noted here, that the rubber roller used as the turntable drive wheel arrangement may be replaced, for example, with material having certain friction toward rim of glass turntable, such as harden plastic or a metal roller with a grinded finish.
Also, optical devices such as an optical prism can be used to deflect the optical axis instead of a mirror.
It is also noted here, that the embodiment described above includes a pair of camera arms and a pair of backlight arms, however, it is also possible to support the backlight unit and camera unit with a single camera arm and a single backlight arm provided that each arm is rigid enough mechanically.
Number | Date | Country | Kind |
---|---|---|---|
0319677.1 | Aug 2003 | GB | national |
Filing Document | Filing Date | Country | Kind | 371c Date |
---|---|---|---|---|
PCT/GB04/03581 | 8/20/2004 | WO | 7/18/2006 |