IMAGE STABILIZATION SYSTEM FOR HANDHELD DEVICES EQUIPPED WITH PICO-PROJECTOR

Information

  • Patent Application
  • 20140071252
  • Publication Number
    20140071252
  • Date Filed
    September 06, 2013
    11 years ago
  • Date Published
    March 13, 2014
    10 years ago
Abstract
It is proposed a method for the projection of a sequence of images onto a projection surface by a handheld device having an embedded projector, the method including, for an input image to be projected of the sequence of images: a step of evaluation wherein a spatial position variation is evaluated with respect to a reference spatial position of the handheld device, and the spatial position variation is defined by a spatial orientation variation and a spatial vector variation, the spatial vector variation being determined by 3 independent coordinates, and at least one of the 3 independent coordinates is provided by a stereoscopic sensor;a step of compensation wherein a compensated image is generated from the input image to be projected depending on the spatial position variation; anda step of projecting onto the projection surface the compensated image.
Description
CROSS-REFERENCE TO RELATED APPLICATIONS

This application claims the benefit and priority of European Patent Application Serial No. 12306080.8, filed on Sep. 7, 2012, which is incorporated by reference herein.


FIELD OF THE INVENTION

The invention relates to the field of image stabilization in pico-projectors and more specifically in handheld devices equipped with projectors and additional sensors.


BACKGROUND

Handheld devices like smartphones are more and more used for the visualization of images or movies. Nevertheless the small size of their screens does not help to watch videos in good conditions. Projection features are being integrated into handheld devices to solve their issue of small screen size. That is why handheld devices like smartphones are more and more equipped with projectors. That way, any wall or surface can be used as a projection surface. This enables a large visual presentation from a very small projector.


Projectors embedded in handheld devices like smartphones are called pico-projectors. The issue of image stabilization does not arise from office projectors as those devices can rest on a stable platform or they can be fixed externally and do not move during a projection session. In the case of pico-projectors, they are sustained by instable object like a hand, so they will suffer vibrations and slight orientation and position changes during the projection, called “hand shake”.


The lack of stability in the projection of a sequence of images can produce several defects. Keystone is a bad orientation of the projector relative to the projection surface; inner rotation is a bad orientation of the projector relative to its own central beam; gap is a bad displacement of the projector relative to the projection area on the projection surface, zoom is a changing distance of the projector with the projection surface. Instability corresponds to a projection session where one or several projected images of the sequence will suffer from one or many of those defects. While the issue of image stabilization due to vibrations in image projection can be solved in office projectors by fixing physically the body of the projector, for Pico-projectors, image stabilization due to hand shake requires image processing techniques.


In the prior art, many correction techniques relate to keystone, but not stabilization. Some of the cited defects are not mentioned. Correcting only the image deformation for example is not sufficient to correct stabilization. Image stabilization requires more data than managing keystone or bad spatial orientation alone. For keystone, knowing the elevation and inclinations angles with the projection surface is necessary.


In some prior art where the image stabilization issue is addressed, there is needed a complete set of motion sensors to compensate for handshake. In a first prior art, for example the patent application US20120113514, a combination of gyroscopes and accelerometers on all three axes is used to compute a corrective signal input to feed a video controller to reduce the apparent motion of the image to be projected. In a second prior art, for example the patent U.S. Pat. No. 6,753,907, an un-calibrated camera is used to observe the projected image; and the image to be displayed is pre-warped so that the distortions induced by the misaligned projection system will be compensated. Generally speaking, techniques with heavy image processing, as they can't be applied in real time during a movie projection, are not image stabilization techniques.


How to manage image stabilization during projection in a real time process and with a low cost approach is still to question in handheld devices. Image stabilization systems are classified into two main categories. The optical or mechanical image stabilizer employs a prism assembly that moves opposite to the shaking of camera for stabilization. The optical image stabilizers are hardware dependent and require built-in devices such as servo motors making it voluminous and costly. The digital or electronic image stabilizer compensates for the image sequence by employing motion sensors to detect the device movement for compensation and performs image compensation through image processing algorithms.


SUMMARY

An object of embodiments of the present invention is to alleviate at least partly the above mentioned drawbacks. More particularly, embodiments of the invention aim at improving image stabilization in pico-projectors during a projection session. It is taken advantage of the several devices and sensors handheld devices can be equipped with for image stabilization. An object of the invention is, in a projection session, to detect a movement and to compensate for it easily and at low cost.


The object of the present invention is achieved by a method for the projection of a sequence of images onto a display surface by a handheld device having an embedded projector, the method comprising, for an image of the sequence of images to be projected: a step of evaluation wherein a spatial position variation is evaluated with respect to a reference spatial position of the handheld device; and the spatial position variation is defined by a spatial orientation variation and a spatial vector variation; the spatial vector variation being determined by 3 independent coordinates; and at least one of the 3 independent coordinates is provided by a stereoscopic sensor embedded in the handheld device a step of compensation wherein a compensated image is generated from the image to be projected depending on the spatial position variation; and a step of projecting onto the display surface the compensated image.


Embodiments can comprise one or more of the following features:

    • The reference spatial position is constant for all images of the sequence of images to be projected.
    • The reference spatial position is variable between two images of the sequence of images to be projected
    • The spatial orientation variation is determined by 3 independent angles.
    • The 3 independent angles are provided by a 3 axes gyroscope.
    • 2 of the 3 independent coordinates are provided by a 2-axes accelerometer.
    • The step of compensation utilizes a graphics processing unit technique to generate the compensated image. This is also achieved with a computer program product comprising a computer readable medium, having thereon a computer program comprising program instructions, the computer program being loadable into a data-processing unit and adapted to cause execution of the method when the computer program is run by the data-processing unit.


The object of the present invention is also achieved with a data storage medium having recorded thereon a computer program comprising instructions for performing the method. The object of the present invention is also achieved by a handheld device having an embedded projector capable of projecting a sequence of images onto a display surface; said handheld device comprising an evaluation unit adapted to evaluate a spatial position variation with respect to a reference spatial position of said handheld device; said spatial position variation being defined by a spatial orientation variation and a spatial vector variation; said spatial vector variation being determined by 3 independent coordinates; and said handheld device comprising a compensation unit adapted to generate a compensated image from an image of the sequence of images to be projected, depending on said spatial position variation; and said handheld device comprising a stereoscopic sensor configured to deliver at least one of the 3 independent coordinates determining said spatial vector variation; and said embedded projector being adapted to project onto said display surface said compensated image.


Embodiments can comprise one or more of the following features:

    • The stereoscopic sensor uses a two-lens stereo cameras
    • The stereoscopic sensor uses two single-lens cameras joined together
    • The spatial orientation variation is determined by 3 independent angles.
    • The handheld device is further comprising a 3 axes gyroscope adapted to deliver the spatial orientation variation
    • The handheld device is configured to utilize a graphics processing unit technique to generate the compensated image.


Further features and advantages of embodiments of the invention will appear from the following description of some embodiments of the invention, given as non-limiting examples, with reference to the accompanying drawings listed hereunder.





BRIEF DESCRIPTION OF THE DRAWINGS


FIG. 1 shows a block diagram of a configuration for projecting an image or frame without image stabilization;



FIG. 2
a to FIG. 2f are a range of figure illustrating defects in a projection process.



FIG. 2
a is an image in a frame correctly projected;



FIG. 2
b is an image in a frame badly projected because of elevation;



FIG. 2
c is an image in a frame badly projected because of inclination;



FIG. 2
d is an image in a frame badly projected because of inner rotation;



FIG. 2
e is an image in a frame badly centered because of gap;



FIG. 2
f is an image of a frame badly displayed because of a zoom issue;



FIG. 3 shows a graphics representative of a view of the projector and a projection surface in a Cartesian coordinate system representative of the reference mark in accordance with an embodiment of the invention;



FIG. 4 shows a block diagram of a configuration for projecting an image frame with image stabilization in accordance with an embodiment of the invention;



FIG. 5 shows a block diagram illustrating an exemplary configuration of an evaluation unit and a compensation unit in accordance with an embodiment of the invention unit; and



FIG. 6 is a functional block of a handheld device according to an embodiment of the present invention.





DETAILED DESCRIPTION

When there is no image stabilization, the quality of the video can decrease. FIG. 1 shows a block diagram of a configuration for projecting an image or frame without image stabilization. FIG. 1 comprises an image 1, a projector 2, and a projection surface 3. Of the projector, the projector image input 21 and the projector image output 22 are represented. If there is no transformation of the image 1 within the projector 2, the image 1 navigates unchanged through the projector 2 until the projector image output 22. But when the position of the projector 2 oscillates, one or several of the resulting images at the projection surface 3 can be distorted or badly positioned. The platform where the camera is fixed is not static. The motion of the platform will cause a deviation between two serial frames. The instability of images frames will make the image displayed on the screen misty. Actually in a projection of a sequence of images through a sequence of frames, some of the frames can be having an unwanted displacement with respect to the complete image sequence before stabilization process is applied. Because of sudden scene differences occurred on certain frames, some corruptions occur on the complete movements of the objects in the video. As result, the quality of the video decreases.



FIG. 2
a is an image of a frame correctly projected. In FIG. 2a, the handheld device is correctly oriented according to the projection area 31 of the projection surface and the projected image 1C which is of good quality.



FIG. 2
b is an image of a frame badly projected because of an arisen elevation of the handheld projector. In FIG. 2b, the handheld device position is not correctly oriented according to the projection surface and the projected image 1C has a bad alignment relative to the projection area 31.



FIG. 2
c is an image of a frame badly projected because of an arisen inclination of the handheld projector. In FIG. 2c, the handheld device position is not correctly oriented according to the projection surface and the projected image 1C has a bad alignment relative to the projection area 31.



FIG. 2
d is an image of a frame badly projected because of an arisen inner rotation. In FIG. 2d, the handheld device position is not correctly oriented according to the projection surface and the projected image 1C has a bad orientation relative to the projection area 31. It comes from a bad axial orientation angle of the handheld device according to the projection surface normal axis and an existing inner rotation angle.



FIG. 2
e is an image of a frame badly centered because of an arisen translation gap. In FIG. 2e, the handheld device has translated according to the projection area 31 in the projection surface and the projected image 1C is not centered.



FIG. 2
f is an image of a frame badly displayed because of a zoom issue. In FIG. 2f, the handheld device has moved away from the projection area 31 on the projection surface 3 and the projected image 1C needs resizing according to a reference image 32. More generally, any bad projection will result from one of the precedent defects, or any combination of two or three defects of sort.


With an image stabilization system, the resulting image at the projection surface is targeting the projection area and is not misty owing to vibration. The goal is to place the projector display in a fixed target area on the projection surface, despite jitter and vibrations during the projection. At each time step the position of the projector is collected and the image to be projected is compensated so as to fulfill the requirement of good projection quality. Several kinds of defects in the projection of images onto the projection surface will need to be compensated. The first defect is the orientation of the pico-projector in respect of the projection surface and the second and third defects are the translation and the remoteness with the target point in the projection surface. The orientation includes elevation, inclination and inner rotation. The translation includes horizontal or vertical gap on the one hand, and depth with the target area on the other hand. Keystone correction needs an estimate of the orientation of the projector relative to the surface. Image stabilization consists in fixing the projection on the surface even under hand jitter. As image stabilization needs both orientation and position of the projector relative to the surface, keystone correction is the easiest case.


Digital image stabilizer can be chosen advantageously. It does not need any mechanical or optical devices and is suitable for handheld devices. Digital stabilization systems use completely electronic processing to control the image stability. It uses more software algorithms and less hardware components to compensate for the disturbances. This makes digital stabilization more portable and cost effective compared with other methods. In digital stabilization, spatial positions variations are obtained by taking consecutive two frames of the sequence and performing a series of operations over the frames. Because of exhaustive image processing operations, spatial position variation evaluation is the most time consuming and difficult part in digital stabilization.


Sensors data and vectors calculations can be used as following. FIG. 3 shows a graphic representative of a view of the handheld device 4 and a projection surface 3 in a Cartesian coordinate system representative of the reference mark in accordance with an embodiment of the invention. The projection surface 3 contains a projected image 1C resulting from the projection by the handheld device 4. According to FIG. 3, the elevation of the handheld device 4 corresponds to a rotation around the X axis or pitch; the inclination of the handheld device 4 corresponds to a rotation around the Y axis or roll; and the inner rotation of the handheld device 4 corresponds to a rotation around the Z axis or yaw. A bad orientation of the handheld device 4 according to the projection surface 3 will correspond to any combination of rotations of the handheld device 4 around those axes. A bad position of the handheld device according to the projection surface 3 will correspond to any combination of translations of the handheld device around towards axes.


With the stereoscopic system, knowing what happened along the Z axis direction is possible. Thus a 2-axis accelerometer is enough for the remaining translation components. In practice, information from 3-axes gyroscopes, 2-axes accelerometers and stereoscopic sensors can be mixed together to enable the pico projector to know the spatial vector variation. A new position of the handheld device 4 corresponds to a translation and a rotation. The translation can be split into firstly the gap to the target of the central beam CB during the projection, and secondly the distance to the projection surface 3. Pico projectors are not supposed to be tied to target on the projection surface 3; it is one of their advantages and an easy to use feature. The distance to the projection surface 3 is compensable with an automatic zooming feature, capable of magnification when the distance is increasing and capable of reduction when the distance is lowering.


The projector can be enriched with an evaluation unit and a compensation unit as following. After spatial position evaluation, motion compensation part is responsible for correcting unintentional motions. It is about the alignment of the frames with respect to the estimated jitter through an inverse transformation process. In this part, same amount of movements are given to the frames in the inverse direction with the jitter in order to obtain stabilized video sequence.



FIG. 4 shows a block diagram of a configuration for projecting an image with image stabilization in accordance with an embodiment of the invention. FIG. 4 comprises an input image 1A, a projector 2, and a projection surface 3. Furthermore, an evaluation unit 23 and a compensation unit 24 are integrated. With the evaluation unit 23 and the compensation unit 24, the output transmitted at the projector 2 is a compensated image 1B according to the input image 1A to enable the resulting image at the projection surface 3 to target the projection area and not to be misty owing to vibration.



FIG. 5 shows a block diagram illustrating an exemplary configuration of an evaluation unit and a compensation unit in accordance with an embodiment of the invention. FIG. 5 comprises the projector image input 21, the compensation unit 24, the projector image output 22, and between the projector image input 21 and the compensation unit 24, the evaluation unit 23. In FIG. 5, the compensation unit 24 is fed in parallel by the input image 1A which is transmitted by the projector image input 21 and by the evaluation unit block 23 output. The compensation unit 24 at the end outputs the compensated image 1B to the projector image output 22. The objective is to keep some kind of history of the spatial positions variation in order to create a stabilized sequence without removing the motion of the camera.


An exemplary embodiment of the method according to the present invention for stabilizing image recordings adapted for projected data will consist firstly by calculating what action is required to keep the image stable from information by motion and proximity sensors and secondly by detecting a movement and compensating for it to get stabilization. The correction can be applied in real time, during the projection of a movie. The whole stabilization is split with a distinction between stabilization in translation and stabilization in rotation along 2 or 3 axes. Hand shaking can be slight flicker or slight vibrations; the amplitude of this movement is often limited to an offset.


The evaluation unit can work as following. Using FIG. 5 notations, the evaluation unit block 23 comprises the reference spatial position 231, the sensing sub unit 232 and the spatial position variation 233. In the evaluation unit 23, the spatial position variation 233 is determined upon the reference spatial position 231 and the current spatial position sensed by the sensors. A handheld device according to the present invention comprises a 3-axes gyroscope sensor, a 2-axes accelerometers, and stereoscopic sensors. The 3-axes gyroscope sensor providing the rotation information is used to measure angles and angles variations. As an orientation sensor, a gyroscope will measure angles evolution around the three axes and according to a reference mark. Practically, the evaluation unit 10 helped with the 3-axes gyroscope sensor senses and checks current position preferably at each image to be projected and saves new values of angles in case of change in a readable medium on the electronics that drives the pico-projector. Other implementations are possible. The 2-axes accelerometer, in measuring non-gravitational accelerations, will measure acceleration and potentially direction. It provides the translation information. Generally, accelerometers can be dedicated either to 2 axes (X and Y) or to 3 axes (X, Y, and Z). It is cost dependent.


To obtain linear movements of the camera, acceleration data are converted into displacement data. Some smartphones are equipped with 3-way axis device which is used to determine the handheld device physical position. The accelerometer can tell when the handheld device is tilted, rotated, or moved. In practice the spatial vector variation is determined by 3 independent coordinates in a coordinate system whereas the spatial orientation variation is determined by 3 independent angles. For spatial bad orientation we need to get an additional rotation angle of the pico-projector around its own central beam axis. Image stabilization will require another triplet corresponding to the 3 independent coordinates of a spatial point in a Cartesian coordinate system representative of the reference mark.


The movement to compensate for can be described in a 6-axis information vector (x, y, z, Rx, Ry, Rz). For each movement, a specific correction will be attributed. For a translation movement towards the X axis, the correction is a Shift on X axis. For a translation movement towards the Y axis, the correction is a shift on Y axis. For a translation movement towards the Z axis, the correction can be a Zoom In or a Zoom Out. For a rotation movement around X axis or Pitch, the keystone correction is to be made by a rotation along vertical axis. For a rotation movement around Y axis or Roll, the keystone correction is to be made by a rotation along horizontal axis. For an inner rotation movement, there is a rotation correction needed along Z axis or Yaw. For a combination of movements, one correction is a combination of associated corrections. For an image of the sequence of images to be projected, a spatial position variation is evaluated with respect to a reference spatial position.


The spatial position variation consists in a spatial orientation variation and a spatial vector variation. Accordingly, the 3 independent angles are preferably provided by a 3 axes gyroscope embedded in the handheld device. The Z-coordinate is provided by a stereoscopic sensor. The X and Y coordinates are provided by a 2-axes accelerometer. Conventional stereo vision is usually achieved with two cameras that are mounted in a known relationship to each other and are synchronized to take images at the same instant. Contrary to a one camera system, stereoscopic system brings the depth information. To measure depth a stereoscopic camera consists of two cameras that capture two different, horizontally shifted perspective viewpoints. This results in a disparity of objects in the recorded scene between the two cameras views depending on their depth. The depth and disparity are related in relationships using the focal length of the cameras and the inter-axial separation between the two lenses of the stereoscopic camera.


There are commonly two possible ways of taking stereoscopic pictures: by using special two-lens stereo cameras which is an optical system with two lens but only one camera or by using systems with two single-lens cameras (which are two separate cameras joined together). Stereoscopic pictures allow us to calculate the distance from the cameras to the chosen object within the picture. The distance is calculated from differences between the pictures and additional technical data like focal length and distance between the cameras.


Practically, the evaluation can be performed through an evaluation unit 10 adapted to evaluate a spatial position variation with respect to a reference spatial position of the handheld device; the spatial position variation being defined by a spatial orientation variation and a spatial vector variation.


The compensation can be performed in several possible implementations. For example one possible implementation is the one that always uses two consecutive frames from the input image sequence to estimate the spatial position variation, which is referred to as frame-to-frame algorithm. In that case, the reference spatial position will be variable between two images of the sequence of images to be projected. Another possible implementation is the one that keeps a reference image and uses it to estimate the variation between the reference and the current input image, which is referred to as the frame-to-reference algorithm. In that case, the reference spatial position will be constant for all images of the sequence of images to be projected. Any combination of those two implementations can of course be used, for example the refreshing of the reference spatial position every n images, with n a constant number. Anyway, in our case, a compensated image is generated from each image to be projected depending on the spatial position variation. At the end, what is projected onto the projection surface is the compensated image instead of the original image.


Any of inverse transformations methods well known can be used to generate a compensated image resulting from the input image to be projected and from sensors data about the spatial position variation. For example graphics processing unit implementations or digital image processing techniques. Generally speaking, for GPU implementations, the spatial transformation can develop a triangle mesh with one triangle defined by 3 vertices or points. The transformation is executed by texture mapping from the rectilinear mesh of the input image to the transformed shape of the destination image. In digital image processing implementation, the spatial transformation consists of spatially defined 2-dimensional image re-sampling or scaling filter. The scaling operation is performed with different scaling ratios in different parts of the image, according to the defined transformation. Practically, the compensation can be performed through a compensation unit 11 adapted to generate a compensated image from an image of the sequence of images to be projected, according to the spatial position variation.


Using FIG. 5 notations, in fact an input image 1A received from the projector image input 21 is transmitted to the compensation unit 24. In parallel the spatial position variation 233 is determined and transmitted to the compensation unit 24. The compensation unit 24 performs the inverse transformation to compensate for image instabilities due to the spatial position variation 233 and gets a compensated image 1B. The inverse transformation is based on the spatial position variation 233 and uses the spatial orientation angles and the spatial position vector coordinates in matrix products and additions. The expressions of the relationships for the inverse transformation include the case where spatial position variation 233 is null, which means that the handheld device 4 has not changed its position according to the reference spatial position. At the end of the compensation process, the compensation unit 24 transmits the compensated image 1B to the projector image output 22. The projector image output 22 projects at the projection surface 3 the compensated image 1B instead of the input image 1A received at the projector image input 21. The image displayed at the projection surface 3 is free of distortion and targets the display area of the projection surface 3.


An exemplary embodiment of the handheld device according to the present invention is as following. FIG. 6 is a functional block of a handheld device according to an embodiment of the present invention. The handheld device 4 of FIG. 6 comprises a projector 2, a stereoscopic sensor 7, a 2-axes accelerometer 8, a 3 axes gyroscope 9, an evaluation unit 10 and a compensation unit 11. Also represented on FIG. 6 is the projection surface 3. The evaluation unit 10 is fed by the stereoscopic sensor 7 and the 2-axes accelerometer 8 and the 3 axes gyroscope 9. Furthermore the evaluation unit 10 outputs to the compensation unit 11. The compensation unit 11 generates a compensated image 1B for the projector 2 with an input image 1A and the input from the evaluation module 10. The projector 2 displays at the projection surface 3 the compensated image 1B received from the compensation unit 11.


The sensors according to the present invention can be connected to the handheld device on chip or they can be externally connected to it, serving as possible extensions of the handheld device. In either case the extension will be considered to be embedded to the handheld device. Presently, pico-projectors or stereoscopic sensors considered here are said to be embedded in the handheld device.


The intelligence of the system can be implemented in a computer readable medium. The handheld device may comprise a computer readable medium such that computers programs are loadable into data-processing units and capable of executing embodiments of the present invention.


The invention has been described with reference to preferred embodiments. However, many variations are possible within the scope of the invention.

Claims
  • 1. A method for the projection of a sequence of images onto a projection surface by a handheld device having an embedded projector, said method comprising, for an input image to be projected of said sequence of images: (a) a step of evaluation wherein: (i) a spatial position variation is evaluated with respect to a reference spatial position of said handheld device;(ii) said spatial position variation is defined by a spatial orientation variation and a spatial vector variation, said spatial vector variation being determined by 3 independent coordinates; and(iii) at least one of said 3 independent coordinates is provided by a stereoscopic sensor;(b) a step of compensation wherein a compensated image is generated from said input image according to said spatial position variation; and(c) a step of projecting onto said projection surface said compensated image.
  • 2. A method according to claim 1 wherein said reference spatial position is constant for all images of said sequence of images.
  • 3. A method according to claim 1 wherein said reference spatial position is variable between two images of said sequence of images.
  • 4. A method according to claim 1 wherein said spatial orientation variation is determined by 3 independent angles.
  • 5. A method according to claim 6 wherein said 3 independent angles are provided by a 3 axes gyroscope.
  • 6. A method according to claim 1 wherein 2 of said 3 independent coordinates are provided by a 2-axes accelerometer.
  • 7. A method according to claim 1 wherein said step of compensation utilizes a graphics processing unit technique to generate said compensated image.
  • 8. A handheld device having an embedded projector capable of projecting a sequence of images onto a projection surface, said handheld device comprising: an evaluation unit adapted to evaluate a spatial position variation with respect to a reference spatial position of said handheld device, said spatial position variation being defined by a spatial orientation variation and a spatial vector variation, and said spatial vector variation being determined by 3 independent coordinates;a compensation unit adapted to generate a compensated image from an input image to be projected of said sequence of images, depending on said spatial position variation; anda stereoscopic sensor configured to provide at least one of the 3 independent coordinates determining said spatial vector variation;said embedded projector being adapted to project onto said projection surface said compensated image.
  • 9. A handheld device according to claim 8 wherein said stereoscopic sensor comprises a two-lens stereo camera.
  • 10. A handheld device according to claim 8 wherein said stereoscopic sensor comprises two single-lens cameras joined together.
  • 11. A handheld device according to claim 8 wherein said spatial orientation variation is determined by 3 independent angles.
  • 12. A handheld device according to claim 8 further comprising a 3 axes gyroscope adapted to provide said spatial orientation variation.
  • 13. A handheld device according to claim 8 configured to utilize a graphics processing unit technique to generate said compensated image.
  • 14. A computer program product, stored in non-transient computer memory, comprising a computer readable medium, having thereon computer program instructions being loadable into a data-processing unit and adapted to cause execution of said program instructions when run by the data-processing unit, said program instructions comprising: (a) evaluation instructions, wherein: (i) a spatial position variation is evaluated with respect to a reference spatial position;(ii) said spatial position variation is defined by a spatial orientation variation and a spatial vector variation, said spatial vector variation being determined by 3 independent coordinates; and(iii) at least one of said 3 independent coordinates is provided by a stereoscopic sensor;(b) compensation instructions generating a compensation image from an input image according to said spatial position variation; and(c) projection instructions projecting a compensated image.
  • 15. The computer program product according to claim 14 wherein the program instructions improve image stabilization in a handheld pico-projector.
Priority Claims (1)
Number Date Country Kind
12306080.8 Sep 2012 EP regional