The present invention relates to projection. More specifically, the present invention relates to a projecting device.
Standard lenses used in cameras and imaging and projection devices may typically focus only at a limited range of distances. Objects located at other distances from the lens will appear blurred. In practice, the decrease in sharpness is gradual on either side of the focused distance, and the interval of distances within which the degraded sharpness is imperceptible under normal viewing conditions is referred to as the Depth Of Field (DOF).
DOF is inversely proportional to the lens aperture diameter. Decreasing the aperture size, a lens is effectively turned into a pencil beam collimator, which in the limit has an infinite DOF (i.e., objects at all distances appear in focus). However, small apertures reduce the amount of light received by the film or the sensor, so it is impractical to achieve large DOFs with standard lenses.
There is thus provided, in accordance with embodiments of the present invention, an imaging device. The imaging device may include at least one single pixel imaging sensor configured to sense image data for a single pixel along a line of sight. The imaging device may also include at least one reoreintable mirror, of which each reorientable mirror is exclusively optically coupled to one or more of said at least one single pixel imaging sensor for deflecting the line of sight of that single pixel imaging sensor. The imaging device may further include a controller for synchronously reorienting each of said at least one reorientable mirror to scan the line of sight across at least a sector of a scene. The imaging device may also include a readout circuit for reading out acquired image data from each of said at least one single pixel imaging sensor. The imaging device may be configured to sample at least two independent dimensions of a four dimensional light field.
Furthermore, in accordance with some embodiments of the present invention, the imaging device may further include at least one collimator each of which is configured to limit reception of incoming light to a predetermined angle of view, for each of said at least one single pixel imaging sensor.
Furthermore, in accordance with some embodiments of the present invention, said at least one single pixel imaging sensor may include a plurality of single pixel imaging sensors.
Furthermore, in accordance with some embodiments of the present invention, the plurality of single pixel imaging sensors may be arranged in a one-dimensional linear arrangement.
Furthermore, in accordance with some embodiments of the present invention, the plurality of single pixel imaging sensors may include single pixel imaging sensors configured to respond to different wavelength ranges of incoming light.
Furthermore, in accordance with some embodiments of the present invention, the plurality of single pixel imaging sensors may be arranged in a two-dimensional array.
Furthermore, in accordance with some embodiments of the present invention, said at least one reoreintable mirror may include a plurality of reorientable mirrors, and wherein the controller is configured to control reorentation of each of said plurality of mirrors about at least two axes.
Furthermore, in accordance with some embodiments of the present invention, said at least two axes may be orthogonal axes.
Furthermore, in accordance with some embodiments of the present invention, a tilting angle of each of said at least one reorientable mirror may be modifiable, so as to allow increasing or decreasing the tilting angle.
Furthermore, in accordance with some embodiments of the present invention, the readout circuit may be configured to sample image data from each of said at least one single pixel imaging sensor, so that the sampled image data is non-linearly related to orientation of said at least one reorientable mirror.
Furthermore, in accordance with some embodiments of the present invention, the imaging device may further include an illumination source for illuminating the scene along the line of sight.
Furthermore, in accordance with some embodiments of the present invention, the imaging device may further include an interferometer optically positioned along the line of sight.
Furthermore, in accordance with some embodiments of the present invention, there is provided a plenoptic imaging device, which may include an array of single pixel imaging sensors each configured to sense image data for a single pixel along a line of sight. The plenoptic imaging device may also include an array of reoreintable mirrors, each of which is optically coupled to one or more of the single pixel imaging sensors of the array for deflecting the line of sight of each of the single pixel imaging sensors. The plenoptic imaging device may further include a controller for synchronously reorienting each of said reorientable mirrors to scan each of the lines of sight across at least a sector of a scene. The plenoptic imaging device may also include a readout circuit for reading out acquired image data from the array of single pixel imaging sensors. The plenoptic imaging device may be configured to sample at least three independent dimensions of a four dimensional light field.
Furthermore, in accordance with some embodiments of the present invention, the array of single pixel imaging sensors may include single pixel imaging sensors configured to respond to different wavelength ranges of incoming light.
Furthermore, in accordance with some embodiments of the present invention, a tilting angle of each of the reorientable mirrors may be modifiable, so as to allow increasing or decreasing the tilting angle.
Furthermore, in accordance with some embodiments of the present invention, the readout circuit may be configured to sample image data from each of the single pixel imaging sensors, so that the sampled image data is non-linearly related to orientation of the reorientable mirrors.
Furthermore, in accordance with some embodiments of the present invention, the plenoptic imaging device may further include an illumination source for illuminating the scene along at least one of the lines of sight.
Furthermore, in accordance with some embodiments of the present invention, there is provided a plenoptic projector device, which may include an array of laser radiation sources each configured to project a light beam along a ray. The plenoptic projector device may also include an array of reoreintable mirrors, each of which is optically coupled to one or more of the laser radiation sources of the array for deflecting the ray of each of the laser radiatino sources. The device may further include a controller for synchronously reorienting each of said reorientable mirrors to sweep a sector in space by each of the rays. The device may still further include a modulation circuit for modulating intensity of the beam of each of the laser radiation sources. The projector device may be configured to reproduce at least three independent dimensions of a four dimensional light field.
Furthermore, in accordance with some embodiments of the present invention, each of the laser radiation sources may include a plurality of laser radiation sources of different wavelengths.
Furthermore, in accordance with some embodiments of the present invention, a tilting angle of each of the reorientable mirrors may be modifiable, so as to allow increasing or decreasing the tilting angle.
Furthermore, in accordance with some embodiments of the present invention, the modulation circuit may be configured to change the intensity of the beam of one or more of the laser radiation sources in a timing sequence that is non-linearly related to orientation of the reorientable mirrors.
The subject matter regarded as the invention is particularly pointed out and distinctly claimed in the concluding portion of the specification. The invention, however, both as to organization and method of operation, together with objects, features, and advantages thereof, may best be understood by reference to the following detailed description when read with the accompanying drawings. It should be noted that the figures are given as examples only and in no way limit the scope of the invention. Like components are denoted by like reference numerals.
Light field is a function that describes the amount of light traveling in every direction through every point in space, and can be parameterized by five parameters (point coordinates x,y,z+direction φ,θ), resulting in a five-dimensional (5D) function L(x,y,z,φ,θ;X), where L represents light intensity from x,y,z in the direction φ,θ at wavelength λ (which is considered here as a parameter rather than additional sixth dimension of the light field). In what follows, we will omit the dependence on the wavelength whenever possible.
Assuming no occlusions or optical interferences are present in space, the amount of light along each direction φ,θ is the same. Thus, the five dimensional representation of the light field function is redundant, and effectively the light field can be described by four parameters, for example, by parametrizing two planes u,v and s,t (u.v being a position coordinate in one plane and s,t being a position coordinate in the other plane).
A plenoptic camera (also known as light-field camera) is a conceptual camera system that allows capturing four-dimensional (4D) light field information of a scene. Typical plenoptic camera designs use a 2D camera array (a plurality of cameras arranged in a two-dimensional matrix) or a microlens array to capture 4D light field.
The resulting light field can be represented as an 2D array of images, each acquired by a single camera of the camera array (or a single lens of the microlens array). In this representation, the coordinates u,v represent the position of the camera in the array, and the coordinates t,s represent the position of a pixel in the image acquired by that camera.
Such light field information can be used in computational photography, e.g., in the following applications:
A. Light field/image based rendering. By extracting appropriate two-dimensional (2D) slices from the 4D light field of a scene, one may produce novel views of the scene (this is generally referred to as image-based rendering). Depending on the parameterization of the light field and slices, these views may be, for example, perspective, orthographic, crossed-slit, multi-perspective, or another type of projection.
B. Synthetic aperture photography. By integrating an appropriate 4D subset of the samples in a light field, one can approximate the view that would be captured by a camera having a finite (i.e., non-pinhole) aperture. Such a view has a finite depth of field. By shearing or warping the light field before performing this integration, one can focus on different fronto-parallel or oblique planes in the scene.
A traditional (non plenoptic) camera can be considered as a trivial setting of the plenoptic camera array with a single camera. Such a camera is capable of capturing only two dimensions of the 4D light field.
Reproducing the 4D light field is conceptually possible by reversing the direction of rays in a plenoptic camera and replacing light sensors by light sources, arranged in a projector array. The reproduced light field when viewed by a human observer allows achieving a full 3D illusion of an object. The trivial setting of such a projection array device is a single projector, capable of reproducing two dimensions of the light field, which are perceived as a 2D image.
In accordance with some embodiments of the present invention it is proposed to replace the camera array by an array of micro-mirrors, each exclusively optically connected to a controlled light source.
In the context of the present invention, the term “light” is understood to mean any electromagnetic radiation, including (but not limited to) visible light, infrared light, ultraviolet light. The term “plenoptic” referring to a device is understood to mean a device capable of acquiring (sampling) or reproducing an approximation of a light field of a scene at a set of directions not lying in a single plane.
In some embodiments of the present invention described hereinafter, imaging devices having a single mirror and one or a plurality of single pixel sensors are described, for simplicity. Devices with a plurality of mirrors are later described.
Theoretically, one can envisage a camera consisting of a single pixel sensor (such as, for example, a photodiode 12 aligned with a pencil collimator 14 (see
By sweeping a desired sector of view, for example in a raster scan, and recording the responses of the photodiode, an image may be obtained. The main property of such an image is that it has a very large (ideally, infinite) DOF. The acquisition of pixels by such a camera may be staged in time, somewhat resembling the action of an electronic rolling shutter implemented in some CMOS sensors.
In accordance with some embodiments of the present invention, an imaging device may include at least one single pixel imaging sensor configured to sense image data for a single pixel along a line of sight. The device may also include at least one reoreintable mirror, of which each reorientable mirror is exclusively optically coupled to one or more of said at least one single pixel imaging sensor for deflecting the line of sight of that single pixel imaging sensor. The device may further include a controller for synchronously reorienting each of said at least one reorientable mirror to scan the line of sight across at least a sector of a scene. The device may also include a readout circuit for reading out acquired image data from each of said at least one single pixel imaging sensor. The imaging device may be configured to sample at least two independent dimensions of a four dimensional light field.
As the components of the theoretical optical system shown in
The micro-mirror device 22 may be designed to rotate about two orthogonal axes 27, so as to allow scanning the imaged sector in a tow dimensional manner.
In some embodiments of the present invention, the micro-mirror device may be designed to rotate about a single axis, and thus facilitate one-dimensional scanning of the imaged object (along a single axis). In other embodiments of the present invention, the micro-mirror device may be designed to rotate about a plurality of orthogonal or non-orthogonal axes.
An example of a micro-mirror device suitable for use in an imaging device in accordance with some embodiments of the present invention may include, for example, MEMS (Micro-Electro-Mechanical-System) scanning micro-mirror manufactured by Microvision™ (Redmond, Wash., US), which is a silicon device at the center of which a tiny mirror is located. The mirror is connected to small flexures allowing it to oscillate. A two-dimensional (2D) MEMS scanning mirror device may rotate about two orthogonal axes to capture an image pixel by pixel. The maximal angle of rotation of the micro-mirror determines the field of view of the imaging device, in accordance with embodiments of the present invention. An exemplary micro-mirror device manufactured by Microvision is capable of rotating the mirror about two perpendicular axes in steps at a frequency of about 12 MHz.
The proposed imaging device does not require lenses, which is likely to significantly reduce the production cost.
The imaging device may further include processor 28, for processing image date received by sensor 12, storage device 24, for storing image data (raw data, or processed data), and for storing software code (e.g., an application for execution by the processor of the imaging device).
The imaging device may also include controller 25, for controlling the operation of micro-mirror device 22 (direction and/or angle of tilt).
Imaging sensor 12 may be configured to sample image date at a frequency matching to the operation frequency of the micro-mirror 22, so as to efficiently and smoothly cover the imaged sector. For example, a video camera scanning at 30 frames per second (fps) with spatial resolution of, for example, 1080*1920, would have to vibrate the mirror at rates of about 30 Hz along one axis and for a spatial resolution of 1080*30, the camera would vibrate at about 30 KHz along the other axis. Scanning in both scanning directions of the moving mirror could reduce scanning rates in half.
In accordance with some embodiments of the present invention, the micro-mirror is configured to rotate about two planes. A pencil collimator is placed between the micro-mirror and the imaging sensor, which in embodied in the form of a single photodiode (one pixel). At every point in time, the imaging device is configured to acquire a single pixel image along a single ray, and multiple pixels are obtained by sweeping the scene with the ray and sampling pixels of the image during the sweep motion. In order to obtain a complete image of an entire scene, the ray sweeps the scene and pixels are sampled in an orderly manner (or at least at a known order)
One simple sweeping scheme may be raster scanning. Other scanning schemes may be used, such as for example as described in the following section.
In another embodiment of the present invention a single mirror may be replaced by an optical system comprising two mirrors each allowed to rotate about a single axis in such a way that together they may deflect light in two independent directions.
In accordance with some embodiments of the present invention, as depicted in
In accordance with some embodiments of the present invention, a control sub-system may be used to control the position of the micro-mirror and to record the signal received from the sensor. In what follows, we assume for simplicity that the optical system implements the moving ray setting (as shown in
A. Raster scan with fixed angular step.
B. Holographic sampling. The scan order may be driven by a holographic sampling hash function, which guarantees that every small subset of pixels covers the region of interest approximately uniformly. In this manner, the imaging device may be made to produce a stream of pixels, accumulation of which progressively improves the image resolution.
C. Variable zoom. By increasing or decreasing the angular step, the imaging device may acquire wider or narrower angular tilt, which is equivalent to zooming out and in, without having to use a costly zoom lens.
D. Region of interest. The micro-mirror may be made to achieve a wide range of possible angular positions, only a subset of which is used during image acquisition, effectively creating a region of interest in which the image is sampled. The simplest form of the region of interest is a rectangle, but other more complex forms consisting of multiple disconnected regions may be achieved. The region of interest may be determined by a higher-level computer vision algorithm such as object detection and tracking operating on a low resolution and low frame rate full angular aperture image.
E. Virtual Pan-Tilt-Zoon (PTZ). By combining holographic sampling with variable zoom techniques (see B and C above), a replacement of a costly PTZ lens may be realized. Moreover, an imaging device according to embodiments of the present invention may have multiple regions of interests with different zoom levels, which is cannot be achieved by standard PTZ lenses. The determination of PTZ settings in an imaging device according to some embodiments of the present invention may be performed by a higher-level computer vision algorithm.
F. Variable resolution. The angular velocity of the mirror may change non-uniformly at different pixels of the image, effectively over-sampling or under-sampling different regions of the imaged object. The resolution of the sampling of the image may be determined by a higher-level image processing or computer vision algorithm operating on a low-resolution image. For example, objects of interest such as faces may be detected in the image and assigned higher resolution. Alternatively, the readout circuit may sample the signal produced by the sensor non-uniformly in time. In other words, the sampling of image data from the sensor (or sensors) is non-linearly related to orientation of the reorientable mirror.
G. Variable frame rate and resolution. The frequency of the micro-mirror displacement may limit the budget (i.e., number) of pixels the imaging device may produce per second. This budget can be allocated in various schemes of space and time producing low-frame rate high-resolution or high-frame rate low-resolution imaging (e.g., video). An imaging device according to embodiments of the present invention may dynamically trade off these parameters either at frame or region or pixel levels. For example, regions with fast motion can be acquired at higher frame rates, while static regions may be acquired at lower frame rates.
H. Variable dynamic range. By controlling exposure time, an imaging device according to embodiments of the present invention may acquire a sequence of frames at different exposures effectively producing a high dynamic range image. Dynamic range may vary in space and time. By “exposure time”, it is meant the time a photodiode is sampled/reset for photons count. Moreover, by slowing down the scanning rate, and extending photon integration time per pixel, one could keep the spatial resolution, while obtaining a sharp and clean image even at low lighting conditions.
I. Variable exposure. By controlling exposure time at pixel level, the imaging device may perform non-linear transformations of the dynamic range and adapt them both in space and time. For example, pixels belonging to dark regions may be assigned higher exposure times, while pixels in bright regions may be exposed for shorter exposure times. The determination of exposure time may be handled by a higher-level image processing algorithm working, e.g., on a lower resolution image.
Various optical devices allow transforming an image to a transformation domain (e.g., Fourier transform). An example of an optical device used to transform an image to a transformation domain is a converging lens. By placing a transforming optical device in front of the imaging device, the sampling performed by the imaging device may become frequency domain sampling. Sampling patterns different from the standard Cartesian sampling are known to be advantageous, such as for example, compressive sensing techniques, reconstruction of a better image from the same number of samples or same quality image from a smaller number of samples. Below, several examples of scan patterns are proposed:
1. Raster scan.
2. Spiral sampling: progressively increasing the sampling frequency and thus the resolution.
3. Holographic sampling in the frequency domain.
4. Polar sampling: similar to one used in CT (Computerized Tomography). Reconstruction is possible by using, for example, the inverse Radon transform, filtered back projection, or convex optimization techniques.
5. Pseudo-polar sampling and reconstruction method introduced by A. Averbuch et al in “Accurate and Fast Discrete Polar Fourier Transform”, Journal on Applied and Computational Harmonic Analysis, Vol. 21, pp. 145-167, 2006.
Other transforms achieved by optical devices may be implemented as well in an imaging device in accordance with some embodiments of the present invention.
Color acquisition by an imaging device in accordance with embodiments of the present invention may be achieved by placing a set of three photodiodes with red, green, and blue filters. Since the physical dimensions of the photodiodes may exceed the width of the ray, a diffuser may be placed between the collimator and the photodiodes to ensure the ray coming out of the collimator is spread over a bigger area. An example of such arrangement is illustrated in
Some embodiments of the present invention may also include a plurality of mirrors and corresponding imaging sensors, as well as illumination sources, optical elements and electronic circuits. This may allow for fuller acquisition or reproduction of light field information as demonstrated henceforth.
In accordance with some embodiments of the present invention, a plenoptic imaging device may include an array of single pixel imaging sensors each configured to sense image data for a single pixel along a line of sight. The plenoptic imaging device may further include an array of reoreintable mirrors, each of which is optically coupled to one or more of the single pixel imaging sensors of the array for deflecting the line of sight of each of the single pixel imaging sensors. The plenoptic imaging device may also include a controller for synchronously reorienting each of said reorientable mirrors to scan each of the lines of sight across at least a sector of a scene. The plenoptic imaging device may also include a readout circuit for reading out acquired image data from the array of single pixel imaging sensors. The plenoptic imaging device may be configured to sample at least three independent dimensions of a four dimensional light field.
Imaging device 180 may include a micro-mirror array 95 (which may be provided with hood 97, which may resemble, for example a lens hood). Micro-mirror array 95 may be connected to micro-mirror steering control 85, in which X modulator 87a, through micro-mirror X control 86a, controls movement of the micro-mirrors of array 95 in X axis, whereas Y modulator 87b, through micro-mirror Y control 86b, controls movement of the micro-mirrors in Y axis (X and Y being orthogonal). The imaging device may further include processing unit 82, which may include operating system 83 and digital signal processor 84. Steering control 85 may be operated by processing unit 82 (e.g. digital signal processor 84).
Light reflected from scene 92, which includes object 94 is collected by tilting the micro-mirrors in a predetermined manner configured to scan a plurality of sectors of the scene by the micro-mirrors. In accordance with some embodiments of the present invention, each micro-mirror may be exclusively assigned a sector. In other embodiments of the present invention, overlapping between sectors in the scanning by the micro-mirrors may occur.
The imaging device may further include readout control 88, which may include photodiode array, e.g., imaging sensors such as, for example, Charge Coupled Device (CCD), Complementary Metal Oxide Semiconductor (CMOS), in which each sensor is exclusively optically coupled to a corresponding mirror of the micro-mirror array 95, and readout arrangement 89 configured to readout image data from the photodiode array and pass it to processing unit 82 (e.g., operating system 83) for processing. The imaging device 180 may thus acquire three or more independent dimensions of the light field of an imaged scene.
The plenoptic imaging device, according to some embodiments of the present invention, effectively samples a 4D light field function of the imaged scene along a set of rays corresponding to each mirror, and in accordance to the chosen scan scheme. In the context of the present invention, “sampling” of a four-dimensional light field L(u,v,s,t;λ) is understood as producing a discrete set of values {L(ui,vi,si,ti,;λi)} for {(ui,vi,si,ti,;λi)} being a set of parameters. By “m independent dimensions”, it is understood that the parameters u,v,s, and t are in turn parametrized as function of m independent parameters {p1, . . . ,pm} sampled at n discrete values {pi}, k=1, . . . , m, i=1, . . . , n, such that the sampled light field is the set {L(u(p1i, . . . , pmi),v(p1i, . . . , pmi),s(p1i, . . . ,pmi),t(p1i, . . . ,pmi);λi)}.
The plenoptic imaging device may include single pixel imaging sensors configured to respond to different wavelength ranges of incoming light.
The tilting angle of each of the reorientable mirrors of the plenoptic imaging device may be modifiable, so as to allow increasing or decreasing the tilting angle.
The readout circuit of the plenoptic imaging device may be configured to sample image data from each of the single pixel imaging sensors, so that the sampled image data is non-linearly related to orientation of the reorientable mirrors. This can be achieved by non-linearly varying the mirror orientation, or non-uniformly sampling the signal read out from the single pixel imaging sensors (since each sampling time of the signal corresponds to a specific mirror orientation), or both. This effectively allows varying the spatial resolution and density of the rays along which the light field is sampled.
The plenoptic imaging device may include an illumination source for illuminating the scene along at least one of the lines of sight.
While the above description (relating to
According to some embodiments of the present invention, a plenoptic projector device may include an array of laser radiation sources each configured to project a light beam along a ray. The plenoptic projector device may also include an array of reoreintable mirrors, each of which is optically coupled to one or more of the laser radiation sources of the array for deflecting the ray of each of the laser radiatino sources. The plenoptic projector device may further include a controller for synchronously reorienting each of said reorientable mirrors to sweep a sector in space by each of the rays. The plenoptic projector device may also include a modulation circuit for modulating intensity of the beam of each of the laser radiation sources. The projector device may be configured to reproduce at least three independent dimensions of a four dimensional light field.
The plenoptic projector device 190 may include processing unit 82, which may include operating system 103 and digital signal processor 84. 4D image data, which may have been acquired using an imaging device, such as the one illustrated in
Thus, light of predetermined characteristics is scanned at predetermined directions by the micro-mirror array 115 so as to present a 4D light field which may be observed by beholder 191.
The plenoptic projector device, according to some embodiments of the present invention, effectively reproduces a 4D light field function approximated by a superposition of light intensities along a set of rays corresponding to each mirror, and in accordance to the chosen scan scheme. “Reproduction” of a four-dimensional light field L(u,v,s,t;λ) is generally understood as filling a region of the space with light beams of different intensities and traveling in different directions in such a way that the intensity of light at a set of points in space parameterized by (u,v,s,t) at wavelength λ is equal to L(u,v,s,t;λ). In the context of the present invention, we will, however, imply approximation of the latter function by a superposition of a set of n rays
L(u,v,s,t;λ)≈I1δ(u−u1, v−v1, s−s1, t−t1)δ(λ−λ1)+ . . .+Inδ(u−un,v−vn, s−sn,t−tn)δ(λ−λn)
where δ denote the Dirac delta function (impulse) or its approximation, {(ui,vi,si,ti,;λi)} is a set of parameters, and the set of intensities {Ii} is the discrete representation of the light field (sampled light field). We will say that a device reproduced “m independent dimensions” of the light field implying that the parameters u, v, s, and t are in turn parametrized by functions of m independent parameters {p1, . . . ,pm} sampled at n discrete values {pki}, k=1, . . . ,m, i=1, . . . ,n.
The controllers driving each of the mirrors effectively produce time sequences of the parameters u, v, s, and t. The modulation circuit effectively produces a time sequence of the intensity values I.
By applying a predetermined projection scheme which includes providing modulated light generated from an array of light emitters (e.g., laser-diodes array 95) and scanned in a predetermined scheme by a micro-mirror array, such as described hereinabove, a light pattern 168 is projected in a three-dimensional manner so as to cause different beholders (represented in the figure in the form of eyes 160 and 162, located at different positions about a virtual visual range 166, and having virtual focal points 164) positioned at different angles with respect to the projector's micro-mirror array 95) to see the light pattern and visualize its three-dimensional aspects (depicted in this figure in the form of different little shapes).
The plenoptic projector device may include laser radiation sources each of which includes a plurality of laser radiation sources of different wavelengths, so as to allow production of different colors. In the context of the present invention, the term “laser radiation source” is not limited to a monochromatic source of coherent light and also includes a plurality of monochromatic light sources producing laser light at different wavelengths optically coupled to produce a single beam of coherent radiation that bears energy at different wavelengths.
The reorientable mirrors of the plenoptic projector device may have a tilting angle which is modifiable, so as to allow increasing or decreasing the tilting angle.
The modulation circuit of the plenoptic projector device may be configured to change the intensity of the beam of one or more of the laser radiation sources in a timing sequence that is non-linearly related to orientation of the reorientable mirrors.
In some embodiments of the above invention, the light source may be a pulsed source, whereas in other embodiments of the present invention the light source may be a continuous light source.
A 3D image of scene 150 may be acquired by a plenoptic imaging device such as plenoptic sensor 152. The 3D image data may be forwarded to image processor 154.
A synthetic 3D image data 153 of a virtual object or scene may be generated by graphic processor 158 and forwarded to image processor 154.
Image processor 154 may combine the 3D image of scene 150 with the 3D image data 153 of the virtual object and plenoptic projector 156 may project the combined 3D image.
Such method 300 may include providing 310 an imaging device that includes one or more single pixel imaging sensors configured to sense image data for a single pixel along a line of sight, one or more reorientable mirrors, of which each reorientable mirror is exclusively optically coupled to one of the single pixel imaging sensors for deflecting the line of sight of that single pixel imaging sensor. The method may also include synchronously exclusively reorienting 320 each of the reorientable mirrors to performing scanning of its line of sight across at least a sector of a scene, and sampling acquired image data from each of the sensors.
Some aspects of the present invention may be embodied as a computer program product saved on one or more non-transitory computer-readable mediums in the form of computer-readable program code embodied thereon. For example, the computer-readable medium may be a computer-readable non-transitory storage medium. A computer-readable storage medium may be, for example, an electronic, optical, magnetic, electromagnetic, infrared, or semiconductor system, apparatus, or device, or any combination thereof.
Computer program code of the above described embodiments of the invention may be written in any suitable programming language. The program code may execute on a single computer, or on a plurality of computers.
The foregoing description of the embodiments of the invention has been presented for the purposes of illustration and description. It is not intended to be exhaustive or to limit the invention to the precise form disclosed. It should be appreciated by persons skilled in the art that many modifications, variations, substitutions, changes, and equivalents are possible in light of the above teaching. It is, therefore, to be understood that the appended claims are intended to cover all such modifications and changes as fall within the true spirit of the invention.
The present invention is a divisional of U.S. patent application Ser. No. 13/023,947, filed on Feb. 9, 2011 and published as U.S. Patent Application Publication No. US 2012/0200829 on Aug. 9, 2012, which is incorporated in its entirety herein by reference.
Number | Date | Country | |
---|---|---|---|
Parent | 13023947 | Feb 2011 | US |
Child | 14450311 | US |