This application claims priority from Korean Patent Application No. 10-2013-0139773, filed on Nov. 18, 2013 in the Korean Intellectual Property Office, the disclosure of which is incorporated herein by reference in its entirety.
1. Field
Apparatuses and methods consistent with exemplary embodiments relate to a camera integrated with a light source, and more particularly, a three-dimensional (3D) depth camera integrated with a light source, which can be miniaturized and allow a user to selectively photograph a desired object.
2. Description of the Related Art
In a related art, game consoles, smart phones, smart televisions (TVs), etc., have been provided with a 3D depth camera for motion recognition and visual communication.
The 3D depth camera 10 shown in
The 3D depth camera 20 shown in
Additionally, with respect to a 3D depth camera emitting infrared light (such as the 3D depth camera 10 shown in
One or more exemplary embodiments may provide a camera integrated with a light source, which can be miniaturized.
Furthermore, one or more exemplary embodiments may also provide a camera integrated with a light source, in which a user can select and photograph even a small object at a long distance.
Also one or more exemplary embodiments may provide a camera integrated with a light source, which is not affected by external noise of ambient light to prevent an error deviation in calculating a depth distance.
According to an aspect of an exemplary embodiment, there is provided a camera integrated with a light source, the camera including: the light source configured to emit light; a scanning mirror configured to scan the light emitted from the light source toward an object; an optical detector configured to detect the light reflected from the object; and a controller configured to transmit a control signal to the scanning mirror for selectively adjusting a scanning angle of the scanning mirror, and to generate a depth image of the object using the reflected light detected by the optical detector.
The light source may include an infrared laser diode.
The infrared laser diode may emit infrared light having a wavelength of 830 m-940 nm.
The light source may emit the light of which a pulse width is modulated in sync with the transmitted control signal.
The camera may further include a collimation lens on an optical path between the light source and the scanning mirror to prevent the light emitted from the light source from being scattered.
The optical detector may receive infrared or near-infrared light.
The optical detector may be configured as a single pixel.
The camera may further include an infrared or near-infrared band pass filter on an optical path to the optical detector.
The optical detector may include at least one of a PIN photodiode, an avalanche photo diode (APD), and a phototransistor.
The camera may further include an infrared or near-infrared band light beam splitter configured to direct the light emitted from the light source toward the scanning mirror.
The camera may further include an optical absorption section, wherein the infrared or near-infrared band light beam splitter may be configured to transmit light except infrared or near-infrared light to the optical absorption section which prevents the transmitted light from reflection.
The scanning minor may include a micro-electromechanical systems (MEMS) minor of which an angle is adjustable in a two-dimensional (2D) direction.
The scanning mirror may be adjustable, in accordance with the control signal, to scan the light within only a partial range of angles from among a full range of angles of the scanning mirror.
The camera may further include a wide-angle extending lens between the scanning mirror and the object to extend a scanning angle of the scanning mirror.
The wide-angle extending lens may include at least one of acryl, glass and sapphire.
One or more sides of the wide-angle extending lens may include at least one of a optical coating for anti-reflection (AR) and a band pass filter (BPF).
The controller may control a modulation speed, a pulse width, and an optical intensity at pulse width modulation (PWM) for the light source.
The controller may determine a distance between the camera and the object based on a light source PWM signal and an optical signal reflected from the object and returning toward the optical detector.
The controller may use time-to-digital conversion to determine the distance.
The controller may convert a depth distance calculated according to scanning sequences into a 2D image format for one frame image.
The controller may transmit the converted frame image to a gesture recognition engine or a host through a parallel data interface or a serial data interface of the controller.
The camera may further include a color image sensor.
The color image sensor may receive light through a wave-dependent splitter.
The wave-dependent splitter may include at least one of a dichroic mirror, an optically coated beam splitter, and a prism lens.
The wave-dependent splitter may refract visible RGB light and transmits infrared and near-infrared light.
The wave-dependent splitter may transmit visible RGB light and refracts infrared and near-infrared light.
The wave-dependent splitter and the wide-angle extending lens for extending a scanning angle may align an optical axis for both the RGB image and the depth image.
According to an aspect of another exemplary embodiment, there is provided a method of operating a camera integrated with a light source, the method including: emitting, by the light source, light; scanning, by a scanning mirror of the camera, the emitted light toward an object; detecting, by an optical detector of the camera, the light reflected from the object; and generating a depth image of the object using the detected light, wherein the scanning comprises selectively adjusting a scanning angle of the scanning mirror.
According to an aspect of another exemplary embodiment, there is provided a method of operating a camera integrated with a light source, the method including: transmitting a first control signal to control the light source to emit light; transmitting a second control signal to selectively adjust a scanning angle of a scanning mirror that scans the light emitted from the light source toward an object; and generating a depth image of the object using detected light reflected from the object.
The above and/or other aspects will become apparent and more readily appreciated from the following description of exemplary embodiments, taken in conjunction with the accompanying drawings, in which:
Below, exemplary embodiments will be described in detail with reference to accompanying drawings. The following exemplary embodiments describe configurations directly related to the present inventive concept, and the descriptions of other configurations may be omitted. However, it will be understood that the omitted configurations are not unnecessary in realizing an apparatus or system to which the present inventive concept is applied. Further, like numerals refer to like elements throughout. Hereinafter, expressions such as “at least one of,” when preceding a list of elements, modify the entire list of elements and do not modify the individual elements of the list.
As shown in
The light source 110 may use a single mode laser diode which can emit light having a near-infrared (NIR) band, for example, 830 nm-940 nm. As shown in
The collimation lens 120 is placed in front of the light source 110 and collimates the light emitted from the light source 110.
The beam splitter 130 is placed in front of the collimation lens 120 and splits infrared or near-infrared light from incident light toward the scanning mirror 150. At this time, light of the incident light other than the infrared or near-infrared light from the collimation lens 120 passes through the beam splitter 130 to the optical absorption section 140. Because the transmitted light may cause reflection, the optical absorption section 140 is placed at an opposite side to the light-incident side of the beam splitter 130. The beam splitter 130 transmits or passes the light reflected and returning from the scanning mirror 150 toward the optical detector 170. Thus, the beam splitter 130 serves as an infrared/near-infrared optical band pass filter through, for example, surface coating.
The scanning minor 150 may employ a micro-electromechanical system (MEMS) two-dimensional (2D) scanning mirror. The scanning mirror 150 in synchronized with an electric X-Y control signal of the controller 180 and causes the light to travel from the light source 110 to a desired 2D direction. Also, the scanning minor 150 reflects the light incident at a specific angle toward an object 1, and returns the light reflected from the object 1 toward the beam splitter 130.
The wide-angle extending lens 160 serves to widen the incident direction and angle of the light output from the scanning minor 150. It is understood that one or more other exemplary embodiments may not implement a wide-angle extending lens 160. For example, if the angle of the scanning mirror 150 is sufficiently large, the wide-angle extending lens 160 may be replaced by a plate that can transmit the infrared band light. The wide-angle extending lens 160 may include acryl, glass, sapphire, etc. Furthermore, optical coating for anti-reflection (AR) or a band pass filter (BPF) may be applied to a single side or both sides of the wide-angle extending lens 160.
The optical detector 170 receives the light reflected from the object 1 and returning through the scanning mirror 150 and the beam splitter 130. The optical detector 170 receives near-infrared or infrared band light. The optical detector 170 may be achieved by a single pixel or photo-detector according to an exemplary embodiment, as opposed to a sensor including 2D pixels for the 3D depth camera. In some exemplary embodiments, an infrared/near-infrared band pass filter may be included in the optical detector 170 and cut off external noise of ambient light. The optical detector 170 may include a PIN photodiode, an avalanche photo diode (APD), a phototransistor, etc.
The controller 180 is configured to drive the light source 110 and to control a modulation speed, a pulse width, an optical pulse intensity, etc., when modulating the optical pulse width of the light source 110. The controller 180 controls the angle of the scanning minor 150 in sync with the pulse width modulation of the light source 110, thereby controlling the angle at which the light is emitted in a 2D space. The controller 180 controls operations of the optical detector 170 that receives light reflected and returning from an object 1.
The controller 180 calculates a distance range between the camera 100 and an object 1 based on a PWM signal of the light source 110 and an optical signal returning from the optical detector 170. At this time, the controller 180 may perform distance calculation through time-to-digital conversion. The controller 180 may convert a depth distance calculated according to scanning sequences into a 2D image format for one frame image. The controller 180 transmits the converted depth frame image to a gesture recognition engine or host through a parallel or serial data interface.
Below, operations of a dynamic range finder will be described with reference to
At a first operation, the controller 180 modulates the light emitted from the light source 110 in accordance with preset PWM conditions, for example, a modulation speed, a duty cycle, a driving current control, etc. According to an exemplary embodiment, a short impulse may be used, although it is understood that one or more other exemplary embodiments are not limited thereto. For example, according to another exemplary embodiment, a general PWM having a duty cycle of about 50% may be used.
At a second operation, the light emitted from the light source 110 is split by the beam splitter 130 so that about half the light travels toward the scanning mirror 150 and the remaining light travels toward the optical absorption section 140.
At a third operation, the light split toward the scanning minor 150 by the beam splitter 130 is refracted to a 2D space of X-Y axes in accordance with the angles of the scanning mirror 150.
At a fourth operation, the angle of the scanning minor 150 is controlled by a control signal level corresponding to the X-Y axes of the mirror 150 applied from the controller 180.
At a fifth operation, the angle of the light refracted by the scanning minor 150 is extended in accordance with a target FoV by the wide-angle extending lens 160. Here, the angle is extended in a case, for example, where the mirror scanning angle is relatively small. However, according to one or more other exemplary embodiments, for example, where the minor angle is proper to the FoV, the wide-angle extending lens 160 may be replaced by a plate.
At a sixth operation, the controller 180 determines a level of an (X, Y) control signal to be applied to the minor, based on information such as a refraction angle of the scanning mirror 150 and a refraction angle of the wide-angle extending lens 160.
At a seventh operation, light source pulses are sequentially emitted through the scanning mirror 150, and the sequentially emitted light is reflected from an object 1 and enters the optical detector 170. That is, the light reflected from the object 1 is received in the optical detector 170 via the wide-angle extending lens 160, the scanning minor 150, and the beam splitter 130 in sequence.
At an eighth operation, the controller 180 converts the incident optical pulses reflected and returning from the object 1 into an electric signal, and performs level normalization.
At a ninth operation, the controller 180 calculates depth data R (k), R(k+1), . . . corresponding to the 2D angle by applying time-to-digital conversion to a time difference between an optical PWM edge and a received optical pulse edge reflected and returning from the object.
At a tenth operation, the controller 180 obtains spherical coordinates based on the distance R(k), R(k+1), . . . between the camera and the object calculated in the ninth operation, and transforms the obtained spherical coordinates into X-Y Cartesian coordinates. For example, R (k) may be transformed into D (X (i), Y (j)).
At an eleventh operation, the controller 180 may performing mapping of a mirror scanning angle (θ, f) to a depth image X-Y Cartesian coordinates D(X, Y). At this time, the mapping may be predetermined in the controller 180 and may undergo calibration.
According to an exemplary embodiment, if a light source PWM speed is 20 MHz, a maximum detectable depth distance is about 7.5 meters. As shown in
As shown in
In a related art 3D depth camera, the field of view (FoV) for emitting the light is fixed in accordance with specifications of a diffuser. If the specifications of the CIS and the lens are determined, the FoV of the light receiver is fixed. Therefore, the related art 3D depth camera cannot detect a small object at a long distance after the specifications of the IR light source and the FoV of the IR receiver are determined for detecting a small object at a short distance. On the other hand, if the specifications of the IR light source and the FoV of the IR receiver are determined for detecting a small object at a long distance, the FoV is narrow at a short distance
However, in the camera 100 integrated with the light source according to an exemplary embodiment, a local FoV and a full FoV can be dynamically controlled and adjusted.
The wave-dependent splitter 255 may employ a dichroic mirror. The wave-dependent splitter 255 is used as an optical filter for splitting the light into RGB light and the infrared/near-infrared light. In
The wave-dependent splitter 255 may employ a beam splitter, a prism lens, etc., coated with optical filters to have characteristics of transmitting spectrums of light as shown in
Thus, the wave-dependent splitter 255 shows different transmittance characteristics in accordance with the wavelengths of the incident light. In
Because a related 3D depth camera is spatially distant from an RGB image sensor module and therefore an RGB image and a depth image are different in an optical axis from each other, pixel registration is separately needed between the RGB image and the depth image.
On the other hand, as shown in
As described above, a camera integrated with a light source according to an exemplary embodiment can be miniaturized since a diffuser and a CIS are not used and an infrared laser diode and a photo detector are integrated with each other.
According to an exemplary embodiment, the camera integrated with the light source can recognize a small object at a long distance and dynamically control the FoV.
Furthermore, according to an exemplary embodiment, the camera integrated with the light source can be minimally affected by external noise of ambient light due to local scanning.
Also, according to an exemplary embodiment, the camera integrated with the light source employs the infrared light source that is not diffused in time order according to coordinates of individual depth data, and thus the intensity of the infrared light is high as compared with external light noise.
While not restricted thereto, an exemplary embodiment can be embodied as computer-readable code on a computer-readable recording medium. For example, a method of operating a camera as described above may be performed by executing instructions recorded on a computer-readable recording medium according to an exemplary embodiment. The computer-readable recording medium is any data storage device that can store data that can be thereafter read by a computer system. Examples of the computer-readable recording medium include read-only memory (ROM), random-access memory (RAM), CD-ROMs, magnetic tapes, floppy disks, and optical data storage devices. The computer-readable recording medium can also be distributed over network-coupled computer systems so that the computer-readable code is stored and executed in a distributed fashion. Also, an exemplary embodiment may be written as a computer program transmitted over a computer-readable transmission medium, such as a carrier wave, and received and implemented in general-use or special-purpose digital computers that execute the programs. Moreover, it is understood that in exemplary embodiments, the controller 180 can include circuitry, a processor, a microprocessor, etc., and may execute a computer program stored in a computer-readable medium.
Although a few exemplary embodiments have been shown and described, it will be appreciated by those skilled in the art that changes may be made in these exemplary embodiments without departing from the principles and spirit of the present inventive concept. Therefore, the foregoing is to be considered as illustrative only. The scope of the invention is defined in the appended claims and their equivalents. Accordingly, all suitable modification and equivalents may fall within the scope of the invention.
Number | Date | Country | Kind |
---|---|---|---|
10-2013-0139773 | Nov 2013 | KR | national |