The present disclosure relates generally to a depth map estimation, and more particularly, to a method and an apparatus for calibrating an active depth sensing apparatus and performing active depth sensing.
Active depth sensing can improve on passive approaches (e.g., depth map estimation from stereo observations) by providing uniquely coded illuminations of the scene so that disparity between projector and observer (and thereby depth) can be estimated, even in homogenous or dimly lit regions. Such depth maps have been utilized in object and facial recognition, and may play a key role in augmented reality and scene reconstruction. Several iterations of active depth sensing devices have been developed, which utilize a projector that illuminates bright dots or speckles in the IR spectrum and are observed by a rigidly mounted infrared camera alongside the projector.
According to one embodiment, a method is provided for recovering a depth map from an observed IR image using an active depth sensing device. Speckles are illuminated in an IR spectrum by a projector of the active depth sensing device. Speckles on a horizon image are captured by an IR camera of the active depth sensing device. Projector distortion is estimated based on speckle locations in the horizon image.
According to one embodiment, an electronic device is provided that includes a projector, an IR camera rigidly mounted with the projector, a processor, and a non-transitory computer readable storage medium storing instructions. When executed, the instructions cause the processor to illuminate speckles in an IR spectrum by the projector, capture, by the IR camera, speckles on a horizon image, and estimate projector distortion based on speckle locations in the horizon image.
The above and other aspects, features, and advantages of certain embodiments of the present disclosure will be more apparent from the following detailed description, taken in conjunction with the accompanying drawings, in which:
Hereinafter, embodiments of the present disclosure are described in detail with reference to the accompanying drawings. It should be noted that the same elements will be designated by the same reference numerals although they are shown in different drawings. In the following description, specific details such as detailed configurations and components are merely provided to assist with the overall understanding of the embodiments of the present disclosure. Therefore, it should be apparent to those skilled in the art that various changes and modifications of the embodiments described herein may be made without departing from the scope of the present disclosure. In addition, descriptions of well-known functions and constructions are omitted for clarity and conciseness. The terms described below are terms defined in consideration of the functions in the present disclosure, and may be different according to users, intentions of the users, or customs. Therefore, the definitions of the terms should be determined based on the contents throughout this specification.
The present disclosure may have various modifications and various embodiments, among which embodiments are described below in detail with reference to the accompanying drawings. However, it should be understood that the present disclosure is not limited to the embodiments, but includes all modifications, equivalents, and alternatives within the scope of the present disclosure.
Although the terms including an ordinal number such as first, second, etc. may be used for describing various elements, the structural elements are not restricted by the terms. The terms are only used to distinguish one element from another element. For example, without departing from the scope of the present disclosure, a first structural element may be referred to as a second structural element. Similarly, the second structural element may also be referred to as the first structural element. As used herein, the term “and/or” includes any and all combinations of one or more associated items.
The terms used herein are merely used to describe various embodiments of the present disclosure but are not intended to limit the present disclosure. Singular forms are intended to include plural forms unless the context clearly indicates otherwise. In the present disclosure, it should be understood that the terms “include” or “have” indicate the existence of a feature, a number, a step, an operation, a structural element, parts, or a combination thereof, and do not exclude the existence or probability of the addition of one or more other features, numerals, steps, operations, structural elements, parts, or combinations thereof.
Unless defined differently, all terms used herein have the same meanings as those understood by a person skilled in the art to which the present disclosure belongs. Terms such as those defined in a generally used dictionary are to be interpreted to have the same meanings as the contextual meanings in the relevant field of art, and are not to be interpreted to have ideal or excessively formal meanings unless clearly defined in the present disclosure.
The electronic device according to one embodiment may be one of various types of electronic devices. The electronic devices may include, for example, a portable communication device (e.g., a smart phone), a computer, a portable multimedia device, a portable medical device, a camera, a wearable device, or a home appliance. According to one embodiment of the disclosure, an electronic device is not limited to those described above.
The terms used in the present disclosure are not intended to limit the present disclosure but are intended to include various changes, equivalents, or replacements for a corresponding embodiment. With regard to the descriptions of the accompanying drawings, similar reference numerals may be used to refer to similar or related elements. A singular form of a noun corresponding to an item may include one or more of the things, unless the relevant context clearly indicates otherwise. As used herein, each of such phrases as “A or B,” “at least one of A and B,” “at least one of A or B,” “A, B, or C,” “at least one of A, B, and C,” and “at least one of A, B, or C,” may include all possible combinations of the items enumerated together in a corresponding one of the phrases. As used herein, terms such as, “1st,” “2nd,” “first,” and “second” may be used to distinguish a corresponding component from another component, but are not intended to limit the components in other aspects (e.g., importance or order). It is intended that if an element (e.g., a first element) is referred to, with or without the term “operatively” or “communicatively”, as “coupled with,” “coupled to,” “connected with,” or “connected to” another element (e.g., a second element), it indicates that the element may be coupled with the other element directly (e.g., wired), wirelessly, or via a third element.
As used herein, the term “module” may include a unit implemented in hardware, software, or firmware, and may interchangeably be used with other terms, such as, for example, “logic,” “logic block,” “part,” and “circuitry.” A module may be a single integral component, or a minimum unit or part thereof, adapted to perform one or more functions. For example, according to one embodiment, a module may be implemented in a form of an application-specific integrated circuit (ASIC).
According to embodiments of the present disclosure, a method is provided for calibrating an active depth sensing device and recovering depth maps from observed IR images. An accurate calibration allows for better modeling of the observations, which in turn, allows for sub-pixel accuracy. Specifically, embodiments of the present disclosure provide the capability of computing accurate and complete depth maps on demand utilizing an active sensing framework, where the reflection of projected light is observed from a nearby IR camera whose image is used to infer depth.
Such a process may have an accuracy of 0.2% to 0.8% in the 0.5 meter (m) to 2 m range, and is computationally efficient being under a computation budget of 1 Gigaflop (Gflop)/frame for the full/image version, and 10 Megaflops (Mflops)/frame for the lightweight/video version. Further, such a process is capable of producing depth maps on demand from any vantage point between the projector and the IR camera, and in particular, from the vantage point of the standard RGB camera.
In order to calibrate a system, known (and usually flat) surfaces are used to estimate intrinsic parameters that best define the inner workings of the system. A parametric forward model is described in detail below, followed by a description of how to estimate the parameters used therein.
Herein, 3-D coordinates are referred to as {right arrow over (X)}=(X, Y, Z), 2D image coordinates are referred to as {right arrow over (x)}=(x, y) (typically in video graphics array (VGA) range: [0 640]×[0 480]), and 2D normalized image coordinates are referred to as
or in some cases the homogenous coordinates
Triangle equalities are definitions.
Unless otherwise specified, all 3D coordinates have an origin at the optical center of the IR camera with a Z-axis coincident with the IR camera's optical axis, an X-axis pointing from the IR camera toward the projector (and presumed orthogonal to the Z-axis), and a Y-axis naturally completing the right-handed coordinate system. Any reference to distance means ∥{right arrow over (X)}∥ while depth denotes the Z component of {right arrow over (X)}.
2-D image coordinate vectors are implicitly converted to 3-D homogenous coordinates when necessary: {right arrow over (p)}=[px, py, 1]T. For example: {right arrow over (q)}=H {right arrow over (p)} are set forth in Equation (1) as:
for 3×3 homography H of the form set forth in Equation (2) as:
The homogenous normalized coordinates [xn, yn, 1]T denote the direction from the camera to 3D point {right arrow over (X)} and that {right arrow over (X)}=Z {right arrow over (x)}n.
The image of the speckle pattern projected onto a surface of infinite depth is referred to as the horizon image (setting aside intensity attenuation effects). There is no parallax (induced disparity) between the projector and IR camera, and the horizon image would appear the same even if viewed from the vantage point of the projector.
All norms are standard Euclidean norms unless described otherwise. The p-norm ball with center x and radius r is written as Bp({right arrow over (x)}, r) and refers to the infinity norm when the subscript is omitted. For example, when referring to integer-valued coordinates {right arrow over (x)}, Σ{right arrow over (x)}∈B({right arrow over (p)},1)( . . . ) refers to a sum over the 4 neighboring pixels surrounding real-valued {right arrow over (p)}. Upper case variables usually denote either matrices or world coordinates, while lower case are more commonly used for image coordinates, subscripts, functions, etc. To prevent confusion between true and estimated quantities, the hat symbol to denote the latter, e.g., Î.
Referring initially to
Referring now to
In
In order to describe how each point in space maps to an image plane, a classic camera model is used. For a 3D point of interest {right arrow over (X)}=(X, Y, Z) (e.g., the location of a speckle reflecting off a surface), the 2D pinhole projection coordinates are given as
The relation between {right arrow over (x)}n and the actual image coordinates {right arrow over (x)} are described below using this classic model. Taking into account the first few terms of lens distortion provides Equation (3) below:
where r2=xn2+yn2, κ1, κ2 are radial distortion coefficients, and K3, K4 are the tangential distortion coefficients. The pixel coordinates are then given as Equation (4) below:
where f is the focal length of the camera in pixels, and (cx, cy) is the principal point. This calibration model can be generalized to a single polynomial in (xn, yn) as set forth in Equation (5) below:
for some vector-valued coefficients {right arrow over (m)}ij.
When much of the lens distortion has been removed in the IR image, {right arrow over (x)}d={right arrow over (x)}n is used and it is assumed that the principal point is the image center {right arrow over (c)}==(cx, cy)=(320, 240), and the primary focus is on estimating the focal length accurately. Accordingly, the following simple calibration model is used as shown in Equation (6), where the focal length is the unknown in the model:
{right arrow over (x)}=K{right arrow over (x)}nf{right arrow over (x)}n+{right arrow over (c)} (6)
With respect to the projector distortion block 202 of
If the locations of these speckles were observed on the horizon image, they would no longer be evenly spaced due to a distortion introduced by the projector. This distortion between the coordinates of the speckles in this pattern image {right arrow over (b)} and the horizon image {right arrow over (p)} is modeled as a polynomial in (bx, by), in a manner similar to how the IR camera distortion was modeled above, as set forth in Equation (7) below:
By the pinhole model, a baseline translation b in a vantage point along the x direction induces a corresponding shift in the apparent location of the object in normalized coordinates set forth in Equation (8) below:
The corresponding effect on pixel coordinates is a shift of magnitude fb/Z as shown in Equation (9) below:
The disparity map
is defined with a focal-length baseline product: C0fb, and with the canonical basis vector e1=(1,0). As set forth below, e1 notation is dropped and {right arrow over (x)}+d({right arrow over (x)}) is used as shorthand for (x+d({right arrow over (x)}), y) or more generally {right arrow over (x)}+c(x+c, y) for any c∈. When the induced displacement is measured in pixels between the observed dot pattern and the horizon dot pattern as the disparity map d({right arrow over (x)}), the depth map can be recovered as
With respect to the application of PSF in the PSF block 208 of
for speckle locations {right arrow over (p)}k and point spread function h({right arrow over (x)}). This PSF h({right arrow over (x)}) is presumed to have local support, reaching its maximum at {right arrow over (x)}={right arrow over (0)} and rapidly decaying for ∥{right arrow over (x)}∥>2 pixels.
With respect to the application of the intensity attenuation in the intensity attenuation block 210 of
where D({right arrow over (x)}) denotes the distance to the speckle point in 3D space, measuring the full distance rather than merely the depth. Typically, a is some negative power close to −2.0 (as in the classic inverse-square power law).
Taking the composition of all functions described above yields the combined forward model as set forth in Equation (12) below:
Given the description of the forward model, an estimation of specific functions that govern the behavior of the model are set forth in detail below.
The focal length is proportional to the amount of shift that a given depth induces between horizon coordinates (disparity zero) and image coordinates. For this reason, given a-priori depth information Z({right arrow over (x)}) and baseline b between the IR camera and the projector, a given focal length candidate f may be evaluated by attempting to shift the perspective of a pair of images I1({right arrow over (x)}), I2({right arrow over (x)}) to the horizon (that is, infinite depth/zero disparity) and observing the similarity via image correlation. The focal length estimate maximizes the similarity between the pair of images warped to the horizon using the focal length parameter, as set forth below in Equation (13):
where the images I1 and I2 are warped to the horizon with candidate f, as set forth below in Equation (14):
Using depth map information, a focal length estimate of f=573 pixels±1% is determined. However, this estimate is only as accurate as the depth maps, and any systemic biases therein will result in corresponding biases in the focal length estimate.
With the intrinsic parameters of the IR camera being known as well as the underlying binary pattern described in
Another challenge is accurately estimating the coordinates of these speckle locations. Because of the aliasing and degradation inherent in the low-resolution VGA image, it is often difficult to determine where one speckle starts and the next one begins. Furthermore, as described above with respect to PSF, speckles that do not closely coincide with the integer lattice of the VGA sampling pattern often have very weak responses, making their locations difficult to reliably estimate.
To mitigate these challenges, a high-resolution version of the horizon image is constructed using a super-resolution approach. Using this high-resolution image, the challenges of tracking and correspondence become more manageable.
To this end, approximately 100 IR frames are captured with corresponding depth map information that can be used to estimate the high-resolution image. These images are captured with enough variation in depth to provide a diverse set of sub-pixel information in the x-direction due to parallax (though no additional information in the y direction). For this reason, the estimated horizon image has twice the resolution in the x direction but the same resolution in the y direction as the original VGA images.
Corresponding depth maps often have “holes”, or missing values where a depth was not estimated. These holes are filled by interpolating the nearby values via Cauchy interpolation.
The high-resolution horizon image estimate is represented as Equation (15) below:
where 1 denotes the VGA image of all ones, the warp operators Wn are defined as the bilinear warp that warps the high-resolution horizon image Ih to the nth image In, and WnT denotes the adjoint operation. For example, supposing the nth warp Wn would cause the pixel In(120,140) to be sampled from the horizon image at location (104.7,140) so that it is represented as 0.7 Ih(104,140)+0.3 Ih(105,140), then the linear adjoint operation WnT would cause the observed pixel In(120,140) to be distributed or pushed to Ih(104,140) and Ih(105,140) with weights 0.7 and 0.3, respectively.
This method of providing estimate of Ihorizon is sufficiently accurate to assist downstream processing for quantities that are of intrinsic interest. A horizon image can be generated directly using the PSF and horizon speckle locations.
In accordance with the projector distortion block 202 of
Finding the approximate location in the horizon image corresponding to each active speckle in the binary pattern image Ib, moves toward estimating the distortion introduced by the projector and thereby a comprehensive list of the precise speckle locations in the horizon image.
The brightest center pixel in the horizon image is manually matched with the center pixel in the pattern image. Proceeding radially outward from there, the speckle locations in the small neighborhood bordering known speckles are iteratively predicted, a 3×3 neighborhood is searched to find the actual speckle locations (to the nearest integer pixel), and these tracked locations are used to update and improve predictions for speckles slightly further away.
This prediction is performed via a coarse warp grid, as set forth in Equation (16) below:
=W({right arrow over (b)}k) (16)
where W is a smooth warp function 48× coarser than the binary pattern that predicts speckle locations in the horizon image via bilinear interpolation of the 4 neighbors in the warp W. After tracking the new locations {right arrow over (p)}k by maximizing the horizon image within a patch of radius 1 surrounding W({right arrow over (b)}k) (3×3 patch), the warp is updated with the newly tracked points by choosing the smoothest warp that best matches the observations, as set forth in Equation (17) below:
where Dx and Dy are the difference operators in the x and y direction (a convolutional kernel of [1 −1]), respectively.
The regularization of derivatives and mixed derivatives ensure smoothness. The first derivatives are allowed to be large as long as local changes in derivatives are not too sudden in any direction. In this way, the warp in each iteration can be extrapolated as the region of confidence grows. This procedure is depicted in
Given this non-parametric warping function W:2→2, the closest polynomial approximation to f0 can be found as shown below in Equation (18):
The warp grid W may be used as data constraints to use robust least squares to estimate the generating coefficients {right arrow over (o)}ij. 5th order polynomial gives sufficiently strong modeling power.
With respect to the PSF block 208 of
{right arrow over (q)}k={right arrow over (p)}k+dProj({right arrow over (p)}kk) (19)
where Dx and Dy are the difference operators in the x and y direction respectively, where the minimization is taken over the sampled h∈33
The forward model may be used to solve the inverse problem: recovering the depth map from the IR observation of an unknown scene.
The forward model is utilized to construct a horizon image Ih to match against. Then the disparity map is estimated, as set forth in Equation (21) below:
where └I┘{right arrow over (x)} is a patch drawn from image I centered at coordinate {right arrow over (x)} (e.g., a 17×17 patch at that location) and where
To evaluate D potential disparities at V locations (e.g., VGA) using a patch size of P results in O (DVP) flops, which can easily surpass a Gflop. There are ways to reduce the computational complexity of this approach by using integral images to compute the D correlations, eliminating the P factor to yield O(DV).
Unless the true disparity is close to some integer, it is possible that none of the candidate └Ih┘{right arrow over (x)}−{right arrow over (d)} shifts match very well due to the localized nature of the point spread function. This problem can be alleviated by sampling {right arrow over (d)} with more granularity (e.g., quarter-pixel shifts). However, this approach may exacerbate the computational bottleneck.
As described in detail below, an approach is set forth to provide an initial estimate of the disparity map at low-resolution among integer disparities, but with some built-in robustness against non-integer-shift mismatch to mitigate the above-described problem.
This approach first defines, for each sampled reference location {right arrow over (x)} in the horizon image, a subspace {right arrow over (x)} that well-approximates all of the potential patches that might be observed in the IR image (in any shift along that row) that are similar to └Ih┘{right arrow over (x)}. In order to estimate the disparity at this location, for every candidate integer disparity d, the distance between └I┘{right arrow over (x)}+{right arrow over (d)} and its closest approximation in the subspace T is minimized, as set forth below in Equation (22):
One way to design such a subspace is to define a random variable R{right arrow over (x)} that describes the natural space of patches similar to └Ih┘{right arrow over (x)} (i.e., corresponding to some modeled space of “real scenes”) and then minimize its approximation error with respect to some projection matrix P{right arrow over (x)}, as set forth below in Equation (23):
where Q{right arrow over (x)} is the correlation matrix of random variable R{right arrow over (x)} and the optimization is taken over all projection matrices P{right arrow over (x)}. This is the standard principal components analysis (PCA) solution. It remains to define R{right arrow over (x)}. One natural design is to take R{right arrow over (x)}=└Ih┘{right arrow over (x)}+{right arrow over (∈)} where g is some zero-mean random shift with variance typically less than a pixel2 in both x and y directions. An 8-dimensional subspace often gives sufficient descriptive power to this random variable, and helps to improve robustness to modeling/calibration inaccuracy in this initial tracking step. The final procedure is then set forth in Equation (24) below:
There are several aspects that demand immediate improvement on this initial estimate. It is low-resolution, integer-valued, contains occasional bad disparity estimates (i.e., far from their true values), and is taken from the vantage point of the projector rather than IR camera.
The problem of “bad” disparity values is addressed by penalizing an assigned disparity value by the L1 distance to its neighbors as set forth below in Equation (25):
Next, low-resolution dProj({right arrow over (x)}) is converted to VGA-resolution d({right arrow over (x)}) while simultaneously changing the vantage point from the projector to the IR camera.
This coarse estimate is refined as set forth below in Equation (26):
d({right arrow over (x)})=arg maxd└I┘{right arrow over (x)}, └Ih┘{right arrow over (x)}−{right arrow over (d)}N (26)
Although model-based tracking offers some potential advantages, it comes with risks as well, and the resulting disparity map estimation will only be as good as the estimated underlying model.
One alternative is to remain agnostic to the precise speckle locations and use the observed image from a flat surface reasonably far away (e.g., 2 meters) as a proxy for the horizon image Ih and then proceed with the NCC-based estimation as in Equation (21). The advantages of this approach are that it is likely to be quite accurate in regions near the reference depth and also gets the horizon point location+point spread function rendering correct due to being an observation itself. A downside is that such a reference image Ih is primarily accurate for approximately-integer disparities, and may suffer from modeling accuracy in between the grid points. Another downside is that this reference image remains blind to what lies beyond the boundaries of the observed VGA image, and may not be useful in generating depth map information near the left and right edges of observed images whose depths differ from the reference depth.
Referring to
The processor 620 may execute, for example, software (e.g., a program 640) to control at least one other component (e.g., a hardware or a software component) of the electronic device 601 coupled with the processor 620, and may perform various data processing or computations. As at least part of the data processing or computations, the processor 620 may load a command or data received from another component (e.g., the sensor module 676 or the communication module 690) in volatile memory 632, process the command or the data stored in the volatile memory 632, and store resulting data in non-volatile memory 634. The processor 620 may include a main processor 621 (e.g., a central processing unit (CPU) or an application processor (AP)), and an auxiliary processor 623 (e.g., a graphics processing unit (GPU), an image signal processor (ISP), a sensor hub processor, or a communication processor (CP)) that is operable independently from, or in conjunction with, the main processor 621. Additionally or alternatively, the auxiliary processor 623 may be adapted to consume less power than the main processor 621, or execute a particular function. The auxiliary processor 623 may be implemented as being separate from, or a part of, the main processor 621.
The auxiliary processor 623 may control at least some of the functions or states related to at least one component (e.g., the display device 660, the sensor module 676, or the communication module 690) among the components of the electronic device 601, instead of the main processor 621 while the main processor 621 is in an inactive (e.g., sleep) state, or together with the main processor 621 while the main processor 621 is in an active state (e.g., executing an application). According to one embodiment, the auxiliary processor 623 (e.g., an image signal processor or a communication processor) may be implemented as part of another component (e.g., the camera module 680 or the communication module 690) functionally related to the auxiliary processor 623.
The memory 630 may store various data used by at least one component (e.g., the processor 620 or the sensor module 676) of the electronic device 601. The various data may include, for example, software (e.g., the program 640) and input data or output data for a command related thereto. The memory 630 may include the volatile memory 632 or the non-volatile memory 634.
The program 640 may be stored in the memory 630 as software, and may include, for example, an operating system (OS) 642, middleware 644, or an application 646.
The input device 650 may receive a command or data to be used by other component (e.g., the processor 620) of the electronic device 601, from the outside (e.g., a user) of the electronic device 601. The input device 650 may include, for example, a microphone, a mouse, or a keyboard.
The sound output device 655 may output sound signals to the outside of the electronic device 601. The sound output device 655 may include, for example, a speaker or a receiver. The speaker may be used for general purposes, such as playing multimedia or recording, and the receiver may be used for receiving an incoming call. According to one embodiment, the receiver may be implemented as being separate from, or a part of, the speaker.
The display device 660 may visually provide information to the outside (e.g., a user) of the electronic device 601. The display device 660 may include, for example, a display, a hologram device, or a projector and control circuitry to control a corresponding one of the display, hologram device, and projector. According to one embodiment, the display device 660 may include touch circuitry adapted to detect a touch, or sensor circuitry (e.g., a pressure sensor) adapted to measure the intensity of force incurred by the touch.
The audio module 670 may convert a sound into an electrical signal and vice versa. According to one embodiment, the audio module 670 may obtain the sound via the input device 650, or output the sound via the sound output device 555 or a headphone of an external electronic device 602 directly (e.g., wired) or wirelessly coupled with the electronic device 601.
The sensor module 676 may detect an operational state (e.g., power or temperature) of the electronic device 601 or an environmental state (e.g., a state of a user) external to the electronic device 601, and then generate an electrical signal or data value corresponding to the detected state. The sensor module 676 may include, for example, a gesture sensor, a gyro sensor, an atmospheric pressure sensor, a magnetic sensor, an acceleration sensor, a grip sensor, a proximity sensor, a color sensor, an infrared (IR) sensor, a biometric sensor, a temperature sensor, a humidity sensor, or an illuminance sensor.
The interface 677 may support one or more specified protocols to be used for the electronic device 601 to be coupled with the external electronic device 602 directly (e.g., wired) or wirelessly. According to one embodiment, the interface 677 may include, for example, a high definition multimedia interface (HDMI), a universal serial bus (USB) interface, a secure digital (SD) card interface, or an audio interface.
A connecting terminal 678 may include a connector via which the electronic device 601 may be physically connected with the external electronic device 602. According to one embodiment, the connecting terminal 578 may include, for example, an HDMI connector, a USB connector, an SD card connector, or an audio connector (e.g., a headphone connector).
The haptic module 679 may convert an electrical signal into a mechanical stimulus (e.g., a vibration or a movement) or an electrical stimulus which may be recognized by a user via tactile sensation or kinesthetic sensation. According to one embodiment, the haptic module 679 may include, for example, a motor, a piezoelectric element, or an electrical stimulator.
The camera module 680 may capture a still image or moving images. According to one embodiment, the camera module 680 may include one or more lenses, image sensors, image signal processors, or flashes.
The power management module 688 may manage power supplied to the electronic device 601. The power management module 688 may be implemented as at least part of, for example, a power management integrated circuit (PMIC).
The battery 689 may supply power to at least one component of the electronic device 601. According to one embodiment, the battery 689 may include, for example, a primary cell which is not rechargeable, a secondary cell which is rechargeable, or a fuel cell.
The communication module 690 may support establishing a direct (e.g., wired) communication channel or a wireless communication channel between the electronic device 601 and the external electronic device (e.g., the electronic device 602, the electronic device 604, or the server 608) and performing communication via the established communication channel. The communication module 690 may include one or more communication processors that are operable independently from the processor 620 (e.g., the AP) and supports a direct (e.g., wired) communication or a wireless communication. According to one embodiment, the communication module 690 may include a wireless communication module 692 (e.g., a cellular communication module, a short-range wireless communication module, or a global navigation satellite system (GNSS) communication module) or a wired communication module 694 (e.g., a local area network (LAN) communication module or a power line communication (PLC) module). A corresponding one of these communication modules may communicate with the external electronic device via the first network 698 (e.g., a short-range communication network, such as Bluetooth™, wireless-fidelity (Wi-Fi) direct, or a standard of the Infrared Data Association (IrDA)) or the second network 699 (e.g., a long-range communication network, such as a cellular network, the Internet, or a computer network (e.g., LAN or wide area network (WAN)). These various types of communication modules may be implemented as a single component (e.g., a single IC), or may be implemented as multiple components (e.g., multiple ICs) that are separate from each other. The wireless communication module 692 may identify and authenticate the electronic device 601 in a communication network, such as the first network 698 or the second network 699, using subscriber information (e.g., international mobile subscriber identity (IMSI)) stored in the subscriber identification module 596.
The antenna module 697 may transmit or receive a signal or power to or from the outside (e.g., the external electronic device) of the electronic device 501. According to one embodiment, the antenna module 697 may include one or more antennas, and, therefrom, at least one antenna appropriate for a communication scheme used in the communication network, such as the first network 698 or the second network 699, may be selected, for example, by the communication module 690 (e.g., the wireless communication module 692). The signal or the power may then be transmitted or received between the communication module 690 and the external electronic device via the selected at least one antenna.
At least some of the above-described components may be mutually coupled and communicate signals (e.g., commands or data) therebetween via an inter-peripheral communication scheme (e.g., a bus, a general purpose input and output (GPIO), a serial peripheral interface (SPI), or a mobile industry processor interface (MIPI)).
According to one embodiment, commands or data may be transmitted or received between the electronic device 601 and the external electronic device 604 via the server 608 coupled with the second network 699. Each of the electronic devices 602 and 604 may be a device of a same type as, or a different type, from the electronic device 601. All or some of operations to be executed at the electronic device 601 may be executed at one or more of the external electronic devices 602, 604, or 608. For example, if the electronic device 601 should perform a function or a service automatically, or in response to a request from a user or another device, the electronic device 601, instead of, or in addition to, executing the function or the service, may request the one or more external electronic devices to perform at least part of the function or the service. The one or more external electronic devices receiving the request may perform the at least part of the function or the service requested, or an additional function or an additional service related to the request, and transfer an outcome of the performing to the electronic device 601. The electronic device 601 may provide the outcome, with or without further processing of the outcome, as at least part of a reply to the request. To that end, a cloud computing, distributed computing, or client-server computing technology may be used, for example.
One embodiment may be implemented as software (e.g., the program 640) including one or more instructions that are stored in a storage medium (e.g., internal memory 636 or external memory 638) that is readable by a machine (e.g., the electronic device 601). For example, a processor of the electronic device 601 may invoke at least one of the one or more instructions stored in the storage medium, and execute it, with or without using one or more other components under the control of the processor. Thus, a machine may be operated to perform at least one function according to the at least one instruction invoked. The one or more instructions may include code generated by a complier or code executable by an interpreter. A machine-readable storage medium may be provided in the form of a non-transitory storage medium. The term “non-transitory” indicates that the storage medium is a tangible device, and does not include a signal (e.g., an electromagnetic wave), but this term does not differentiate between where data is semi-permanently stored in the storage medium and where the data is temporarily stored in the storage medium.
According to one embodiment, a method of the disclosure may be included and provided in a computer program product. The computer program product may be traded as a product between a seller and a buyer. The computer program product may be distributed in the form of a machine-readable storage medium (e.g., a compact disc read only memory (CD-ROM)), or be distributed (e.g., downloaded or uploaded) online via an application store (e.g., Play Store™), or between two user devices (e.g., smart phones) directly. If distributed online, at least part of the computer program product may be temporarily generated or at least temporarily stored in the machine-readable storage medium, such as memory of the manufacturer's server, a server of the application store, or a relay server.
According to one embodiment, each component (e.g., a module or a program) of the above-described components may include a single entity or multiple entities. One or more of the above-described components may be omitted, or one or more other components may be added. Alternatively or additionally, a plurality of components (e.g., modules or programs) may be integrated into a single component. In this case, the integrated component may still perform one or more functions of each of the plurality of components in the same or similar manner as they are performed by a corresponding one of the plurality of components before the integration. Operations performed by the module, the program, or another component may be carried out sequentially, in parallel, repeatedly, or heuristically, or one or more of the operations may be executed in a different order or omitted, or one or more other operations may be added.
Although certain embodiments of the present disclosure have been described in the detailed description of the present disclosure, the present disclosure may be modified in various forms without departing from the scope of the present disclosure. Thus, the scope of the present disclosure shall not be determined merely based on the described embodiments, but rather determined based on the accompanying claims and equivalents thereto.
This application is based on and claims priority under 35 U.S.C. § 119(e) to a U.S. Provisional Patent Application filed on Oct. 19, 2018 in the United States Patent and Trademark Office and assigned Ser. No. 62/748,023, the contents of which are incorporated herein by reference.
Number | Name | Date | Kind |
---|---|---|---|
6285959 | Greer | Sep 2001 | B1 |
6310650 | Johnson et al. | Oct 2001 | B1 |
8681224 | Mallon et al. | Mar 2014 | B2 |
9066084 | Zalevsky et al. | Jun 2015 | B2 |
9329263 | Haynes et al. | May 2016 | B2 |
9794545 | Bronstein et al. | Oct 2017 | B2 |
9813692 | Zabatani et al. | Nov 2017 | B2 |
9824461 | Bronstein et al. | Nov 2017 | B2 |
20150304617 | Chang | Oct 2015 | A1 |
20160245641 | Kowdle et al. | Aug 2016 | A1 |
20170035268 | Kumar | Feb 2017 | A1 |
20170256069 | Link | Sep 2017 | A1 |
20170278268 | Russell | Sep 2017 | A1 |
20190033448 | Molnar et al. | Jan 2019 | A1 |
Number | Date | Country |
---|---|---|
2 701 387 | Feb 2014 | EP |
WO 2018205036 | Nov 2018 | WO |
Entry |
---|
H. Yamazoe et al., “Easy depth sensor calibration”, 2012, 21st International Conference on Pattern Recognition, p. 465-468. (Year: 2012). |
A. R. Orghidan et al., “Projector-camera calibration for 3D reconstruction using vanishing points”, 2012, 2012 International Conference on 3D Imaging, p. 1-6. (Year: 2012). |
Raposo, Carolina et al., Fast and Accurate Calibration of a Kinect Sensor, 2013 International Conference on 3D Vision—3DV 2013, Seattle, WA, 2013, pp. 342-349. |
Michael J. Landau et al., “Simulating Kinect Infrared and Depth Images”, IEEE Transactions on Cybernetics, vol. 46, No. 12, Dec. 2016, 14 pages. |
Number | Date | Country | |
---|---|---|---|
20200126246 A1 | Apr 2020 | US |
Number | Date | Country | |
---|---|---|---|
62748023 | Oct 2018 | US |