The subject matter disclosed herein relates to generating an image of a scene. In particular, the subject matter disclosed herein relates to methods, systems, and computer-readable storage media for creating three-dimensional images of a scene.
Stereoscopic, or three-dimensional, imagery is based on the principle of human vision. Two separate detectors detect the same object or objects in a scene from slightly different positions and/or angles and project them onto two planes. The resulting images are transferred to a processor which combines them and gives the perception of the third dimension, i.e. depth, to a scene.
Many techniques of viewing stereoscopic images have been developed and include the use of colored or polarizing filters to separate the two images, temporal selection by successive transmission of images using a shutter arrangement, or physical separation of the images in the viewer and projecting them separately to each eye. In addition, display devices have been developed recently that are well-suited for displaying stereoscopic images. For example, such display devices include digital still cameras, personal computers, digital picture frames, set-top boxes, high-definition televisions (HDTVs), and the like.
The use of digital image capture devices, such as digital still cameras, digital camcorders (or video cameras), and phones with built-in cameras, for use in capturing digital images has become widespread and popular. Because images captured using these devices are in a digital format, the images can be easily distributed and edited. For example, the digital images can be easily distributed over networks, such as the Internet. In addition, the digital images can be edited by use of suitable software on the image capture device or a personal computer.
Digital images captured using conventional image capture devices are two-dimensional. It is desirable to provide methods and systems for using conventional devices for creating three-dimensional images.
Methods, systems, and computer program products for creating three-dimensional images of a scene are disclosed herein. According to one aspect, a method includes receiving a plurality of images of a scene. The method also includes determining attributes of the plurality of images. Further, the method includes determining, based on the attributes of all captured images, a pair of images from among the plurality of images for use in creating a three-dimensional image.
According to another aspect, a user can, by use of the subject matter disclosed herein, use an image capture device for capturing a plurality of different images of the same scene and for creating a three-dimensional, or stereoscopic, image of the scene. The subject matter disclosed herein includes a process for creating three-dimensional images. The generation process can include identification of suitable pairs of images, registration, rectification, color correction, transformation, depth adjustment, and motion detection and removal. The functions of the subject matter disclosed herein can be implemented in hardware and/or software that can be executed on an image capture device or a suitable display device. For example, the functions can be implemented using a digital still camera, a personal computer, a digital picture frame, a set-top box, an HDTV, and the like.
According to another aspect, a system for creating a three-dimensional image of a scene is disclosed. The system may include a memory having stored therein computer-executable instructions. The system may also include a computer processor that executes the computer-executable instructions. Further, the system may include an image generator configured to receive a plurality of images of a scene. The image generator may also be configured to determine attributes of the plurality of images. The image generator may also be configured to determine, based on the attributes, a pair of images from among the plurality of images for use in generating a three-dimensional image.
According to another aspect, the pair of images comprises different perspective views of the scene.
According to another aspect, the image generator may be configured to compare color attributes of the images; determine images in which the color attributes are within a predetermined color threshold; and modify the images within the predetermined color threshold for creating the three-dimensional image.
According to another aspect, the image generator is configured to: identify similar objects in the plurality of images; and compare images including the similar objects for occlusion.
According to another aspect, the image generator is configured to: identify similar objects in the plurality of images; and compare vertical displacement of the similar objects with respect to a predetermined threshold level.
According to another aspect, the image generator is configured to compare color differences between images.
According to another aspect, the image generator is configured to apply horizontal and vertical edge line analysis to images.
According to another aspect, the image generator is configured to: apply a Hough transform to identify lines within the images; and determine perspective view changes between the images based on the identified lines.
According to another aspect, the three-dimensional image includes a stereoscopic pair representing a left view image and a right view image.
According to another aspect, the image generator is configured to automatically determine the left view image and the right view image for creating the stereoscopic pair.
According to another aspect, the image generator is configured to: receive the left view image and the right view image from different sources; and automatically create the three-dimensional image using the left view image and the right view image.
According to another aspect, the image generator is configured to determine interest points within the left view image and the right view image for one or more of rectification and registration.
According to another aspect, the image generator is configured to: perform horizontal and vertical edge detection; filter for strong edges of a minimum length; and identify crossing points of the strong edges.
According to another aspect, the image generator is configured to: determine a parallax disparity between the left view image and the right view image; determine whether the parallax disparity meets a predetermined criteria; and, if the parallax disparity does not meet the predetermined criteria, adjust an attribute of at least one pixel in one of the left view image and the right view image such that the parallax disparity meets the predetermined criteria.
According to another aspect, the image generator is configured to crop one of the left view image and the right view image.
According to another aspect, the image generator is configured to: determine a parallax disparity between at least a portion of an object in the left view image and the right view image; determine whether the parallax disparity is greater than a predetermined threshold level; and, if the parallax disparity is greater than the predetermined threshold level, remove the object from the left view image and the right view image.
According to another aspect, the system further includes an image capture device for capturing the plurality of images of the scene.
According to another aspect, the image capture device is one of a digital still camera, a video camera, a mobile phone, and a smart phone for capturing the plurality of images of the scene.
According to another aspect, the image generator is configured to generate a three-dimensional image of the scene using the pair of images.
According to another aspect, the image generator is configured to implement: one or more of registration, rectification, color correction, matching edges of the pair of images, transformation, depth adjustment, motion detection, and removal of moving objects.
According to another aspect, the image generator is configured to display the three-dimensional image on a suitable three-dimensional image display.
According to another aspect, the image generator is configured to display the three-dimensional image on a device such as a digital still camera, a computer, a video camera, a digital picture frame, a set-top box, phone, or a high-definition television.
According to an aspect, a system for generating a three-dimensional video of a scene is disclosed. The system may include a memory having stored therein computer-executable instructions. The system may also include a computer processor that executes the computer-executable instructions. Further, the system may include an image generator configured to: receive a video sequence comprising a plurality of still images of a scene; determine attributes of the plurality of still images; and determine, based on the attributes, a plurality of stereoscopic image pairs from among the plurality of still images for use in generating a three-dimensional video.
According to another aspect, the computer processor and memory are configured to adjust at least one of a parallax and perspective of each image pair such that at least one of a predetermined orientation and depth results.
According to another aspect, the computer processor and memory are configured to sequence stereoscopic image pairs in the three-dimensional video according to an order of the video sequence of the still images.
According to another aspect, the computer processor and memory are configured to compress the image pairs of the three-dimensional video.
The foregoing summary, as well as the following detailed description of various embodiments, is better understood when read in conjunction with the appended drawings. For the purposes of illustration, there is shown in the drawings exemplary embodiments; however, the invention is not limited to the specific methods and instrumentalities disclosed. In the drawings:
The subject matter of the present invention is described with specificity to meet statutory requirements. However, the description itself is not intended to limit the scope of this patent. Rather, the inventors have contemplated that the claimed subject matter might also be embodied in other ways, to include different steps or elements similar to the ones described in this document, in conjunction with other present or future technologies. Moreover, although the term “step” may be used herein to connote different aspects of methods employed, the term should not be interpreted as implying any particular order among or between various steps herein disclosed unless and except when the order of individual steps is explicitly described.
Methods, systems, and computer program products for creating one or more three-dimensional images of a scene are disclosed herein. The three-dimensional images can be viewed or displayed on a stereoscopic display. The three-dimensional images may also be viewed or displayed on any other display capable of presenting three-dimensional images to a person using other suitable equipment, such as, but not limited to, three-dimensional glasses. In addition, the functions and methods described herein may be implemented on a device capable of capturing still images, displaying three-dimensional images, and executing computer executable instructions on a processor. The device may be, for example, a digital still camera, a video camera (or camcorder), a personal computer, a digital picture frame, a set-top box, an HDTV, a phone, or the like. Such devices may be capable of presenting three-dimensional images to a person without additional equipment, or if used in combination with other suitable equipment such as three-dimensional glasses. The functions of the device may include methods for rectifying and registering at least two images, matching the color and edges of the images, identifying moving objects, removing or adding moving objects from or to the images to equalize them, altering a perceived depth of objects, and any final display-specific transformation to create a single, high-quality three-dimensional image. The techniques described herein may be applied to still-captured images and video images, which can be thought of as a series of images; hence for the purpose of generalization the majority of the description herein is limited to still-captured image processing.
It should be noted that any of the processes and steps described herein may be implemented in an automated fashion. For example, any of the methods and techniques described herein may be automatically implemented without user input after the capture of a plurality of images.
Referring to
The memory 104 and the CPU 106 may be operable together to implement an image generator function 114 for creating three-dimensional images in accordance with embodiments of the present invention. The image generator function 114 may generate a three-dimensional image of a scene using two or more images of the scene captured by the device 100.
The method of
The method of
The generated two or more images may also be suitably processed 206. For example, the images may be corrected and adjusted for display as described herein.
The method of
Although the above examples are described for use with a device capable of capturing images, embodiments of the present invention described herein are not so limited. Particularly, the methods described herein for creating a three-dimensional image of a scene may for example be implemented in any suitable system including a memory and computer processor. The memory may have stored therein computer-executable instructions. The computer processor may execute the computer-executable instructions. The memory and computer processor may be configured for implementing methods in accordance with embodiments of the present invention described herein.
Images suitable for use as a three-dimensional image may be captured by a user using any suitable technique. For example,
In another example,
The distance between positions at which images are captured (the stereo baseline) for creating a three-dimensional image can affect the quality of the three-dimensional image. The optimal stereo baseline between the camera positions can vary anywhere between 3 centimeters (cm) and several feet, dependent upon a variety of factors, including the distance of the closest objects in frame, the lens focal length or other optics properties of the camera, the camera crop factor (dependent on sensor size), the size and resolution of the display on which the images will be viewed, and the distance from the display at which viewers will view the images. A general recommendation is that the stereo baseline should not exceed the distance defined by the following equation:
where B is the stereo baseline separation in inches, D is the distance in feet to the nearest object in frame, F is the focal length of the lens in millimeters (mm), and C is the camera crop factor relative to a full frame (36×24 square mm.) digital sensor (which approximates the capture of a 35 mm analog camera). In the examples provided herein, it is assumed that at least two images have been captured, at least two of which can be interpreted as a stereoscopic pair.
The identification of stereo pairs in 302 is bypassed in the cases where the user has manually selected the image pair for 3D image registration. This bypass can also be triggered if a 3D-enabled capture device is used that identifies the paired images prior to the registration process. Returning to
A preliminary, quick analysis may be utilized for determining whether images among the plurality of captured images are similar enough to warrant a more detailed analysis. This analysis may be performed by, for example, the image generator function 114 shown in
The method of
The method of
Image pair is not stereoscopic=ABS(AVm−AVm+1)>ThresholdAV
OR
For all k, ABS(SAVk,m−SAVk,m−1)>ThresholdSAV
OR
ABS(MAXm−MAXm+1)>ThresholdMAX
OR
ABS(MINm−MINm+1)>ThresholdMIN
ThresholdAV, ThresholdSAV, ThresholdMAX, and ThresholdMIN are threshold value levels for the average, segmented average, maximum and minimum, respectively. These equations can be applied to all or at least some of the colors.
The method of
Referring again to
Using the results of the motion estimation process used for object similarity evaluation, vertical displacement can be assessed. Vertical motion vector components are indicative of vertical parallax between the images, which when large can indicate a poor image pair. Vertical parallax must be corrected via rectification and registration to allow for comfortable viewing, and this correction will reduce the size of the overlapping region of the image in proportion to the original amount of vertical parallax.
Using the motion vectors from the similarity of objects check, color data may be compared to search for large changes between images. Such large changes can represent a color difference between the images regardless of similar luminance.
The method of
Referring to
For each vertical edge in one image, determine the closest edge in the other image, subject to meeting criteria for length, slope and curvature. For distance, use the distance between the primary points. If this distance is larger than ε, it is deemed that no edge matches, and this edge contributes ε to the cost function. The end result of the optimization is the determination of δ, the optimal shift between the two images based on this vertical edge matching. In box 622, the same optimization process from box 620 is repeated; this time, however, is for horizontal edge matching, and utilizes the vertical δ already determined from box 620.
In an example for block 622, the following equation may be used:
Block 624 then uses the calculated horizontal and vertical δ's to match each edge with its closest edge that meets the length, slope and curvature criteria. In an example for block 624, the following equation may be used:
Ci,j={0 otherwise1 if P
The output of this stage is the matrix C, which has 1 in location i,j if edge i and j are matching edges and otherwise 0. This matrix is then pruned in Box 626 so that no edge is matched with multiple other edges. In the event of multiple matches, the edge match with minimal distance is used. Finally, in Box 628, the edge matches are broken down into regions of the image. The set of matching edges within each region are then characterized by the mean shift, and this mean shift is then the characteristic shift of the region. By examining the direction of the shifts of each subregion, it is thus possible to determine which picture is left and which is right. It is also possible to determine whether the second captured picture was captured with a focal axis parallel to the first picture. If not, there is some amount of toe-in or toe-out which can be characterized by the directional shifts of the subregions.
Referring to
A Hough transform can be applied 306 to identify lines in the two images of the potential stereoscopic pair. Lines that are non-horizontal, non-vertical, and hence indicate some perspective in the image can be compared between the two images to search for perspective changes between the two views that may indicate a perspective change or excessive toe-in during capture of the pair.
The aforementioned criteria may be applied to scaled versions of the original images for reducing computational requirements. The results of each measurement may be gathered, weighted, and combined to make a final decision regarding the probable quality of a given image pair as a stereoscopic image pair.
The method of
At step 804, color segmentation is performed on the objects. At step 806, the bounding box of 8×8 blocks for each object in each image may be identified. At step 810, images may be partitioned into N×N blocks. At step 812, blocks with high information content may be selected. At step 813, the method includes performing motion estimation on blocks in L relative to R image (accumulate motion vectors for L/R determination. These steps may be considered Techniques 1, 2, and 3.
At step 814, edge detection may be performed on left/right images. Next, at step 816, vertical and horizontal lines in left/right images may be identified and may be classified by length, location, and slope. At step 818, a Hough transform may be performed on the left/right images. Next, at step 820, the method includes analyzing Hough line slope for left/right images and identifying non-vertical and non-horizontal lines.
Referring to
At step 822, the following calculations may be performed for all objects or blocks of interest and lines:
At step 824, a weighted average of the above measures may be performed to determine whether images are a pair or not. Next, at step 826, average motion vector direction may be used to determine left/right images.
Referring again to
For a stereo pair of left and right view images, the method of
For a stereo pair of left and right view images with a set of identified interest points, rectification 318 may be performed on the stereo pair of images. Using the interest point set for the left view image, motion estimation techniques (as described in stereo pair identification above) and edge matching techniques are applied to find the corresponding points in the right view image.
and the fundamental matrix equation
rightptsT*F*leftpts=0
is solved or approximated to determine the 3×3 fundamental matrix, F, and epipoles, e1 and e2. The camera epipoles are used with the interest point set to generate a pair of rectifying homographies. It can be assumed that the camera properties are consistent between the two captured images. The respective homographies are then applied to the right and left images, creating the rectified images. The overlapping rectangular region of the two rectified images is then identified, the images are cropped to this rectangle, and the images are resized to their original dimensions, creating the rectified image pair, right_r and left_r. The rectified image pair can be defined by the following equations:
right—r=cropped(F*right)
left—r=cropped(F*left)
For the stereo pair of “left_r” and “right_r” images, registration is next performed on the stereo pair. A set of interest points is required, and the interest point set selected for rectification (or a subset thereof) may be translated to positions relative to the output of the rectification process by applying the homography of the rectification step to the points. Optionally, a second set of interest points may be identified for the left_r image, and motion estimation and edge matching techniques may be applied to find the corresponding points in the right_r image. The interest point selection process for the registration operation is the same as that for rectification. Again, the N corresponding interest points are made into a 3×N set of point values as set forth in the following equations:
and the following matrix equation
left—rpts=Tr*right—rpts
is approximated for a 3×3 linear conformal transformation, Tr, which may incorporate both translation on the X and Y axes and rotation in the X/Y plane. The transform Tr is applied to the right_r image to create the image “Right′” as defined by the following equation:
Right′=Tr*right—r,
where right_r is organized as a 3×N set of points (xir, yir, 1) for i=1 to image_rows*image cols.
Finally, the second set of interest points for the left_r image may be used to find correspondence in the Right′ image, the set of points as set forth in the following equations:
is identified and composed, and the equation
Right′pts=Tl*left—rpts
is approximated for a second linear conformal transformation, Tl. The transform Tl is applied to the left_r image to create the image “Left′”, as defined by the following equation:
Left′=Tl*left—r
“Right′” and “Left′” images represent a rectified, registered stereoscopic pair.
The method of
The method of
After steps 1010 and 1014 of
Returning now to
For a stereoscopic pair of registered “Left′” and “Right′” images, the screen plane of the stereoscopic image can be altered 336, or relocated, to account for disparities measured as greater than a viewer can resolve. This is performed by scaling the translational portion of transforms that created the registered image views by a percent offset and re-applying the transforms to the original images. For example, if the initial left image transform is as follows:
for scaling factor S, X/Y rotation angle θ, and translational offsets Tx and Ty, the adjustment transform becomes
where Xscale and Yscale are determined by the desired pixel adjustment relative to the initial transform adjustment, i.e.,
Only in rare occurrences will Yscale be other than zero, and only then as a corrective measure for any noted vertical parallax. Using the altered transform, a new registered image view is created, e.g. the following:
Left′=Tlalt*left—r
Such scaling effectively adds to or subtracts from the parallax for each pixel, effectively moving the point of now parallax forward or backward in the scene. The appropriate scaling is determined by the translational portion of the transform and the required adjustment.
At step 338 of
Since moving an object region in the image may result in a final image that has undefined pixel values, a pixel-fill process is required to ensure that all areas of the resultant image have defined pixel values after object movement. An exemplary procedure for this is described below. Other processes, both more or less complex, may be applied.
Rul=(xl,yu); the upper left coordinate
Rll=(xl,yl); the lower left coordinate
Rur=(xr,yu); the upper right coordinate
Rlr=(xr,yl); the lower right coordinate
For a large or complex object, multiple rectangular regions may need to be defined and moved, but the process executes identically for each region.
In an example of defining left/right bounds of a region M for left/right motion, the region M is the region to which the altered transform can be applied. This process first assesses the direction of movement to occur and defines one side of region M. If the intended movement is to the right, then the right bounding edge of region M is defined by the following coordinate pair in the appropriate left_r or right_r image (whichever is to be adjusted):
Mur=(xr+P,yu); upper right
Mlr=(xr+P,yl); lower right
If movement is to the left, the left bounding edge of region M is defined as:
Mul=(xl−P,yu); upper left
Mll=(xl−P,yl); lower left
P is an extra number of pixels for blending purposes. The scaled version of the registration transform matrix Talt is provided 1104. The inverse of the altered transform (assumed already calculated as above for movement of the screen plane for the whole image) may then be applied 1106 to the opposite edge of the region R to get the other edge of region M. For the sake of example, assume that the movement of R is intended to be to the right, and that the left image is to be altered (meaning Tlalt has been created for the intended movement). Since the right side of M is already known, the other side can now be determined as:
Mul=Tlalt−1*Rul+(P,0); upper right
Mll=Tlalt−1*Ru+(P,0); lower right
Again, P is an extra number of pixels for blending, and Tlalt−1 is the inverse transform of Tlalt. Note that P is added after the transform application, and only to the X coordinates. The region to be moved is now defined as the pixels within the rectangle defined by M.
The method also includes applying 1108 the inverse transform of Tlalt to the image to be transformed for blocks in the region M. For example, from this point, one of two operations can be used, depending on a measurement of the uniformity (texture) or the area defined by the coordinates Mul, Mll, Rul, and Rll (remembering again that the region would be using other coordinates for a movement to the left). Uniformity is measured by performing a histogram analysis on the RGB values for the pixels in this area. If the pixel variation is within a threshold, the area is deemed uniform, and the movement of the region is affected by applying the following equation: Left′=Tlalt*left_r, for left_r εM. This is the process shown in the example method of
Left′=Tlalt*left—r, for the left—r region defined by Rul, Rll, Mur, and Mlr.
The method of
The method of
d=Rul(x)−Mul(x)
for the x-coordinates of Rul and Mul, and then proceeds to determine an interpolated gradient between the two pixel positions to fill in the missing values. For simplicity of implementation, the interpolation is always performed on a power of two, meaning that the interpolation will produce one of 1, 2, 4, 8, 16, etc. pixels as needed between the two defined pixels. Pixel regions that are not a power of two are mapped to the closest power of two, and either pixel repetition or truncation of the sequence is applied to fit. As an example, if Rul(x)=13 and Mul(x)=6, then d=7, and the following intermediate pixel gradient is calculated for a given row, j, in the region:
Since only 7 values are needed, p8 would go unused in this case, such that the following assignments would be made:
(x6,yj)=p1
(x7,yj)=p2
(x8,yj)=p3
(x9,yj)=p4
(x10,yj)=p5
(x11,yj)=p6
(x12,yj)=p7.
This process can repeat for each row in the empty region.
A weighted averaging the outer P “extra” pixels on each side of the rectangle with the pixel data currently in those positions is performed to blend the edges.
As an alternative to the procedure of applying movement and pixel blending to alter the parallax of an object, the disparity map calculated using the two views, “Left′” and “Right′,” can be altered for the region M to reduce the disparity values in that region, and then applied to one of the “Left′” or “Right′” single image views to create a new view (e.g., “Left_disparity”). The result of this process is a new stereo pair (e.g., “Left′” and “Left_disparity”) that recreates the depth of the original pair, but with lesser parallax for the objects within the region M. Once created in this manner, the “disparity” view becomes the new opposite image to the original, or for example, a created “Left_disparity” image becomes the new “Right′” image.
Returning to
The method of
For objects where motion is indicated and where the motion of an object is below the acceptable disparity threshold, identify the most suitable image to copy the object from, copy the object to the left and right target images and adjust the disparities as shown in the attached figure. The more frames that are captured, the less estimation is needed to determine the rightmost pixel of the right view. Most of occluded pixels can be extracted from the leftmost images. For an object that is moving in and out of the scene between the first and last picture, identify the object and completely remove it from the first picture if there is enough data in the captured sequence of images to fill in the missing pixels.
For objects where motion is indicated and where the motion is above the acceptable disparity, identify the most suitable picture from which to extract the target object and extrapolate the proper disparity information from the remaining captured pictures.
The actual object removal process involves identifying N×N blocks, with N empirically determined, to make up a bounding region for the region of “infinite” parallax, plus an additional P pixels (for blending purposes), determining the corresponding position of those blocks in the other images using the parallax values of the surrounding P pixels that have a similar gradient value (meaning that high gradient areas are extrapolated from similar edge areas and low gradient areas are extrapolated from similar surrounding flat areas), copying the blocks/pixels from the opposite locations to the intended new location, and performing a weighted averaging of the outer P “extra” pixels with the pixel data currently in those positions to blend the edges. If it is determined to remove an object, fill-in data is generated 346. Otherwise, the method proceeds to step 348.
The movement of the object 1304 is such that the disparity is unacceptable and should be corrected. In this example, the image obtained from position 1300 can be utilized for creating a three-dimensional image, and the image obtained from position 1302 can be altered for use together with the other image in creating the three-dimensional image. To correct, the object 1304 may be moved to the left (as indicated by direction arrow 1312 in
Another example of a process for adding/removing objects from a single image is illustrated in
Referring to
Referring to
As an alternative to the procedure of identifying bounding regions of 8×8 blocks around objects to be added or removed in a view, the disparity map calculated using multiple views, “Left”, “Right”, and/or the images in between, can be applied to one of the “Left” or “Right” single image views to create a new view (e.g., “Left_disparity”). The result of this process is a new stereo pair (e.g., “Left′” and “Left_disparity”) that effectively recreates the depth of the original pair, but without object occlusions, movement, additions, or removals. Once created in this manner, the “disparity” view becomes the new opposite image to the original, or for example, a created “Left_disparity” image becomes the new “Right′” image. Effectively, this procedure mimics segmented object removal and/or addition, but on a full image scale.
Returning to
For a finalized, color corrected, motion corrected stereoscopic image pair, the “Left′” and “Right′” images are ordered and rendered to a display as a stereoscopic image. The format is based on the display parameters. Rendering can require interlacing, anamorphic compression, pixel alternating, and the like.
For a finalized, color corrected, motion corrected stereoscopic image pair, the “Left′” view may be compressed as the base image and the “Right′” image may be compressed as the disparity difference from the “Left′” using a standard video codec, differential JPEG, or the like.
The method of
When a video sequence is captured with lateral camera motion as described above, stereoscopic pairs can be found within the sequence of resulting images. Stereoscopic pairs are identified based on their distance from one another determined by motion analysis (e.g., motion estimation techniques). Each pair represents a three-dimensional picture or image, which can be viewed on a suitable stereoscopic display. If the camera does not have a stereoscopic display, the video sequence can be analyzed and processed on any suitable display device. If the video sequence is suitable for creating three-dimensional content (e.g., one or more three-dimensional images), it is likely that there are many potential stereoscopic pairs, as an image captured at a given position may form a pair with images captured at several other positions. The image pairs can be used to create three-dimensional still images or re-sequenced to create a three-dimensional video.
When creating three-dimensional still images, the user can select which images to use from the potential pairs, thereby adjusting both the perspective and parallax of the resulting images to achieve the desired orientation and depth.
Another method of creating a three-dimensional sequence includes creating stereoscopic pairs by grouping the first and last images in the sequence, followed by the second and next-to-last images, and so on until all images have been used. During playback this creates the effect of the camera remaining still while the depth of the scene decreases over time due to decreasing parallax. The three-dimensional images can also be sequenced in the opposite order so that the depth of the scene increases over time.
The generation and presentation, such as display, of three-dimensional images of a scene in accordance with embodiments of the present invention may be implemented by a single device or combination of devices. In one or more embodiments of the present invention, images may be captured by a camera such as, but not limited to, a digital camera. The camera may be connected to a personal computer for communication of the captured images to the personal computer. The personal computer may then generate one or more three-dimensional images in accordance with embodiments of the present invention. After generation of the three-dimensional images, the personal computer may communicate the three-dimensional images to the camera for display on a suitable three-dimensional display. The camera may include a suitable three-dimensional display. Also, the camera may be in suitable electronic communication with a high-definition television for display of the three-dimensional images on the television. The communication of the three-dimensional images may be, for example, via an HDMI connection.
In one or more other embodiments of the present invention, three-dimensional images may be generated by a camera and displayed by a separate suitable display. For example, the camera may capture conventional two-dimensional images and then use the captured images to generate three-dimensional images. The camera may be in suitable electronic communication with a high-definition television for display of the three-dimensional images on the television. The communication of the three-dimensional images may be, for example, via an HDMI connection.
The subject matter disclosed herein may be implemented by a digital still camera, a video camera, a mobile phone, a smart phone, phone, and the like. In order to provide additional context for various aspects of the disclosed subject matter,
Generally, however, program modules include routines, programs, objects, components, data structures, etc. that perform particular tasks or implement particular data types. The operating environment 1800 is only one example of a suitable operating environment and is not intended to suggest any limitation as to the scope of use or functionality of the subject matter disclosed herein. Other well known computer systems, environments, and/or configurations that may be suitable for use with the invention include but are not limited to, personal computers, hand-held or laptop devices, multiprocessor systems, microprocessor-based systems, programmable consumer electronics, network PCs, minicomputers, mainframe computers, distributed computing environments that include the above systems or devices, and the like.
With reference to
The system bus 1908 can be any of several types of bus structure(s) including the memory bus or memory controller, a peripheral bus or external bus, and/or a local bus using any variety of available bus architectures including, but not limited to, 11-bit bus, Industrial Standard Architecture (ISA), Micro-Channel Architecture (MCA), Extended ISA (EISA), Intelligent Drive Electronics (IDE), VESA Local Bus (VLB), Peripheral Component Interconnect (PCI), Universal Serial Bus (USB), Advanced Graphics Port (AGP), Personal Computer Memory Card International Association bus (PCMCIA), and Small Computer Systems Interface (SCSI).
The system memory 1906 includes volatile memory 1910 and nonvolatile memory 1912. The basic input/output system (BIOS), containing the basic routines to transfer information between elements within the computer 1902, such as during start-up, is stored in nonvolatile memory 1912. By way of illustration, and not limitation, nonvolatile memory 1912 can include read only memory (ROM), programmable ROM (PROM), electrically programmable ROM (EPROM), electrically erasable ROM (EEPROM), or flash memory. Volatile memory 1910 includes random access memory (RAM), which acts as external cache memory. By way of illustration and not limitation, RAM is available in many forms such as synchronous RAM (SRAM), dynamic RAM (DRAM), synchronous DRAM (SDRAM), double data rate SDRAM (DDR SDRAM), enhanced SDRAM (ESDRAM), Synchlink DRAM (SLDRAM), and direct Rambus RAM (DRRAM).
Computer 1902 also includes removable/nonremovable, volatile/nonvolatile computer storage media.
It is to be appreciated that
A user enters commands or information into the computer 1902 through input device(s) 1926. Input devices 1926 include, but are not limited to, a pointing device such as a mouse, trackball, stylus, touch pad, keyboard, microphone, joystick, game pad, satellite dish, scanner, TV tuner card, digital camera, digital video camera, web camera, and the like. These and other input devices connect to the processing unit 1904 through the system bus 1908 via interface port(s) 1928. Interface port(s) 1928 include, for example, a serial port, a parallel port, a game port, and a universal serial bus (USB). Output device(s) 1930 use some of the same type of ports as input device(s) 1926. Thus, for example, a USB port may be used to provide input to computer 1902 and to output information from computer 1902 to an output device 1930. Output adapter 1932 is provided to illustrate that there are some output devices 1930 like monitors, speakers, and printers among other output devices 1930 that require special adapters. The output adapters 1932 include, by way of illustration and not limitation, video and sound cards that provide a means of connection between the output device 1930 and the system bus 1908. It should be noted that other devices and/or systems of devices provide both input and output capabilities such as remote computer(s) 1934.
Computer 1902 can operate in a networked environment using logical connections to one or more remote computers, such as remote computer(s) 1934. The remote computer(s) 1934 can be a personal computer, a server, a router, a network PC, a workstation, a microprocessor based appliance, a peer device or other common network node and the like, and typically includes many or all of the elements described relative to computer 1902. For purposes of brevity, only a memory storage device 1936 is illustrated with remote computer(s) 1934. Remote computer(s) 1934 is logically connected to computer 1902 through a network interface 1938 and then physically connected via communication connection 1940. Network interface 1938 encompasses communication networks such as local-area networks (LAN) and wide-area networks (WAN). LAN technologies include Fiber Distributed Data Interface (FDDI), Copper Distributed Data Interface (CDDI), Ethernet/IEEE 1102.3, Token Ring/IEEE 1102.5 and the like. WAN technologies include, but are not limited to, point-to-point links, circuit switching networks like Integrated Services Digital Networks (ISDN) and variations thereon, packet switching networks, and Digital Subscriber Lines (DSL).
Communication connection(s) 1940 refers to the hardware/software employed to connect the network interface 1938 to the bus 1908. While communication connection 1940 is shown for illustrative clarity inside computer 1902, it can also be external to computer 1902. The hardware/software necessary for connection to the network interface 1938 includes, for exemplary purposes only, internal and external technologies such as, modems including regular telephone grade modems, cable modems and DSL modems, ISDN adapters, and Ethernet cards.
The various techniques described herein may be implemented with hardware or software or, where appropriate, with a combination of both. Thus, the methods and apparatus of the disclosed embodiments, or certain aspects or portions thereof, may take the form of program code (i.e., instructions) embodied in tangible media, such as floppy diskettes, CD-ROMs, hard drives, or any other machine-readable storage medium, wherein, when the program code is loaded into and executed by a machine, such as a computer, the machine becomes an apparatus for practicing the invention. In the case of program code execution on programmable computers, the computer will generally include a processor, a storage medium readable by the processor (including volatile and non-volatile memory and/or storage elements), at least one input device and at least one output device. One or more programs are preferably implemented in a high level procedural or object oriented programming language to communicate with a computer system. However, the program(s) can be implemented in assembly or machine language, if desired. In any case, the language may be a compiled or interpreted language, and combined with hardware implementations.
The described methods and apparatus may also be embodied in the form of program code that is transmitted over some transmission medium, such as over electrical wiring or cabling, through fiber optics, or via any other form of transmission, wherein, when the program code is received and loaded into and executed by a machine, such as an EPROM, a gate array, a programmable logic device (PLD), a client computer, a video recorder or the like, the machine becomes an apparatus for practicing the invention. When implemented on a general-purpose processor, the program code combines with the processor to provide a unique apparatus that operates to perform the processing of the present invention.
While the embodiments have been described in connection with the preferred embodiments of the various figures, it is to be understood that other similar embodiments may be used or modifications and additions may be made to the described embodiment for performing the same function without deviating therefrom. Therefore, the disclosed embodiments should not be limited to any single embodiment, but rather should be construed in breadth and scope in accordance with the appended claims.
This application claims the benefit of U.S. provisional patent application No. 61/230,131, filed Jul. 31, 2009, the disclosure of which is incorporated herein by reference in its entirety. The disclosures of the following U.S. provisional patent applications, commonly owned and simultaneously filed Jul. 31, 2009, are all incorporated by reference in their entirety: U.S. provisional patent application No. 61/230,133; and U.S. provisional patent application No. 61/230,138.
Number | Name | Date | Kind |
---|---|---|---|
3503316 | Takano | Mar 1970 | A |
3953869 | Wah Lo | Apr 1976 | A |
4661986 | Adelson | Apr 1987 | A |
4956705 | Wright | Sep 1990 | A |
4980762 | Heeger | Dec 1990 | A |
5043806 | Choquet | Aug 1991 | A |
5151609 | Nakagawa | Sep 1992 | A |
5305092 | Mimura | Apr 1994 | A |
5369735 | Thier | Nov 1994 | A |
5444479 | Femekes | Aug 1995 | A |
5511153 | Azabayejani | Apr 1996 | A |
5603687 | Hori | Feb 1997 | A |
5613048 | Chen | Mar 1997 | A |
5652616 | Chen | Jul 1997 | A |
5673081 | Yamashita | Sep 1997 | A |
5678089 | Bacs, Jr. | Oct 1997 | A |
5682437 | Okino | Oct 1997 | A |
5719954 | Onda | Feb 1998 | A |
5734743 | Matsugu et al. | Mar 1998 | A |
5748199 | Palm | May 1998 | A |
5777666 | Tanase | Jul 1998 | A |
5808664 | Yamashita | Sep 1998 | A |
5874988 | Gu | Feb 1999 | A |
5883695 | Paul | Mar 1999 | A |
5953054 | Mercier | Sep 1999 | A |
5963247 | Banitt | Oct 1999 | A |
5991551 | Bacs, Jr. | Nov 1999 | A |
6018349 | Szeliski | Jan 2000 | A |
6023588 | Ray | Feb 2000 | A |
6031538 | Chupeau | Feb 2000 | A |
6047078 | Kang | Apr 2000 | A |
6064759 | Buckley | May 2000 | A |
6075905 | Herman | Jun 2000 | A |
6094215 | Sundahl | Jul 2000 | A |
6215516 | Ma | Apr 2001 | B1 |
6215899 | Morimura et al. | Apr 2001 | B1 |
6240198 | Rehg | May 2001 | B1 |
6246412 | Shum | Jun 2001 | B1 |
6269172 | Rehg | Jul 2001 | B1 |
6269175 | Hanna et al. | Jul 2001 | B1 |
6278460 | Myers | Aug 2001 | B1 |
6314211 | Kim | Nov 2001 | B1 |
6324347 | Bacs, Jr. | Nov 2001 | B1 |
6381302 | Berestov | Apr 2002 | B1 |
6384859 | Matsumoto | May 2002 | B1 |
6385334 | Saneyoshi | May 2002 | B1 |
6414709 | Palm | Jul 2002 | B1 |
6434278 | Hashimoto | Aug 2002 | B1 |
6445833 | Murata | Sep 2002 | B1 |
6496598 | Harman | Dec 2002 | B1 |
6512892 | Montgomery | Jan 2003 | B1 |
6556704 | Chen | Apr 2003 | B1 |
6559846 | Uyttendaele | May 2003 | B1 |
6611268 | Szeliski | Aug 2003 | B1 |
6661913 | Zhang | Dec 2003 | B1 |
6677981 | Mancuso | Jan 2004 | B1 |
6677982 | Chen | Jan 2004 | B1 |
6686926 | Kaye | Feb 2004 | B1 |
6747610 | Taima | Jun 2004 | B1 |
6750904 | Lambert | Jun 2004 | B1 |
6760488 | Moura | Jul 2004 | B1 |
6798406 | Jones | Sep 2004 | B1 |
6862364 | Berestov | Mar 2005 | B1 |
6927769 | Roche, Jr. | Aug 2005 | B2 |
6947059 | Pierce | Sep 2005 | B2 |
6967659 | Jayavant | Nov 2005 | B1 |
6970591 | Lyons | Nov 2005 | B1 |
6978051 | Edwards | Dec 2005 | B2 |
7027642 | Rubbert | Apr 2006 | B2 |
7046840 | Chang et al. | May 2006 | B2 |
7054478 | Harman | May 2006 | B2 |
7068275 | Nakamura | Jun 2006 | B2 |
7081892 | Alkouh | Jul 2006 | B2 |
7103213 | Hirvonen | Sep 2006 | B2 |
7108657 | Irion | Sep 2006 | B2 |
7116323 | Kaye | Oct 2006 | B2 |
7116324 | Kaye | Oct 2006 | B2 |
7126598 | Oh | Oct 2006 | B2 |
7164790 | Zhang | Jan 2007 | B2 |
7180536 | Wolowelsky | Feb 2007 | B2 |
7181061 | Kawano | Feb 2007 | B2 |
7196730 | Mihelcic | Mar 2007 | B2 |
7197192 | Edwards | Mar 2007 | B2 |
7203356 | Gokturk | Apr 2007 | B2 |
7215809 | Sato | May 2007 | B2 |
7218757 | Franz | May 2007 | B2 |
7224357 | Chen | May 2007 | B2 |
7224382 | Baker | May 2007 | B2 |
7245768 | Harman | Jul 2007 | B1 |
7260243 | Shibayama | Aug 2007 | B2 |
7321374 | Naske | Jan 2008 | B2 |
7349006 | Sato | Mar 2008 | B2 |
7373017 | Edwards | May 2008 | B2 |
7397481 | Endo | Jul 2008 | B2 |
7400782 | Zhou | Jul 2008 | B2 |
7409105 | Jiu | Aug 2008 | B2 |
7466336 | Regan | Dec 2008 | B2 |
7483590 | Nielsen | Jan 2009 | B2 |
7489812 | Fox | Feb 2009 | B2 |
7508977 | Lyons | Mar 2009 | B2 |
7512883 | Wallick | Mar 2009 | B2 |
7515759 | Sun | Apr 2009 | B2 |
7538876 | Hewitt | May 2009 | B2 |
7551770 | Harman | Jun 2009 | B2 |
7557824 | Holliman | Jul 2009 | B2 |
7573475 | Sullivan | Aug 2009 | B2 |
7573489 | Davidson | Aug 2009 | B2 |
7580463 | Routhier | Aug 2009 | B2 |
7605776 | Satoh | Oct 2009 | B2 |
7616886 | Matsumura | Nov 2009 | B2 |
7619656 | Ben-Ezra | Nov 2009 | B2 |
7639838 | Nims | Dec 2009 | B2 |
7643062 | Silverstein | Jan 2010 | B2 |
7680323 | Nichani | Mar 2010 | B1 |
7693221 | Routhier | Apr 2010 | B2 |
7701506 | Silverbrook | Apr 2010 | B2 |
7705970 | Piestun | Apr 2010 | B2 |
7711181 | Kee | May 2010 | B2 |
7711201 | Wong | May 2010 | B2 |
7711221 | Burgi | May 2010 | B2 |
7768702 | Hirose | Aug 2010 | B2 |
7817187 | Silsby | Oct 2010 | B2 |
7844001 | Routhier | Nov 2010 | B2 |
7857455 | Cowen | Dec 2010 | B2 |
7873207 | Tsubaki | Jan 2011 | B2 |
7876948 | Wetzel | Jan 2011 | B2 |
20020106120 | Brandenburg | Aug 2002 | A1 |
20020158873 | Wiliamson | Oct 2002 | A1 |
20020190991 | Efran | Dec 2002 | A1 |
20020191841 | Harman | Dec 2002 | A1 |
20030002870 | Baron | Jan 2003 | A1 |
20030030636 | Yamaoka | Feb 2003 | A1 |
20030151659 | Kawano | Aug 2003 | A1 |
20030152264 | Perkins | Aug 2003 | A1 |
20040100565 | Chen | May 2004 | A1 |
20040136571 | Hewitson | Jul 2004 | A1 |
20040218269 | Divelbiss et al. | Nov 2004 | A1 |
20050041123 | Ansari | Feb 2005 | A1 |
20060008268 | Suwa | Jan 2006 | A1 |
20060203335 | Martin | Sep 2006 | A1 |
20060210111 | Cleveland | Sep 2006 | A1 |
20060221072 | Se et al. | Oct 2006 | A1 |
20060221179 | Seo et al. | Oct 2006 | A1 |
20060222260 | Sambongi | Oct 2006 | A1 |
20070024614 | Tam | Feb 2007 | A1 |
20070064098 | Tran | Mar 2007 | A1 |
20070146232 | Redert | Jun 2007 | A1 |
20070165129 | Hill et al. | Jul 2007 | A1 |
20070165942 | Jin | Jul 2007 | A1 |
20070168820 | Kutz | Jul 2007 | A1 |
20070189599 | Ryu et al. | Aug 2007 | A1 |
20070189747 | Ujisato | Aug 2007 | A1 |
20070291143 | Barbieri | Dec 2007 | A1 |
20080024614 | Li et al. | Jan 2008 | A1 |
20080031327 | Wang | Feb 2008 | A1 |
20080056609 | Rouge | Mar 2008 | A1 |
20080062254 | Edwards | Mar 2008 | A1 |
20080080852 | Chen | Apr 2008 | A1 |
20080117280 | Schowengerdt | May 2008 | A1 |
20080117289 | Schowegerdt | May 2008 | A1 |
20080150945 | Wang | Jun 2008 | A1 |
20080158345 | Schklair | Jul 2008 | A1 |
20080180550 | Gulliksson | Jul 2008 | A1 |
20080240607 | Sun | Oct 2008 | A1 |
20080252725 | Lanfermann | Oct 2008 | A1 |
20080317379 | Steinberg | Dec 2008 | A1 |
20090061381 | Durbin | Mar 2009 | A1 |
20090080036 | Paterson | Mar 2009 | A1 |
20090116732 | Zhou et al. | May 2009 | A1 |
20090141967 | Hiroshi | Jun 2009 | A1 |
20090154793 | Shin | Jun 2009 | A1 |
20090154823 | Ben-Ezra | Jun 2009 | A1 |
20090290013 | Hayashi | Nov 2009 | A1 |
20090290037 | Pore | Nov 2009 | A1 |
20090295907 | Kim | Dec 2009 | A1 |
20100039502 | Higgins | Feb 2010 | A1 |
20100097444 | Lablans | Apr 2010 | A1 |
20100128109 | Banks | May 2010 | A1 |
20100134598 | St. Pierre | Jun 2010 | A1 |
20100142824 | Lu | Jun 2010 | A1 |
20100165152 | Lim | Jul 2010 | A1 |
20100182406 | Benitez | Jul 2010 | A1 |
20100201682 | Quan et al. | Aug 2010 | A1 |
20100208942 | Porter | Aug 2010 | A1 |
20100220932 | Dong-Qing et al. | Sep 2010 | A1 |
20100238327 | Griffith | Sep 2010 | A1 |
20100239158 | Rouge | Sep 2010 | A1 |
20100309286 | Chen | Dec 2010 | A1 |
20100309288 | Stettner | Dec 2010 | A1 |
20110018975 | Chen | Jan 2011 | A1 |
20110025825 | McNamer et al. | Feb 2011 | A1 |
20110025829 | McNamer | Feb 2011 | A1 |
20110025830 | McNamer et al. | Feb 2011 | A1 |
20110050853 | Zhang | Mar 2011 | A1 |
20110050859 | Kimmel | Mar 2011 | A1 |
20110050864 | Bond | Mar 2011 | A1 |
20110169921 | Lee et al. | Jul 2011 | A1 |
20110255775 | McNamer et al. | Oct 2011 | A1 |
20120105602 | McNamer et al. | May 2012 | A1 |
20120314036 | Dahi et al. | Dec 2012 | A1 |
Number | Date | Country |
---|---|---|
1056049 | Nov 2000 | EP |
1056049 | Nov 2000 | EP |
1240540 | Feb 2011 | EP |
1240540 | Feb 2011 | EP |
2405764 | Sep 2005 | GB |
2455316 | Jun 2008 | GB |
2455316 | Jun 2009 | GB |
10-0653965 | Dec 2006 | KR |
2005025239 | Mar 2005 | WO |
WO2005025239 | Mar 2005 | WO |
2006062325 | Jun 2006 | WO |
WO2006062325 | Jun 2006 | WO |
2008016882 | Feb 2008 | WO |
2008075276 | Jun 2008 | WO |
WO2008075276 | Jun 2008 | WO |
2009122200 | Oct 2009 | WO |
WO2009122200 | Oct 2009 | WO |
2010024479 | Mar 2010 | WO |
WO2010024479 | Mar 2010 | WO |
2010052741 | May 2010 | WO |
WO2010052741 | May 2010 | WO |
2010147609 | Dec 2010 | WO |
WO2010147609 | Dec 2010 | WO |
2011014419 | Feb 2011 | WO |
2011014420 | Feb 2011 | WO |
2011014421 | Feb 2011 | WO |
2011017308 | Feb 2011 | WO |
WO2011014419 | Feb 2011 | WO |
WO2011014420 | Feb 2011 | WO |
WO2011014421 | Feb 2011 | WO |
WO2011017308 | Feb 2011 | WO |
WO2012016549 | May 2012 | WO |
Entry |
---|
Related application: Tassos Markas et al.; Methods, Systems, and Computer-Readable Storage Media for Identifying a Rough Depth Map in a Scene and for Determining a Stereo-Base Distance for Three-Dimensional (3D) Content Creation filed Dec. 27, 2011 as U.S. Appl. No. 13/337,676 (not yet published). |
Related application: Bahram Dahi et al.; Primary and Auxiliary Image Capture Devices for Image Processing and Related Methods filed May 25, 2011 as U.S. Appl. No. 13/115,589 and corresponding non-final Office Action mailed Oct. 11, 2011 and Examiner Interview dated Oct. 11, 2011 and Applicant's response filed Nov. 9, 2011; and corresponding final rejection mailed Dec. 2, 2011 and amendment after final mailed by Applicant on Jan. 2, 2012; corresponding Applicant Initiated Interview summary dated Jan. 10, 2012; corresponding Advisory Action mailed Jan. 12, 2012; and Applicant's Request for Continued Examination filed Jan. 20, 2012. |
Related application: Search Report and Written Opinion for PCT International Patent Application No. PCT/US10/43022 filed Jul. 23, 2010. |
Related application: Search Report and Written Opinion for PCT International Patent Application No. PCT/US10/43023 filed Jul. 23, 2010. |
Related application: Search Report and Written Opinion for PCT International Patent Application No. PCT/US10/43025 filed Jul. 23, 2010. |
Related application: Written Opinion/IPRP for PCT International Patent Application No. PCT/US11/59057 filed Nov. 3, 2011. |
Related application: Tassos Markas et al.; Methods, Systems, and Computer-Readable Storage Media for Identifying a Rough Depth Map in a Scene and for Determining a Stereo-Base Distance for Three-Dimensional (3D) Content Creation filed Dec. 27, 2011 as PCT/US11/67349 (not yet published). |
Related application: Bahram Dahi et al.; Primary and Auxiliary Image Capture Devices for Image Processing and Related Methods filed Dec. 9, 2011 as PCT/US11/64050 (not yet published). |
Chen, Shu-Ching et al., Video Scene Change Detection Method Using Unsupervised Segmentation and Object Tracking, IEEE International Conference on Multimedia and Expo (CME), pp. 57-60, 2001. |
Nöll, Tobias et al., Markerless Camera Pose Estimation—An Overview, Visualization of Large and Unstructured Data Sets—IRTG Workshop, pp. 45-54, 2010. |
Chen, Shenchang Eric et al., View Interpolation for Image Synthesis, Proceedings of ACM SIGGRAPH, pp. 279-288, 1993. |
McMillan, Jr., Leonard, An Image-Based Approach to Three-Dimensional Computer Graphics, PhD. Dissertation submitted to the University of North Carolina at Chapel Hill, Apr. 1997. |
Related U.S. Appl. No. 12/842,171, filed Jul. 23, 2010, and titled Methods, Systems, and Computer-Readable Storage Media for Selecting Image Capture Positions to Generate Three-Dimensional Images. |
Related U.S. Appl. No. 12/842,257, filed Jul. 23, 2010, and titled Methods, Systems, and Computer-Readable Storage Media for Generating Stereoscopic Content via Depth Map Creation. |
Search Report and Written Opinion for PCT international application No. PCT/US2010/043022. |
Search Results for PCT international application No. PCT/US2010/043023. |
Search Results for PCT international application No. PCT/US2010/043025. |
Michael McNamer et al; “Methods, Systems, and Computer-Readable Storage Media for Generating Three-Dimensional Images of a Scene,” filed Jul. 31, 2009. U.S. Appl. No. 61/230,131. |
Michael McNamer et al; “Methods, Systems, and Computer-Readable Storage Media for Generating Three-Dimensional Images of a Scene,” filed Jul. 23, 2010. U.S. Appl. No. 12/842,084. |
Michael McNamer et al; “Methods, Systems, and Computer-Readable Storage Media for Selecting Image Capture Positions to Generate Three-Dimensional Images,” filed Jul. 31, 2009. U.S. Appl. No. 61/230,133. |
Michael McNamer et al; “Methods, Systems, and Computer-Readable Storage Media for Selecting Image Capture Positions to Generate Three-Dimensional Images,” filed Jul. 23, 2010. Utility U.S. Appl. No. 12/842,171. |
Michael McNamer et al; “Methods, Systems, and Computer-Readable Storage Media for Generating Stereoscopic Content via a Depth Map Creation,” filed Jul. 31, 2009. U.S. Appl. No. 61/230,138. |
Michael McNamer et al; “Methods, Systems, and Computer-Readable Storage Media for Generating Stereoscopic Content via a Depth Map Creation,” filed Jul. 23, 2010. Utility U.S. Appl. No. 12/842,257. |
Michael McNamer et al, “Methods, Systems, and Computer-Readable Storage Media for Generating Three-Dimensional (3D) Images of a Scene,” filed May 26, 2011, Utility U.S. Appl. No. 13/115,459. |
Dahi et al, “Primary and Auxiliary Image Capture Devices for Image Processing and Related Methods,” filed May 26, 2011, Utility U.S. Appl. No. 13/115,589. |
ISR and Written Opinion mailed Jul. 18, 2012 for related application PCT/US11/64050 filed Dec. 9, 2011 and published as WO2012/091878 on Jul. 5, 2012. |
Int'l. Preliminary Report on Patentability mailed May 16, 2013, for related application No. PCT/US2011/059057 filed Nov. 3, 2011. |
Notice of Allowance mailed Mar. 26, 2013 for related U.S. Appl. No. 12/842,171. |
Related U.S. Appl. No. 13/865,127, filed Apr. 17, 2013; not yet public. |
Related PCT Patent Application Serial No. PCT/US13/37010 filed Apr. 17, 2013; not yet public. |
Related U.S. Appl. No. 13/865,312, filed Apr. 18, 2013; not yet public. |
Mahajan et al., “Moving Gradients: A Path-Based Method for Plausible Image Interpolation”, Proceedings of ACM SIGGRAPH 2009, vol. 28, Issue 3 (Aug. 2009). |
Notice of Allowance mailed Apr. 9, 2013 for related U.S. Appl. No. 13/584,744. |
Issue Notification mailed Apr. 24, 2013 for related U.S. Appl. No. 13/584,744. |
Related U.S. Appl. No. 13/888,737, filed May 7, 2013; not yet public. |
Joly, Phillippe et al., Efficient automatic analysis of camera work and microsegmentation of video using spatiotemporal images, 1996, Signal Process Image Communication, pp. 295-307. |
Krotkov, Eric et al., An Agile Stereo Camera System for Flexible Image Acquisition, 1988, IEEE, vol. 4, pp. 108-113. |
U.S. Non-Final Office Action dated Nov. 26, 2012 for Related U.S. Appl. No. 12/842,171, filed Jul. 23, 2010. |
Non-Final Office Action mailed Dec. 5, 2012 for related U.S. Appl. No. 12/842,257, filed Jul. 23, 2010. |
Related U.S. Appl. No. 13/662,443, filed Oct. 27, 2012; not yet public. |
Related U.S. Appl. No. 13/751,203, filed Jan. 28, 2013; not yet public. |
Applicant-Initiated Interview Summary mailed Feb. 8, 2013 for related U.S. Appl. No. 12/842,171. |
Related application: Bahram Dahi et al.; Primary and Auxiliary Image Capture Devices for Image Processing and Related Methods filed Aug. 13, 2012 as U.S. Appl. No. 13/584,744, not yet public. |
Int'l. Search Report and Written Opinion mailed Sep. 7, 2012 for related application PCT/US2011/067349 filed Dec. 27, 2011 and published as WO2012/092246 on Jul. 5, 2012. |
Number | Date | Country | |
---|---|---|---|
20110025825 A1 | Feb 2011 | US |
Number | Date | Country | |
---|---|---|---|
61230131 | Jul 2009 | US |