Systems and methods for feature tracking

Information

  • Patent Grant
  • 9208608
  • Patent Number
    9,208,608
  • Date Filed
    Monday, February 25, 2013
    11 years ago
  • Date Issued
    Tuesday, December 8, 2015
    9 years ago
Abstract
A computer-implemented method for processing, by a graphical processor unit (GPU), a plurality of images of a user. A plurality of features detected by the GPU in a first image of the plurality of images of the user is selected. Each selected feature includes one or more pixels. In a second image of the plurality of images of the user, a search is performed for the plurality of features selected in the first image. A variance is calculated, on the GPU, for each selected feature found in the second image. The variance indicates a degree to which a portion of the second image varies from a corresponding portion of the first image. The calculated variance is stored in a variance file.
Description
BACKGROUND

The use of computer systems and computer-related technologies continues to increase at a rapid pace. This increased use of computer systems has influenced the advances made to computer-related technologies. Indeed, computer systems have increasingly become an integral part of the business world and the activities of individual consumers. Computers have opened up an entire industry of internet shopping. In many ways, online shopping has changed the way consumers purchase products. For example, a consumer may want to know what they will look like in and/or with a product. On the webpage of a certain product, a photograph of a model with the particular product may be shown. However, users may want to see more accurate depictions of themselves in relation to various products.


SUMMARY

According to at least one embodiment, a computer-implemented method for processing, by a graphical processor unit (GPU), a plurality of images of a user. A plurality of features detected by the GPU in a first image of the plurality of images of the user may be selected. Each selected feature may include one or more pixels. In a second image of the plurality of images of the user, a search may be performed for the plurality of features selected in the first image. A variance may be calculated, on the GPU, for each selected feature found in the second image. The variance may indicate a degree to which a portion of the second image varies from a corresponding portion of the first image. The calculated variance may be stored in a variance file.


In one embodiment, one or more patches may be selected from among the selected features in the first image based on the calculated variance of each selected feature. Each patch may include a square area of pixels centered on one of the selected features of the first image. In some configurations, upon selecting the one or more patches in the first image, the first image may be removed from memory. In some embodiments, the one or more patches may be selected based on a predetermined threshold of calculated variance. In one embodiment, the one or more patches may be selected based on a predetermined number of patches. In some configurations, each variance may be divided into first and second elements. The first element may be stored in a first file and the second element may be stored in a second file.


In one embodiment, a cross-correlation algorithm may be performed on a GPU to determine how a first patch, selected among the one or more patches, is positioned in first and second sample images of the plurality of images of the user. In some embodiments, performing the cross-correlation algorithm on the GPU may include determining a pose of the user in the first and second sample images, performing a fast Fourier transform (FFT) on the first patch, and performing the FFT on the first and second sample images. The first sample image may be placed in the real element of a complex number and the second sample image may be placed in the imaginary element of the complex number. The FFT of the first patch may be stored in a third file. In some configurations, performing the cross-correlation algorithm on the GPU may include multiplying element-wise the FFT of the first patch by the FFT of the first and second sample images, calculating an inverse FFT of the multiplied FFTs, resulting in a first score for the first sample image and a second score for the second sample image, and normalizing the result of the cross-correlation by dividing both first and second scores by the calculated variance stored in the variance file.


In one embodiment, the cross-correlation of each selected patch may be performed on the GPU simultaneously. In some configurations, on the GPU, a second cross-correlation algorithm may be performed on third and fourth sample images of the plurality of images of the user using the FFT of the first patch stored in the third file to determine how the first patch is positioned in the third and fourth sample images. In some embodiments, a position of the selected feature of the first patch may be determined as a point in a virtual three-dimensional (3-D) space.


A computing device configured to process, by a graphical processor unit (GPU), a plurality of images of a user is also described. The device may include a processor and memory in electronic communication with the processor. The memory may store instructions that are executable by the GPU to select a plurality of features detected by the GPU in a first image of the plurality of images of the user. Each selected feature may include one or more pixels. In one embodiment, the instructions may be executable by the GPU to search, in a second image of the plurality of images of the user, for the plurality of features selected in the first image. In some configurations, the instructions may be executable by the GPU to calculate variance, on the GPU, for each selected feature found in the second image. The variance may indicate a degree to which a portion of the second image varies from a corresponding portion of the first image. In one embodiment, the instructions may be executable by the GPU to store the calculated variance in a variance file.


A computer-program product to process, by a graphical processor unit (GPU), a plurality of images of a user is also described. The computer-program product may include a non-transitory computer-readable medium that stores instructions. The instructions may be executable by the GPU to select a plurality of features detected by the GPU in a first image of the plurality of images of the user. Each selected feature may include one or more pixels. In one embodiment, the instructions may be executable by the GPU to search, in a second image of the plurality of images of the user, for the plurality of features selected in the first image. In some configurations, the instructions may be executable by the GPU to calculate variance, on the GPU, for each selected feature found in the second image. The variance may indicate a degree to which a portion of the second image varies from a corresponding portion of the first image. In one embodiment, the instructions may be executable by the GPU to store the calculated variance in a variance file.


According to at least one embodiment, a computer-implemented method for processing, by a graphical processor unit (GPU), a plurality of images of a user. A plurality of features detected by the GPU in a first image of the plurality of images of the user may be selected. Each selected feature may include one or more pixels. In a second image of the plurality of images of the user, a search may be performed for the plurality of features selected in the first image. A variance may be calculated, on the GPU, for each selected feature found in the second image. The variance may indicate a degree to which a portion of the second image varies from a corresponding portion of the first image. The calculated variance may be stored in a variance file.


Features from any of the above-mentioned embodiments may be used in combination with one another in accordance with the general principles described herein. These and other embodiments, features, and advantages will be more fully understood upon reading the following detailed description in conjunction with the accompanying drawings and claims.





BRIEF DESCRIPTION OF THE DRAWINGS

The accompanying drawings illustrate a number of exemplary embodiments and are a part of the specification. Together with the following description, these drawings demonstrate and explain various principles of the instant disclosure.



FIG. 1 is a block diagram illustrating one embodiment of an environment in which the present systems and methods may be implemented;



FIG. 2 is a block diagram illustrating another embodiment of an environment in which the present systems and methods may be implemented;



FIG. 3 is a block diagram illustrating one example of a graphical processor unit (GPU);



FIG. 4 is a block diagram illustrating one example of a feature detection module;



FIG. 5 is a block diagram illustrating one example of a cross correlation module;



FIG. 6 is a diagram illustrating an example of a device for capturing an image of a user;



FIG. 7 illustrates an example arrangement of features detected in the depicted images of a user;



FIG. 8 is a flow diagram illustrating one embodiment of a method for detecting features in images;



FIG. 9 is a flow diagram illustrating one embodiment of a method for performing cross-correlation algorithms on a GPU;



FIG. 10 is a flow diagram illustrating one embodiment of a method for performing a cross-correlation algorithm on two images simultaneously; and



FIG. 11 depicts a block diagram of a computer system suitable for implementing the present systems and methods.





While the embodiments described herein are susceptible to various modifications and alternative forms, specific embodiments have been shown by way of example in the drawings and will be described in detail herein. However, the exemplary embodiments described herein are not intended to be limited to the particular forms disclosed. Rather, the instant disclosure covers all modifications, equivalents, and alternatives falling within the scope of the appended claims.


DETAILED DESCRIPTION OF EXEMPLARY EMBODIMENTS

The systems and methods described herein relate to processing, by a graphical processor unit (GPU), a plurality of images of a user. Specifically, the systems and methods described herein relate to feature detection and normalized cross-correlation (i.e., template matching) of a set of images. Feature detection may include performing a “computer vision algorithm” to detect an “interesting” part of an image. Features may be used as a starting point in a computer vision algorithm. A desirable property of a feature detector may be repeatability, i.e., whether the same feature may be detected in one or more different images of the same scene. Typically, when working with feature detection and template matching of images, the image data may be copied from the GPU to the central processing unit (CPU), back to the GPU, and so forth. Copying image data back and forth between the GPU and CPU causes a bottleneck on the CPU, wasting valuable resources and computing cycles. On the CPU, feature detection of each image, as well as the cross-correlation of each interest point (i.e., corners, blobs, and/or points used in image analysis to detect distinguishable features of an image) on an image, may be performed one at a time, one after the other. However, because the results of one cross correlation do not depend on the results from another, feature detection and cross-correlation may be performed on the GPU simultaneously. Using a GPU instead of the CPU, the fast Fourier transform (FFT), element-wise multiplication, inverse FFT, and normalization may be performed simultaneously, all in parallel, for every interest point of a given image.


Additionally, in a typical implementation, the variance (degree of variation between pixels of two or more images) may not be calculated because of the computing costs involved calculating variance on the CPU. However, calculating variance on a GPU may be relatively fast compared to the CPU, and the results of the variance may be stored and used later in cross-correlation. However, rounding errors in the cross-correlation may be most affected when the variance of an interest point, or an extracted patch of the interest point, is low. Thus, in a GPU implementation, the most “trackable” features in an image may be those that have the highest variance.



FIG. 1 is a block diagram illustrating one embodiment of an environment 100 in which the present systems and methods may be implemented. In some embodiments, the systems and methods described herein may be performed on a single device (e.g., device 102). For example, a GPU 104 may be located on the device 102. Examples of devices 102 include mobile devices, smart phones, tablet computing devices, personal computing devices, computers, servers, etc.


In some configurations, a device 102 may include a GPU 104, a camera 106, and a display 108. In one example, the device 102 may be coupled to a database 110. In one embodiment, the database 110 may be internal to the device 102. In another embodiment, the database 110 may be external to the device 102. In some configurations, the database 110 may include variance data 112 and FFT patch data 114.


In one embodiment, the GPU 104 may enable feature detection and normalized cross-correlation to be performed in efficient, parallel operations. In some configurations, the GPU 104 may obtain multiple images of the user. For example, the GPU 104 may capture multiple images of a user via the camera 106. For instance, the GPU 104 may capture a video (e.g., a 5 second video) via the camera 106. In some configurations, the GPU 104 may use variance data 112 and FFT patch data 114 in relation to feature detection, cross-correlation, and 3-D modeling of a user. For example, the GPU may detect a degree of variation between a feature, or interest point, detected in a first image, and the same feature detected in a second image. The GPU 104 may store the detected variance in the variance data 112. In some configurations, the GPU 104 may generate a patch of an interest point. A patch may be a square set of pixels (e.g., 10 px by 10 px square of pixels) centered on the interest point. In some embodiments, the GPU 104 may perform an FFT algorithm on the patch. The GPU 104 may store the FFT of the patch in the FFT patch data 114. The GPU 104 may use, and reuse, the stored variance and FFT of the patch in relation to performing feature detection and feature tracking cross-correlation of one or more images.



FIG. 2 is a block diagram illustrating another embodiment of an environment 200 in which the present systems and methods may be implemented. In some embodiments, a device 102-a may communicate with a server 206 via a network 204. Example of networks 204 include, local area networks (LAN), wide area networks (WAN), virtual private networks (VPN), wireless networks (using 802.11, for example), cellular networks (using 3G and/or LTE, for example), etc. In some configurations, the network 204 may include the internet. In some configurations, the device 102-a may be one example of the device 102 illustrated in FIG. 1. For example, the device 102-a may include the camera 106, the display 108, and an application 202. It is noted that in some embodiments, the device 102-a may not include a GPU 104. In some embodiments, both a device 102-a and a server 206 may include a GPU 104 where at least a portion of the functions of the GPU 104 are performed separately and/or concurrently on both the device 102-a and the server 206.


In some embodiments, the server 206 may include the GPU 104 and may be coupled to the database 110. For example, the GPU 104 may access the variance data 112 and the FFT patch data 114 in the database 110 via the server 206. The database 110 may be internal or external to the server 206. In some embodiments, the database 110 may be accessible by the device 102-a and/or the server 206 over the network 204.


In some configurations, the application 202 may capture multiple images via the camera 106. For example, the application 202 may use the camera 106 to capture a video. Upon capturing the multiple images, the application 202 may process the multiple images to generate image data. In some embodiments, the application 202 may transmit one or more images to the server 206. Additionally or alternatively, the application 202 may transmit to the server 206 the image data or at least one file associated with the image data.


In some configurations, the GPU 104 may process multiple images of a user to detect features in an image, track the same features among multiple images, and determine a point in a virtual 3-D space corresponding to the tracked features. In some embodiments, the application 202 may process one or more image captured by the camera 106 in order to generate a 3-D model of a user.



FIG. 3 is a block diagram illustrating one example of a GPU 104-a. The GPU 104-a may be one example of the GPU 104 depicted in FIGS. 1 and/or 2. As depicted, the GPU 104-a may include a feature detection module 302 and a cross-correlation module 304.


In some configurations, the feature detection module 302 may examine a pixel of an image to determine whether the pixel includes a feature of interest. In some embodiments, the feature detection module 302 detects a face and/or head of a user in an image. In some embodiments, the feature detection module 302 detects features of the user's head and/or face. In some embodiments, the feature detection module 302 may detect an edge, corner, interest point, blob, and/or ridge in an image of a user. An edge may be points of an image where there is a boundary (or an edge) between two image regions, or a set of points in the image which have a relatively strong gradient magnitude. Corners and interest points may be used interchangeably. An interest point may refer to a point-like feature in an image, which has a local two dimensional structure. In some embodiments, the feature detection module 302 may search for relatively high levels of curvature in an image gradient to detect an interest point and/or corner (e.g., corner of an eye, corner of a mouth). Thus, the feature detection module 302 may detect in an image of a user's face the corners of the eyes, eye centers, pupils, eye brows, point of the nose, nostrils, corners of the mouth, lips, center of the mouth, chin, ears, forehead, cheeks, and the like. A blob may include a complementary description of image structures in terms of regions, as opposed to corners that may be point-like in comparison. Thus, in some embodiments, the feature detection module 302 may detect a smooth, non-point-like area (i.e., blob) in an image. Additionally, or alternatively, in some embodiments, the feature detection module 302 may detect a ridge of points in the image. In some embodiments, the feature detection module 302 may extract a local image patch around a detected feature in order to track the feature in other images.


In some embodiments, the cross-correlation module 304 may process a feature detected by the feature detection module 302. The cross-correlation module 304 may track a feature detected by the feature detection module 302 and determine a position of a point in a virtual 3-D space corresponding to the tracked feature. Operations of the respective feature detection and cross-correlation modules 302 and 304 are discussed in further detail below.



FIG. 4 is a block diagram illustrating one example of a feature detection module 302-a. The feature detection module 302-a may be one example of the feature detection module 302 illustrated in FIG. 3. As depicted, the feature detection module 302 may include a selection module 402, a comparing module 404, a variance module 406, and a patch module 408.


In one embodiment, the selection module 402 may select a feature detected by the GPU 104 in a first image. The first image may be one of multiple images. The multiple images may include images of a user. Each detected feature may include one or more pixels. For example, the detected features may include edges, interest points, corners, blobs, ridges, and/or any other visual or pixel feature contained in an image (e.g., corners of the eyes, eye centers, point of the nose, corners of the mouth, etc.). In some configurations, the comparing module 404 may search a second image of the plurality of images of the user to find a match for the feature detected in the first image. For example, upon detecting the corner of a user's eye, the selection module 402 may select the detected corner of the user's eye as a feature to be tracked in subsequent images. The comparing module 404 may compare the detected corner of the user's eye in image 1 to the corresponding detected corner of the user's eye in image 2.


In some embodiments, the variance module 406 may calculate a variance for each selected feature found in the second image. The variance may indicate a degree a portion of the second image varies from a corresponding portion of the first image. For example, upon comparing the detected corner of the user's eye in image 1 to the corresponding detected corner of the user's eye in image 2, the variance module 406 may determine the degree of variation between the corner of the user's eye in image 2 and the corner of the user's eye in image 1. In some embodiments, the variance module 406 may write the calculated variance to a file and store the file in a storage medium. In some embodiments, the variance module 406 may store the variance in the variance data 112.


In some configurations, the variance module 406 may divide a calculated variance into first and second elements. The variance module 406 may store the first element in a first file and the second element in a second file within the variance data 112. In some embodiments, the variance module 406 divides the calculated variance based on a low order element of the variance and a high order element of the variance. For example, if the variance were calculated to be a decimal number equal to 123,456, then the variance module 406 may divide the hundreds element (i.e., low order element) from the thousands element (i.e., high order element). Storing the calculated variance as 3-digit decimal numbers, the variance module 406 may store the thousands element “123” in a first high order element file in the variance data 112, and may store the hundreds element “456” in a low order element file in the variance data 112. In some embodiments, the GPU 102 and the elements of the GPU 102 may access the high and low order element files stored in the variance data 112 in order to implement the calculated variance in one or more processes related to feature detection, detection tracking, cross-correlation, and generation of a 3-D model.


In one embodiment, the patch module 408 may select, or extract, from among the selected features in the first image one or more patches based on the calculated variance of each selected feature. Each extracted patch may include a square area of pixels centered on one of the selected features, or interest points, of the first image. Upon selecting and extracting the one or more patches in the first image, in some embodiments, the patch module 408 may remove the first image from memory. In some configurations, the patch module 408 may select and extract the one or more patches based on a predetermined threshold of calculated variance. Because a relatively low variance may cause a rounding error in the normalization of a cross-correlation process, in some embodiments, the patch module 408 may extract only those patches where the associated variance is above a certain level of variance. In some embodiments, the patch module 408 may select the one or more patches based on a predetermined number of patches.



FIG. 5 is a block diagram illustrating one example of a cross correlation module 304-a. The cross-correlation module 304-a may be one example of the cross-correlation module 304 illustrated in FIG. 3. As depicted, the cross-correlation module 304-a may include a pose detection module 502, a FFT module 504, a multiplication module 506, and a normalizing module 508.


In some configurations, the cross-correlation module 304-a may perform a cross-correlation algorithm to determine how a patch, selected among the one or more patches, is positioned in first and second sample images of the plurality of images of the user. In some embodiments, the cross-correlation module 304-a may determine a position of the selected feature of the first patch as a point in a virtual three-dimensional (3-D) space. In some configurations, as part of the process to perform the cross-correlation algorithm, the pose detection module 502 may determine a pose of the user in the first and second sample images. The FFT module 504 may perform a FFT on the first patch. In some embodiments, the FFT of the first patch may be written to a file and stored in the FFT patch data 114.


In some embodiments, the FFT module 504 may perform a FFT on the first and second sample images. The FFT module 504 may place the first sample image in a real element of a complex number (e.g., “a” in (a+bi)), and may place the second sample image in an imaginary element of the complex number (e.g., “b” in (a+bi)). In some configurations, the multiplication module 506 may multiply element-wise the FFT of the first patch by the FFT of the first and second sample images. The FFT module 504 may calculate an inverse FFT of the multiplied FFTs, resulting in a first score for the first sample image and a second score for the second sample image. In some embodiments, the normalization module 508 may normalize the result of the cross-correlation by dividing both first and second scores by the calculated variance stored in the variance file. For example, the normalization module 508 may access the high and low order element files stored in the variance data 112 to divide the first and second scores by the calculated variance.


In some embodiments, the cross correlation module 304-a may perform the cross-correlation of every selected patch of a given image simultaneously. For example, the patch module 408 may extract 32 patches from image 1. The 32 patches may be cross-correlated simultaneously with image 2 and image 3 by placing image 2 and 3 in the elements of a complex number, performing the FFT of each patch, and normalizing the result by the stored variance. The FFTs of each patch may be stored by the patch module 408 in the FFT patch data 114 for subsequent cross-correlation operations. In some embodiments, the cross-correlation module 304-a may perform a second cross-correlation algorithm on images 4 and 5 using the same FFT patches stored in the FFT patch data 114.



FIG. 6 is a diagram 600 illustrating an example of a device 102-b for capturing an image 604 of a user 602. The device 102-b may be one example of the device 102 illustrated in FIGS. 1 and/or 2. As depicted, the device 102-b may include a camera 106-b, and display 108-b. The camera 106-b and display 108-b may be examples of the respective camera 106 and display 108 illustrated in FIGS. 1 and/or 2.


In one embodiment, the user may operate the device 102-b. For example, the application 202 may allow the user to interact with and/or operate the device 102-b. In one embodiment, the camera 106-a may allow the user to capture an image 604 of the user 602. The GPU 104 may perform feature detection and feature tracking in relation to the images of the user captured by the device 102-b. Additionally, or alternatively, the GPU 104 may perform a normalized cross-correlation algorithm on one or more of the images of the user to track one or more features detected in each image. The GPU 104 may determine a position of the selected feature as one or more points in a virtual three-dimensional (3-D) space to enable the generation of a 3-D model of the user.



FIG. 7 illustrates an example arrangement 700 of a feature 706 detected in the depicted images 702 and 704. As depicted, the example arrangement 700 may include a first image of the user 702 and a second image of the user 704. In some embodiments, the feature detection module 302 may detect a feature 708 on the first image of the user 702 (e.g., an eye of the user). The selection module 402 may select the feature 708. The patch module 408 may extract the selected feature 708 as a patch 706. Upon extracting the patch 706, in some embodiments, the patch module 408 may remove the image 702 from memory. The comparing module 404 may compare the patch 706 to the second image of the user 704. Upon finding a match between the patch 706 and the detected feature 710 of the of the second image 704, the variance module 406 may measure the degree of variation between the patch 706 and the detected feature 710. For example, the variation module 406 may detect that the detected feature 710 has shifted in one or more directions and/or changed shape (e.g., change in the shape of the eye due to rotation of the face) in relation to the patch 706. In some embodiments, the variance module 406 may store the variance in the variance data 112. The stored variance may be used to normalize a cross-correlation of the patch 706 with one or more images. In some embodiments, the FFT module 504 may take the FFT of the patch 706 and store the FFT of the patch 706 in the FFT patch data 114. In some embodiments, the cross-correlation module 304 may use the stored FFT of the patch 706 in the performance of multiple cross-correlation algorithms.



FIG. 8 is a flow diagram illustrating one embodiment of a method 800 for detecting features in images. In some configurations, the method 800 may be implemented by the GPU 104 illustrated in FIGS. 1, 2, and/or 3. In some configurations, the method 800 may be implemented by the application 202 illustrated in FIG. 2.


At block 802, a plurality of features detected by the GPU in a first image of the plurality of images of the user may be selected. Each selected feature may include one or more pixels. At block 804, in a second image of the plurality of images of the user, a search may be performed for the plurality of features selected in the first image.


At block 806, a variance may be calculated, on the GPU, for each selected feature found in the second image. At block 808, the variance may indicate a degree to which a portion of the second image varies from a corresponding portion of the first image. The calculated variance may be stored in a variance file.



FIG. 9 is a flow diagram illustrating one embodiment of a method 900 for performing cross-correlation algorithms on a GPU. In some configurations, the method 900 may be implemented by the GPU 104 illustrated in FIGS. 1, 2, and/or 3. In some configurations, the method 900 may be implemented by the application 202 illustrated in FIG. 2.


At block 902, one or more patches may be selected from among selected features in a first image based on a calculated variance of each selected feature. At block 904, a first cross-correlation algorithm may be performed on a GPU using a FFT of a first patch stored in a file to determine how the first patch is positioned in first and second sample images. At block 906, a second cross-correlation algorithm may be performed on the GPU on third and fourth sample images using the FFT of the first patch stored in the file to determine how the first patch is positioned in third and fourth sample images.



FIG. 10 is a flow diagram illustrating one embodiment of a method 1000 for performing a cross-correlation algorithm on two images simultaneously. In some configurations, the method 1000 may be implemented by the GPU 104 illustrated in FIGS. 1, 2, and/or 3. In some configurations, the method 1000 may be implemented by the application 202 illustrated in FIG. 2.


At block 1002, a pose of the user in the first and second sample images may be determined. At block 1004, a FFT may be performed on a patch extracted from an image. At block 1006, a FFT may be performed on the first and second sample images. The first sample image may be placed in a real element of a complex number and the second sample image may be placed in an imaginary element of the complex number.


At block 1008, the FFT of the first patch may be multiplied element-wise by the FFT of the first and second sample images. At block 1010, an inverse FFT of the multiplied FFTs may be calculated, resulting in a first score for the first sample image and a second score for the second sample image. At block 1012, the result of the cross-correlation may be normalized by dividing both first and second scores by the calculated variance stored in the variance file.



FIG. 11 depicts a block diagram of a computer system 1100 suitable for implementing the present systems and methods. The depicted computer system 1100 may be one example of a server 206 depicted in FIG. 2. Alternatively, the system 1100 may be one example of a device 102 depicted in FIGS. 1, 2, and/or 6. Computer system 1100 includes a bus 1102 which interconnects major subsystems of computer system 1100, such as a GPU 1104, a system memory 1106 (typically RAM, but which may also include ROM, flash RAM, or the like), an input/output controller 1108, an external audio device, such as a speaker system 1110 via an audio output interface 1112, an external device, such as a display screen 1114 via display adapter 1116, serial ports 1118 and mouse 1146, a keyboard 1122 (interfaced with a keyboard controller 1124), multiple USB devices 1126 (interfaced with a USB controller 1128), a storage interface 1130, a host bus adapter (HBA) interface card 1136A operative to connect with a Fibre Channel network 1138, a host bus adapter (HBA) interface card 1136B operative to connect to a SCSI bus 1140, and an optical disk drive 1142 operative to receive an optical disk 1144. The GPU 1104 may be one example of the GPU 104 depicted in FIGS. 1, 2, and/or 3. Also included are a mouse 1146 (or other point-and-click device, coupled to bus 1102 via serial port 1118), a modem 1148 (coupled to bus 1102 via serial port 1120), and a network interface 1150 (coupled directly to bus 1102).


Bus 1102 allows data communication between GPU 1104 and system memory 1106, which may include read-only memory (ROM) or flash memory (neither shown), and random access memory (RAM) (not shown), as previously noted. The RAM is generally the main memory into which the operating system and application programs are loaded. The ROM or flash memory can contain, among other code, the Basic Input-Output system (BIOS) which controls basic hardware operation such as the interaction with peripheral components or devices. For example, one or more instructions related to the operations of the GPU 1104 to implement the present systems and methods may be stored within the system memory 1106. Applications resident with computer system 1100 are generally stored on and accessed via a non-transitory computer readable medium, such as a hard disk drive (e.g., fixed disk 1152), an optical drive (e.g., optical drive 1142), or other storage medium. Additionally, applications can be in the form of electronic signals modulated in accordance with the application and data communication technology when accessed via network modem 1148 or interface 1150.


Storage interface 1130, as with the other storage interfaces of computer system 1100, can connect to a standard computer readable medium for storage and/or retrieval of information, such as a fixed disk drive 1152. Fixed disk drive 1152 may be a part of computer system 1100 or may be separate and accessed through other interface systems. Modem 1148 may provide a direct connection to a remote server via a telephone link or to the Internet via an internet service provider (ISP). Network interface 1150 may provide a direct connection to a remote server via a direct network link to the Internet via a POP (point of presence). Network interface 1150 may provide such connection using wireless techniques, including digital cellular telephone connection, Cellular Digital Packet Data (CDPD) connection, digital satellite data connection or the like.


Many other devices or subsystems (not shown) may be connected in a similar manner (e.g., document scanners, digital cameras and so on). Conversely, all of the devices shown in FIG. 11 need not be present to practice the present systems and methods. The devices and subsystems can be interconnected in different ways from that shown in FIG. 11. The operation of at least some of the computer system 1100 such as that shown in FIG. 11 is readily known in the art and is not discussed in detail in this application. Code to implement the present disclosure can be stored in a non-transitory computer-readable medium such as one or more of system memory 1106, fixed disk 1152, or optical disk 1144. The operating system provided on computer system 1100 may be MS-DOS®, MS-WINDOWS®, OS/2®, UNIX®, Linux®, or another known operating system.


Moreover, regarding the signals described herein, those skilled in the art will recognize that a signal can be directly transmitted from a first block to a second block, or a signal can be modified (e.g., amplified, attenuated, delayed, latched, buffered, inverted, filtered, or otherwise modified) between the blocks. Although the signals of the above described embodiment are characterized as transmitted from one block to the next, other embodiments of the present systems and methods may include modified signals in place of such directly transmitted signals as long as the informational and/or functional aspect of the signal is transmitted between blocks. To some extent, a signal input at a second block can be conceptualized as a second signal derived from a first signal output from a first block due to physical limitations of the circuitry involved (e.g., there will inevitably be some attenuation and delay). Therefore, as used herein, a second signal derived from a first signal includes the first signal or any modifications to the first signal, whether due to circuit limitations or due to passage through other circuit elements which do not change the informational and/or final functional aspect of the first signal.


While the foregoing disclosure sets forth various embodiments using specific block diagrams, flowcharts, and examples, each block diagram component, flowchart step, operation, and/or component described and/or illustrated herein may be implemented, individually and/or collectively, using a wide range of hardware, software, or firmware (or any combination thereof) configurations. In addition, any disclosure of components contained within other components should be considered exemplary in nature since many other architectures can be implemented to achieve the same functionality.


The process parameters and sequence of steps described and/or illustrated herein are given by way of example only and can be varied as desired. For example, while the steps illustrated and/or described herein may be shown or discussed in a particular order, these steps do not necessarily need to be performed in the order illustrated or discussed. The various exemplary methods described and/or illustrated herein may also omit one or more of the steps described or illustrated herein or include additional steps in addition to those disclosed.


Furthermore, while various embodiments have been described and/or illustrated herein in the context of fully functional computing systems, one or more of these exemplary embodiments may be distributed as a program product in a variety of forms, regardless of the particular type of computer-readable media used to actually carry out the distribution. The embodiments disclosed herein may also be implemented using software modules that perform certain tasks. These software modules may include script, batch, or other executable files that may be stored on a computer-readable storage medium or in a computing system. In some embodiments, these software modules may configure a computing system to perform one or more of the exemplary embodiments disclosed herein.


The foregoing description, for purpose of explanation, has been described with reference to specific embodiments. However, the illustrative discussions above are not intended to be exhaustive or to limit the invention to the precise forms disclosed. Many modifications and variations are possible in view of the above teachings. The embodiments were chosen and described in order to best explain the principles of the present systems and methods and their practical applications, to thereby enable others skilled in the art to best utilize the present systems and methods and various embodiments with various modifications as may be suited to the particular use contemplated.


Unless otherwise noted, the terms “a” or “an,” as used in the specification and claims, are to be construed as meaning “at least one of.” In addition, for ease of use, the words “including” and “having,” as used in the specification and claims, are interchangeable with and have the same meaning as the word “comprising.” In addition, the term “based on” as used in the specification and the claims is to be construed as meaning “based at least upon.”

Claims
  • 1. A computer-implemented method for processing, by a graphical processor unit (GPU), a plurality of images of a user, the method comprising: selecting, via a device with a central processing unit (CPU) and the GPU, a plurality of features detected by the GPU in a first image of the plurality of images of the user, wherein each feature comprises one or more pixels;in a second image of the plurality of images of the user, searching for the plurality of features selected in the first image;calculating, on the GPU, a variance for each selected feature found in the second image, wherein the variance indicates a degree a portion of the second image varies from a corresponding portion of the first image;storing the calculated variance in a variance file; andperforming on the GPU a cross-correlation algorithm, wherein performing the cross-correlation algorithm on the GPU comprises: performing, via the GPU, a fast Fourier transform (FFT) on first and second images from the plurality of images of the user.
  • 2. The method of claim 1, further comprising: selecting from among the selected features in the first image one or more patches based on the calculated variance of each selected feature, wherein each patch comprises a square area of pixels centered on one of the selected features of the first image.
  • 3. The method of claim 2, further comprising: upon selecting the one or more patches in the first image, removing the first image from memory.
  • 4. The method of claim 2, wherein selecting the one or more patches further comprises: selecting the one or more patches based on a predetermined threshold of calculated variance.
  • 5. The method of claim 2, wherein selecting the one or more patches further comprises: selecting the one or more patches based on a predetermined number of patches.
  • 6. The method of claim 1, further comprising: dividing each variance into first and second elements; andstoring the first element in a first file and the second element in a second file.
  • 7. The method of claim 2, further comprising: performing on the GPU the cross-correlation algorithm to determine how a first patch, selected among the one or more patches, is positioned in the first and second images of the plurality of images of the user.
  • 8. The method of claim 7, wherein performing the cross-correlation algorithm on the GPU comprises: determining a pose of the user in the first and second images;performing the FFT on the first patch, wherein the FFT of the first patch is stored in a third file;wherein performing the FFT on the first and second images includes placing the first image in a real element of a complex number and placing the second image in an imaginary element of the complex number;multiplying element-wise the FFT of the first patch by the FFT of the first and second images;calculating an inverse FFT of the multiplied FFTs, resulting in a first score for the first image and a second score for the second image; andnormalizing the result of the cross-correlation by dividing both first and second scores by the calculated variance stored in the variance file.
  • 9. The method of claim 8, further comprising: performing on the GPU a second cross-correlation algorithm on third and fourth images of the plurality of images of the user using the FFT of the first patch stored in the third file to determine how the first patch is positioned in the third and fourth images.
  • 10. The method of claim 7, wherein performing on the GPU the cross-correlation algorithm further comprises performing on the GPU the cross-correlation of each selected patch simultaneously.
  • 11. The method of claim 1, further comprising: determining a position of the selected feature of the first patch as a point in a virtual three-dimensional (3-D) space.
  • 12. A computing device configured to process, by a graphical processor unit (GPU), a plurality of images of a user, comprising: the GPU;memory in electronic communication with the GPU;instructions stored in the memory, the instructions being executable by the GPU to: select, via a device with a central processing unit (CPU) and the GPU, a plurality of features detected by the GPU in a first image of the plurality of images of the user, wherein each feature comprises one or more pixels;in a second image of the plurality of images of the user, search for the plurality of features selected in the first image;calculate, on the GPU, a variance for each selected feature found in the second image, wherein the variance indicates a degree a portion of the second image varies from a corresponding portion of the first image;storing the calculated variance in a variance file; andperform on the GPU a cross-correlation algorithm, wherein performing the cross-correlation algorithm on the GPU comprises instructions executable by the GPU to: perform, via the GPU, a fast Fourier transform (FFT) on first and second images from the plurality of images of the user, placing the first image in a real element of a complex number and placing the second image in an imaginary element of the complex number.
  • 13. The computing device of claim 12, wherein the instructions are executable by the GPU to: select from among the selected features in the first image one or more patches based on the calculated variance of each selected feature, wherein each patch comprises a square area of pixels centered on one of the selected features of the first image;upon selecting the one or more patches in the first image, remove the first image from memory.
  • 14. The computing device of claim 13, wherein the instructions are executable by the GPU to: select the one or more patches based on a predetermined threshold of calculated variance.
  • 15. The computing device of claim 13, wherein the instructions are executable by the GPU to: perform on the GPU the cross-correlation algorithm to determine how a first patch, selected among the one or more patches, is positioned in the first and second images of the plurality of images of the user.
  • 16. The computing device of claim 15, wherein performing the cross-correlation algorithm on the GPU comprises instructions executable by the GPU to: determine a pose of the user in the first and second images;perform the FFT on the first patch, wherein the FFT of the first patch is stored in a third file;multiply element-wise the FFT of the first patch by the FFT of the first and second images;calculate an inverse FFT of the multiplied FFTs, resulting in a first score for the first image and a second score for the second image; andnormalize the result of the cross-correlation by dividing both first and second scores by the calculated variance stored in the variance file.
  • 17. The computing device of claim 16, wherein the instructions are executable by the GPU to: perform on the GPU a second cross-correlation algorithm on third and fourth images of the plurality of images of the user using the FFT of the first patch stored in the third file to determine how the first patch is positioned in the third and fourth images.
  • 18. The computing device of claim 15, wherein the instructions are executable by the GPU to: perform on the GPU the cross-correlation of each selected patch simultaneously.
  • 19. A computer-program product for processing, by a graphical processor unit (GPU), a plurality of images of a user, the computer-program product comprising a non-transitory computer-readable medium storing instructions thereon, the instructions being executable by the GPU to: select, via a device with a central processing unit (CPU) and the GPU, a plurality of features detected by the GPU in a first image of the plurality of images of the user, wherein each feature comprises one or more pixels;in a second image of the plurality of images of the user, search for the plurality of features selected in the first image;calculate, on the GPU, a variance for each selected feature found in the second image, wherein the variance indicates a degree a portion of the second image varies from a corresponding portion of the first image;store the calculated variance in a variance file; andperform on the GPU a cross-correlation algorithm, wherein performing the cross-correlation algorithm on the GPU comprises instructions executable by the GPU to: perform, via the GPU, a fast Fourier transform (FFT) on first and second images from the plurality of images of the user, placing the first image in a real element of a complex number and placing the second image in an imaginary element of the complex number.
  • 20. The computer-program product of claim 19, wherein the instructions are executable by the GPU to: perform on the GPU the first cross-correlation algorithm of each selected patch simultaneously to determine how each patch is positioned in the first and second images of the plurality of images of the user;perform on the GPU a second cross-correlation algorithm of each selected patch simultaneously to determine how each patch is positioned in third and fourth images of the plurality of images of the user; andnormalize the results of the first and second cross-correlations by dividing the results by the calculated variance stored in the variance file.
RELATED APPLICATIONS

This application claims priority to U.S. Provisional Application No. 61/650,983, entitled SYSTEMS AND METHODS TO VIRTUALLY TRY-ON PRODUCTS, filed on May 23, 2012; and U.S. Provisional Application No. 61/735,951, entitled SYSTEMS AND METHODS TO VIRTUALLY TRY-ON PRODUCTS, filed on Dec. 11, 2012, which is incorporated herein in its entirety by this reference.

US Referenced Citations (485)
Number Name Date Kind
3927933 Humphrey Dec 1975 A
4370058 Trötscher et al. Jan 1983 A
4467349 Maloomian Aug 1984 A
4522474 Slavin Jun 1985 A
4534650 Clerget et al. Aug 1985 A
4539585 Spackova et al. Sep 1985 A
4573121 Saigo et al. Feb 1986 A
4613219 Vogel Sep 1986 A
4698564 Slavin Oct 1987 A
4724617 Logan et al. Feb 1988 A
4730260 Mori et al. Mar 1988 A
4781452 Ace Nov 1988 A
4786160 Fürter Nov 1988 A
4845641 Ninomiya et al. Jul 1989 A
4852184 Tamura et al. Jul 1989 A
4957369 Antonsson Sep 1990 A
5139373 Logan et al. Aug 1992 A
5255352 Falk Oct 1993 A
5257198 van Schoyck Oct 1993 A
5280570 Jordan Jan 1994 A
5281957 Schoolman Jan 1994 A
5428448 Albert-Garcia Jun 1995 A
5485399 Saigo et al. Jan 1996 A
5550602 Braeuning Aug 1996 A
5592248 Norton et al. Jan 1997 A
5631718 Markovitz et al. May 1997 A
5666957 Juto Sep 1997 A
5682210 Weirich Oct 1997 A
5720649 Gerber et al. Feb 1998 A
5724522 Kagami et al. Mar 1998 A
5774129 Poggio et al. Jun 1998 A
5809580 Arnette Sep 1998 A
5844573 Poggio et al. Dec 1998 A
5880806 Conway Mar 1999 A
5908348 Gottschald Jun 1999 A
5974400 Kagami et al. Oct 1999 A
5980037 Conway Nov 1999 A
5983201 Fay Nov 1999 A
5987702 Simioni Nov 1999 A
5988862 Kacyra et al. Nov 1999 A
D417883 Arnette Dec 1999 S
6016150 Lengyel et al. Jan 2000 A
6018339 Stevens Jan 2000 A
D420037 Conway Feb 2000 S
D420379 Conway Feb 2000 S
D420380 Simioni et al. Feb 2000 S
6024444 Little Feb 2000 A
D421764 Arnette Mar 2000 S
D422011 Conway Mar 2000 S
D422014 Simioni et al. Mar 2000 S
D423034 Arnette Apr 2000 S
D423552 Flanagan et al. Apr 2000 S
D423553 Brune Apr 2000 S
D423554 Conway Apr 2000 S
D423556 Conway Apr 2000 S
D423557 Conway Apr 2000 S
D424094 Conway May 2000 S
D424095 Brune et al. May 2000 S
D424096 Conway May 2000 S
D424589 Arnette May 2000 S
D424598 Simioni May 2000 S
D425542 Arnette May 2000 S
D425543 Brune May 2000 S
D426568 Conway Jun 2000 S
D427225 Arnette Jun 2000 S
D427227 Conway Jun 2000 S
6072496 Guenter et al. Jun 2000 A
6095650 Gao et al. Aug 2000 A
6102539 Tucker Aug 2000 A
D430591 Arnette Sep 2000 S
D432156 Conway et al. Oct 2000 S
D433052 Flanagan Oct 2000 S
6132044 Sternbergh Oct 2000 A
6139141 Zider Oct 2000 A
6139143 Brune et al. Oct 2000 A
6142628 Saigo Nov 2000 A
6144388 Bornstein Nov 2000 A
D434788 Conway Dec 2000 S
D439269 Conway Mar 2001 S
6208347 Migdal et al. Mar 2001 B1
6222621 Taguchi et al. Apr 2001 B1
6231188 Gao et al. May 2001 B1
6233049 Kondo et al. May 2001 B1
6246468 Dimsdale Jun 2001 B1
6249600 Reed et al. Jun 2001 B1
6281903 Martin et al. Aug 2001 B1
6305656 Wemyss Oct 2001 B1
6307568 Rom Oct 2001 B1
6310627 Sakaguchi Oct 2001 B1
6330523 Kacyra et al. Dec 2001 B1
6356271 Reiter et al. Mar 2002 B1
6377281 Rosenbluth et al. Apr 2002 B1
6386562 Kuo May 2002 B1
6415051 Callari et al. Jul 2002 B1
6419549 Shirayanagi Jul 2002 B2
6420698 Dimsdale Jul 2002 B1
6434278 Hashimoto Aug 2002 B1
6456287 Kamen et al. Sep 2002 B1
6466205 Simpson et al. Oct 2002 B2
6473079 Kacyra et al. Oct 2002 B1
6492986 Metaxas et al. Dec 2002 B1
6493073 Epstein Dec 2002 B2
6508553 Gao et al. Jan 2003 B2
6512518 Dimsdale Jan 2003 B2
6512993 Kacyra et al. Jan 2003 B2
6516099 Davison et al. Feb 2003 B1
6518963 Waupotitsch et al. Feb 2003 B1
6527731 Weiss et al. Mar 2003 B2
6529192 Waupotitsch Mar 2003 B1
6529626 Watanabe et al. Mar 2003 B1
6529627 Callari et al. Mar 2003 B1
6533418 Izumitani et al. Mar 2003 B1
6535223 Foley Mar 2003 B1
6556196 Blanz et al. Apr 2003 B1
6563499 Waupotitsch et al. May 2003 B1
6583792 Agnew Jun 2003 B1
6624843 Lennon Sep 2003 B2
6634754 Fukuma et al. Oct 2003 B2
6637880 Yamakaji et al. Oct 2003 B1
6647146 Davison et al. Nov 2003 B1
6650324 Junkins Nov 2003 B1
6659609 Mothes Dec 2003 B2
6661433 Lee Dec 2003 B1
6664956 Erdem Dec 2003 B1
6668082 Davison et al. Dec 2003 B1
6671538 Ehnholm et al. Dec 2003 B1
6677946 Ohba Jan 2004 B1
6682195 Dreher Jan 2004 B2
6692127 Abitbol et al. Feb 2004 B2
6705718 Fossen Mar 2004 B2
6726463 Foreman Apr 2004 B2
6734849 Dimsdale et al. May 2004 B2
6736506 Izumitani et al. May 2004 B2
6760488 Moura et al. Jul 2004 B1
6775128 Leitao Aug 2004 B2
6785585 Gottschald Aug 2004 B1
6791584 Xie Sep 2004 B1
6792401 Nigro et al. Sep 2004 B1
6807290 Liu et al. Oct 2004 B2
6808381 Foreman et al. Oct 2004 B2
6817713 Ueno Nov 2004 B2
6825838 Smith et al. Nov 2004 B2
6847383 Agnew Jan 2005 B2
6847462 Kacyra et al. Jan 2005 B1
6876755 Taylor et al. Apr 2005 B1
6893245 Foreman et al. May 2005 B2
6903746 Fukushima et al. Jun 2005 B2
6907310 Gardner et al. Jun 2005 B2
6922494 Fay Jul 2005 B1
6943789 Perry et al. Sep 2005 B2
6944327 Soatto Sep 2005 B1
6950804 Strietzel Sep 2005 B2
6961439 Ballas Nov 2005 B2
6965385 Welk et al. Nov 2005 B2
6965846 Krimmer Nov 2005 B2
6968075 Chang Nov 2005 B1
6980690 Taylor et al. Dec 2005 B1
6999073 Zwern et al. Feb 2006 B1
7003515 Glaser et al. Feb 2006 B1
7016824 Waupotitsch et al. Mar 2006 B2
7034818 Perry et al. Apr 2006 B2
7043059 Cheatle et al. May 2006 B2
7051290 Foreman et al. May 2006 B2
7062722 Carlin et al. Jun 2006 B1
7069107 Ueno Jun 2006 B2
7095878 Taylor et al. Aug 2006 B1
7103211 Medioni et al. Sep 2006 B1
7116804 Murase et al. Oct 2006 B2
7133048 Brand Nov 2006 B2
7152976 Fukuma et al. Dec 2006 B2
7154529 Hoke et al. Dec 2006 B2
7156655 Sachdeva et al. Jan 2007 B2
7184036 Dimsdale et al. Feb 2007 B2
7209557 Lahiri Apr 2007 B2
7212656 Liu et al. May 2007 B2
7212664 Lee et al. May 2007 B2
7215430 Kacyra et al. May 2007 B2
7218323 Halmshaw et al. May 2007 B1
7219995 Ollendorf et al. May 2007 B2
7224357 Chen et al. May 2007 B2
7234937 Sachdeva et al. Jun 2007 B2
7242807 Waupotitsch et al. Jul 2007 B2
7290201 Edwards Oct 2007 B1
7310102 Spicer Dec 2007 B2
7324110 Edwards et al. Jan 2008 B2
7415152 Jiang et al. Aug 2008 B2
7421097 Hamza et al. Sep 2008 B2
7426292 Moghaddam et al. Sep 2008 B2
7434931 Warden et al. Oct 2008 B2
7436988 Zhang et al. Oct 2008 B2
7441895 Akiyama et al. Oct 2008 B2
7450737 Ishikawa et al. Nov 2008 B2
7489768 Strietzel Feb 2009 B1
7492364 Devarajan et al. Feb 2009 B2
7508977 Lyons et al. Mar 2009 B2
7523411 Carlin Apr 2009 B2
7530690 Divo et al. May 2009 B2
7532215 Yoda et al. May 2009 B2
7533453 Yancy May 2009 B2
7540611 Welk et al. Jun 2009 B2
7557812 Chou et al. Jul 2009 B2
7563975 Leahy et al. Jul 2009 B2
7573475 Sullivan et al. Aug 2009 B2
7573489 Davidson et al. Aug 2009 B2
7587082 Rudin et al. Sep 2009 B1
7609859 Lee et al. Oct 2009 B2
7630580 Repenning Dec 2009 B1
7634103 Rubinstenn et al. Dec 2009 B2
7643685 Miller Jan 2010 B2
7646909 Jiang et al. Jan 2010 B2
7651221 Krengel et al. Jan 2010 B2
7656402 Abraham et al. Feb 2010 B2
7657083 Parr et al. Feb 2010 B2
7663648 Saldanha et al. Feb 2010 B1
7665843 Xie Feb 2010 B2
7689043 Austin et al. Mar 2010 B2
7699300 Iguchi Apr 2010 B2
7711155 Sharma et al. May 2010 B1
7717708 Sachdeva et al. May 2010 B2
7720285 Ishikawa et al. May 2010 B2
D616918 Rohrbach Jun 2010 S
7736147 Kaza et al. Jun 2010 B2
7755619 Wang et al. Jul 2010 B2
7756325 Vetter et al. Jul 2010 B2
7760923 Walker et al. Jul 2010 B2
7768528 Edwards et al. Aug 2010 B1
D623216 Rohrbach Sep 2010 S
7804997 Geng et al. Sep 2010 B2
7814436 Schrag et al. Oct 2010 B2
7830384 Edwards et al. Nov 2010 B1
7835565 Cai et al. Nov 2010 B2
7835568 Park et al. Nov 2010 B2
7845797 Warden et al. Dec 2010 B2
7848548 Moon et al. Dec 2010 B1
7852995 Strietzel Dec 2010 B2
7856125 Medioni et al. Dec 2010 B2
7860225 Strietzel Dec 2010 B2
7860301 Se et al. Dec 2010 B2
7876931 Geng Jan 2011 B2
7896493 Welk et al. Mar 2011 B2
7907774 Parr et al. Mar 2011 B2
7929745 Walker et al. Apr 2011 B2
7929775 Hager et al. Apr 2011 B2
7953675 Medioni et al. May 2011 B2
7961914 Smith Jun 2011 B1
8009880 Zhang et al. Aug 2011 B2
8026916 Wen Sep 2011 B2
8026917 Rogers et al. Sep 2011 B1
8026929 Naimark Sep 2011 B2
8031909 Se et al. Oct 2011 B2
8031933 Se et al. Oct 2011 B2
8059917 Dumas et al. Nov 2011 B2
8064685 Solem et al. Nov 2011 B2
8070619 Edwards Dec 2011 B2
8073196 Yuan et al. Dec 2011 B2
8090160 Kakadiaris et al. Jan 2012 B2
8113829 Sachdeva et al. Feb 2012 B2
8118427 Bonnin et al. Feb 2012 B2
8126242 Brett et al. Feb 2012 B2
8126249 Brett et al. Feb 2012 B2
8126261 Medioni et al. Feb 2012 B2
8130225 Sullivan et al. Mar 2012 B2
8131063 Xiao et al. Mar 2012 B2
8132123 Schrag et al. Mar 2012 B2
8144153 Sullivan et al. Mar 2012 B1
8145545 Rathod et al. Mar 2012 B2
8155411 Hof et al. Apr 2012 B2
8160345 Pavlovskaia et al. Apr 2012 B2
8177551 Sachdeva et al. May 2012 B2
8182087 Esser et al. May 2012 B2
8194072 Jones et al. Jun 2012 B2
8199152 Sullivan et al. Jun 2012 B2
8200502 Wedwick Jun 2012 B2
8204299 Arcas et al. Jun 2012 B2
8204301 Xiao et al. Jun 2012 B2
8204334 Bhagavathy et al. Jun 2012 B2
8208717 Xiao et al. Jun 2012 B2
8212812 Tsin et al. Jul 2012 B2
8217941 Park et al. Jul 2012 B2
8218836 Metaxas et al. Jul 2012 B2
8224039 Ionita et al. Jul 2012 B2
8243065 Kim Aug 2012 B2
8248417 Clifton Aug 2012 B1
8260006 Callari et al. Sep 2012 B1
8260038 Xiao et al. Sep 2012 B2
8260039 Shiell et al. Sep 2012 B2
8264504 Naimark Sep 2012 B2
8269779 Rogers et al. Sep 2012 B2
8274506 Rees Sep 2012 B1
8284190 Muktinutalapati et al. Oct 2012 B2
8286083 Barrus et al. Oct 2012 B2
8289317 Harvill Oct 2012 B2
8290769 Taub et al. Oct 2012 B2
8295589 Ofek et al. Oct 2012 B2
8300900 Lai et al. Oct 2012 B2
8303113 Esser et al. Nov 2012 B2
8307560 Tulin Nov 2012 B2
8330801 Wang et al. Dec 2012 B2
8346020 Guntur Jan 2013 B2
8355079 Zhang et al. Jan 2013 B2
8372319 Liguori et al. Feb 2013 B2
8374422 Roussel Feb 2013 B2
8385646 Lang et al. Feb 2013 B2
8391547 Huang et al. Mar 2013 B2
8459792 Wilson Jun 2013 B2
8605942 Takeuchi Dec 2013 B2
8605989 Rudin et al. Dec 2013 B2
8743051 Moy et al. Jun 2014 B1
8813378 Grove Aug 2014 B2
20010023413 Fukuma et al. Sep 2001 A1
20010026272 Feld et al. Oct 2001 A1
20010051517 Strietzel Dec 2001 A1
20020010655 Kjallstrom Jan 2002 A1
20020105530 Waupotitsch et al. Aug 2002 A1
20020149585 Kacyra et al. Oct 2002 A1
20030001835 Dimsdale et al. Jan 2003 A1
20030030904 Huang Feb 2003 A1
20030071810 Shoov et al. Apr 2003 A1
20030110099 Trajkovic et al. Jun 2003 A1
20030112240 Cerny Jun 2003 A1
20040004633 Perry et al. Jan 2004 A1
20040090438 Alliez et al. May 2004 A1
20040217956 Besl et al. Nov 2004 A1
20040223631 Waupotitsch et al. Nov 2004 A1
20040257364 Basler Dec 2004 A1
20050053275 Stokes Mar 2005 A1
20050063582 Park et al. Mar 2005 A1
20050111705 Waupotitsch et al. May 2005 A1
20050128211 Berger et al. Jun 2005 A1
20050162419 Kim et al. Jul 2005 A1
20050190264 Neal Sep 2005 A1
20050208457 Fink et al. Sep 2005 A1
20050226509 Maurer et al. Oct 2005 A1
20060012748 Periasamy et al. Jan 2006 A1
20060017887 Jacobson et al. Jan 2006 A1
20060067573 Parr et al. Mar 2006 A1
20060127852 Wen Jun 2006 A1
20060161474 Diamond et al. Jul 2006 A1
20060212150 Sims, Jr. Sep 2006 A1
20060216680 Buckwalter et al. Sep 2006 A1
20070013873 Jacobson et al. Jan 2007 A9
20070104360 Huang et al. May 2007 A1
20070127848 Kim et al. Jun 2007 A1
20070160306 Ahn et al. Jul 2007 A1
20070183679 Moroto et al. Aug 2007 A1
20070233311 Okada et al. Oct 2007 A1
20070262988 Christensen Nov 2007 A1
20080084414 Rosel et al. Apr 2008 A1
20080112610 Israelsen et al. May 2008 A1
20080136814 Chu et al. Jun 2008 A1
20080152200 Medioni et al. Jun 2008 A1
20080162695 Muhn et al. Jul 2008 A1
20080163344 Yang Jul 2008 A1
20080170077 Sullivan et al. Jul 2008 A1
20080201641 Xie Aug 2008 A1
20080219589 Jung et al. Sep 2008 A1
20080240588 Tsoupko-Sitnikov et al. Oct 2008 A1
20080246759 Summers Oct 2008 A1
20080271078 Gossweiler et al. Oct 2008 A1
20080278437 Barrus et al. Nov 2008 A1
20080278633 Tsoupko-Sitnikov et al. Nov 2008 A1
20080279478 Tsoupko-Sitnikov et al. Nov 2008 A1
20080280247 Sachdeva et al. Nov 2008 A1
20080294393 Laake et al. Nov 2008 A1
20080297503 Dickinson et al. Dec 2008 A1
20080310757 Wolberg et al. Dec 2008 A1
20090010507 Geng Jan 2009 A1
20090040216 Ishiyama Feb 2009 A1
20090123037 Ishida May 2009 A1
20090129402 Moller et al. May 2009 A1
20090132371 Strietzel et al. May 2009 A1
20090135176 Snoddy et al. May 2009 A1
20090135177 Strietzel et al. May 2009 A1
20090144173 Mo et al. Jun 2009 A1
20090153552 Fidaleo et al. Jun 2009 A1
20090153553 Kim et al. Jun 2009 A1
20090153569 Park et al. Jun 2009 A1
20090154794 Kim et al. Jun 2009 A1
20090184960 Carr et al. Jul 2009 A1
20090185763 Park et al. Jul 2009 A1
20090219281 Maillot Sep 2009 A1
20090279784 Arcas et al. Nov 2009 A1
20090296984 Nijim et al. Dec 2009 A1
20090304270 Bhagavathy et al. Dec 2009 A1
20090310861 Lang et al. Dec 2009 A1
20090316945 Akansu Dec 2009 A1
20090316966 Marshall et al. Dec 2009 A1
20090324030 Frinking et al. Dec 2009 A1
20090324121 Bhagavathy et al. Dec 2009 A1
20100030578 Siddique et al. Feb 2010 A1
20100134487 Lai et al. Jun 2010 A1
20100138025 Morton et al. Jun 2010 A1
20100141893 Altheimer et al. Jun 2010 A1
20100145489 Esser et al. Jun 2010 A1
20100166978 Nieminen Jul 2010 A1
20100179789 Sachdeva et al. Jul 2010 A1
20100191504 Esser et al. Jul 2010 A1
20100198817 Esser et al. Aug 2010 A1
20100209005 Rudin et al. Aug 2010 A1
20100277476 Johanson et al. Nov 2010 A1
20100293192 Suy et al. Nov 2010 A1
20100293251 Suy et al. Nov 2010 A1
20100302275 Saldanha et al. Dec 2010 A1
20100329568 Gamliel et al. Dec 2010 A1
20110001791 Kirshenboim et al. Jan 2011 A1
20110025827 Shpunt et al. Feb 2011 A1
20110026606 Bhagavathy et al. Feb 2011 A1
20110026607 Bhagavathy et al. Feb 2011 A1
20110029561 Slaney et al. Feb 2011 A1
20110040539 Szymczyk et al. Feb 2011 A1
20110043540 Fancher et al. Feb 2011 A1
20110043610 Ren et al. Feb 2011 A1
20110071804 Xie Mar 2011 A1
20110075916 Knothe et al. Mar 2011 A1
20110096832 Zhang et al. Apr 2011 A1
20110102553 Corcoran et al. May 2011 A1
20110115786 Mochizuki May 2011 A1
20110148858 Ni et al. Jun 2011 A1
20110157229 Ni et al. Jun 2011 A1
20110158394 Strietzel Jun 2011 A1
20110166834 Clara Jul 2011 A1
20110188780 Wang et al. Aug 2011 A1
20110208493 Altheimer et al. Aug 2011 A1
20110211816 Goedeken et al. Sep 2011 A1
20110227923 Mariani et al. Sep 2011 A1
20110227934 Sharp Sep 2011 A1
20110229659 Reynolds Sep 2011 A1
20110229660 Reynolds Sep 2011 A1
20110234581 Eikelis et al. Sep 2011 A1
20110234591 Mishra et al. Sep 2011 A1
20110249136 Levy et al. Oct 2011 A1
20110262717 Broen et al. Oct 2011 A1
20110279634 Periyannan et al. Nov 2011 A1
20110292034 Corazza et al. Dec 2011 A1
20110293247 Bhagavathy et al. Dec 2011 A1
20110304912 Broen et al. Dec 2011 A1
20110306417 Sheblak et al. Dec 2011 A1
20120002161 Altheimer et al. Jan 2012 A1
20120008090 Atheimer et al. Jan 2012 A1
20120013608 Ahn et al. Jan 2012 A1
20120016645 Altheimer et al. Jan 2012 A1
20120021835 Keller et al. Jan 2012 A1
20120038665 Strietzel Feb 2012 A1
20120075296 Wegbreit et al. Mar 2012 A1
20120079377 Goossens Mar 2012 A1
20120082432 Ackley et al. Apr 2012 A1
20120114184 Barcons-Palau et al. May 2012 A1
20120114251 Solem et al. May 2012 A1
20120121174 Bhagavathy et al. May 2012 A1
20120130524 Clara et al. May 2012 A1
20120133640 Chin et al. May 2012 A1
20120133850 Broen et al. May 2012 A1
20120147324 Marin et al. Jun 2012 A1
20120158369 Bachrach et al. Jun 2012 A1
20120162218 Kim et al. Jun 2012 A1
20120166431 Brewington et al. Jun 2012 A1
20120170821 Zug et al. Jul 2012 A1
20120176380 Wang et al. Jul 2012 A1
20120177283 Wang et al. Jul 2012 A1
20120183202 Wei et al. Jul 2012 A1
20120183204 Aarts et al. Jul 2012 A1
20120183238 Savvides et al. Jul 2012 A1
20120192401 Pavlovskaia et al. Aug 2012 A1
20120206610 Wang et al. Aug 2012 A1
20120219195 Wu et al. Aug 2012 A1
20120224629 Bhagavathy et al. Sep 2012 A1
20120229758 Marin et al. Sep 2012 A1
20120256906 Ross et al. Oct 2012 A1
20120263437 Barcons-Palau et al. Oct 2012 A1
20120288015 Zhang et al. Nov 2012 A1
20120294369 Bhagavathy et al. Nov 2012 A1
20120294530 Bhaskaranand Nov 2012 A1
20120299914 Kilpatrick et al. Nov 2012 A1
20120306874 Nguyen et al. Dec 2012 A1
20120307074 Bhagavathy et al. Dec 2012 A1
20120314023 Barcons-Palau et al. Dec 2012 A1
20120320153 Barcons-Palau et al. Dec 2012 A1
20120321128 Medioni et al. Dec 2012 A1
20120323581 Strietzel et al. Dec 2012 A1
20130027657 Esser et al. Jan 2013 A1
20130070973 Saito et al. Mar 2013 A1
20130088490 Rasmussen et al. Apr 2013 A1
20130187915 Lee et al. Jul 2013 A1
20130201187 Tong et al. Aug 2013 A1
20130271451 Tong et al. Oct 2013 A1
Foreign Referenced Citations (68)
Number Date Country
10007705 Sep 2001 DE
0092364 Oct 1983 EP
0359596 Mar 1990 EP
0994336 Apr 2000 EP
1011006 Jun 2000 EP
1136869 Sep 2001 EP
1138253 Oct 2001 EP
0444902 Jun 2002 EP
1450201 Aug 2004 EP
1728467 Dec 2006 EP
1154302 Aug 2009 EP
2966038 Apr 2012 FR
2449855 Dec 2008 GB
2003345857 Dec 2003 JP
2004272530 Sep 2004 JP
2005269022 Sep 2005 JP
20000028583 May 2000 KR
200000051217 Aug 2000 KR
20040097200 Nov 2004 KR
20080086945 Sep 2008 KR
20100050052 May 2010 KR
WO 9300641 Jan 1993 WO
WO 9604596 Feb 1996 WO
WO 9740342 Oct 1997 WO
WO 9740960 Nov 1997 WO
WO 9813721 Apr 1998 WO
WO 9827861 Jul 1998 WO
WO 9827902 Jul 1998 WO
WO 9835263 Aug 1998 WO
WO 9852189 Nov 1998 WO
WO 9857270 Dec 1998 WO
WO 9956942 Nov 1999 WO
WO 9964918 Dec 1999 WO
WO 0000863 Jan 2000 WO
WO 0016683 Mar 2000 WO
WO 0045348 Aug 2000 WO
WO 0049919 Aug 2000 WO
WO 0062148 Oct 2000 WO
WO 0064168 Oct 2000 WO
WO 0123908 Apr 2001 WO
WO 0132074 May 2001 WO
WO 0135338 May 2001 WO
WO 0161447 Aug 2001 WO
WO 0167325 Sep 2001 WO
WO 0174553 Oct 2001 WO
WO 0178630 Oct 2001 WO
WO 0188654 Nov 2001 WO
WO 0207845 Jan 2002 WO
WO 0241127 May 2002 WO
WO 03079097 Sep 2003 WO
WO 03084448 Oct 2003 WO
WO 2007012261 Feb 2007 WO
WO 2007017751 Feb 2007 WO
WO 2007018017 Feb 2007 WO
WO 2008009355 Jan 2008 WO
WO 2008009423 Jan 2008 WO
WO 2008135178 Nov 2008 WO
WO 2009023012 Feb 2009 WO
WO 2009043941 Apr 2009 WO
2010039976 Apr 2010 WO
2010042990 Apr 2010 WO
WO 2011012743 Feb 2011 WO
WO 2011095917 Aug 2011 WO
WO 2011134611 Nov 2011 WO
WO 2011147649 Dec 2011 WO
WO 2012051654 Apr 2012 WO
WO 2012054972 May 2012 WO
WO 2012054983 May 2012 WO
Non-Patent Literature Citations (28)
Entry
(Tracker, “Tracker Help”, Nov. 2009).
(Sudipta Sinha, “GPU-based Video Feature Tracking and Matching”, May 2006,http://frahm.web.unc.edu/files/2014/01/GPU-based-Video-Feature-Tracking-And-Matching.pdf).
PCT International Search Report for PCT International Patent Application No. PCT/US2013/042504, mailed Aug. 19, 2013.
PCT International Search Report for PCT International Patent Application No. PCT/US2013/042509, mailed Sep. 2, 2013.
PCT International Search Report for PCT International Patent Application No. PCT/US2013/042514, mailed Aug. 30, 2013.
PCT International Search Report for PCT International Patent Application No. PCT/US2013/042517, mailed Aug. 29, 2013.
3D Morphable Model Face Animation, http://www.youtube.com/watch?v=nice6NYb—WA, Apr. 20, 2006.
Visionix 3D iView, Human Body Measurement Newsletter, vol. 1., No. 2, Sep. 2005, pp. 2 and 3.
Blaise Aguera y Areas demos Photosynth, May 2007. Ted.com, http://www.ted.com/talks/blaise—aguera—y—arcas—demos—photosynth.html.
ERC Tecnology Leads to Eyeglass “Virtual Try-on” System, Apr. 20, 2012, http://showcase.erc-assoc.org/accomplishments/microelectronic/imsc6-eyeglass.htm.
Information about Related Patents and Patent Applications.
U.S. Appl. No. 13/775,785, filed Feb. 25, 2013, Systems and Methods for Adjusting a Virtual Try-On.
U.S. Appl. No. 13/775,764, filed Feb. 25, 2013, Systems and Methods for Feature Tracking.
U.S. Appl. No. 13/774,995, filed Feb. 22, 2013, Systems and Methods for Scaling a Three-Dimensional Model.
U.S. Appl. No. 13/774,985, filed Feb. 22, 2013, Systems and Methods for Generating a 3-D Model of a Virtual Try-On Product.
U.S. Appl. No. 13/774,983, filed Feb. 22, 2013, Systems and Methods for Generating a 3-D Model of a User for a Virtual Try-On Product.
U.S. Appl. No. 13/774,978, filed Feb. 22, 2013, Systems and Methods for Efficiently Processing Virtual 3-D Data.
U.S. Appl. No. 13/774,958, filed Feb. 22, 2013, Systems and Methods for Rendering Virtual Try-On Products.
U.S. Appl. No. 13/706,909, filed Dec. 6, 2012, Systems and Methods for Obtaining a Pupillary Distance Measurement Using a Mobile Computing Device.
PCT International Search Report for PCT International Patent Application No. PCT/US2012/068174, mailed Mar. 7, 2013.
PCT International Search Report for PCT International Patent Application No. PCT/US2013/042512, mailed Sep. 6, 2013.
PCT International Search Report for PCT International Patent Application No. PCT/US2013/042529, mailed Sep. 17, 2013.
PCT International Search Report for PCT International Patent Application No. PCT/US2013/042525, mailed Sep. 17, 2013.
PCT International Search Report for PCT International Patent Application No. PCT/US2013/042520, mailed Sep. 27, 2013.
Dror et al., Recognition of Surface Relfectance Properties form a Single Image under Unknown Real-World Illumination, IEEE, Proceedings of the IEEE Workshop on Identifying Objects Across Variations in Lighting: Psychophysics & Computation, Dec. 2011.
Simonite, 3-D Models Created by a Cell Phone, Mar. 23, 2011, url: http://www.technologyreview.com/news/423386/3-d-models-created-by-a-cell-phone/.
Fidaleo, Model-Assisted 3D Face Reconstruction from Video, AMFG'07 Analysis and Modeling of Faces and Gestures Lecture Notes in Computer Science vol. 4778, 2007, pp. 124-138.
Garcia-Mateos, Estimating 3D facial pose in video with just three points, CVPRW '08 Computer vision and Pattern Recognition Workshops, 2008.
Related Publications (1)
Number Date Country
20130315487 A1 Nov 2013 US
Provisional Applications (2)
Number Date Country
61650983 May 2012 US
61735951 Dec 2012 US