Despite increasing access to technology, people in the modern world are increasingly busy. For many, however, attention to one's appearance remains a high priority. Many people continue to invest time in maintaining and augmenting their wardrobes, shopping for special outfits, etc. In some cases, the investment in time has to do with going to a retail establishment to try on and purchase clothing and accessories. The process of selecting the right article in the right size by trying on a series of candidate articles can be very time consuming. Online shopping provides a faster alternative to the conventional retail setting. Despite its advantages, however, online shopping presents certain drawbacks. One drawback is that it may be difficult for a person to visualize how a given article would look if worn by that person—owing to the rich variation in body size and shape, hair and skin color, etc., in the human population.
One embodiment of this disclosure provides a method to help a user visualize how a wearable article will look on the user's body. Enacted on a computing system, the method includes receiving an image of the user's body from an image-capture component. Based on the image, a posable, three-dimensional, virtual avatar is constructed to substantially resemble the user. In this example method, data is obtained that identifies the wearable article as being selected for the user. This data includes a plurality of metrics that at least partly define the wearable article. Then, a virtualized form of the wearable article is attached to the avatar, which is provided to a display component for the user to review.
This Summary is provided to introduce, in simplified form, a selection of concepts that are further described in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter. Furthermore, the claimed subject matter is not limited to implementations that solve any disadvantages noted in any part of this disclosure.
As used in this disclosure, an ‘avatar’ is a posable, three-dimensional, computer representation of a human subject. An avatar can be used in various applications—e.g., video games, interactive training software, physical therapy, or in a retail setting. More generally, an avatar can be used wherever a virtual representation of the subject is desired.
One example application of an avatar in a retail setting is to enable a customer to ‘try on’, virtually, all manner of wearable articles. Such articles may include clothing, eyewear, footwear, accessories, prostheses, jewelry, tattoos, and/or make-up, as examples. Having augmented her avatar with such articles in virtualized form, the customer may be able to predict how she might look when wearing the corresponding actual articles. This approach can be used to pre-screen the articles prior to an actual visit to the fitting room, to save time. In addition, the customer may elect to share with others the images of her avatar with the attached virtual articles. In some scenarios, the sharing can be done remotely—via email or cell phone, for example—with friends or family members not physically present. In this manner, the customer may profit from another's counsel before making a decision to purchase an article. In an avatar-based on-line retail experience, the entire process of selecting an article, trying the article on, and then purchasing the article can be conducted in the privacy of the customer's home or workplace.
In the embodiment illustrated in
In one embodiment, image data from a pair of stereoscopic cameras may be co-registered and mathematically combined to yield a depth map. In other embodiments, user-input device 14A may be configured to project a structured infrared illumination comprising numerous, discrete features (e.g., lines or dots). The depth camera may be configured to image the structured illumination reflected from the subject. Based on the spacings between adjacent features in the various regions of the imaged subject, a depth map of the subject may be constructed.
In other embodiments, user-input device 14A may be configured to project a pulsed infrared illumination. A pair of cameras may be configured to detect the pulsed illumination reflected from the subject. Both cameras may include an electronic shutter synchronized to the pulsed illumination, but the integration times for the cameras may differ such that a pixel-resolved time-of-flight of the pulsed illumination, from the source to the subject and then to the cameras, is discernible from the relative amounts of light received in corresponding pixels of the two cameras.
The configurations described above enable various methods for helping a user to visualize how a wearable article will look on the user's body, for mediating an on-line retail experience, etc. Some such methods are now described, by way of example, with continued reference to the above configurations. It will be understood, however, that the methods here described, and others within the scope of this disclosure, may be enabled by different configurations as well. The methods may be entered upon any time system 10 is operating, and may be executed repeatedly.
At 30 of method 28, an image of the user's body is received from an image-capture component, such as user-input device 14A. The image includes at least a portion of the user's body, but may include the user's head and/or face as well. In the embodiments here contemplated, the image may be one of a set of images of the user received from the image-capture component. The set may include one or more depth maps.
At 32 a posable, three-dimensional, virtual avatar is constructed to substantially resemble the user. The avatar may be constructed based at least partly on the image(s) received at 30.
In one embodiment, each joint may be assigned various parameters—e.g., Cartesian coordinates specifying joint position, angles specifying joint rotation, and additional parameters specifying a conformation of the corresponding body part (hand open, hand closed, etc.). The virtual skeleton may take the form of a data structure including any or all of these parameters for each joint. In this manner, the metrical data defining the virtual skeleton—its size, shape, orientation, position, etc.—may be assigned to the joints.
Returning now to
Continuing in
At 46 the set of characteristic metrics harvested in this manner is provided as input to an algorithm configured to output a virtual body mesh resembling the subject's body. In other words, the algorithm computes the virtual body mesh as a function of the characteristic metrics. In some embodiments, the algorithm used for this purpose may be an algorithm trained using machine learning. More particularly, the algorithm may be one that has been trained using an actual human model in a range of poses and a range of human models in a single pose, as described in further detail in PCT Application PCT/CN2012/077303, the entirety of which is incorporated by reference.
Continuing in
To construct the virtual head mesh, a virtual head-mesh template may be deformed to minimize distance between points on the second depth map and corresponding points on the virtual head mesh. The virtual head mesh may then be augmented with color and texture derived from one or more image captures from the color camera of the user-input device. In this manner, the virtual head mesh may be personalized to resemble the actual human subject—viz., it may present facial features resembling those of the subject both in shape and in skin color/texture.
At 50 the virtual body mesh is connected to the virtual head mesh. In this step, the head of the virtual body template mesh is first cut out, and then is connected to the virtual head template mesh by triangulating the two open boundaries of the template meshes. The connected model is then stored in the system and loaded when the virtual body mesh and the virtual head mesh are ready. The two template meshes are replaced by two virtual meshes, respectively, since they have the same connectivities. The scale of the virtual head mesh is adjusted according to the proportion consistent with the virtual body mesh. The vertices around the neck are also smoothed, while the other vertices are held fixed. In this manner, a geometrically realistic and seamless head/body mesh may be constructed.
At 52 the virtual body mesh is augmented with a skin color and/or skin texture appropriate for the subject. In some embodiments, the skin color and/or skin texture may be selected based on color-image data from user-input device 14A—such as color image data from a region that includes the subject's face. In other words, the skin color and/or skin texture applied to the virtual body mesh may be synthesized to match that of the subject's face. In one embodiment, the system first selects a body texture image in a pre-designed database and then modulates the low frequency color component so that its overall color is consistent with the color of the face skin.
Returning now to
In method 58A, the wearable article is selected for the user via input from another person. In one embodiment, the other person may be a friend or relative of the user. In another embodiment, the other person may be selected by a local or remote computing system, such as computing system 16A, based on suitable criteria. For example, the other person may be selected because that person exhibits purchasing behaviors similar to that of the user—a history of purchasing the same actual items, or items of the same style or price range, for example.
At 60 a mechanism is provided, via computing system 16A, for the user to request such input from the other person. In any convenient manner, the user may open a channel to receive from the other person suggestions that identify one or more wearable articles. The input from the other person may take several different forms. One example form of input may include an indication of one or more outfits worn by the other person, and suggested for the user. When a suggestion is received through an open channel, the suggested wearable article may, at 62, be selected for the user by computing system 16A. At 64 relevant data—e.g., metrical data, material data, color data, etc.—that define the wearable article—are retrieved. In one embodiment, the data may be retrieved from a database maintained by a vendor or manufacturer of the article. From 64 the method returns.
In some embodiments, the various channels that transmit suggestions for wearable articles and/or requests for suggestions can be email channels, text- or picture-messaging channels, tweets, and/or communications channels of one or more social networks.
Turning now to
Returning now to the method as illustrated, at 66 the inventory accessible to the user is parsed for a wearable article. In subsequent decision blocks of method 58B, the wearable article is analyzed with respect to selected criteria. It will be understood that the criteria chosen for inclusion in
At 68 of method 58B, it is determined whether the article is appropriate for current or predicted weather conditions. If the wearable article is appropriate for the weather conditions, then the method advances to 70. However, if the wearable article is not appropriate, then the method returns to 66, where the inventory is again parsed for the next wearable article. Accordingly, the selection engine may be configured to select the wearable article based at least partly on current or predicted weather in the user's locality.
At 70 it is determined whether the wearable article is available to be worn. If the article is available to be worn, then the method advances to 72. However, if the article is not available to be worn, then the method returns to 66. In this manner, the inventory of articles accessible to the user may exclude articles currently unavailable for wearing. Accordingly, the illustrated method may cooperate with an associated method to track a plurality of articles in the user's wardrobe to determine which items are currently available. An article may be unavailable because it is in the laundry or at the cleaner's, on loan to another person, etc.
In a related embodiment, an associated method may track when each article in the user's wardrobe was last cleaned, and thereby exclude articles that are in need of cleaning. In another related embodiment, an associated method may track articles in the user's wardrobe to determine how frequently each item has been worn. Accordingly, the inventory of articles accessible to the user may exclude articles worn by the user above a threshold frequency. This feature is borne out at 72 of method 58B, where it is determined whether the wearable article has been worn too much. If the wearable article has not been worn too much, then the method advances to 74. However, if the article has been worn too much, then the method returns to 66.
At 74 it is determined whether the wearable article meets selected match criteria. Thus, the wearable article referred to hereinabove may be a wearable first article; the selection engine may be configured to select the first article based on one or more of color, style, and/or brand matching with respect to a wearable second article. In one embodiment, the wearable second article may be an article already selected by the user for wearing and/or purchase. In another embodiment, the wearable second article may be an article already in a wardrobe of the user. Accordingly, the selection engine may be configured to select the first article from an inventory of articles for sale to the user based on one or, more of color, style, and/or brand matching with respect to a second article already in the user's wardrobe.
Furthermore, the second article can be one of a plurality of articles in the user's wardrobe to which the first article is matched. Method 58B can be used, in a sense, to fill gaps in the user's wardrobe. This feature can be enacted according to an algorithm that searches for a set of articles found in the wardrobe of the user. The target domain for the search may include the wardrobes of selected other users on a network. If the set is found in the wardrobes of a significant number of other users, together with an additional article not included in the set, then the algorithm may return this additional item as the selected first wearable article of method 58B.
Continuing in method 58B, if the wearable article does meet the match criteria, then the method advance to 64. Otherwise, the method returns to 66. At 64 data on the wearable article is retrieved, in the manner described above for method 58A. From 64 the method returns.
No aspect of methods 58A or 58B should be understood in a limiting sense, for a wearable article may be selected for the user in other ways too. For instance, the wearable article may be selected through an automated service of a vendor of the wearable article. In some embodiments, the user may select from a plurality of such services offered by the vendor—services associated with wearable articles endorsed by different celebrities (e.g., the Joe Hammerswing line, the Bonnie Belle line, etc.).
Returning again to
At 78 a virtualized form of wearable article is attached to the avatar. At this stage of processing, the metrical data for the selected article may be adapted based on the underlying topology of the avatar. Material data for the wearable article may be used to determine how the wearable article will conform to the topology. For instance, the resiliency of a fabric may be used as an indication of the degree to which the fabric will stretch to accommodate the user's body. In embodiments where a plurality of wearable articles are to be reviewed simultaneously, virtualized forms of a first, second, third wearable article, etc., may be attached to the avatar.
At 80 the avatar, with the virtualized form of the wearable article(s) attached, is provided to a display component—e.g., display 18A of
In the approach contemplated herein, virtually any input mechanism may be used to specify the initial and subsequent desired poses for the avatar. Such mechanisms may include spoken commands directing body movement, or selection from among a menu of body movements and/or gestures via a user interface. In yet another embodiment, real-time skeletal tracking with user-input device 14A may guide the movement of the animated avatar. More particularly, the user may specify the movement of the avatar simply by moving her own body in the manner desired for the avatar. The user-input device may be configured to track the movement of the user's body, and provide a stream of gesture data to the personal computer on which method 88 is enacted.
In certain embodiments, the user may optionally authorize that the presentation of her avatar, with wearable article(s) attached, be shared with at least one other person. Accordingly, the user may request that the other person post a comment concerning how the user's avatar looks while wearing the item. By tracking the other person's comment, the user may profit from the other person's advice.
Returning again to
The subsequent steps of method 28 may be enacted once the user has agreed to the selection, has received the selected wearable article (if the article was purchased) and put on the article. At 98 one or more images of user are received while the user is wearing the actual article is received. At 100 these images are saved for subsequent review by the user. In this manner, the user can maintain a record of how she actually looks in each item of her wardrobe.
Aspects of this disclosure are set forth by example, with reference to the illustrated embodiments described above. Components, process steps, and other elements that may be substantially the same in one or more embodiments are identified coordinately and are described with minimal repetition. It will be noted, however, that elements identified coordinately may also differ to some degree. It will be further noted that the drawing figures included in this disclosure are schematic and generally not drawn to scale. Rather, the various drawing scales, aspect ratios, and numbers of components shown in the figures may be purposely distorted to make certain features or relationships easier to see.
In the methods illustrated and/or described herein, some of the indicated process steps may be omitted without departing from the scope of this disclosure. Likewise, the indicated sequence of the process steps may not always be required to achieve the intended results, but is provided for ease of illustration and description. One or more of the illustrated actions, functions, or operations may be performed repeatedly, depending on the particular strategy being used.
In some embodiments, the methods and processes described above may be tied to a computing system of one or more computing devices. In particular, such methods and processes may be implemented as a computer-application program or service, an application-programming interface (API), a library, and/or other computer-program product.
Computing system 16 includes a logic subsystem 20 and a storage subsystem 22. Computing system 16 may optionally include a display subsystem 18, input subsystem 14, communication subsystem 102, and/or other components not shown in
Logic subsystem 20 includes one or more physical devices configured to execute instructions. For example, the logic subsystem may be configured to execute instructions that are part of one or more applications, services, programs, routines, libraries, objects, components, data structures, or other logical constructs. Such instructions may be implemented to perform a task, implement a data type, transform the state of one or more components, or otherwise arrive at a desired result.
The logic subsystem may include one or more processors configured to execute software instructions. Additionally or alternatively, the logic subsystem may include one or more hardware or firmware logic machines configured to execute hardware or firmware instructions. The processors of the logic subsystem may be single-core or multi-core, and the programs executed thereon may be configured for sequential, parallel or distributed processing. The logic subsystem may optionally include individual components that are distributed among two or more devices, which can be remotely located and/or configured for coordinated processing. Aspects of the logic subsystem may be virtualized and executed by remotely accessible networked computing devices configured in a cloud-computing configuration.
Storage subsystem 22 includes one or more physical, non-transitory, devices configured to hold data and/or instructions executable by the logic subsystem to implement the herein-described methods and processes. When such methods and processes are implemented, the state of storage subsystem 22 may be transformed—e.g., to hold different data.
Storage subsystem 22 may include removable media and/or built-in devices. Storage subsystem 22 may include optical memory devices (e.g., CD, DVD, HD-DVD, Blu-Ray Disc, etc.), semiconductor memory devices (e.g., RAM, EPROM, EEPROM, etc.) and/or magnetic memory devices (e.g., hard-disk drive, floppy-disk drive, tape drive, MRAM, etc.), among others. Storage subsystem 22 may include volatile, nonvolatile, dynamic, static, read/write, read-only, random-access, sequential-access, location-addressable, file-addressable, and/or content-addressable devices.
It will be appreciated that storage subsystem 22 includes one or more physical, non-transitory devices. However, in some embodiments, aspects of the instructions described herein may be propagated in a transitory fashion by a pure signal—e.g., an electromagnetic or optical signal, etc.—that is not held by a physical device for a finite duration. Furthermore, data and/or other forms of information pertaining to the present disclosure may be propagated by a pure signal.
In some embodiments, aspects of logic subsystem 20 and of storage subsystem 22 may be integrated together into one or more hardware-logic components through which the functionally described herein may be enacted, at least in part. Such hardware-logic components may include field-programmable gate arrays (FPGAs), program- and application-specific integrated circuits (PASIC/ASICs), program- and application-specific standard products (PSSP/ASSPs), system-on-a-chip (SOC) systems, and complex programmable logic devices (CPLDs), for example.
The terms ‘module,’ ‘program,’ and ‘engine’ may be used to describe an aspect of computing system 16 implemented to perform a particular function. In some cases, a module, program, or engine may be instantiated via logic subsystem 20 executing instructions held by storage subsystem 22. Accordingly,
It will be appreciated that a ‘service’, as used herein, is an application program executable across multiple user sessions. A service may be available to one or more system components, programs, and/or other services. In some implementations, a service may run on one or more server-computing devices.
When included, display subsystem 18 may be used to present a visual representation of data held by storage subsystem 22. This visual representation may take the form of a graphical user interface (GUI). As the herein described methods and processes change the data held by the storage subsystem, and thus transform the state of the storage subsystem, the state of display subsystem 18 may likewise be transformed to visually represent changes in the underlying data. Display subsystem 18 may include one or more display devices utilizing virtually any type of technology. Such display devices may be combined with logic subsystem 20 and/or storage subsystem 22 in a shared enclosure, or such display devices may be peripheral display devices.
When included, communication subsystem 102 may be configured to communicatively couple computing system 16 with one or more other computing devices. Communication subsystem 102 may include wired and/or wireless communication devices compatible with one or more different communication protocols. As non-limiting examples, the communication subsystem may be configured for communication via a wireless telephone network, or a wired or wireless local- or wide-area network. In some embodiments, the communication subsystem may allow computing system 16 to send and/or receive messages to and/or from other devices via a network such as the Internet.
Finally, it will be noted that the subject matter of the present disclosure includes all novel and non-obvious combinations and sub-combinations of the various processes, systems and configurations, and other features, functions, acts, and/or properties disclosed herein, as well as any and all equivalents thereof.
Number | Date | Country | Kind |
---|---|---|---|
PCT/CN2012/077303 | Jun 2012 | WO | international |
This application is a continuation-in-part of U.S. patent application Ser. No. 12/752,917 filed 1 Apr. 2010 and entitled MOTION-BASED INTERACTIVE SHOPPING ENVIRONMENT, and of Patent Cooperation Treaty (PCT) Application PCT/CN2012/077303 filed 21 Jun. 2012 and entitled AVATAR CONSTRUCTION USING DEPTH CAMERA. The entirety of these applications is hereby incorporated by reference herein for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
4627620 | Yang | Dec 1986 | A |
4630910 | Ross et al. | Dec 1986 | A |
4645458 | Williams | Feb 1987 | A |
4695953 | Blair et al. | Sep 1987 | A |
4702475 | Elstein et al. | Oct 1987 | A |
4711543 | Blair et al. | Dec 1987 | A |
4751642 | Silva et al. | Jun 1988 | A |
4796997 | Svetkoff et al. | Jan 1989 | A |
4809065 | Harris et al. | Feb 1989 | A |
4817950 | Goo | Apr 1989 | A |
4843568 | Krueger et al. | Jun 1989 | A |
4893183 | Nayar | Jan 1990 | A |
4901362 | Terzian | Feb 1990 | A |
4925189 | Braeunig | May 1990 | A |
5101444 | Wilson et al. | Mar 1992 | A |
5148154 | MacKay et al. | Sep 1992 | A |
5184295 | Mann | Feb 1993 | A |
5229754 | Aoki et al. | Jul 1993 | A |
5229756 | Kosugi et al. | Jul 1993 | A |
5239463 | Blair et al. | Aug 1993 | A |
5239464 | Blair et al. | Aug 1993 | A |
5288078 | Capper et al. | Feb 1994 | A |
5295491 | Gevins | Mar 1994 | A |
5320538 | Baum | Jun 1994 | A |
5347306 | Nitta | Sep 1994 | A |
5385519 | Hsu et al. | Jan 1995 | A |
5405152 | Katanics et al. | Apr 1995 | A |
5417210 | Funda et al. | May 1995 | A |
5423554 | Davis | Jun 1995 | A |
5454043 | Freeman | Sep 1995 | A |
5469740 | French et al. | Nov 1995 | A |
5495576 | Ritchey | Feb 1996 | A |
5516105 | Eisenbrey et al. | May 1996 | A |
5524637 | Erickson et al. | Jun 1996 | A |
5534917 | MacDougall | Jul 1996 | A |
5563988 | Maes et al. | Oct 1996 | A |
5577981 | Jarvik | Nov 1996 | A |
5580249 | Jacobsen et al. | Dec 1996 | A |
5594469 | Freeman et al. | Jan 1997 | A |
5597309 | Riess | Jan 1997 | A |
5616078 | Oh | Apr 1997 | A |
5617312 | Iura et al. | Apr 1997 | A |
5638300 | Johnson | Jun 1997 | A |
5641288 | Zaenglein | Jun 1997 | A |
5682196 | Freeman | Oct 1997 | A |
5682229 | Wangler | Oct 1997 | A |
5690582 | Ulrich et al. | Nov 1997 | A |
5703367 | Hashimoto et al. | Dec 1997 | A |
5704837 | Iwasaki et al. | Jan 1998 | A |
5715834 | Bergamasco et al. | Feb 1998 | A |
5875108 | Hoffberg et al. | Feb 1999 | A |
5877803 | Wee et al. | Mar 1999 | A |
5913727 | Ahdoot | Jun 1999 | A |
5933125 | Fernie | Aug 1999 | A |
5980256 | Carmein | Nov 1999 | A |
5989157 | Walton | Nov 1999 | A |
5995649 | Marugame | Nov 1999 | A |
6005548 | Latypov et al. | Dec 1999 | A |
6009210 | Kang | Dec 1999 | A |
6054991 | Crane et al. | Apr 2000 | A |
6066075 | Poulton | May 2000 | A |
6072494 | Nguyen | Jun 2000 | A |
6073489 | French et al. | Jun 2000 | A |
6077201 | Cheng et al. | Jun 2000 | A |
6098458 | French et al. | Aug 2000 | A |
6100896 | Strohecker et al. | Aug 2000 | A |
6101289 | Kellner | Aug 2000 | A |
6128003 | Smith et al. | Oct 2000 | A |
6130677 | Kunz | Oct 2000 | A |
6141463 | Covell et al. | Oct 2000 | A |
6147678 | Kumar et al. | Nov 2000 | A |
6152856 | Studor et al. | Nov 2000 | A |
6159100 | Smith | Dec 2000 | A |
6173066 | Peurach et al. | Jan 2001 | B1 |
6181343 | Lyons | Jan 2001 | B1 |
6188777 | Darrell et al. | Feb 2001 | B1 |
6195104 | Lyons | Feb 2001 | B1 |
6215890 | Matsuo et al. | Apr 2001 | B1 |
6215898 | Woodfill et al. | Apr 2001 | B1 |
6226396 | Marugame | May 2001 | B1 |
6229913 | Nayar et al. | May 2001 | B1 |
6256033 | Nguyen | Jul 2001 | B1 |
6256400 | Takata et al. | Jul 2001 | B1 |
6283860 | Lyons et al. | Sep 2001 | B1 |
6289112 | Jain et al. | Sep 2001 | B1 |
6299308 | Voronka et al. | Oct 2001 | B1 |
6308565 | French et al. | Oct 2001 | B1 |
6316934 | Amorai-Moriya et al. | Nov 2001 | B1 |
6363160 | Bradski et al. | Mar 2002 | B1 |
6384819 | Hunter | May 2002 | B1 |
6411744 | Edwards | Jun 2002 | B1 |
6430997 | French et al. | Aug 2002 | B1 |
6476834 | Doval et al. | Nov 2002 | B1 |
6496598 | Harman | Dec 2002 | B1 |
6503195 | Keller et al. | Jan 2003 | B1 |
6539931 | Trajkovic et al. | Apr 2003 | B2 |
6570555 | Prevost et al. | May 2003 | B1 |
6633294 | Rosenthal et al. | Oct 2003 | B1 |
6640202 | Dietz et al. | Oct 2003 | B1 |
6661918 | Gordon et al. | Dec 2003 | B1 |
6681031 | Cohen et al. | Jan 2004 | B2 |
6714665 | Hanna et al. | Mar 2004 | B1 |
6731799 | Sun et al. | May 2004 | B1 |
6738066 | Nguyen | May 2004 | B1 |
6765726 | French et al. | Jul 2004 | B2 |
6788809 | Grzeszczuk et al. | Sep 2004 | B1 |
6801637 | Voronka et al. | Oct 2004 | B2 |
6873723 | Aucsmith et al. | Mar 2005 | B1 |
6876496 | French et al. | Apr 2005 | B2 |
6901379 | Balter et al. | May 2005 | B1 |
6937742 | Roberts et al. | Aug 2005 | B2 |
6950534 | Cohen et al. | Sep 2005 | B2 |
7003134 | Covell et al. | Feb 2006 | B1 |
7036094 | Cohen et al. | Apr 2006 | B1 |
7038855 | French et al. | May 2006 | B2 |
7039676 | Day et al. | May 2006 | B1 |
7042440 | Pryor et al. | May 2006 | B2 |
7050606 | Paul et al. | May 2006 | B2 |
7058204 | Hildreth et al. | Jun 2006 | B2 |
7060957 | Lange et al. | Jun 2006 | B2 |
7062454 | Giannini et al. | Jun 2006 | B1 |
7113918 | Ahmad et al. | Sep 2006 | B1 |
7121946 | Paul et al. | Oct 2006 | B2 |
7133839 | Inoue et al. | Nov 2006 | B2 |
7170492 | Bell | Jan 2007 | B2 |
7184048 | Hunter | Feb 2007 | B2 |
7202898 | Braun et al. | Apr 2007 | B1 |
7222078 | Abelow | May 2007 | B2 |
7227526 | Hildreth et al. | Jun 2007 | B2 |
7259747 | Bell | Aug 2007 | B2 |
7308112 | Fujimura et al. | Dec 2007 | B2 |
7317836 | Fujimura et al. | Jan 2008 | B2 |
7346561 | Devitt et al. | Mar 2008 | B1 |
7348963 | Bell | Mar 2008 | B2 |
7359121 | French et al. | Apr 2008 | B2 |
7367887 | Watabe et al. | May 2008 | B2 |
7379563 | Shamaie | May 2008 | B2 |
7379566 | Hildreth | May 2008 | B2 |
7389591 | Jaiswal et al. | Jun 2008 | B2 |
7412077 | Li et al. | Aug 2008 | B2 |
7421093 | Hildreth et al. | Sep 2008 | B2 |
7430312 | Gu | Sep 2008 | B2 |
7436496 | Kawahito | Oct 2008 | B2 |
7450736 | Yang et al. | Nov 2008 | B2 |
7452275 | Kuraishi | Nov 2008 | B2 |
7460690 | Cohen et al. | Dec 2008 | B2 |
7489812 | Fox et al. | Feb 2009 | B2 |
7536032 | Bell | May 2009 | B2 |
7555142 | Hildreth et al. | Jun 2009 | B2 |
7560701 | Oggier et al. | Jul 2009 | B2 |
7570805 | Gu | Aug 2009 | B2 |
7574020 | Shamaie | Aug 2009 | B2 |
7576727 | Bell | Aug 2009 | B2 |
7590262 | Fujimura et al. | Sep 2009 | B2 |
7593552 | Higaki et al. | Sep 2009 | B2 |
7598942 | Underkoffler et al. | Oct 2009 | B2 |
7607509 | Schmiz et al. | Oct 2009 | B2 |
7620202 | Fujimura et al. | Nov 2009 | B2 |
7668340 | Cohen et al. | Feb 2010 | B2 |
7680298 | Roberts et al. | Mar 2010 | B2 |
7683954 | Ichikawa et al. | Mar 2010 | B2 |
7684592 | Paul et al. | Mar 2010 | B2 |
7701439 | Hillis et al. | Apr 2010 | B2 |
7702130 | Im et al. | Apr 2010 | B2 |
7704135 | Harrison, Jr. | Apr 2010 | B2 |
7710391 | Bell et al. | May 2010 | B2 |
7729530 | Antonov et al. | Jun 2010 | B2 |
7746345 | Hunter | Jun 2010 | B2 |
7760182 | Ahmad et al. | Jul 2010 | B2 |
7809167 | Bell | Oct 2010 | B2 |
7834846 | Bell | Nov 2010 | B1 |
7852262 | Namineni et al. | Dec 2010 | B2 |
RE42256 | Edwards | Mar 2011 | E |
7898522 | Hildreth et al. | Mar 2011 | B2 |
8035612 | Bell et al. | Oct 2011 | B2 |
8035614 | Bell et al. | Oct 2011 | B2 |
8035624 | Bell et al. | Oct 2011 | B2 |
8072470 | Marks | Dec 2011 | B2 |
9098873 | Geisner et al. | Aug 2015 | B2 |
20010026272 | Feld et al. | Oct 2001 | A1 |
20010034668 | Whitworth | Oct 2001 | A1 |
20020090985 | Tochner et al. | Jul 2002 | A1 |
20020121980 | Wan et al. | Sep 2002 | A1 |
20040085310 | Snuffer | May 2004 | A1 |
20040155962 | Marks | Aug 2004 | A1 |
20050128211 | Berger | Jun 2005 | A1 |
20060287877 | Wannier et al. | Dec 2006 | A1 |
20070225859 | Kolesnychenko et al. | Sep 2007 | A1 |
20070229498 | Matusik et al. | Oct 2007 | A1 |
20080026838 | Dunstan et al. | Jan 2008 | A1 |
20080071559 | Arrasvuori | Mar 2008 | A1 |
20080163344 | Yang | Jul 2008 | A1 |
20080252637 | Berndt et al. | Oct 2008 | A1 |
20090158220 | Zalewski et al. | Jun 2009 | A1 |
20090163262 | Kang | Jun 2009 | A1 |
20090215533 | Zalewski et al. | Aug 2009 | A1 |
20100030578 | Siddique et al. | Feb 2010 | A1 |
20100070384 | Kruusmaa et al. | Mar 2010 | A1 |
20100097375 | Tadaishi et al. | Apr 2010 | A1 |
20100111370 | Black et al. | May 2010 | A1 |
20100134490 | Corazza et al. | Jun 2010 | A1 |
20100302138 | Poot et al. | Dec 2010 | A1 |
20110022965 | Lawrence | Jan 2011 | A1 |
20110025689 | Perez et al. | Feb 2011 | A1 |
20110043610 | Ren et al. | Feb 2011 | A1 |
20110148864 | Lee et al. | Jun 2011 | A1 |
20110157306 | Lin et al. | Jun 2011 | A1 |
20110234581 | Eikelis et al. | Sep 2011 | A1 |
20110234589 | Lee et al. | Sep 2011 | A1 |
20110292036 | Sali et al. | Dec 2011 | A1 |
20120019517 | Corazza et al. | Jan 2012 | A1 |
20120070070 | Litvak | Mar 2012 | A1 |
20120086783 | Sareen | Apr 2012 | A1 |
20120113106 | Choi et al. | May 2012 | A1 |
20120183238 | Savvides et al. | Jul 2012 | A1 |
Number | Date | Country |
---|---|---|
1928908 | Mar 2007 | CN |
101254344 | Sep 2008 | CN |
102201099 | Sep 2011 | CN |
102221886 | Oct 2011 | CN |
102222329 | Oct 2011 | CN |
102222347 | Oct 2011 | CN |
102508867 | Jun 2012 | CN |
0583061 | Feb 1994 | EP |
08044490 | Feb 1996 | JP |
2004502997 | Jan 2004 | JP |
2013533537 | Aug 2013 | JP |
9310708 | Jun 1993 | WO |
9717598 | May 1997 | WO |
9944698 | Sep 1999 | WO |
2009026726 | Mar 2009 | WO |
2013189058 | Dec 2013 | WO |
Entry |
---|
“Fast Penetration Resolving for Multi-Layered Virtual Garment Dressing,” by Yueqi Zhong, Textile Research Journal, vol. 79 (9): 815-821 (Jun. 2009). |
Kanade et al., “A Stereo Machine for Video-rate Dense Depth Mapping and Its New Applications”, IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Jun. 1996, pp. 196-202,The Robotics Institute, Carnegie Mellon University, Pittsburgh, PA. |
Miyagawa et al., “CCD-Based Range Finding Sensor”, Oct. 1997, pp. 1648-1652, vol. 44 No. 10, IEEE Transactions on Electron Devices. |
Rosenhahn et al., “Automatic Human Model Generation”, Sep. 2005, pp. 41-48, University of Auckland (CITR), New Zealand. |
Aggarwal et al., “Human Motion Analysis: A Review”, IEEE Nonrigid and Articulated Motion Workshop, Jun. 1997, University of Texas at Austin, Austin, TX., 13 pages. |
Shao et al., “An Open System Architecture for a Multimedia and Multimodal User Interface”, Aug. 24, 1998, Japanese Society for Rehabilitation of Persons with Disabilities (JSRPD), Japan. 8 pages. |
Kohler, “Special Topics of Gesture Recognition Applied in Intelligent Home Environments”, In Proceedings of the Gesture Workshop, Sep. 1998, pp. 285-296, Germany. |
Kohler, “Vision Based Remote Control in Intelligent Home Environments”, University of Erlangen-Nuremberg/Germany, 1996, pp. 147-154, Germany. |
Kohler, “Technical Details and Ergonomical Aspects of Gesture Recognition applied in Intelligent Home Environments”, 1997, Germany. |
Hasegawa et al., “Human-Scale Haptic Interaction with a Reactive Virtual Human in a Real-Time Physics Simulator”, Jul. 2006, vol. 4, No. 3, Article 6C, ACM Computers in Entertainment, New York, NY. |
Qian et al., “A Gesture-Driven Multimodal Interactive Dance System”, Jun. 2004, pp. 1579-1582, IEEE International Conference on Multimedia and Expo (ICME), Taipei, Taiwan. |
Zhao, “Dressed Human Modeling, Detection, and Parts Localization”, 2001, The Robotics Institute, Carnegie Mellon University, Pittsburgh, PA. |
He, “Generation of Human Body Models”, Apr. 2005, University of Auckland, New Zealand. |
Isard et al., “CONDENSATION—Conditional Density Propagation for Visual Tracking”, Aug. 1998, pp. 5-28, International Journal of Computer Vision 29(1), Netherlands. |
Livingston, “Vision-based Tracking with Dynamic Structured Light for Video See-through Augmented Reality”, 1998, University of North Carolina at Chapel Hill, North Carolina, USA. 145 pages. |
Wren et al., “Pfinder: Real-Time Tracking of the Human Body”, MIT Media Laboratory Perceptual Computing Section Technical Report No. 353, Jul. 1997, vol. 19, No. 7, pp. 780-785, IEEE Transactions on Pattern Analysis and Machine Intelligence, Caimbridge, MA. |
Breen et al., “Interactive Occlusion and Collusion of Real and Virtual Objects in Augmented Reality”, Technical Report ECRC-95-02, Feb. 1995, European Computer-Industry Research Center GmbH, Munich, Germany, 22 pages. |
Freeman et al., “Television Control by Hand Gestures”, Dec. 1994, Mitsubishi Electric Research Laboratories, TR94-24, Caimbridge, MA, 7 pages. |
Hongo et al., “Focus of Attention for Face and Hand Gesture Recognition Using Multiple Cameras”, Mar. 2000, pp. 156-161, 4th IEEE International Conference on Automatic Face and Gesture Recognition, Grenoble, France. |
Pavlovic et al., “Visual Interpretation of Hand Gestures for Human-Computer Interaction: A Review”, Jul. 1997, pp. 677-695, vol. 19, No. 7, IEEE Transactions on Pattern Analysis and Machine Intelligence. |
Azarbayejani et al., “Visually Controlled Graphics”, Jun. 1993, vol. 15, No. 6, IEEE Transactions on Pattern Analysis and Machine Intelligence, 4 pages. |
Granieri et al., “Simulating Humans in VR”, The British Computer Society, Oct. 1994, Academic Press, 15 pages. |
Brogan et al., “Dynamically Simulated Characters in Virtual Environments”, Sep./Oct. 1998, pp. 2-13, vol. 18, Issue 5, IEEE Computer Graphics and Applications. |
Fisher et al., “Virtual Environment Display System”, ACM Workshop on Interactive 3D Graphics, Oct. 1986, Chapel Hill, NC, 12 pages. |
“Virtual High Anxiety”, Tech Update, Aug. 1995, pp. 22. |
Sheridan et al., “Virtual Reality Check”, Technology Review, Oct. 1993, pp. 22-28, vol. 96, No. 7. |
Stevens, “Flights into Virtual Reality Treating Real World Disorders”, The Washington Post, Mar. 27, 1995, Science Psychology, 2 pages. |
“Simulation and Training”, 1994, Division Incorporated, 6 pages. |
Aitpayev, et al., “Creation of 3D Human Avatar using Kinect”, Retrieved at <<http://www.asian-transactions.org/Journals/Vol01Issue06/ATFECM/ATFECM-30120064.pdf>>, Proceedings: Asian Transactions on Fundamentals of Electronics, Communication & Multimedia (ATFECM), Jan. 2012, vol. 01 Issue 05, pp. 1-3. |
Anguelov, D. et al., “SCAPE: Shape Completion and Animation of People”, Proceedings of ACM SIGGRAPH 2005, pp. 408-416, Jul. 2005, 9 pages. |
“Behind an Inspired Room. Or, how I revamped a living room from 793 miles away!”, Blog website, InspiredRoomDesign.com, Sep. 15, 2009, Inspired Room Design, LLC, Site by BrightBridge Studios, http://inspiredroomdesign.com/blog.php?post=19. |
State Intellectual Property Office of China, Office Action issued in Chinese Patent Application No. 201110087429.5, Jul. 30, 2013, 13 pages. |
State Intellectual Property Office of China, Office Action issued in Chinese Patent Application No. 201110087429.5, Dec. 4, 2012, 6 pages. |
Jojic, “A Framework for Garment Shopping Over the Internet”, to appear in May 1999 in Handbook of Electronic Commerce, edited by Mike Shaw, Springer Verlag, 22 pages. |
“Sears Transforms the Online Shopping Experience with Help From IBM and My Virtual Model”, Press Release, Mar. 9, 2009, http://www-01.ibm.com/software/success/cssdb.nsf/CS/CPOR-7PW4TU?OpenDocument&Site=default&cty=en—us, 3 pages. |
Toyama, K. et al., “Probabilistic Tracking in a Metric Space”, Eighth International Conference on Computer Vision, Vancouver, Canada, vol. 2, Jul. 2001, 8 pages. |
Zhou, S. et al., “Parametric Reshaping of Human Bodies in Images”, Proceedings of ACM SIGGRAPH 2010, Article No. 126, Jul. 2010, 10 pages. |
State Intellectual Property Office of China, International Search Report of PCT/CN2012/077303, WIPO, Mar. 28, 2013, 3 pages. |
ISA European Patent Office, International Search Report and Written Opinion Issued in PCT/US2013/053115, Feb. 11, 2015, Netherlands, 9 Pages. |
Javed, et al., “Plan for Integration”, Retrieved at <<http://taik.aalto.fi/fi/research/designium/publications/smart—store—s53—96—n.pdf>>, Retrieved Date: Dec. 27, 2011, pp. 53-96. |
Ashdown, et al., “S08-CR03 Development of Visual Fit Assessment Tool for Apparel Firm”, Retrieved at <<http://www.ntcresearch.org/pdf-rpts/AnRp08/S08-CR03-A8.pdf>>, In National Textile Center Annual Report, Nov. 2008, pp. 1-7. |
Hauswiesner, et al., “Free Viewpoint Virtual Try-On With Commodity Depth Cameras”, Retrieved at <<http://www.icg.tugraz.at/kinectclothpaper>>, In Proceedings of the 10th International Conference on Virtual Reality Continuum and Its Applications in Industry, Dec. 11-12, 2011, pp. 8. |
“AdDude's Insights”, Retrieved at <<http://lifson.com/?tag=augmented-reality>>, Apr. 9, 2010, pp. 13. |
Mhatre, Nachiket., “Kinect Delivers Augmented Reality Shopping”, Retrieved at <<http://www.techtree.com/India/News/Kinect—Delivers—Augmented—Reality—Shopping/551-115193-585.html>>, May 11, 2011, pp. 3. |
Ferenstein, Gregory., “KinectShop: The Next Generation of Shopping”, Retrieved at <<http://www.fastcompany.com/1758674/the-next-generation-of-shopping-kinectshop-exclusive>>, Jun. 9, 2011, pp. 4. |
“Fits.me Virtual Fitting Room”, Retrieved at <<http://www.fits.me/>>, Dec. 27, 2011, pp. 1. |
Paul, R. et al., “Brave New Cyberworld: The Employer's Legal Guide to the Interactive Internet,” Computer and Internet Lawyer, vol. 27, No. 3, Mar. 2010, 27 pages. |
State Intellectual Property Office of the People's Republic of China, First Office Action Issued in Application No. 201380040978.4, Sep. 1, 2016, 18 pages. |
Number | Date | Country | |
---|---|---|---|
20120299912 A1 | Nov 2012 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 12752917 | Apr 2010 | US |
Child | 13565586 | US |