On-line shopping for commodity items such as books and tools can be accomplished with little anxiety about whether the item will be suitable after delivery to the consumer. However, personal items are not in that category of safe purchases, with concerns about color, texture, and fit not addressed until the item is delivered. 3D modeling has been proposed as a solution to such concerns, for example, for previewing the fit and drape of clothing. But such modeling requires complex and data-intense processing on even high-end platforms, particularly for netbook and handheld devices.
A system for displaying 3D objects uses a hierarchy of computing platforms to divide and process three dimensional (3D) model data before rendering final images for display on simple user devices including, but not limited to netbooks, mobile phones, gaming devices, laptop and desktop computers, using only an out-of-the-box web browser. Different backgrounds and lighting conditions are supported, and in the case of clothing, different styles of clothing as well as fabric type, color, and patterns can be simulated on an animated mannequins. Unlike slower ray-tracing rendering used in feature films, the 3D images can be calculated, rendered, and delivered at frame rates at or near full motion video even on limited function viewing platforms.
Once rendered, the final still frames or animations can be delivered to multiple platforms, allowing users to share an experience, such as apparel and accessory selection.
The mannequins may be selected from a palette of mannequins representing different body styles or may be customized to a person's exact measurements. Modeling the physics of a fabric allow the motion of the mannequin to present a user with the fit, flow, and drape of a garment over a body in motion from different viewing positions and in different lighting conditions. A more complete discussion of this process is available in the above-referenced patent application.
As opposed to shopping in a mall environment, a user of the system can view an apparel item in an appropriate setting and lighting condition, such as a swimsuit at the beach in bright sun or an evening gown worn at a ballroom under dimmed lights. Additionally, a user can view apparel items from a retailer in combination with other garments or accessories already owned by the user or available from another retailer.
In the case of clothing, a virtual closet of clothes and accessories may be built for use in mix and match planning with clothing already owned or contemplated for purchase. A virtual clothing environment also allows a person to mix and match apparel and accessories with friends and family.
The technique is also applicable to other 3D modeling applications, such as furniture in a room, window dressings, interior/exterior colors on an automobile, etc., where lighting, fabric/surface characteristics, viewing angle, and background play a role in overall perception.
Although the following text sets forth a detailed description of numerous different embodiments, it should be understood that the legal scope of the description is defined by the words of the claims set forth at the end of this disclosure. The detailed description is to be construed as exemplary only and does not describe every possible embodiment since describing every possible embodiment would be impractical, if not impossible. Numerous alternative embodiments could be implemented, using either current technology or technology developed after the filing date of this patent, which would still fall within the scope of the claims.
It should also be understood that, unless a term is expressly defined in this patent using the sentence “As used herein, the term ‘_’ is hereby defined to mean . . . ” or a similar sentence, there is no intent to limit the meaning of that term, either expressly or by implication, beyond its plain or ordinary meaning, and such term should not be interpreted to be limited in scope based on any statement made in any section of this patent (other than the language of the claims). To the extent that any term recited in the claims at the end of this patent is referred to in this patent in a manner consistent with a single meaning, that is done for sake of clarity only so as to not confuse the reader, and it is not intended that such claim term by limited, by implication or otherwise, to that single meaning. Finally, unless a claim element is defined by reciting the word “means” and a function without the recital of any structure, it is not intended that the scope of any claim element be interpreted based on the application of 35 U.S.C. §112, sixth paragraph.
Much of the inventive functionality and many of the inventive principles are best implemented with or in software programs or instructions and integrated circuits (ICs) such as application specific ICs. It is expected that one of ordinary skill, notwithstanding possibly significant effort and many design choices motivated by, for example, available time, current technology, and economic considerations, when guided by the concepts and principles disclosed herein will be readily capable of generating such software instructions and programs and ICs with minimal experimentation. Therefore, in the interest of brevity and minimization of any risk of obscuring the principles and concepts in accordance to the present invention, further discussion of such software and ICs, if any, will be limited to the essentials with respect to the principles and concepts of the preferred embodiments.
With reference to
A series of system busses may couple various system components including a high speed system bus 123 between the processor 120, the memory/graphics interface 121 and the I/O interface 122, a front-side bus 124 between the memory/graphics interface 121 and the system memory 130, and an advanced graphics processing (AGP) bus 125 between the memory/graphics interface 121 and the graphics processor 190. The system bus 123 may be any of several types of bus structures including, by way of example, and not limitation, such architectures include Industry Standard Architecture (ISA) bus, Micro Channel Architecture (MCA) bus and Enhanced ISA (EISA) bus. As system architectures evolve, other bus architectures and chip sets may be used but often generally follow this pattern. For example, companies such as Intel and AMD support the Intel Hub Architecture (IHA) and the Hypertransport™ architecture, respectively.
The computer 110 typically includes a variety of computer readable media. Computer readable media can be any available media that can be accessed by computer 110 and includes both volatile and nonvolatile media, removable and non-removable media. By way of example, and not limitation, computer readable media may comprise a computer storage media. Computer storage media includes both volatile and nonvolatile, removable and non-removable media implemented in any method or technology for storage of information such as computer readable instructions, data structures, program modules or other data. Computer storage media includes, but is not limited to, RAM, ROM, EEPROM, flash memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical disk storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information and which can accessed by computer 110.
The system memory 130 includes computer storage media in the form of volatile and/or nonvolatile memory such as read only memory (ROM) 131 and random access memory (RAM) 132. The system ROM 131 may contain permanent system data 143, such as identifying and manufacturing information. In some embodiments, a basic input/output system (BIOS) may also be stored in system ROM 131. RAM 132 typically contains data and/or program modules that are immediately accessible to and/or presently being operated on by processor 120. By way of example, and not limitation,
The I/O interface 122 may couple the system bus 123 with a number of other busses 126, 127 and 128 that couple a variety of internal and external devices to the computer 110. A serial peripheral interface (SPI) bus 126 may connect to a basic input/output system (BIOS) memory 133 containing the basic routines that help to transfer information between elements within computer 110, such as during start-up.
A super input/output chip 160 may be used to connect to a number of ‘legacy’ peripherals, such as floppy disk 152, keyboard/mouse 162, and printer 196, as examples. The super I/O chip 160 may be connected to the I/O interface 122 with a bus 127, such as a low pin count (LPC) bus, in some embodiments. Various embodiments of the super I/O chip 160 are widely available in the commercial marketplace.
In one embodiment, bus 128 may be a Peripheral Component Interconnect (PCI) bus, or a variation thereof, may be used to connect higher speed peripherals to the I/O interface 122. A PCI bus may also be known as a Mezzanine bus. Variations of the PCI bus include the Peripheral Component Interconnect-Express (PCI-E) and the Peripheral Component Interconnect-Extended (PCI-X) busses, the former having a serial interface and the latter being a backward compatible parallel interface. In other embodiments, bus 128 may be an advanced technology attachment (ATA) bus, in the form of a serial ATA bus (SATA) or parallel ATA (PATA).
The computer 110 may also include other removable/non-removable, volatile/nonvolatile computer storage media. By way of example only,
Removable media, such as a universal serial bus (USB) memory 153, firewire (IEEE 1394), or CD/DVD drive 156 may be connected to the PCI bus 128 directly or through an interface 150. A storage media 154 similar to that described below with respect to
The drives and their associated computer storage media discussed above and illustrated in
The computer 110 may operate in a networked environment using logical connections to one or more remote computers, such as a remote computer 180 via a network interface controller (NIC) 170. The remote computer 180 may be a personal computer, a server, a router, a network PC, a peer device or other common network node, and typically includes many or all of the elements described above relative to the computer 110. The logical connection between the NIC 170 and the remote computer 180 depicted in
In some embodiments, the network interface may use a modem (not depicted) when a broadband connection is not available or is not used. It will be appreciated that the network connection shown is exemplary and other means of establishing a communications link between the computers may be used.
As used herein, the term scene is defined to mean a collection of viewable elements and conditions used in created a final rendered image. The collection may include a set or setting, such as an office, an entertainment venue, a beach, a street, etc. The collection may also include a mannequin and pose. The mannequin may be selected from a palette of mannequins to match a users general body measurements or the mannequin may be generated from a given set of body measurements. The collection may also include one or more apparel items and accessories, for example, a dress, skirt and top, pants, shirt, necklace, bracelet, belt, shoes, etc. The collection may further include a light type or lighting condition, such as, but not limited to, sunny, bright, dim, afternoon, fluorescent, etc., and a camera view, that is, a point from which to generate the image.
In operation, a client device, such as the smart phone 204 may initiate a browsing session to select and display a scene. The scene information may be transmitted using predefined references, such as set 2, pose 4, apparel 10 (for example, from a personal closet), camera view orthogonal distances as x-coordinate, y-coordinate, z-coordinate, corresponding to a location relative to a center of the selected set. The request may be carried over the network 208 to a first tier of processing that separates the scene into component elements. The component elements may be further processed at the same or different servers available via the network 208. Renderings of the component elements are returned to a server for combination back into the scene and may be flattened to a 2D image for transmission back to the smart phone 204, where the image may be displayed.
The selected scene may include animation information that is used to generate a series of requests that are processed in real time or near real time so that an animated sequence may be presented on the smart phone 204. The animated sequence allows a user to view, for example, the drape, flow, and color of the selected apparel as it would appear not just in one pose but in motion. The process is discussed in more detail below.
In the exemplary embodiment of
At a logical tier below the application and composition server 306 or servers, may be individual base render servers. For example, a server 308 may be used to render a selected set. The server 308 may use a database 310 of predetermined set types. Another exemplary server 312 may be used to render a mannequin and pose selected from a mannequin/pose database 314. Yet another exemplary server 316 may be used to render a garment and accessories, selected from a corresponding apparel and accessory database 318. The composition and base render servers are particular examples of processing resources that may be used to calculate desired results. Other examples of processing resources may be dedicated processors of a multi-processor computer or separate processes running on a single computer or server.
The apparel and accessory database 318 may include separate tables, or similar representations, of a particular user's apparel inventory 320 and one or more retailer of apparel inventories 322, 324. Additional user apparel inventories (not depicted) may be accessible to a particular user given the correct permissions.
As illustrated in
The depiction of the hierarchical embodiment of
In the exemplary embodiment of
As discussed below, the application server may split the scene data received via network 506 into sets of data used to render a particular element of the scene. For example, a set server 512, using set descriptive data from set database 514 may render the set using the user selected set, light type, and camera view information. A mannequin/pose server 516, using mannequin/pose database 518, may render a selected mannequin in a selected pose according to the user selected mannequin, pose, light, and camera view information. An apparel/accessories server 520 may access an apparel and accessory database 522 that may include one or more user and retailer apparel inventories, for example user apparel inventory 524, a first retailer apparel inventory 526, and another retailer apparel inventory 528. The apparel/accessories server 520 may use the database 522 to render the apparel and accessories selected by the user in view of the selected light type and camera view.
In this exemplary embodiment, the rendered outputs from each of the servers 512, 516, 520 may be returned to the composition server 510 for combining and flattening from three dimensions (3D) to two dimensions (2D). The composition server 510 may then send the final image to a browser on the smart phone 502 for viewing by the user. Alternatively, or in addition to, sending the image to the smart phone 502, the image may be sent to another device, such as tablet 504, for viewing. The exemplary second device, tablet 504, may provide a higher resolution display or may be used by another person with whom the original user wishes to share a view the final image.
The scene options may be presented to a user via web browser hosted by a web page supported by an application server, such as application server 508, or by an application/composition server, such as application/composition server 306. The web browser may also collect inputs from the user related to the scene options, such a set, a mannequin, a pose of the mannequin, apparel, accessories (if any), light type, camera view, etc. The camera view may be expressed relative to the set as a side displacement (x), a front-to-back displacement (y), and a vertical or height displacement (z) from an initial position of the mannequin on the set.
The same web browser used to collect scene inputs may also be used as a viewing resource for the display of the image resulting from the rendering processes, although more than one browser window can be dedicated to the scene input collection and the viewing resource. In some embodiments, such as when another user is invited to share a view or animation, the two functions may be supported on different browsers on different platforms.
Optionally, in one embodiment, animation inputs may also be collected with the scene information. Animation inputs may be selected from a predetermined track or may be traced using the web browser. The animation inputs may include route and body motions selected to show the color response, drape, and flow of a item of apparel for the given set and lighting conditions.
At block 604, after the scene inputs are collected at an application/composition server 306, different groups of data may be generated. A first data group including the set, the light type, and the camera view may be generated. A second data group including the mannequin, the pose, the light type, and the camera view may be generated. And a third data group including the apparel, the light type, and the camera view may also be generated. The scene inputs may include metadata as well, such as the pixel dimensions and color depth of the end viewing area, so that the remaining steps can tailor their respective outputs to the target viewing area and capability.
At block 606, the first, second, and third data groups may be sent to the respective set server 308, mannequin/pose server 312, and apparel/accessories server 316 by the application/composition server 306.
At block 608, the set processing resource 608 may generate a first base rendering of the set from the first data group. At block 610, the mannequin/pose server 312 may generate a second base rendering of the mannequin at a given pose, which may include a displacement position from an initial position, from the second data group. At block 612, the apparel/accessories server 316 may generate a third base rendering of the apparel and any accessories from the third data group. Rendering involves determining a color for each pixel in the viewing frame. Numerous rendering techniques are known and applicable, such as various forms of scanline rendering or pixel-by-pixel rendering. Rather than attempting to render both moving and stationary elements in the same pass, in this embodiment, elements are sorted and rendered by their type. That is, the stationary set, the moving mannequin with a relatively constant surface, and the cloth of the apparel, which may have folds or color changes based on light angle, are all calculated separately on different processing resources. Mannequin and set images may create reflections in elements of the set. These reflections may also be calculated during the respective mannequin and apparel base rendering processes.
At block 614, the separate base first base rendering of the set, the second base rendering of the mannequin, and the third base rendering of the apparel may be sent to a composition processing resource. The composition processing resource may be the same process that collected the scene inputs at block 604, but may be different.
At block 616, the composition processing resource may generate a composite rendering including the first, second, and third base renderings. The composite rendering may be accomplished by simply overlaying the three renderings. Reflections and overlaps may be accommodated by setting different levels of transparency for any element through which another element may be seen.
At block 618, the combined rendering may be flattened, that is, the 3D rendering may be projected onto a 2D surface and the rendered 2D image captured.
At block 620, the flattened, composite 2D image may be sent to a viewing resource. The viewing resource may be a handheld device or other display-capable computing platform. At block 622, the viewing resource may display the composite 2D image, for example, using a web browser. In other embodiments, the image may be delivered to more than one viewing resource for joint viewing by more than one user.
When an animation sequence is selected, as described above with respect to block 602, the process may return to block 604 where a next frame of the animation is queued and the activities of blocks 604 through 622 may be repeated. The process may be repeated in real time with respect to a frame rate of displaying the composite 2D image at the viewing resource, so that minimal or no buffering in the viewing resource is required. Because not buffering kept to an absolute minimum, for example, one frame, the viewing resource may have only minimal memory and associated memory management capabilities. Because the images arrive already flattened and, optionally, sized to a display area, complex image processing at the viewing resource is minimized or eliminated, unlike a dedicated gaming system or high end computer, although those machines can also be used as viewing resources.
Because the animation process is built to the ‘weakest link,’ that is, a low function graphics display and the more compute intensive processes are offloaded and, optionally, distributed, partial animation of greater than 3 frames per second may be supported, while in some cases, full motion animation of 10-30 frames per second may be achieved.
The ability to capture complex scene information, divided among a number of services and or services on different servers, allows complex, customized animations to be requested and viewed from a very simple platform such as a common web browser. The ability to generate data sets with overlapping information, such as lighting type and camera view, which are then separately rendered and later combined allows a speed improvement over ray-tracing algorithms of several orders of magnitude. This speed improvement enables users with very simple platforms, such as smart phones, to create customized full-motion animations in real time. When applied to a shopping situation the user benefits from being able to view a selected item of apparel or accessory in a variety of settings and lighting conditions as well as from different angles or ‘camera views.’ A retailer, particularly an online retailer, benefits from being able to present a user with a more complete understanding of an item contemplated for purchase as well as being able to suggest complementary accessories for different types of use.
The same technology may be easily applied to related online shopping experiences. For example, customized rooms may be furnished with online 3-D models of furniture and appliances for viewing in a variety of lighting conditions and from a variety of angles, using even simple platforms such as smart phones.
Although the foregoing text sets forth a detailed description of numerous different embodiments of the invention, it should be understood that the scope of the invention is defined by the words of the claims set forth at the end of this patent. The detailed description is to be construed as exemplary only and does not describe every possibly embodiment of the invention because describing every possible embodiment would be impractical, if not impossible. Numerous alternative embodiments could be implemented, using either current technology or technology developed after the filing date of this patent, which would still fall within the scope of the claims defining the invention.
Thus, many modifications and variations may be made in the techniques and structures described and illustrated herein without departing from the spirit and scope of the present invention. Accordingly, it should be understood that the methods and apparatus described herein are illustrative only and are not limiting upon the scope of the invention.
This application is a related to U.S. patent application Ser. No. 12/652,351, titled “Automated Generation of Garment Construction Specification,” filed on Jan. 5, 2010, which is hereby incorporated by reference for all purposes.