Embodiments of the present disclosure relate generally to computer technology and, more particularly, but not by way of limitation, to regional item recommendations.
Many regions of the world have unique environments, styles, and customs. Local residents typically wear apparel adapted to the region in which they reside. Visitors from other regions can stand out as non-congruent with local style or can be ill equipped to deal with the region's environment.
Various ones of the appended drawings merely illustrate example embodiments of the present disclosure and should not be considered as limiting its scope.
The headings provided herein are merely for convenience and do not necessarily affect the scope or meaning of the terms used.
The description that follows includes systems, methods, techniques, instruction sequences, and computing machine program products that embody illustrative embodiments of the disclosure. In the following description, for the purposes of explanation, numerous specific details are set forth in order to provide an understanding of various embodiments of the inventive subject matter. It will be evident, however, to those skilled in the art, that embodiments of the inventive subject matter may be practiced without these specific details. In general, well-known instruction instances, protocols, structures, and techniques are not necessarily shown in detail.
In various example embodiments, systems and methods for providing regional item recommendations are provided. In an example embodiment, an application server receives a destination geolocation. The destination geolocation can be a specific location, a city, a region, a country, and the like. In further embodiments, the application server receives activities or an indication of an activity along with the destination geolocation from the user (e.g., clubbing, site seeing, skiing, and so forth).
Subsequently, the application server retrieves data associated with the destination from a plurality of sources. In an example embodiment, the application server retrieves the destination data from social network servers, government servers, web camera servers, commercial servers, and so forth. The destination data can include images, data associated with the images, social network data, purchase history data associated with the destination geolocation, population demographic information, climate information, weather forecast information, and so forth.
The application server extracts a destination characteristic from the destination data. The destination characteristic indicates an affinity score (e.g., a metric or rating indicating a likelihood or probability that a particular person of the destination geolocation desires or has demand for a particular item) for apparel associated with the destination geolocation. For example, the destination characteristic can indicate that a particular style or cut of jeans are popular in an area surrounding the destination geolocation.
The application server then determines a candidate apparel item based on the extracted destination characteristic. The candidate apparel item, for instance, can be clothing that conforms to social norms of the area around the destination geolocation or apparel that is suited for the area around the destination geolocation (e.g., winter clothing for cold regions).
The application server identifies an item listing corresponding to the candidate apparel item from an online marketplace application server. The application server can recommend the item listing to the user, for example, by causing presentation of the item listing on a user interface of the user device.
In further embodiments, the application server retrieves user data corresponding to the user from the plurality of sources (e.g., social network servers, user input into a user interface, online marketplace servers). The user data can include user demographic information (e.g., age, gender, income level), purchase histories, prior destinations of the user, social network data (e.g., posts to a social network and social network contacts including data associated with the contact), and other data. In some embodiments, the user data can also include sensor data indicating user apparel item usage. In these embodiments, the sensor data includes tag detections indicating use, by the user, of a particular apparel item. For example, the sensor data include tag detections such as RFID tag detections (e.g., a RFID tag embedded in an article of clothing), smart tag detections (e.g., near field communication (NFC) tags), or Quick Response (QR) code detections. In some embodiments, the application server extracts a user characteristic from the user data and uses the user characteristic, in part, to determine the candidate apparel item.
In some embodiments, the application server provides the user various options associated with the item listing. In an example embodiment, the options include an option to deliver a purchased item of the item listing to the destination geolocation. In a further embodiment, the options include an option to automatically list the purchased item of the item listing for sale on an online marketplace at a determined date, such as upon the user leaving the destination.
In a specific example, the application server receives a destination geolocation such as Playa Azul beach in Cozumel, Mexico. The application server retrieves destination data associated with Mexico, Cozumel, Playa Azul beach, the region surrounding Cozumel, and the region surrounding Playa Azul beach. The destination data may indicate that the destination is a beach in a warm climate. Based on an analysis of the destination data in this example, the application server retrieves item listings of beachwear items that are recommended to the user for purchase. The user may have provided an activity or the destination data may have indicated an activity associated with the destination, for example, snorkeling. Based on the activity provided by the user or indicated in the destination data, listings of snorkeling equipment items may be retrieved from the application server and may be included in the candidate items. The application server may also retrieve user data that indicates the user is a woman. Based on the user data and the destination data, item listings of beachwear suits for women may be retrieved from the application server and included in the candidate items. In further embodiments of this example, the destination data associated with the user data and the destination data may be retrieved. For example, the application server can retrieve, and cause presentation to the user, data from social network profiles of people located near the destination geolocation that have similar demographic characteristics to the user. In this example, the application server can retrieve pictures of woman (demographically similar to the user) that live near Cozumel and present them to the user to assist the user in evaluating the item listings for purchase (e.g., evaluating a style of locals near Cozumel to better fit in when visiting Cozumel). In still further embodiments of this example, the user data may indicate, for example, that the user has already purchased goggles but does not have a snorkel. Based on the user data, goggles may not be included and snorkels may be included in the item listings.
With reference to
In various implementations, the client device 110 comprises a computing device that includes at least a display and communication capabilities that provide access to the networked system 102 via the network 104. The client device 110 comprises, but is not limited to, a remote device, work station, computer, general purpose computer, Internet appliance, hand-held device, wireless device, portable device, wearable computer, cellular or mobile phone, Personal Digital Assistant (PDA), smart phone, tablet, ultrabook, netbook, laptop, desktop, multi-processor system, microprocessor-based or programmable consumer electronic, game consoles, set-top box, network Personal Computer (PC), mini-computer, and so forth. In an example embodiment, the client device 110 comprises one or more of a touch screen, accelerometer, gyroscope, biometric sensor, camera, microphone, Global Positioning System (GPS) device, and the like.
The client device 110 communicates with the network 104 via a wired or wireless connection. For example, one or more portions of the network 104 comprises an ad hoc network, an intranet, an extranet, a Virtual Private Network (VPN), a Local Area Network (LAN), a wireless LAN (WLAN), a Wide Area Network (WAN), a wireless WAN (WWAN), a Metropolitan Area Network (MAN), a portion of the Internet, a portion of the Public Switched Telephone Network (PSTN), a cellular telephone network, a wireless network, a Wireless Fidelity (WI-FI®)) network, a Worldwide Interoperability for Microwave Access (WiMax) network, another type of network, or any suitable combination thereof.
In some example embodiments, the client device 110 includes one or more of the applications (also referred to as “apps”) such as, but not limited to, web browsers, book reader apps (operable to read e-books), media apps (operable to present various media forms including audio and video), fitness apps, biometric monitoring apps, messaging apps, electronic mail (email) apps, and e-commerce site apps (also referred to as “marketplace apps”). In some implementations, the client application(s) 114 include various components operable to present information to the user and communicate with networked system 102. In some embodiments, if the e-commerce site application is included in the client device 110, then this application is configured to locally provide the user interface and at least some of the functionalities with the application configured to communicate with the networked system 102, on an as needed basis, for data or processing capabilities not locally available (e.g., access to a database of items available for sale, to authenticate a user, to verify a method of payment). Conversely, if the e-commerce site application is not included in the client device 110, the client device 110 can use its web browser to access the e-commerce site (or a variant thereof) hosted on the networked system 102.
The web client 112 accesses the various systems of the networked system 102 via the web interface supported by a web server 122. Similarly, the programmatic client 116 and client application(s) 114 accesses the various services and functions provided by the networked system 102 via the programmatic interface provided by an Application Program Interface (API) server 120. The programmatic client 116 can, for example, be a seller application (e.g., the Turbo Lister application developed by EBAY® Inc., of San Jose, Calif.) to enable sellers to author and manage listings on the networked system 102 in an off-line manner, and to perform batch-mode communications between the programmatic client 116 and the networked system 102.
Users (e.g., the user 106) comprise a person, a machine, or other means of interacting with the client device 110. In some example embodiments, the user is not part of the network architecture 100, but interacts with the network architecture 100 via the client device 110 or another means. For instance, the user provides input (e.g., touch screen input or alphanumeric input) to the client device 110 and the input is communicated to the networked system 102 via the network 104. In this instance, the networked system 102, in response to receiving the input from the user, communicates information to the client device 110 via the network 104 to be presented to the user. In this way, the user can interact with the networked system 102 using the client device 110.
The API server 120 and the web server 122 are coupled to, and provide programmatic and web interfaces respectively to, one or more application server(s) 140. The application server(s) 140 can host one or more publication system(s) 142, payment system(s) 144, and a regional apparel system 150, each of which comprises one or more modules or applications and each of which can be embodied as hardware, software, firmware, or any combination thereof. The application server(s) 140 are, in turn, shown to be coupled to one or more database server(s) 124 that facilitate access to one or more information storage repositories or database(s) 126. In an example embodiment, the database(s) 126 are storage devices that store information to be posted (e.g., publications or listings) to the publication system(s) 142. The database(s) 126 also stores digital good information in accordance with some example embodiments.
Additionally, a third party application 132, executing on third party server(s) 130, is shown as having programmatic access to the networked system 102 via the programmatic interface provided by the API server 120. For example, the third party application 132, utilizing information retrieved from the networked system 102, supports one or more features or functions on a website hosted by the third party. The third party website, for example, provides one or more promotional, marketplace, or payment functions that are supported by the relevant applications of the networked system 102.
The publication system(s) 142 provides a number of publication functions and services to the users that access the networked system 102. The payment system(s) 144 likewise provides a number of functions to perform or facilitate payments and transactions. While the publication system(s) 142 and payment system(s) 144 are shown in
In some implementations, the regional apparel system 150 provides functionality to identify and recommend apparel items associated with a particular a geolocation. In some example embodiments, the regional apparel system 150 communicates with the client device 110, the third party server(s) 130, the publication system(s) 142 (e.g., retrieving listings), and the payment system(s) 144 (e.g., purchasing a listing). In an alternative example embodiment, the regional apparel system 150 is a part of the publication system(s) 142. The regional apparel system 150 will be discussed further in connection with
Further, while the client-server-based network architecture 100 shown in
The presentation module 210 provides various presentation and user interface functionality operable to interactively present (or cause presentation) and receive information from the user. For instance, the presentation module 210 can cause presentation of the item listings along with various images and data associated with the item listings (e.g., price, description, images, availability date, currency, dimensions, weight, or condition). In other instances, the presentation module 210 generates and causes presentation of user interfaces to recommend item listings, present social media information (e.g., images and postings), present inventory data of the user, and so forth. In various implementations, the presentation module 210 presents or causes presentation of information (e.g., visually displaying information on a screen, acoustic output, haptic feedback). Interactively presenting information is intended to include the exchange of information between a particular device and the user. The user may provide input to interact with the user interface in many possible manners such as alphanumeric, point based (e.g., cursor), tactile, or other input (e.g., touch screen, tactile sensor, light sensor, infrared sensor, biometric sensor, microphone, gyroscope, accelerometer, or other sensors), and the like. It will be appreciated that the presentation module 210 provides many other user interfaces to facilitate functionality described herein. Further, it will be appreciated that “presenting,” as used herein, is intended to include communicating information or instructions to a particular device that is operable to perform presentation based on the communicated information or instructions.
The communication module 220 provides various communications functionality and web services. For example, the communication module 220 provides network communication such as communicating with the networked system 102, the client device 110, and the third party server(s) 130. In various example embodiments, the network communication can operate over wired or wireless modalities. Web services are intended to include retrieving information from the third party server(s) 130, the database(s) 126, and the application server(s) 140. In some implementations, information retrieved by the communication module 220 comprises data associated with the user (e.g., user profile information from an online account, social network service data associated with the user), data associated with one or more items listed on an e-commerce website (e.g., images of the item, reviews of the item, item price), or other data to facilitate the functionality described herein.
The data module 230 provides functionality to access or retrieve data from a variety of sources. For instance, the data module 230 can retrieve destination data and user data from the databases 126, the third party servers 130, the client device 110, and other sources. In various specific examples, the data module 230 accesses or retrieves the destination data and the user data from social networking severs (e.g., posts, likes, follows, friends, or check-ins) and online market place servers (e.g., current item listings, purchase histories, and account settings). The destination data includes, for example, images (e.g., images with a geotag within a distance of the destination geolocation), social media data (e.g., posts of member of the social network near the destination geolocation), climate information, weather forecast data, purchase histories of other users, and so forth. The user data includes, for example, demographic data (e.g., age, gender, socio-economic status), geolocation data (e.g., as determined by a Global Positioning System (GPS) component of a mobile device of the user), social media data (e.g., posts, likes, follows, or check-ins of the user), a purchase history of the user, and so on.
The style module 240 provides functionality to determine candidate apparel items (e.g., suggested or recommended items for the user) based on various analyses using various data. For example, the style module 240 extracts the destination characteristic and the user characteristic respectively from the destination data and the user data. Characteristics, as used herein, are intended to include traits, qualities, actions, activities, attitudes, habits, behaviors, and the like pertaining to a person or a geolocation as is suitable. The style module 240 can perform other analyses such as identifying social networking service profiles of members that are similar to the user (e.g., similar demographics).
The fitment module 250 provides functionality to determine garment or apparel sizes of the user and identify and recommend item listings based on the determine apparel size of the user. For instance, the fitment module 250 can analyze a purchase history of the user to determine a user apparel size of the user. The fitment module 250 can then analyze purchase histories of other users to determine, estimate, or gauge an appropriate or candidate garment or apparel size for a particular item listing (e.g., other users that are of a same or similar size as the user frequently purchased a particular piece of apparel in a particular size).
The purchase module 260 provides functionality to facilitate purchase of an item corresponding to an item listing. In some embodiments, the purchase module 260 provides functionality to identify delivery options for the purchase. In further embodiments, the purchase module 260 generates an option for automatic sale for an item of an item listing according to various pieces of data. For example, the user may be planning a ski trip and the purchase module 260 can facilitate purchase of items for the ski trip, have the purchased items delivered to the ski trip destination, and provide the user with an option to automatically list the items for sale when the ski trip is over.
In a specific example, the user is at the geolocation 304 and is interested in traveling to the geolocation 308 (the destination geolocation) for a skiing trip. In this example, the geolocation 308 is in a different country than the geolocation 304. The user device 316 of the user provides an indication of the geolocation 308 to the regional apparel system 150. For example, the user may input the geolocation 308 at a user interface configured to receive a destination geolocation. Subsequently, the regional apparel system 150 retrieves destination data corresponding to the geolocation 308. The destination data can include a wide range of data indicating various social norms, activities, popular apparel, and so forth, associated with the geolocation 308. For example, the destination data includes purchase histories of user that reside within a distance of the geolocation 308. The regional apparel system 150 extracts a destination characteristic from the destination data. The destination characteristic indicates an affinity or preference for certain apparel associated with the geolocation 308. For example, residents within a distance of the geolocation 308 that purchase ski equipment also purchase a particular brand of jeans. The regional apparel system 150 then determines a candidate apparel item based on the extracted destination characteristic. The regional apparel system 150 identifies item listings corresponding to the candidate apparel item and causes presentation of the identified item listings on the user interface 318 of the user device 316. In this way, the regional apparel system 150 recommends the items 310 to the user based on the geolocation 308. The item listings may be culturally appropriate, popular in the region, or otherwise suited to the geolocation 308.
At operation 410, the communication module 220 receives an indication of a destination geolocation from a user device of a user. For example, the user can input an indication of the destination geolocation (e.g., the geolocation 308) into a user interface configured to receive the destination geolocation. In other examples, the regional apparel system 150 infers the destination geolocation based on user activity such as performing searches associated with a particular geolocation.
At operation 420, the data module 230 retrieves or accesses the destination data corresponding to the destination geolocation. The data module 230 retrieves or accesses the destination data from a wide variety of sources such as social network servers, government servers, online marketplace servers, the user device, and so on. The destination data includes, for example, images (e.g., images with a geotag near the destination geolocation), social media data (e.g., posts of member of the social network near the destination geolocation), climate information, weather forecast data, purchase histories of other users (e.g., purchases from online marketplaces), search histories (e.g., search histories originating from near the destination geolocation and associated with apparel), and so forth. The destination data can be associated with the destination geolocation, an area surrounding the destination geolocation, a region surrounding the destination geolocation, a specific location near the destination geolocation, and so on.
At operation 430, the style module 240 extracts a destination characteristic from the destination data. The destination characteristic indicates an affinity score or rating (e.g., a metric indicating a likelihood or probability that a particular person associated with the destination geolocation desires or has demand for a particular item) for apparel associated with the destination geolocation. For instance, if the destination characteristic indicates the destination geolocation is in a warm climate, then the destination characteristic may indicate a high affinity score for apparel or garments suited for a warm climate (e.g., as determined by a portion of a population of the destination geolocation that purchases such apparel as compared to other populations). In another example, the destination characteristic indicates a particular activity associated with the destination geolocation (e.g., prolific ski slopes near the destination geolocation) and consequently indicates an affinity or preference for apparel associated with the particular activity. In yet another example, the destination characteristic indicates popularity of a particular piece of apparel based on trends indicated by social media data included in the destination data (e.g., numerous social media posts associated with a particular piece of apparel for users near the destination geolocation or a spike in purchases for a particular piece of apparel near the destination geolocation). In still another example, the destination characteristic indicates a particular piece of apparel is popular among residents near the destination geolocation according to purchases made by residents within a distance of the destination geolocation. In other examples, the destination characteristic indicates an affinity score for at least one of a color, pattern, design, style, or type of apparel. In an example, the destination characteristic indicates that a particular feature of a garment (e.g., button style or placement on a garment) is popular in an area surrounding the destination geolocation.
At operation 440, the style module 240 determines a candidate apparel item based on the extracted destination characteristic. The candidate apparel item can comprise equipment (e.g., skiing equipment), items for activities (e.g., concert tickets), medicine (e.g., motion sickness medicine for certain activities such as a theme park visit), clothing, garments, and so forth. In a specific example, if the user is planning a business trip to a particular foreign nation that has particular social norms for business attire, the style module 240 determines candidate apparel according to the destination characteristic that indicates a particular social norm for business attire.
At operation 450, the data module 230 identifies an item listing corresponding to the candidate apparel item. For instance, the data module 230 performs a search or a lookup for item listings on an online marketplace server for item listings that are associated with the candidate apparel item (e.g., item listings for items that are the same or similar to the candidate item). For instance, if the candidate apparel item is a pair of jeans with a particular cut and style, the data module 230 identifies item listings for items of the same or similar jeans with the particular cut and style.
At operation 460, the presentation module 210 causes presentation of the item listing on a user interface of the user device. As discussed further in connection with
In the user interface diagram 500, user interface element 520 provides an option to sort the item listings 530, or otherwise navigate the item listings 530, according to various schemes such as sorting based on recentness (e.g., based on temporal information corresponding to respective item listings 530), item price, distance from the user, relevance, geolocation, fitment (e.g., how well a particular piece of apparel may fit the user), style (how well a particular piece of apparel conforms to a style of the user or a region), or other metrics. In an example embodiment, the item listings 530 include various portions of item information such as an item image, price, merchant, brand, or other information retrieved from the publication system(s) 142. In some implementations, activating a particular item listing presents additional information corresponding to the particular item listing. In an example embodiment, the user interface 510 includes an option to purchase the item corresponding to the item listing 530 (e.g., activating a particular user interface element facilitates a transaction for the item, for example, using the payment system(s) 144).
At operation 610, the data module 230 retrieves a destination image associated with the destination geolocation. For instance, the data module 230 can perform a web search for images associated with the destination geolocation (e.g., an image with a description that mentions the destination geolocation or an image with a geotag near the destination geolocation).
At operation 620, the style module 240 determines the destination image comprises an image of a person associated with the destination geolocation. For instance, the style module 240 uses image recognition techniques to determine that the destination image includes a person (e.g., detecting a face within an image as a proxy for determining the image includes a person). In other instances, the style module 240 uses metadata associated with a particular image to determine that the image comprise a person (e.g., an image that includes tags corresponding to members of a particular social network service). In some embodiments, the style module 240 determines that the destination image is associated with a particular piece of apparel or particular activity (e.g., images of people wearing winter clothing or going skiing).
At operation 630, the presentation module 210 causes presentation of the destination image on the user interface of the user device to assist the user in evaluation a style of the item listing. For instance, the destination images can provide insight into local styles and customs associated with the destination geolocation.
At operation 810, the style module 240 identifies a similar profile among a plurality of member profiles of a social network service. The similar profile corresponds to a member of the social network service that is similar to the user. For example, the style module 240 may identifies a similar profile of a member having the same or similar demographical information as the user (e.g., same age or gender).
At operation 820, the style module 240 determines the similar profile includes member geolocation data indicating the member of the similar profile resides within a distance of the destination geolocation. For instance, the similar profile may indicate that the member of the similar profile resides at a geolocation within a distance of the destination geolocation.
At operation 830, the data module 230 accesses the member image included in the similar profile. The member image being an image of the member of the similar profile.
At operation 840, the presentation module 210 causes presentation of the member image on a user interface of the user device. In some instances, the style module 240 ranks or sorts a plurality of member images by date, similarity to the user, type of apparel associated with the member image, and causes presentation of a highest ranking member image (e.g., a recent member image of a member that is most similar to the user).
At operation 1010, the communication module 220 receives an indication of a destination time or time range of the user. The destination time comprises a time when the user plans to be in a vicinity of the destination geolocation. In an example, the communication module 220 can receive the indication of the destination time from a user input of the destination time. In another example, the style module 240 infers the destination time based on various data associated with the user. For instance, if a user purchase history indicates an airline ticket for specified dates, the style module 240 can infer the destination time from the purchase history.
In further embodiments, the data module 230 uses the destination time to retrieve the destination data. For example, given the destination time that is in summer, the data module 230 may retrieve the destination data including date associated with summer time (e.g., summer clothing).
At operation 1020, the purchase module 260 identifies a delivery option for the item listing. The delivery option includes delivery parameters according to the destination geolocation and the destination time. For example, if the user is going on vacation and has provided the destination geolocation for the vacation, it may be more convenient for the user to simply have vacation items delivered to the vacation destination. In some embodiments, purchase module 260 provides a number of delivery choices for the item listing to the user. The delivery choices can include delivering the items on a particular date, a particular type of delivery (e.g., 24 hour courier), a method of delivery (e.g., delivery to the front desk of a hotel with instructions to be delivered to a room on a date), and other delivery choices. In further embodiments, a dynamic delivery choice may be provided to the user. The dynamic delivery choice may locate the user at the destination to make the delivery with or without location information provided by the user. For example, the user may provide a particular city as the destination. The dynamic delivery choice may then attempt to locate the user within the city. The user may be found by GPS on a mobile device, for example. The dynamic delivery may have a default delivery location if the user cannot be located (e.g., the user's home or a specific location near the destination).
At operation 1030, the purchase module 260 causes presentation of the delivery option with the delivery parameters on the user interface of the user device. The user can indicate a selection of a particular delivery option and the purchase module 260 can facilitate delivery of the item of the item listing according to the delivery parameters of the selected delivery option.
At operation 1210, the communication module 220 receives an indication of a return time or time range of the user. The return time comprises a time when the user plans to leave a vicinity of the destination geolocation. For example, the user can input a return time into a user interface configured to receive the return time. In another example, the style module 240 infers the return time based on various data such as date indicated by an airline ticket.
At operation 1220, the purchase module 260 generates an automatic sale option for the item listing according to the destination geolocation and the return item. The automatic sale option specifies instructions to automatically list an apparel item of the item listing for sale on behalf of the user. For example, if the user is going on vacation and purchases the items in the item listings specifically for the vacation and the items are no longer needed after the vacation, it may be convenient to automatically list those items for sale after or during the vacation. In some embodiments, the user may provide a date of sale and the items may be listed to the online marketplace on a date corresponding to the date of sale.
At operation 1230, the presentation module 210 causes presentation of the automatic sale option on the user interface of the user device. The user can provide an indication of a selection of the automatic sale option. In response to the communication module 220 receiving the selection of the automatic sale option, the purchase module 260 can schedule a listing of the purchased item for sale at a date in the future (e.g., a time at or after the return time). The purchase module 260 can automatically generate the listing (e.g., images and description) using information from the original item listing the user purchased.
At operation 1410, the data module 230 accesses user data corresponding to the user. The user data can include user demographic information (e.g., age, gender, income level), purchase histories, prior destinations of the user (e.g., as determined by stored data from a GPS component of a mobile device of the user), social network data (e.g., posts to a social network and social network contacts including data associated with the contact), and other data. The user data can also include sensor data indicating user apparel item usage. The sensor data includes tag detections indicating use, by the user, of a particular apparel item. For example, the sensor data include tag detections such as RFID tag detections (e.g., a RFID tag embedded in an article of clothing), smart tag detections (e.g., near field communication (NFC) tags), or bar code (e.g., a QR code) detections.
At operation 1420, the style module 240 extracts a user characteristic from the user data. The user characteristic indicates a user affinity score or rating (e.g., a probability or likelihood that the user desires or has demand for a particular item) for apparel associated with the user. For example, if the user has often purchased items in a particular color or brand, the user characteristic may indicate a higher user affinity score for apparel in that color or brand. In another example, the style module 240 can extract a particular user characteristic that indicates a purpose or reason for a visit to the destination geolocation (e.g., a skiing trip or a business trip). In this example, the particular user characteristic can be indicative of a particular type of apparel (e.g., business attire for a business trip).
At operation 1430, the style module 240 determines the candidate item based on the extracted user characteristic and the extracted destination characteristic. For instance, the style module 240 can determine a candidate item based on the user's age, past purchases, recent usage patterns (e.g., wearing a particular style of clothing more often in recent weeks), or other user characteristics.
At operation 1610, the fitment module 250 determines the item listing comprises a cross-border item listing based on a user geolocation indicated by the user data and an item listing geolocation indicated by the item listing. In a cross-border item listing, the user geolocation can be in a different country than the geolocation indication by the item listing. In such cross-border item listings, sizes of apparel can deviate from an expectation of the user (e.g., a user location in the United States may have a different expectation for a size large than a user located in Japan). To better provide recommendations to the user, the fitment module 250 can translate sizes of the user to sizes associated with a particular piece of apparel.
At operation 1620, the fitment module 250 extracts a user apparel size from the user data. For instance, the purchase history or inventory data of the user can indicate a garment size the user typically purchases for various types of apparel.
At operation 1630, the data module 230 accesses apparel size data corresponding to the cross-border item listing. For example, the data module 230 can access purchase histories of other users that are of a similar physical size to the user and that have purchased a same or similar item corresponding to the cross-border item listing. For example, the fitment module 250 identifies other users of a same physical size by comparing the purchase histories of the user and the other users and finding a match in garment size for particular pieces of apparel (e.g., buying a particular brand and style of jeans in a same length and waist size). Once the fitment module 250 identifies other users of a same or similar physical size to the user, the data module 230 can then identify users among the other users that have purchased the item corresponding to the cross-border item listing.
At operation 1640, the fitment module 250 maps the user apparel size to a candidate apparel size for the cross-border item listing using the apparel size data. For instance, the fitment module 250 determines that the other users, that are of a same or similar physical size as compared to the user, often purchase the item of the cross-border item listing in a particular size. Using such a determination, the fitment module 250 can map the user apparel size to the candidate apparel size.
At operation 1650, the presentation module 210 causes presentation of the cross-border item listing including the candidate apparel size that maps to the user apparel size. For instance, the presentation module 210 can cause presentation of a most frequently purchased size of the item corresponding to the cross-border item listing for users that are of a same or similar physical size as the user. In this way, the regional apparel system 150 can translate apparel sizes for apparel that is from another country or region.
Certain embodiments are described herein as including logic or a number of components, modules, or mechanisms. Modules can constitute either software modules (e.g., code embodied on a machine-readable medium or in a transmission signal) or hardware modules. A “hardware module” is a tangible unit capable of performing certain operations and can be configured or arranged in a certain physical manner. In various example embodiments, one or more computer systems (e.g., a standalone computer system, a client computer system, or a server computer system) or one or more hardware modules of a computer system (e.g., a processor or a group of processors) is configured by software (e.g., an application or application portion) as a hardware module that operates to perform certain operations as described herein.
In some embodiments, a hardware module is implemented mechanically, electronically, or any suitable combination thereof. For example, a hardware module can include dedicated circuitry or logic that is permanently configured to perform certain operations. For example, a hardware module can be a special-purpose processor, such as a Field-Programmable Gate Array (FPGA) or an Application Specific Integrated Circuit (ASIC). A hardware module may also include programmable logic or circuitry that is temporarily configured by software to perform certain operations. For example, a hardware module can include software encompassed within a general-purpose processor or other programmable processor. It will be appreciated that the decision to implement a hardware module mechanically, in dedicated and permanently configured circuitry, or in temporarily configured circuitry (e.g., configured by software) can be driven by cost and time considerations.
Accordingly, the phrase “hardware module” should be understood to encompass a tangible entity, be that an entity that is physically constructed, permanently configured (e.g., hardwired), or temporarily configured (e.g., programmed) to operate in a certain manner or to perform certain operations described herein. As used herein, “hardware-implemented module” refers to a hardware module. Considering embodiments in which hardware modules are temporarily configured (e.g., programmed), each of the hardware modules need not be configured or instantiated at any one instance in time. For example, where a hardware module comprises a general-purpose processor configured by software to become a special-purpose processor, the general-purpose processor may be configured as respectively different special-purpose processors (e.g., comprising different hardware modules) at different times. Software can accordingly configure a particular processor or processors, for example, to constitute a particular hardware module at one instance of time and to constitute a different hardware module at a different instance of time.
Hardware modules can provide information to, and receive information from, other hardware modules. Accordingly, the described hardware modules can be regarded as being communicatively coupled. Where multiple hardware modules exist contemporaneously, communications can be achieved through signal transmission (e.g., over appropriate circuits and buses) between or among two or more of the hardware modules. In embodiments in which multiple hardware modules are configured or instantiated at different times, communications between such hardware modules may be achieved, for example, through the storage and retrieval of information in memory structures to which the multiple hardware modules have access. For example, one hardware module performs an operation and stores the output of that operation in a memory device to which it is communicatively coupled. A further hardware module can then, at a later time, access the memory device to retrieve and process the stored output. Hardware modules can also initiate communications with input or output devices, and can operate on a resource (e.g., a collection of information).
The various operations of example methods described herein can be performed, at least partially, by one or more processors that are temporarily configured (e.g., by software) or permanently configured to perform the relevant operations. Whether temporarily or permanently configured, such processors constitute processor-implemented modules that operate to perform one or more operations or functions described herein. As used herein, “processor-implemented module” refers to a hardware module implemented using one or more processors.
Similarly, the methods described herein can be at least partially processor-implemented, with a particular processor or processors being an example of hardware. For example, at least some of the operations of a method can be performed by one or more processors or processor-implemented modules. Moreover, the one or more processors may also operate to support performance of the relevant operations in a “cloud computing” environment or as a “software as a service” (SaaS). For example, at least some of the operations may be performed by a group of computers (as examples of machines including processors), with these operations being accessible via a network (e.g., the Internet) and via one or more appropriate interfaces (e.g., an Application Program Interface (API)).
The performance of certain of the operations may be distributed among the processors, not only residing within a single machine, but deployed across a number of machines. In some example embodiments, the processors or processor-implemented modules are located in a single geographic location (e.g., within a home environment, an office environment, or a server farm). In other example embodiments, the processors or processor-implemented modules are distributed across a number of geographic locations.
In various implementations, the operating system 1804 manages hardware resources and provides common services. The operating system 1804 includes, for example, a kernel 1820, services 1822, and drivers 1824. The kernel 1820 acts as an abstraction layer between the hardware and the other software layers consistent with some embodiments. For example, the kernel 1820 provides memory management, processor management (e.g., scheduling), component management, networking, and security settings, among other functionality. The services 1822 can provide other common services for the other software layers. The drivers 1824 are responsible for controlling or interfacing with the underlying hardware, according to some embodiments. For instance, the drivers 1824 can include display drivers, camera drivers, BLUETOOTH® drivers, flash memory drivers, serial communication drivers (e.g., Universal Serial Bus (USB) drivers), WI-FI® drivers, audio drivers, power management drivers, and so forth.
In some embodiments, the libraries 1806 provide a low-level common infrastructure utilized by the applications 1810. The libraries 1806 can include system libraries 1830 (e.g., C standard library) that can provide functions such as memory allocation functions, string manipulation functions, mathematic functions, and the like. In addition, the libraries 1806 can include API libraries 1832 such as media libraries (e.g., libraries to support presentation and manipulation of various media formats such as Moving Picture Experts Group-4 (MPEG4), Advanced Video Coding (H.264 or AVC), Moving Picture Experts Group Layer-3 (MP3), Advanced Audio Coding (AAC), Adaptive Multi-Rate (AMR) audio codec, Joint Photographic Experts Group (JPEG or JPG), or Portable Network Graphics (PNG)), graphics libraries (e.g., an OpenGL framework used to render in two dimensions (2D) and three dimensions (3D) in a graphic context on a display), database libraries (e.g., SQLite to provide various relational database functions), web libraries (e.g., WebKit to provide web browsing functionality), and the like. The libraries 1806 can also include a wide variety of other libraries 1834 to provide many other APIs to the applications 1810.
The frameworks 1808 provide a high-level common infrastructure that can be utilized by the applications 1810, according to some embodiments. For example, the frameworks 1808 provide various graphic user interface (GUI) functions, high-level resource management, high-level location services, and so forth. The frameworks 1808 can provide a broad spectrum of other APIs that can be utilized by the applications 1810, some of which may be specific to a particular operating system or platform.
In an example embodiment, the applications 1810 include a home application 1850, a contacts application 1852, a browser application 1854, a book reader application 1856, a location application 1858, a media application 1860, a messaging application 1862, a game application 1864, and a broad assortment of other applications such as a third party application 1866. According to some embodiments, the applications 1810 are programs that execute functions defined in the programs. Various programming languages can be employed to create one or more of the applications 1810, structured in a variety of manners, such as object-oriented programming languages (e.g., Objective-C, Java, or C++) or procedural programming languages (e.g., C or assembly language). In a specific example, the third party application 1866 (e.g., an application developed using the ANDROID™ or IOS™ software development kit (SDK) by an entity other than the vendor of the particular platform) may be mobile software running on a mobile operating system such as IOS™, ANDROID™, WINDOWS® PHONE, or another mobile operating system. In this example, the third party application 1866 can invoke the API calls 1812 provided by the operating system 1804 to facilitate functionality described herein.
In various embodiments, the machine 1900 comprises processors 1910, memory 1930, and I/O components 1950, which can be configured to communicate with each other via a bus 1902. In an example embodiment, the processors 1910 (e.g., a Central Processing Unit (CPU), a Reduced Instruction Set Computing (RISC) processor, a Complex Instruction Set Computing (CISC) processor, a Graphics Processing Unit (GPU), a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Radio-Frequency Integrated Circuit (RFIC), another processor, or any suitable combination thereof) include, for example, a processor 1912 and a processor 1914 that may execute the instructions 1916. The term “processor” is intended to include multi-core processors that may comprise two or more independent processors (also referred to as “cores”) that can execute instructions contemporaneously. Although
The memory 1930 comprises a main memory 1932, a static memory 1934, and a storage unit 1936 accessible to the processors 1910 via the bus 1902, according to some embodiments. The storage unit 1936 can include a machine-readable medium 1938 on which are stored the instructions 1916 embodying any one or more of the methodologies or functions described herein. The instructions 1916 can also reside, completely or at least partially, within the main memory 1932, within the static memory 1934, within at least one of the processors 1910 (e.g., within the processor's cache memory), or any suitable combination thereof, during execution thereof by the machine 1900. Accordingly, in various embodiments, the main memory 1932, the static memory 1934, and the processors 1910 are considered machine-readable media 1938.
As used herein, the term “memory” refers to a machine-readable medium 1938 able to store data temporarily or permanently and may be taken to include, but not be limited to, random-access memory (RAM), read-only memory (ROM), buffer memory, flash memory, and cache memory. While the machine-readable medium 1938 is shown in an example embodiment to be a single medium, the term “machine-readable medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, or associated caches and servers) able to store the instructions 1916. The term “machine-readable medium” shall also be taken to include any medium, or combination of multiple media, that is capable of storing instructions (e.g., instructions 1916) for execution by a machine (e.g., machine 1900), such that the instructions, when executed by one or more processors of the machine 1900 (e.g., processors 1910), cause the machine 1900 to perform any one or more of the methodologies described herein. Accordingly, a “machine-readable medium” refers to a single storage apparatus or device, as well as “cloud-based” storage systems or storage networks that include multiple storage apparatus or devices. The term “machine-readable medium” shall accordingly be taken to include, but not be limited to, one or more data repositories in the form of a solid-state memory (e.g., flash memory), an optical medium, a magnetic medium, other non-volatile memory (e.g., Erasable Programmable Read-Only Memory (EPROM)), or any suitable combination thereof. The term “machine-readable medium” specifically excludes non-statutory signals per se.
The I/O components 1950 include a wide variety of components to receive input, provide output, produce output, transmit information, exchange information, capture measurements, and so on. In general, it will be appreciated that the I/O components 1950 can include many other components that are not shown in
In some further example embodiments, the I/O components 1950 include biometric components 1956, motion components 1958, environmental components 1960, or position components 1962, among a wide array of other components. For example, the biometric components 1956 include components to detect expressions (e.g., hand expressions, facial expressions, vocal expressions, body gestures, or eye tracking), measure biosignals (e.g., blood pressure, heart rate, body temperature, perspiration, or brain waves), identify a person (e.g., voice identification, retinal identification, facial identification, fingerprint identification, or electroencephalogram based identification), and the like. The motion components 1958 include acceleration sensor components (e.g., an accelerometer), gravitation sensor components, rotation sensor components (e.g., a gyroscope), and so forth. The environmental components 1960 include, for example, illumination sensor components (e.g., a photometer), temperature sensor components (e.g., one or more thermometers that detect ambient temperature), humidity sensor components, pressure sensor components (e.g., a barometer), acoustic sensor components (e.g., one or more microphones that detect background noise), proximity sensor components (e.g., infrared sensors that detect nearby objects), gas sensor components (e.g., machine olfaction detection sensors, gas detection sensors to detect concentrations of hazardous gases for safety or to measure pollutants in the atmosphere), or other components that may provide indications, measurements, or signals corresponding to a surrounding physical environment. The position components 1962 include location sensor components (e.g., a Global Positioning System (GPS) receiver component), altitude sensor components (e.g., altimeters or barometers that detect air pressure from which altitude may be derived), orientation sensor components (e.g., magnetometers), and the like.
Communication can be implemented using a wide variety of technologies. The I/O components 1950 may include communication components 1964 operable to couple the machine 1900 to a network 1980 or devices 1970 via a coupling 1982 and a coupling 1972, respectively. For example, the communication components 1964 include a network interface component or another suitable device to interface with the network 1980. In further examples, communication components 1964 include wired communication components, wireless communication components, cellular communication components, Near Field Communication (NFC) components, BLUETOOTH® components (e.g., BLUETOOTH® Low Energy), WI-FI® components, and other communication components to provide communication via other modalities. The devices 1970 may be another machine or any of a wide variety of peripheral devices (e.g., a peripheral device coupled via a Universal Serial Bus (USB)).
Moreover, in some embodiments, the communication components 1964 detect identifiers or include components operable to detect identifiers. For example, the communication components 1964 include Radio Frequency Identification (RFID) tag reader components, NFC smart tag detection components, optical reader components (e.g., an optical sensor to detect one-dimensional bar codes such as a Universal Product Code (UPC) bar code, multi-dimensional bar codes such as a Quick Response (QR) code, Aztec Code, Data Matrix, Dataglyph, MaxiCode, PDF417, Ultra Code, Uniform Commercial Code Reduced Space Symbology (UCC RSS)-2D bar codes, and other optical codes), acoustic detection components (e.g., microphones to identify tagged audio signals), or any suitable combination thereof. In addition, a variety of information can be derived via the communication components 1964, such as location via Internet Protocol (IP) geo-location, location via WI-FI® signal triangulation, location via detecting a BLUETOOTH® or NFC beacon signal that may indicate a particular location, and so forth.
In various example embodiments, one or more portions of the network 1980 can be an ad hoc network, an intranet, an extranet, a virtual private network (VPN), a local area network (LAN), a wireless LAN (WLAN), a wide area network (WAN), a wireless WAN (WWAN), a metropolitan area network (MAN), the Internet, a portion of the Internet, a portion of the Public Switched Telephone Network (PSTN), a plain old telephone service (POTS) network, a cellular telephone network, a wireless network, a WI-FI® network, another type of network, or a combination of two or more such networks. For example, the network 1980 or a portion of the network 1980 may include a wireless or cellular network, and the coupling 1982 may be a Code Division Multiple Access (CDMA) connection, a Global System for Mobile communications (GSM) connection, or another type of cellular or wireless coupling. In this example, the coupling 1982 can implement any of a variety of types of data transfer technology, such as Single Carrier Radio Transmission Technology (1×RTT), Evolution-Data Optimized (EVDO) technology, General Packet Radio Service (GPRS) technology, Enhanced Data rates for GSM Evolution (EDGE) technology, third Generation Partnership Project (3GPP) including 3G, fourth generation wireless (4G) networks, Universal Mobile Telecommunications System (UMTS), High Speed Packet Access (HSPA), Worldwide Interoperability for Microwave Access (WiMAX), Long Term Evolution (LTE) standard, others defined by various standard-setting organizations, other long range protocols, or other data transfer technology.
In example embodiments, the instructions 1916 are transmitted or received over the network 1980 using a transmission medium via a network interface device (e.g., a network interface component included in the communication components 1964) and utilizing any one of a number of well-known transfer protocols (e.g., Hypertext Transfer Protocol (HTTP)). Similarly, in other example embodiments, the instructions 1916 are transmitted or received using a transmission medium via the coupling 1972 (e.g., a peer-to-peer coupling) to the devices 1970. The term “transmission medium” shall be taken to include any intangible medium that is capable of storing, encoding, or carrying the instructions 1916 for execution by the machine 1900, and includes digital or analog communications signals or other intangible media to facilitate communication of such software.
Furthermore, the machine-readable medium 1938 is non-transitory (in other words, not having any transitory signals) in that it does not embody a propagating signal. However, labeling the machine-readable medium 1938 “non-transitory” should not be construed to mean that the medium is incapable of movement; the medium should be considered as being transportable from one physical location to another. Additionally, since the machine-readable medium 1938 is tangible, the medium may be considered to be a machine-readable device.
Throughout this specification, plural instances may implement components, operations, or structures described as a single instance. Although individual operations of one or more methods are illustrated and described as separate operations, one or more of the individual operations may be performed concurrently, and nothing requires that the operations be performed in the order illustrated. Structures and functionality presented as separate components in example configurations may be implemented as a combined structure or component. Similarly, structures and functionality presented as a single component may be implemented as separate components. These and other variations, modifications, additions, and improvements fall within the scope of the subject matter herein.
Although an overview of the inventive subject matter has been described with reference to specific example embodiments, various modifications and changes may be made to these embodiments without departing from the broader scope of embodiments of the present disclosure. Such embodiments of the inventive subject matter may be referred to herein, individually or collectively, by the term “invention” merely for convenience and without intending to voluntarily limit the scope of this application to any single disclosure or inventive concept if more than one is, in fact, disclosed.
The embodiments illustrated herein are described in sufficient detail to enable those skilled in the art to practice the teachings disclosed. Other embodiments may be used and derived therefrom, such that structural and logical substitutions and changes may be made without departing from the scope of this disclosure. The Detailed Description, therefore, is not to be taken in a limiting sense, and the scope of various embodiments is defined only by the appended claims, along with the full range of equivalents to which such claims are entitled.
As used herein, the term “or” may be construed in either an inclusive or exclusive sense. Moreover, plural instances may be provided for resources, operations, or structures described herein as a single instance. Additionally, boundaries between various resources, operations, modules, engines, and data stores are somewhat arbitrary, and particular operations are illustrated in a context of specific illustrative configurations. Other allocations of functionality are envisioned and may fall within a scope of various embodiments of the present disclosure. In general, structures and functionality presented as separate resources in the example configurations may be implemented as a combined structure or resource. Similarly, structures and functionality presented as a single resource may be implemented as separate resources. These and other variations, modifications, additions, and improvements fall within a scope of embodiments of the present disclosure as represented by the appended claims. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense.
This application is a continuation of U.S. patent application Ser. No. 14/579,936, filed Dec. 22, 2014, which claims the benefit of U.S. Provisional Application No. 61/921,349, filed Dec. 27, 2013. Each of the aforementioned applications is hereby incorporated by reference in its entirety.
Number | Name | Date | Kind |
---|---|---|---|
3823308 | Goldberg | Jul 1974 | A |
3852571 | Hall et al. | Dec 1974 | A |
3896266 | Waterbury | Jul 1975 | A |
4471216 | Herve | Sep 1984 | A |
5053606 | Kimizu | Oct 1991 | A |
5255352 | Falk | Oct 1993 | A |
5416306 | Imahata | May 1995 | A |
5495568 | Beavin | Feb 1996 | A |
5532464 | Josephson et al. | Jul 1996 | A |
5679938 | Templeton et al. | Oct 1997 | A |
5708422 | Blonder et al. | Jan 1998 | A |
5718178 | Smith | Feb 1998 | A |
5750972 | Botvin | May 1998 | A |
5770843 | Rose et al. | Jun 1998 | A |
5793028 | Wagener et al. | Aug 1998 | A |
5817482 | Bandman et al. | Oct 1998 | A |
5870456 | Rogers | Feb 1999 | A |
5883810 | Franklin et al. | Mar 1999 | A |
5903878 | Talati et al. | May 1999 | A |
5907801 | Albert et al. | May 1999 | A |
5907832 | Pieterse et al. | May 1999 | A |
5930769 | Rose | Jul 1999 | A |
5963917 | Ogram | Oct 1999 | A |
5987500 | Arunachalam | Nov 1999 | A |
6029150 | Kravitz | Feb 2000 | A |
6052675 | Checchio | Apr 2000 | A |
6175655 | Georg et al. | Jan 2001 | B1 |
6212556 | Arunachalam | Apr 2001 | B1 |
6226624 | Watson et al. | May 2001 | B1 |
6266649 | Linden et al. | Jul 2001 | B1 |
6310627 | Sakaguchi | Oct 2001 | B1 |
6415199 | Liebermann | Jul 2002 | B1 |
6490534 | Pfister | Dec 2002 | B1 |
6497359 | Chihara | Dec 2002 | B1 |
6546309 | Gazzuolo | Apr 2003 | B1 |
6643385 | Bravomalo | Nov 2003 | B1 |
6813838 | McCormick | Nov 2004 | B2 |
6836765 | Sussman | Dec 2004 | B1 |
7242999 | Wang | Jul 2007 | B2 |
7308332 | Okada et al. | Dec 2007 | B2 |
7328119 | Pryor et al. | Feb 2008 | B1 |
7354411 | Perry et al. | Apr 2008 | B2 |
7398133 | Wannier et al. | Jul 2008 | B2 |
7430537 | Templeton et al. | Sep 2008 | B2 |
7436976 | Levy et al. | Oct 2008 | B2 |
7548794 | Vandergriff et al. | Jun 2009 | B2 |
7574653 | Croney et al. | Aug 2009 | B2 |
7647041 | Gonsalves et al. | Jan 2010 | B2 |
7663648 | Saldanha et al. | Feb 2010 | B1 |
7714912 | Faisman et al. | May 2010 | B2 |
8024260 | Hogl et al. | Sep 2011 | B1 |
8032943 | DeMello et al. | Oct 2011 | B2 |
8073740 | Siegel | Dec 2011 | B1 |
8090465 | Zeng | Jan 2012 | B2 |
8269778 | Baraff et al. | Sep 2012 | B1 |
8359247 | Vock | Jan 2013 | B2 |
8525828 | Bates | Sep 2013 | B1 |
8655053 | Hansen | Feb 2014 | B1 |
8659596 | Corazza et al. | Feb 2014 | B2 |
8704832 | Taylor et al. | Apr 2014 | B2 |
8711175 | Aarabi | Apr 2014 | B2 |
8736606 | Ramalingam | May 2014 | B2 |
8738292 | Faaborg | May 2014 | B1 |
8749556 | De Aguiar et al. | Jun 2014 | B2 |
8797328 | Corazza et al. | Aug 2014 | B2 |
8935611 | Oberbrunner et al. | Jan 2015 | B2 |
8970585 | Weaver | Mar 2015 | B2 |
9098813 | Konig et al. | Aug 2015 | B1 |
9098873 | Geisner et al. | Aug 2015 | B2 |
9241184 | Weerasinghe | Jan 2016 | B2 |
9378593 | Chhugani et al. | Jun 2016 | B2 |
9420319 | Story, Jr. et al. | Aug 2016 | B1 |
9460342 | Freund et al. | Oct 2016 | B1 |
9465572 | Yamat et al. | Oct 2016 | B2 |
9691161 | Yalniz et al. | Jun 2017 | B1 |
9717982 | Quinn et al. | Aug 2017 | B2 |
9905019 | Applegate et al. | Feb 2018 | B2 |
9923622 | Jactat et al. | Mar 2018 | B2 |
9940749 | Chen et al. | Apr 2018 | B2 |
9953460 | Chhugani et al. | Apr 2018 | B2 |
10068371 | Su et al. | Sep 2018 | B2 |
10089680 | Lin et al. | Oct 2018 | B2 |
10176515 | Lutnick | Jan 2019 | B2 |
10366439 | Leonard et al. | Jul 2019 | B2 |
20010023417 | Stefik et al. | Sep 2001 | A1 |
20010026272 | Feld et al. | Oct 2001 | A1 |
20020004763 | Lam | Jan 2002 | A1 |
20020035793 | Byrd | Mar 2002 | A1 |
20020077837 | Krueger et al. | Jun 2002 | A1 |
20020126328 | Lehmeier et al. | Sep 2002 | A1 |
20020174360 | Ikeda | Nov 2002 | A1 |
20030101105 | Vock | May 2003 | A1 |
20030139896 | Dietz et al. | Jul 2003 | A1 |
20040049309 | Gardner et al. | Mar 2004 | A1 |
20040083142 | Kozzinn | Apr 2004 | A1 |
20040127277 | Walker et al. | Jul 2004 | A1 |
20050289081 | Sporny | Dec 2005 | A1 |
20060020482 | Coulter | Jan 2006 | A1 |
20060059054 | Adiseshan | Mar 2006 | A1 |
20060202986 | Okada et al. | Sep 2006 | A1 |
20070005174 | Thomas | Jan 2007 | A1 |
20070124215 | Simmons | May 2007 | A1 |
20070182736 | Weaver | Aug 2007 | A1 |
20070199076 | Rensin et al. | Aug 2007 | A1 |
20070223704 | Brickell et al. | Sep 2007 | A1 |
20070232272 | Gonsalves et al. | Oct 2007 | A1 |
20070250203 | Yamamoto et al. | Oct 2007 | A1 |
20080022086 | Ho et al. | Jan 2008 | A1 |
20080140650 | Stackpole | Jun 2008 | A1 |
20080163344 | Yang | Jul 2008 | A1 |
20080201228 | Gillet et al. | Aug 2008 | A1 |
20080201638 | Nair | Aug 2008 | A1 |
20080221403 | Fernandez | Sep 2008 | A1 |
20080312765 | Gardiner et al. | Dec 2008 | A1 |
20080312998 | Templeton et al. | Dec 2008 | A1 |
20090002224 | Khatib et al. | Jan 2009 | A1 |
20090018803 | Ko et al. | Jan 2009 | A1 |
20090029337 | Nasci et al. | Jan 2009 | A1 |
20090115777 | Reyers | May 2009 | A1 |
20090144639 | Nims et al. | Jun 2009 | A1 |
20090193675 | Sieber | Aug 2009 | A1 |
20090248537 | Sarkeshik | Oct 2009 | A1 |
20090276300 | Shaw et al. | Nov 2009 | A1 |
20090287452 | Stanley et al. | Nov 2009 | A1 |
20090293116 | Demello et al. | Nov 2009 | A1 |
20100030578 | Siddique et al. | Feb 2010 | A1 |
20100049633 | Wannier et al. | Feb 2010 | A1 |
20100082360 | Chien et al. | Apr 2010 | A1 |
20100097395 | Chang et al. | Apr 2010 | A1 |
20100191770 | Cho et al. | Jul 2010 | A1 |
20100280920 | Scott et al. | Nov 2010 | A1 |
20100305909 | Wolper et al. | Dec 2010 | A1 |
20100306082 | Wolper et al. | Dec 2010 | A1 |
20100313141 | Yu et al. | Dec 2010 | A1 |
20100332567 | Samadani | Dec 2010 | A1 |
20110022372 | Isogai et al. | Jan 2011 | A1 |
20110022965 | Lawrence et al. | Jan 2011 | A1 |
20110063208 | Van den eerenbeemd et al. | Mar 2011 | A1 |
20110099122 | Bright et al. | Apr 2011 | A1 |
20110145093 | Paradise et al. | Jun 2011 | A1 |
20110184831 | Dalgleish | Jul 2011 | A1 |
20110191070 | Ramalingam | Aug 2011 | A1 |
20110231278 | Fries | Sep 2011 | A1 |
20110292034 | Corazza et al. | Dec 2011 | A1 |
20110298897 | Sareen et al. | Dec 2011 | A1 |
20120030060 | Lu et al. | Feb 2012 | A1 |
20120030062 | Stauffer et al. | Feb 2012 | A1 |
20120054059 | Rele | Mar 2012 | A1 |
20120078145 | Malhi et al. | Mar 2012 | A1 |
20120095589 | Vapnik | Apr 2012 | A1 |
20120233003 | Calman et al. | Sep 2012 | A1 |
20120259720 | Nuzzi | Oct 2012 | A1 |
20120281019 | Tamstorf et al. | Nov 2012 | A1 |
20120299912 | Kapur et al. | Nov 2012 | A1 |
20120308087 | Chao et al. | Dec 2012 | A1 |
20120309520 | Evertt et al. | Dec 2012 | A1 |
20120310791 | Weerasinghe | Dec 2012 | A1 |
20130024301 | Mikan et al. | Jan 2013 | A1 |
20130071584 | Bell | Mar 2013 | A1 |
20130108121 | De Jong | May 2013 | A1 |
20130110482 | Ellens et al. | May 2013 | A1 |
20130173226 | Reed et al. | Jul 2013 | A1 |
20130215113 | Corazza et al. | Aug 2013 | A1 |
20130215116 | Siddique et al. | Aug 2013 | A1 |
20130246222 | Weerasinghe | Sep 2013 | A1 |
20130254025 | Liu | Sep 2013 | A1 |
20130258045 | Wojciech | Oct 2013 | A1 |
20130268399 | Lu et al. | Oct 2013 | A1 |
20130317944 | Huang et al. | Nov 2013 | A1 |
20140035913 | Higgins et al. | Feb 2014 | A1 |
20140040041 | Ohnemus et al. | Feb 2014 | A1 |
20140095348 | Goulart | Apr 2014 | A1 |
20140114620 | Grinspun et al. | Apr 2014 | A1 |
20140114884 | Daway | Apr 2014 | A1 |
20140129381 | Fries | May 2014 | A1 |
20140129390 | Mauge et al. | May 2014 | A1 |
20140164902 | Sager | Jun 2014 | A1 |
20140176565 | Adeyoola et al. | Jun 2014 | A1 |
20140180864 | Orlov et al. | Jun 2014 | A1 |
20140236703 | Nordstrand | Aug 2014 | A1 |
20140236753 | Abhyanker | Aug 2014 | A1 |
20140257993 | Paolini | Sep 2014 | A1 |
20140267717 | Pitzer et al. | Sep 2014 | A1 |
20140270540 | Spector et al. | Sep 2014 | A1 |
20140279200 | Hosein et al. | Sep 2014 | A1 |
20140279289 | Steermann | Sep 2014 | A1 |
20140282721 | Kuncl et al. | Sep 2014 | A1 |
20140313192 | Corazza et al. | Oct 2014 | A1 |
20140333614 | Black et al. | Nov 2014 | A1 |
20140368499 | Kaur | Dec 2014 | A1 |
20150130795 | Chhugani et al. | May 2015 | A1 |
20150134302 | Chhugani et al. | May 2015 | A1 |
20150134493 | Su et al. | May 2015 | A1 |
20150134494 | Su et al. | May 2015 | A1 |
20150134495 | Naware et al. | May 2015 | A1 |
20150134496 | Grinblat et al. | May 2015 | A1 |
20150154691 | Curry et al. | Jun 2015 | A1 |
20150186977 | Leonard et al. | Jul 2015 | A1 |
20150366504 | Connor | Dec 2015 | A1 |
20160029706 | Braverman | Feb 2016 | A1 |
20160035061 | Gadre et al. | Feb 2016 | A1 |
20160063588 | Gadre et al. | Mar 2016 | A1 |
20160088284 | Sareen et al. | Mar 2016 | A1 |
20160092956 | Su et al. | Mar 2016 | A1 |
20160117749 | Desmarais et al. | Apr 2016 | A1 |
20160155186 | Su et al. | Jun 2016 | A1 |
20160165988 | Glasgow et al. | Jun 2016 | A1 |
20160165989 | Glasgow et al. | Jun 2016 | A1 |
20160171583 | Glasgow et al. | Jun 2016 | A1 |
20160180447 | Kamalie et al. | Jun 2016 | A1 |
20160180449 | Naware et al. | Jun 2016 | A1 |
20160180562 | Naware et al. | Jun 2016 | A1 |
20160210602 | Siddique et al. | Jul 2016 | A1 |
20160239889 | Nuzzi | Aug 2016 | A1 |
20160247017 | Sareen et al. | Aug 2016 | A1 |
20160249699 | Inghirami | Sep 2016 | A1 |
20160292779 | Rose et al. | Oct 2016 | A1 |
20160292915 | Chhugani et al. | Oct 2016 | A1 |
20170004567 | Dutt et al. | Jan 2017 | A1 |
20180350140 | Su et al. | Dec 2018 | A1 |
20190057428 | Su et al. | Feb 2019 | A1 |
Number | Date | Country |
---|---|---|
199928111 | Jan 2000 | AU |
200197079 | Jun 2002 | AU |
2012240481 | Aug 2015 | AU |
2015255283 | Apr 2017 | AU |
102842089 | Dec 2012 | CN |
103455501 | Dec 2013 | CN |
103605832 | Feb 2014 | CN |
4425271 | Jan 1996 | DE |
19628045 | Jan 1998 | DE |
19634418 | Mar 1998 | DE |
19922150 | Nov 2000 | DE |
19926472 | Dec 2000 | DE |
10022973 | Feb 2001 | DE |
0527639 | Feb 1993 | EP |
0216521 | Nov 1993 | EP |
0519843 | Feb 1994 | EP |
0400911 | Nov 1996 | EP |
0848360 | Jun 1998 | EP |
2091015 | Aug 2009 | EP |
2187325 | May 2010 | EP |
2012A000628 | May 2014 | IT |
2004519748 | Jul 2004 | JP |
2008257747 | Oct 2008 | JP |
2010124604 | Jun 2010 | JP |
1020030097465 | Dec 2003 | KR |
1020100015465 | Feb 2010 | KR |
1020100053646 | May 2010 | KR |
1020100058356 | Jun 2010 | KR |
101606623 | Mar 2016 | KR |
101775855 | Sep 2017 | KR |
1995006294 | Mar 1995 | WO |
1995016971 | Jun 1995 | WO |
1996038813 | Dec 1996 | WO |
1997010560 | Mar 1997 | WO |
1997013228 | Apr 1997 | WO |
1999008242 | Feb 1999 | WO |
1999066436 | Dec 1999 | WO |
2000077754 | Dec 2000 | WO |
2002005224 | Jan 2002 | WO |
2002005224 | Jul 2003 | WO |
2010060113 | May 2010 | WO |
2012110828 | Aug 2012 | WO |
2012138483 | Oct 2012 | WO |
2013188908 | Dec 2013 | WO |
2014076633 | May 2014 | WO |
2014182545 | Nov 2014 | WO |
2016106126 | Jun 2016 | WO |
2016106193 | Jun 2016 | WO |
2016106216 | Jun 2016 | WO |
2016106216 | Aug 2016 | WO |
2016160776 | Oct 2016 | WO |
Entry |
---|
Non Final Office Action received for U.S. Appl. No. 16/167,867, dated Jun. 26, 2020, 25 pages. |
U.S. Appl. No. 14/579,936, filed Dec. 22, 2014. |
Advisory Action received for U.S. Appl. No. 14/579,936, dated Sep. 12, 2018, 3 pages. |
Applicant-Initiated Interview Summary received for U.S. Appl. No. 14/579,936, dated Jan. 14, 2019, 3 pages. |
Applicant-Initiated Interview Summary received for U.S. Appl. No. 14/579,936, dated Aug. 15, 2018, 3 pages. |
Applicant-Initiated Interview Summary received for U.S. Appl. No. 14/579,936, dated Sep. 12, 2018, 3 pages. |
Final Office Action received for U.S. Appl. No. 14/579,936, dated Jun. 27, 2018, 18 pages. |
Final Office Action received for U.S. Appl. No. 14/579,936, dated Jul. 10, 2017, 25 pages. |
Non-Final Office Action received for U.S. Appl. No. 14/579,936, dated Jan. 8, 2018, 15 pages. |
Non-Final Office Action received for U.S. Appl. No. 14/579,936, dated Mar. 24, 2017, 36 pages. |
Non-Final Office Action received for U.S. Appl. No. 14/579,936, dated Oct. 25, 2018, 21 pages. |
Notice of Allowability received for U.S. Appl. No. 14/579,936, dated Apr. 9, 2019, 6 pages. |
Notice of Allowance received for U.S. Appl. No. 14/579,936, dated Mar. 15, 2019, 9 pages. |
Response to Advisory Action filed on Sep. 27, 2018, for U.S. Appl. No. 14/579,936, dated Sep. 12, 2018, 22 pages. |
Response to Final Office Action filed on Aug. 27, 2018, for U.S. Appl. No. 14/579,936, dated Jun. 27, 2018, 20 pages. |
Response to Final Office Action filed on Dec. 1, 2017, for U.S. Appl. No. 14/579,936, dated Jul. 10, 2017, 16 pages. |
Response to Non-Final Office Action filed on Jan. 25, 2019, for U.S. Appl. No. 14/579,936, dated Oct. 25, 2018, 20 pages. |
Response to Non Final Office Action filed on May 31, 2017, for U.S. Appl. No. 14/579,936, dated Mar. 24, 2017, 19 pages. |
Response to Non-Final Office Action filed on Apr. 4, 2018, for U.S. Appl. No. 14/579,936, dated Jan. 8, 2018, 29 pages. |
Automated Clearing House Rules, “A Complete Guide to Rules and Regulations Governing the ACH Network”, The National Autorned Clearing House Association, 1999, 627 pages. |
Harwood et al., “The Use of the Kawabata Evaluation System For Product Development And Quality Control”, Journal of the Society of Dyers and Colourists, vol. 106, No. 2, Feb. 1990, pp. 64-68. |
Horne, “Letter from Gowling Lafleur Henderson LLP to Park, Vaughan and Fleming LLP”, Jul. 21, 2004, 3 pages. |
Jassim, “Semi-Optimal Edge Detector based on Simple Standard Deviation with Adjusted Thresholding”, International Journal of Computer Applications 68, No. 2, Apr. 2013, pp. 43-48. |
McDonnell et al., “Pipeline for Populating Games with Realistic Crowds”, Int. J. Intell Games & Simulation 4, No. 2, Oct. 31, 2018, 1-15 pp. |
Oregon, “ACH Debit Electronic Funds Transfer: Program Guide”, Retrieved from the Internet URL: <http://www.dor.state.or.us/19forms/206-029.pdf>, Feb. 1999, 8 pages. |
Telstra Corporation Ltd, “In the Matter of Australian Application Serial No. 2001271968, in the name of PayPal, Inc-and-In the Matter of Opposition to the Application by Telstra Corporation Limited” “Statement of Grounds of Opposition and Particulars Relating to Each Ground”, Sep. 17, 2007, 1-8 pp. |
Applicant-Initiated Interview Summary received for U.S. Appl. No. 14/503,309, dated Mar. 20, 2018, 3 pages. |
Non-Compliant Notice for U.S. Appl. No. 14/503,309, dated Dec. 21, 2017, 6 pages. |
Non-Final Office Action received for U.S. Appl. No. 14/503,309, dated May 23, 2018, 14 pages. |
Response to Non-Compliant Amendment filed on Apr. 18, 2018, for U.S. Appl. No. 14/503,309, dated Dec. 21, 2017, 10 pages. |
Response to Restriction Requirement filed on Nov. 13, 2017, for U.S. Appl. No. 14/503,309, dated Sep. 13, 2017, 10 pages. |
Restriction Requirement received for U.S. Appl. No. 14/503,309, dated Sep. 13, 2017, 7 pages. |
A World Away, “Selena Mell Sends Home Her First Impressions Of A New Life In The United Arab Emirates”, Community, Nov. 25, 2011, 3 pages. |
Basenese, “Virtual Fitting Rooms . . . Coming to a Store Near You”, Retrieved from the Internet URL :<https://www.wallstreetdaily.com/2011/07/07/virtual-fitting-rooms-fits-me/>, Aug. 13, 2014, 2 pages. |
Binkley, “The Goal: A Perfect First-Time Fit: True Fit Is Online Retailers Latest Attempt to Help Consumers Buy Right Size; No Tape Measures”, Retrieved from the Internet URL: <http://online.wsj.cominewslarticies/SB10001424052702304724404577293593210807790#printMode>, Mar. 23, 2012, 4 pages. |
Bossard et al., “Apparel Classification with Style”, Retrieved from the Internet URL :<http://www.vision.ee.ethz.ch/˜lbossard/bossard_accv12_apparel-classification-with-style.pdf>, 2012, pp. 1-14. |
Bryant, “Fits.me Launches Robot to Help Women Size Up Clothes Online”, Retrieved from the Internet URL : <https://thenextweb.com/eu/2011/06/10/fits-me-launches-robot-to-help-women-size-up-clothes-online/#. tnw_BRXPLr8L>, Jun. 10, 2011, 4 pages. |
Chang, “Virtual Dressing Rooms Hanging the Shape of Clothes Shopping”, Retrieved from the Internet URL: <https://phys.org/news/2012-07-virtual-rooms.html>, Jul. 23, 2012, 3 pages. |
Chang, “Virtual Fitting Rooms Changing the Clothes Shopping Experience”, Retrieved from the Internet URL :<http://articles.latimes.com/2012/jul/13/business/la-fi-virtual-dressing-room-20120714>, Jul. 13, 2012, 2 pages. |
Cheng et al., “A 3D Virtual Show Room for Online Apparel Retail Shop”, In Proceedings: APSIPA ASC 2009: Asia-Pacific Signal and Information Processing Association, Annual Summit and Conference, Retrieved from the Internet URL :< http://www.apsipa.org>, Oct. 4-7, 2009, pp. 193-199. |
Criminisi et al., “Single View Metrology”, International Journal of Computer Vision, vol. 40, Issue 2, 2000, pp. 123-148. |
Feng et al., “A Deformation Transformer for Real-Time Cloth Animation”, Retrieved from the Internet URL :<http://i.cs.hku.hk/˜yzyu/publication/dtcloth-sig2010.pdf>, Jul. 2010, pp. 1-8. |
Fuhrmann et al., “Interaction-Free Dressing of Virtual Humans”, Computers & Graphics 27, No. 1, 2003, pp. 71-82. |
Gioberto, “Garment-Integrated Wearable Sensing for Knee Joint Monitoring”, Proceedings of the 2014 ACM International Symposium on Wearable Computers: Adjunct Program, Sep. 13-17, 2014, pp. 113-118. |
Gioberto et al., “Overlock-Stitched Stretch Sensors: Characterization and Effect of Fabric Property”, Journal of Textile and Apparel, Technology and Management, vol. 8, Issue 3, 2013, 14 pages. |
Hughes et al., “Physical Simulation for Animation and Visual Effects: Parallelization and Characterization for Chip Multiprocessors”, In ACM SIGARCH Computer Architecture News, vol. 35, No. 2, May 2007, 12 pages. |
Jojic et al., “A Framework for Garment Shopping over the Internet”, 2000, 22 pages. |
Karsch et al., “Rendering Synthetic Objects into Legacy Photographs”, ACM Transactions on Graphics (TOG). vol. 30, No. 6, 2011, 12 pages. |
Krashinsky, “Vector-Thread Architecture and Implementation”, Retrieved from the Internet URL :<http://scale.eecs.berkeley.edu/papers/krashinsky-phd.pdf>, May 25, 2007, pp. 1-186. |
Kristensen et al., “Towards a Next Generation Universally Accessible ‘Online Shopping-for-Apparel’ System”, Retrieved from the Internet URL :<http://vbn.aau.dk/files/78490156/VDR_paper_from_HCII2013_V03_LNCS8006_978_3_642_39264_1.pdf>, 2013, pp. 418-427. |
Li et al., “Wearable Sensors in Intelligent Clothing for Measuring Human Body Temperature Based on Optical Fiber Bragg Grating”, Retrieved from the Internet URL: <http://ro.uow.edu.au/eispapers/298>, Optics Express, vol. 20, Issue 11, May 9, 2012, pp. 11740-11752. |
Lim et al., “Characterization of Noise in Digital Photographs for Image Processing”, Retrieved from the Internet URL : <https://www.spiedigitallibrary.org/conference-proceedings-of-spie/6069/60690O/Characterization-of-noise-in-digital-photographs-for-image-processing/10.1117/12.655915.short>, Feb. 10, 2006, 11 pages. |
Luo et al., “Reactive 2D/3D Garment Pattern Design Modification”, Computer-Aided Design, vol. 37, No. 6, May 2005, pp. 623-630. |
Niceinteractive, “Virtual Dressing Room”, Retrieved from the Internet URL: <https://www.youtube.com/watch?v=UhOzN2z3wtl>, Sep. 3, 2012, 2 pages. |
O'Brien, “Fits.me—Imitates Ladies of All Shapes and Sixes, Tries Clothes on for you (video)”, Retrieved from the Internet URL: <https://www.engadget.com/2011/06/13/fits-me-imitates-ladies-of-all-shapes-and-sizes-tries-clothes-o/>, Accessed on Aug. 13, 2014, 10 pages. |
Okreylos, “3D Video Capture with Three Kinects”, Retrieved from the Internet URL: <https://www.youtube.com/watch?v=Ghgbycqb92c>, May 13, 2014, 3 pages. |
Photoshop, “Placing An Image Inside of Another With Photoshop CS6”, Retrieved from the Internet URL: <http://www.photoshopessentials.com/photo-effects/placing-an-image-inside-another-with-photoshop-cs6/>, Sep. 9, 2014, 8 pages. |
Rudolph et al., “A Simple Load Balancing Scheme for Task Allocation in Parallel Machines”, Retrieved from the Internet URL: <http://people.csail.mit.edu/rudolph/Autobiography/LoadBalancing.pdf>, 1991, pp. 237-245. |
Satish et al., “IEEE Xplore Abstact—Can Traditional Programming Bridge the Ninja Performance Gap for Parallel computing applications?”,39th Annual ISCA, Retrieved from the Internet URL: <http://ieeexplore.ieee.org/xpl/articleDetails.jsp?reload=true&arnumbe-r=6237038> 2012, 12 pages. |
Selle et al., “Robust High-Resolution Cloth Using Parallelism, History-Based Collisions and Accurate Friction”, IEEE Transactions on Visualization and Computer Graphics, vol. 15, No. 2, Mar.-Apr. 2009, pp. 339-350. |
Styku, “Styku Startup Revolutionizes Apparel Shopping, Reduces Returns with Virtual Fitting Room”, Retrieved from the Internet URL :<file:///C:/Users/swadhwa/Downloads/Styku_Kinect_CaseStudy%20(1).pdf>, Nov. 6, 2012, 7 pages. |
Yang et al., “Detailed Garment Recovery from a Single-View Image”, Retrieved from the Internet URL :<https://arxiv.org/pdf/1608.01250.pdf>, 2016, pp. 1-13. |
“02Micro Inc Files Taiwan Patent Application for Method and Device for Electronic Fitting,” Global IP News, Electronics Patent News, Dec. 26, 2013, 1 Page. |
Final Office Action Received for U.S. Appl. No. 16/167,867, dated Nov. 9, 2020, 33 Pages. |
Response to Non-Final Office Action filed on Sep. 28, 2020 for U.S. Appl. No. 16/167,867, dated Jun. 26, 2020, 17 pages. |
Notice Of Allowance received for U.S. Appl. No. 16/167,867, dated Feb. 26, 2021, 11 Pages. |
Number | Date | Country | |
---|---|---|---|
20190295147 A1 | Sep 2019 | US |
Number | Date | Country | |
---|---|---|---|
61921349 | Dec 2013 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14579936 | Dec 2014 | US |
Child | 16440760 | US |