Social network applications often enable users to give reviews of events and destinations. To give reviews on a social network site, a user often takes steps to find the appropriate review web pages in order to enter a review. For example, to review an event or destination, a user typically performs an online search for a web page for reviews, which involves the user typing key words into a search field, pressing a button to initiate the search, and then perusing the search results. The user may then access a web page that permits reviews of events and/or destinations by clicking on a link on the search results page. The user may then enter review information.
Embodiments generally relate to a system of obtaining online user reviews of events and/or destinations based on mobile device information. In one embodiment, a method includes receiving location data corresponding to a location of a device, and receiving temporal data corresponding to a time when the device was at the location. The method also includes obtaining event information based on the location data and the temporal data, and generating a user interface on the device to permit reviews of an event associated with the event information. The method also includes generating prompts that elicit reviews from the user, and which are based on the location data and the temporal data.
With further regard to the method, in one embodiment, the location data and the temporal data of the device are provided by activity data associated with one or more actions of the user. In some embodiments, the activity data may include check-in information associated with a device (e.g., mobile phone, etc.). As such, the location data and the temporal data may be determined from check-in information. In some embodiments, the activity data may include payment information associated with a device. As such, the location data and temporal data may be determined from payment information. In an embodiment, the method includes modifying the prompts based on a duration of the device at a particular location, where the duration is determined at least in part from the location data and the temporal data. In an embodiment, the method includes modifying the prompts based on determining a number of check-ins input from the device. In an embodiment the method includes modifying the prompts based on determining responses from the device. In an embodiment, the prompts include questions.
In another embodiment the method includes both receiving location data corresponding to a location of a device, where the location data is provided by activity data associated with one or more actions of a user in a social network, and receiving temporal data corresponding to a time when the device was at the location, where the temporal data is provided by activity data associated with one or more actions of the user. The method also includes obtaining event information based on the location data and the temporal data, where the obtaining of event information includes identifying one or more possible events associated with the location data and the temporal data. The method also includes generating a user interface on the device to permit reviews of an event associated with the event information, and generating prompts based on the location data and the temporal data.
In one embodiment, a system includes one or more processors, and logic encoded in one or more tangible media for execution by the one or more processors. When executed, the logic is operable to perform operations including: receiving location data corresponding to a location of a device, receiving temporal data corresponding to a time when the device was at the location, obtaining event information based on the location data and the temporal data, and generating a user interface on the device to permit reviews of an event associated with the event information, and generating prompts based on the location data and the temporal data.
With further regard to the system, in one embodiment, the temporal data of the device is provided by activity data associated with one or more actions of the user. In another embodiment, the location data and the temporal data are determined from check-in information associated with the device. In an embodiment, the location data is determined from payment information associated with the device. In an embodiment, the logic when executed is further operable to perform operations including modifying the prompts based on a duration of the device at a particular location, where the duration is determined at least in part from the location data and the temporal data. In an embodiment, the logic when executed is further operable to perform operations including modifying the prompts based on determining a number of check-ins input from the device. In an embodiment, the prompts include questions.
Embodiments described herein provide a system of obtaining online user reviews of events and/or destinations (e.g., social functions and/or geographical locations that a user may visit, attend, or participate in, etc.) based on mobile device information. In various embodiments, such mobile device information may include location data and time data, which may be gathered in order to generate pertinent prompts for event and/or destination participants. In some embodiments, a system receives location data corresponding to a location of a mobile device. In some embodiments, the system receives temporal data corresponding to a time when the mobile device is at a particular location. The location data and time data may be provided by activity data associated with one or more actions of the user.
The activity data may be determined, for example, from check-in information (e.g., services that allow a user to check-in to a particular location or venue) associated with the device. The system may then determine event information based on the location data and the temporal data. For example, the system may identify one or more possible events associated with the location data and the temporal data. The system may then generate a user interface on the device to permit participant reviews of an event. The system may generate one or more prompts for the reviews based on the location data and the temporal data. For example, the system may generate a first set of prompts for a user in the case that location data and temporal data indicate that a mobile device is present at a location for an entire day (e.g., the user being an employee at the particular location). In another embodiment, the system may generate a second, different set of prompts for a user in the case that location data and temporal data indicate a first check-in for 10 minutes (e.g., the user being a customer).
Embodiments prompt the user to review an event or destination, and may provide appropriate prompts. Embodiments may elicit more event and destination reviews from participants, and may tailor prompts to encourage a range of reviews (e.g., a range of negative to positive reviews). Sometimes a user may not think to post a review unless something out of the ordinary occurred at the event or destination. For example, many users do not think to review many of their favorite destinations or events, and may only post reviews when they have a negative experience with an event or destination.
In various embodiments, an event may be any type of function, celebration, class, ceremony, experience, etc., and such events may occur at respective destinations. The term “event” is hereafter used to connote either an event or a destination. Non-limiting examples may include a gym class, movie, concert, meal at a restaurant, speech, wedding, sports game, match, symphony, store, amusement park trip, birthday party, etc. Examples may also include various combinations of possible events. For example, a user may attend a gym class, movie, have a meal at a restaurant, etc.
System 102 may represent multiple systems, server devices, and databases. Also, there may be any number of client devices 110, 120, 130, and 140. In other embodiments, network environment 100 may not have all of the components shown and/or may have other elements including other types of elements instead of, or in addition to, those shown herein. In various embodiments, users U1, U2, U3, and U4 may communicate with each other, and may interact with system 102 using respective client devices 110, 120, 130, and 140.
Location data may be obtained using any suitable approach. For example, a global positioning system (GPS), which may be associated with a client device 110, 120, 130, or 140, may maintain a log file of geographic coordinates. In some implementations, the log file is accessible to the user only, and cannot be obtained from the cellular telephone absent an affirmative act by the user and/or the user's consent. In some implementations, the user may permit the location log file to be maintained on a location database. In another implementation, location data may be obtained from a geo-location database that stores both a user's time and location. Accordingly, the client devices 110, 120, 130, and 140 may obtain user location data from the location database.
In block 204, system 102 receives temporal data corresponding to a time at which a client device 110, 120, 130, or 140 is at the location. In various embodiments, system 102 may receive temporal data from the temporal database. In another embodiment, temporal data may be obtained from a geo-location database that stores both a user's time and location. In various embodiments, system 102 may receive temporal data from the check-in database containing check-in information. For example, some social networking services or check-in services allow a user to check-in to a location. The embodiments described herein also apply to users checking in from any suitable computing device. For example, a user may indicate presence through checking in to a location using a computer terminal or by paying for an event at a ticket kiosk. In some implementations the kiosk or computer terminal may send the check-in information to the user's mobile device via system 102. This check-in information may be stored, for example, on the user's mobile computing device in which the check-in occurred, or the information may be available, with the user's permission, from the user's account. Example accounts may include a user's account with the social network system and/or with a check-in service.
Also, a user may access services by signing into a single account with a unified sign-in system that permits access to multiple services and can be used with any type of client device. For example, the user may sign in to their single account in the unified sign-in system and access various services and applications such as email, web bookmarks, photo sharing applications, search history, review pages, check-in services, or mobile payment services. In the foregoing example, the user could sign in using any type of client device: non-limiting examples include a mobile computing device, cell phone, tablet, laptop, mobile internet device, desktop, or wearable computer.
Temporal data may be obtained using any suitable approach. For example a global positioning system (GPS), which is associated with the client device 110, 120, 130, and 140, may maintain a log file of times when a mobile device was at corresponding locations. In some implementations, the log file is accessible to the user only, and cannot be obtained from the cellular telephone absent an affirmative act by the user and/or the user's consent. In another embodiment, the user may permit the temporal log file to be maintained on a temporal database, accordingly, system 102 may obtain user temporal data from the location database.
For ease of illustration, some embodiments are described in the context of a mobile device. In various embodiments, system 102 may obtain location data and temporal data from any suitable database such as an event database. In some implementations, the event database could be contained within social network database 106, or any other suitable database, or from a combination of any suitable client devices (e.g., client devices 110, 120, 130, and 140).
Referring still to
For example, if system 102 finds possible events (e.g., a movie, a concert, etc.) in a search, and system 102 determines that a device is present at the location and time of the events, system 102 may postulate or infer that the user associated with the device is attending one of the events and might like to review the event.
In block 207, system 102 correlates received location data, temporal data, and event information, if any. For example, system 102 may determine event information solely from check-in information. In another example, system 102 may determine that there are no events taking place based on the present location and time.
In block 208, system 102 generates a user interface on the device to permit reviews of an event. In various embodiments, the user interface includes brief descriptions of events that may be presented on a device to a user; and the user may select which of those events to review, if any. In some embodiments, a user interface may include a thumbnail image of an event. For example, a user interface may provide a picture of a movie poster, a restaurant sign, or a concert poster.
Referring to
Referring to
In an example scenario, the user may check-in at a gym, where event 314 may be a pilates class, event 316 may be a spinning class, and event 318 may be a yoga class. In some implementations, system 102 may use suggestions in the form of event buttons 304, 306, and 308 to be displayed, effectively asking the user to review the pilates class, spinning class, yoga class, etc. In another example scenario, the user may check-in at a movie theater, where event 314 may be a first movie playing at the theater, event 316 may be a second movie playing at the theater, and event 318 may be a third movie playing at the theater. In one example scenario, the user may check-in at a dining complex with multiple restaurants, where event 314 may be a meal at an Italian restaurant, event 316 may be a meal at a French restaurant, and event 318 may be a meal at a Japanese restaurant.
In another example scenario, a user may check-in at a venue that includes a gym, movie theater, and restaurant. In some implementations, based on location data and temporal data, system 102 may present the user with event buttons asking the user to review event 314 a movie, event 316 a yoga class, event 318 a meal at a Japanese restaurant, etc.
In some implementations, if system 102 provides multiple suggestions, system 102 may enable the user to accept one of the prompts by selecting or clicking one of the event buttons. In one embodiment, if there are numerous possible events, system 102 may provide a subset of all possible events in order to simplify the event selection process. In various implementations, the subset provided may be based on any combination of historical user check-in information, temporal data, location data, payment information, search information, etc.
For example, if system 102 determines that over the course of a year the user has never attended a yoga class and user is near the yoga class at the time that it is occurring, the yoga class could be eliminated from the possible events presented for review. In another example, if the user utilizes a mobile payment service to purchase a ticket for a particular movie then system 102 may determine the exact movie the user attended and invite the user to review that movie. In another scenario, if while eating a meal at an Italian restaurant, the user conducts a search for “whole wheat pasta” system 102 may correlate this search with menu data for the restaurant and determine that the user may have tried a particular whole wheat rigatoni dish and invite the user to review that dish.
In one implementation, system 102 may enable the user to manually type in an event and select it. In one implementation, system 102 may enhance manually typed events by providing auto-complete suggestions associated with the event information while the user types.
In block 210, system 102 generates event review prompts that elicit reviews from a user. In some implementations, review prompts may generated or synthesized from any suitable algorithm. In some embodiments, prompts may be obtained from any suitable database. Prompts contain various forms such as radio buttons and checkboxes. For example, review prompts and prompt buttons may correspond to events, based in part on location data and temporal data. Prompts and the generation of prompts are further described with reference to
In some implementations, an event participant may choose one event from GUI 300 of
There are also many ways to answer specific prompts that may be simply represented with text that has no additional functionality. For example, instead of a prompt button, simple text in separate paragraphs may be displayed. After each prompt, input elements such as check boxes, text fields, touch screen numerical rating scales, etc. may be displayed to prompt the user's response to each respective prompt.
Referring to
In some implementations, any one or more of prompts 414, 416, and 418, may be presented as questions. In some embodiments, any one or more of prompts 414, 416, and 418 may elicit reviews from a user. Prompts 414, 416, and 418 may be displayed to elicit the user's response and guide the user in an event review. Many iterations are possible which share similar functionality of guiding a user to select a question and capturing that user's input in response to the question prompts. For example, a user may click a question prompt button to answer a question associated with an event review or provide a direct answer to a question to an input field associated with the review.
In one embodiment, prompts 414, 416, and 418, may be generated based on the duration of the user at a particular location. For example, if system 102 determines that a device checks in to a location every morning at 9:00 a.m. and effectively is not present any longer around 6:00 p.m., system 102 may determine that the user associated with the device is an employee at that location, such as a trainer at a gym or a worker at a post office. Accordingly, system 102 will generate prompts for the employee that would be different from a customer of the post office as compared to a worker. For example, system 102 would generate a prompt for the customer that relates to how he or she feels about waiting in line.
In another example implementation, system 102 may generate specific review prompts, based on frequency of visits or attendance. For example, if system 102 determines a device has checked into a movie for a second time, prompts may be focused on how much a user liked the movie the second time. In another example, if system 102 determines a that device has checked-in to an amusement park weekly during the course of a summer, system 102 may generate prompts about new rides in the beginning of the summer, but may not generate prompts about new rides when the user checks-in at the end of the summer.
In various embodiments, prompts may be generated based on predetermined information about a device's home location. For example, if a user consents, a device may include as part of a profile, the device's base or home location. System 102 may determine, based on location data and temporal data, that the device is not proximate its home location and that the user of the device is attending an event while on business or personal travel. System 102 may then generate specific prompts designed to fit the user's traveling profile. For example the prompt may state, “Was this your favorite restaurant on your first trip to Paris?”
In another embodiment, system 102 may generate prompts based on determining that a device is at an event in its home location. For example, system 102 may then generate a prompt such as, “As a Chicagoan, do you feel this restaurant is the best Italian restaurant in Chicago?”
In another example embodiment, system 102 may generate different prompts over multiple visits to an event based on a combination of one or more of the foregoing factors and a user's past responses to prompts. System 102 may also generate prompts by correlating information from one or more of the foregoing factors.
For example, system 102 may present more general prompts on the first check-in of a device at a sports game such as: “What did you think of the stadium?”; “What do you think of the team?”; or “What did you think of the game?” In subsequent visits, based on mobile payment information, if a user has user has consented to the use of such information, then system 102 may ask, “What did you think of the hot dogs?” Or if in past responses to, “What do you think of the team?”, the user has repeatedly responded with the same response, for example “great”, system 102 may stop asking that prompt.
In another example scenario, system 102 may determine based on the duration of a user's stay at a restaurant that the user has ordered take-out. For example, system 102 may evaluate the time that the user is at the location based on location data and temporal data. If the time falls within a predetermined range (e.g., less than 10 minutes), system 102 may determine that user is likely ordering take-out. System 102 may correlate this take-out determination with mobile payment information. For example, system does not confuse a scenario where a user has simply stopped by the restaurant to peruse a menu. Once system 102 determines that an order is take-out, prompts may focus, for example, more on issues associated with take-out (e.g., whether the packaging is leak proof, utensils are provided, and hotness of food is proper when picked up). In another embodiment, system 102 may determine that a user has used a mobile or phone based ordering system and requested home delivery of a restaurant food order. Accordingly, upon delivery of the food order, system 102 may generate prompts that relate to delivery times and courtesy of the delivery person.
By virtue of the of the foregoing process, a user is prompted to review an event and is guided by context specific prompts while the user is attending or has just completed an event. In some embodiments, prompts may be generated through a search corresponding to check-in information. Some services, such as location based check-in services and/or social networking services, allow a user to check-in to a location (e.g., an event). This information may be available, with the user's permission from their account or may be stored (e.g., on the user's mobile computing device).
In some implementations, the check-in information may contain geographic coordinates, which identify an event or site (e.g., a particular movie theater, etc.) or specific location within the event or site. The check-in information may also contain the time at which the user checked in. This information may be used to fetch pertinent review prompts.
In another example scenario, a user may check-in at a store. While browsing at the store the user may perform searches on a mobile device for red shoes. System 102 may correlate search data with store inventory data. As such, system 102 may present the user with prompts about particular merchandise (e.g., brands of shoes that include red color pairs in particular styles, etc.).
In an example scenario, user may check into an event such as a concert on a cell phone and indicate that they do not wish to receive review prompts during the concert. If user consents, system 102 could prompt user for a review of the concert later in time. For example, when user logs onto their desktop or tablet computer at home or while traveling the next day, they could receive an email with review questions about the concert.
Responses to prompts may be presented in various ways. For example, system 102 may generate multiple choice response buttons (not shown) on GUI 400 that correspond to prompts. In an example, response buttons could include ranges of answers such as very satisfied, satisfied, neutral, and dissatisfied. In another example, prompts could ask for numerical ratings in a pre-defined range such as 1-30 for various characteristics. For example, for a restaurant, users could be asked to provide a rating for food, decor, and service. In another example embodiment, prompts could cue the user to enter an overall rating for a place or event. In another embodiment, prompts could cue the user to enter ratings for categories of appeal, facilities, and service.
In one embodiment, system 102 may enable the user to manually type in responses to prompts. In one embodiment, manually typed responses be enhanced by system 102 providing auto-complete suggestions of known responses from response information while the user types.
Although the steps, operations, or computations described herein may be presented in a specific order, the order may be changed in particular embodiments. Other orderings of the steps are possible, depending on the particular implementation. In some particular embodiments, multiple steps shown as sequential in this specification may be performed at the same time.
While system 102 is described as performing the steps as described in the embodiments herein, any suitable component or combination of components of system 102 or any suitable processor or processors associated with system 102 may perform the steps described.
In situations where system 102 collects and/or uses personal information about users, or may make use of personal information, the users may be provided with an opportunity to control whether programs or features collect user information (e.g., information about a user's social network, social actions or activities, profession, a user's preferences, or a user's current location), or to control whether and/or how to receive content from the content server (not shown) that may be more relevant to the user. In addition, certain data may be treated in one or more ways before it is stored or used, so that personally identifiable information is removed. For example, a user's identity may be treated so that no personally identifiable information can be determined for the user, or a user's geographic location may be generalized where location information is obtained (such as to a city, ZIP code, or state level), so that a particular location cannot be determined. Thus, the user may have control over how information is collected by system 102 and used by a content server.
For ease of illustration,
Although the description has been described with respect to particular embodiments thereof, these particular embodiments are merely illustrative, and not restrictive. Concepts illustrated in the examples may be applied to other examples and embodiments.
Note that the functional blocks, methods, devices, and systems described in the present disclosure may be integrated or divided into different combinations of systems, devices, and functional blocks as would be known to those skilled in the art.
Any suitable programming languages and programming techniques may be used to implement the routines of particular embodiments. Different programming techniques may be employed such as procedural or object-oriented. The routines may execute on a single processing device or multiple processors. Although the steps, operations, or computations may be presented in a specific order, the order may be changed in different particular embodiments. In some particular embodiments, multiple steps shown as sequential in this specification may be performed at the same time.
A “processor” includes any suitable hardware and/or software system, mechanism or component that processes data, signals or other information. A processor may include a system with a general-purpose central processing unit, multiple processing units, dedicated circuitry for achieving functionality, or other systems. Processing need not be limited to a geographic location, or have temporal limitations. For example, a processor may perform its functions in “real-time,” “offline,” in a “batch mode,” etc. Portions of processing may be performed at different times and at different locations, by different (or the same) processing systems. A computer may be any processor in communication with a memory. The memory may be any suitable processor-readable storage medium, such as random-access memory (RAM), read-only memory (ROM), magnetic or optical disk, or other tangible media suitable for storing instructions for execution by the processor.
This application is a continuation of U.S. application Ser. No. 13/658,490 filed Oct. 23, 2012, the contents of which is incorporated herein by reference in its entirety.
| Number | Name | Date | Kind |
|---|---|---|---|
| 6421358 | Stimmel et al. | Jul 2002 | B1 |
| 6513014 | Walker et al. | Jan 2003 | B1 |
| 6557041 | Mallart | Apr 2003 | B2 |
| 6940998 | Garoutte | Sep 2005 | B2 |
| 6947976 | Devitt et al. | Sep 2005 | B1 |
| 7298960 | Taylor | Nov 2007 | B1 |
| 7489946 | Srinivasan et al. | Feb 2009 | B2 |
| 7512829 | Mital et al. | Mar 2009 | B2 |
| 7539747 | Lucovsky et al. | May 2009 | B2 |
| 7552068 | Brinkerhoff | Jun 2009 | B1 |
| 7620902 | Manion et al. | Nov 2009 | B2 |
| 7804426 | Etcheson | Sep 2010 | B2 |
| 7812998 | Miers | Oct 2010 | B2 |
| 7844482 | Larson | Nov 2010 | B1 |
| 7849481 | Moon et al. | Dec 2010 | B2 |
| 7903904 | Loeb et al. | Mar 2011 | B1 |
| 7917859 | Singh et al. | Mar 2011 | B1 |
| 8332512 | Wu et al. | Dec 2012 | B1 |
| 8359285 | Dicker | Jan 2013 | B1 |
| 8359540 | Darling | Jan 2013 | B2 |
| 8380039 | Luo et al. | Feb 2013 | B2 |
| 8412773 | Chapweske et al. | Apr 2013 | B1 |
| 8630494 | Svendsen | Jan 2014 | B1 |
| 8730397 | Zhang | May 2014 | B1 |
| 8761523 | Cok et al. | Jun 2014 | B2 |
| 8914483 | Dixon et al. | Dec 2014 | B1 |
| 9021045 | Pennington | Apr 2015 | B2 |
| 20030099379 | Monk et al. | May 2003 | A1 |
| 20030214128 | Roberts et al. | Nov 2003 | A1 |
| 20040073615 | Darling | Apr 2004 | A1 |
| 20050010470 | Marino | Jan 2005 | A1 |
| 20050021822 | Cherkasova et al. | Jan 2005 | A1 |
| 20050105396 | Schybergson | May 2005 | A1 |
| 20050187943 | Finke-Anlauff et al. | Aug 2005 | A1 |
| 20060112080 | Chen et al. | May 2006 | A1 |
| 20060125930 | Mindrum et al. | Jun 2006 | A1 |
| 20060159007 | Frutiger et al. | Jul 2006 | A1 |
| 20060234769 | Srinivasan et al. | Oct 2006 | A1 |
| 20060287105 | Willis | Dec 2006 | A1 |
| 20070022447 | Arseneau et al. | Jan 2007 | A1 |
| 20070219949 | Mekikian | Sep 2007 | A1 |
| 20070260361 | Etcheson | Nov 2007 | A1 |
| 20070266047 | Cortes et al. | Nov 2007 | A1 |
| 20070294177 | Volk et al. | Dec 2007 | A1 |
| 20080005761 | Repasi et al. | Jan 2008 | A1 |
| 20080086511 | Takao | Apr 2008 | A1 |
| 20080114844 | Sanchez et al. | May 2008 | A1 |
| 20080133697 | Stewart et al. | Jun 2008 | A1 |
| 20080136930 | Nagai | Jun 2008 | A1 |
| 20080144135 | Miers | Jun 2008 | A1 |
| 20080184139 | Stewart et al. | Jul 2008 | A1 |
| 20080186926 | Baio et al. | Aug 2008 | A1 |
| 20080189175 | Chan | Aug 2008 | A1 |
| 20080201299 | Lehikoinen et al. | Aug 2008 | A1 |
| 20080301736 | Heilbron et al. | Dec 2008 | A1 |
| 20090019902 | Baranek | Jan 2009 | A1 |
| 20090070412 | D'Angelo et al. | Mar 2009 | A1 |
| 20090171873 | Dolin et al. | Jul 2009 | A1 |
| 20090191902 | Osbourne | Jul 2009 | A1 |
| 20090248692 | Tsukagoshi | Oct 2009 | A1 |
| 20090248703 | Tsukagoshi et al. | Oct 2009 | A1 |
| 20090276531 | Myka et al. | Nov 2009 | A1 |
| 20090319472 | Jain et al. | Dec 2009 | A1 |
| 20100088182 | Ryder | Apr 2010 | A1 |
| 20100088336 | Johnston et al. | Apr 2010 | A1 |
| 20100124378 | Das et al. | May 2010 | A1 |
| 20100138295 | Caron | Jun 2010 | A1 |
| 20100158315 | Martin | Jun 2010 | A1 |
| 20100169389 | Weber et al. | Jul 2010 | A1 |
| 20100198880 | Petersen | Aug 2010 | A1 |
| 20100241945 | Chen et al. | Sep 2010 | A1 |
| 20100250633 | Hannuksela et al. | Sep 2010 | A1 |
| 20100262916 | Jones et al. | Oct 2010 | A1 |
| 20100290699 | Adam et al. | Nov 2010 | A1 |
| 20100315664 | Miers | Dec 2010 | A1 |
| 20100318611 | Curtin et al. | Dec 2010 | A1 |
| 20110013845 | Tu et al. | Jan 2011 | A1 |
| 20110066743 | Hurley et al. | Mar 2011 | A1 |
| 20110072035 | Gaucas et al. | Mar 2011 | A1 |
| 20110080424 | Peters et al. | Apr 2011 | A1 |
| 20110099199 | Stalenhoef et al. | Apr 2011 | A1 |
| 20110137709 | Mayer et al. | Jun 2011 | A1 |
| 20110138003 | Yoon et al. | Jun 2011 | A1 |
| 20110150324 | Ngan et al. | Jun 2011 | A1 |
| 20110161423 | Pratt et al. | Jun 2011 | A1 |
| 20110167136 | Naimark et al. | Jul 2011 | A1 |
| 20110183651 | Mundy | Jul 2011 | A1 |
| 20110208702 | Minde et al. | Aug 2011 | A1 |
| 20110211737 | Krupka et al. | Sep 2011 | A1 |
| 20110214077 | Singh et al. | Sep 2011 | A1 |
| 20110231240 | Schoen et al. | Sep 2011 | A1 |
| 20110238631 | Cortes et al. | Sep 2011 | A1 |
| 20110252340 | Thomas | Oct 2011 | A1 |
| 20110270923 | Jones et al. | Nov 2011 | A1 |
| 20110276396 | Rathod | Nov 2011 | A1 |
| 20110276513 | Erhart et al. | Nov 2011 | A1 |
| 20110295667 | Butler | Dec 2011 | A1 |
| 20110296536 | Muller et al. | Dec 2011 | A1 |
| 20120007995 | Barrett | Jan 2012 | A1 |
| 20120016948 | Sinha | Jan 2012 | A1 |
| 20120023129 | Vedula et al. | Jan 2012 | A1 |
| 20120027256 | Kiyohara | Feb 2012 | A1 |
| 20120030194 | Jain | Feb 2012 | A1 |
| 20120047147 | Redstone et al. | Feb 2012 | A1 |
| 20120072428 | Kao et al. | Mar 2012 | A1 |
| 20120079396 | Neer et al. | Mar 2012 | A1 |
| 20120082378 | Peters et al. | Apr 2012 | A1 |
| 20120092685 | Barrett | Apr 2012 | A1 |
| 20120100869 | Liang et al. | Apr 2012 | A1 |
| 20120102409 | Fan et al. | Apr 2012 | A1 |
| 20120109901 | Mase | May 2012 | A1 |
| 20120114296 | Luo et al. | May 2012 | A1 |
| 20120122554 | Paquet et al. | May 2012 | A1 |
| 20120123867 | Hannan | May 2012 | A1 |
| 20120130823 | Levin | May 2012 | A1 |
| 20120136689 | Ickman et al. | May 2012 | A1 |
| 20120153016 | Slaby et al. | Jun 2012 | A1 |
| 20120158871 | Amano et al. | Jun 2012 | A1 |
| 20120179664 | Auerbach et al. | Jul 2012 | A1 |
| 20120213404 | Steiner | Aug 2012 | A1 |
| 20120214568 | Herrmann | Aug 2012 | A1 |
| 20120220314 | Altman et al. | Aug 2012 | A1 |
| 20120246003 | Hart | Sep 2012 | A1 |
| 20120251011 | Gao et al. | Oct 2012 | A1 |
| 20120278387 | Garcia et al. | Nov 2012 | A1 |
| 20120324589 | Nukala et al. | Dec 2012 | A1 |
| 20130006882 | Galliani | Jan 2013 | A1 |
| 20130013683 | Elliott | Jan 2013 | A1 |
| 20130039547 | Liu et al. | Feb 2013 | A1 |
| 20130066963 | Odio et al. | Mar 2013 | A1 |
| 20130101220 | Bosworth et al. | Apr 2013 | A1 |
| 20130117365 | Padmanabhan et al. | May 2013 | A1 |
| 20130124311 | Sivanandan | May 2013 | A1 |
| 20130129232 | Cok et al. | May 2013 | A1 |
| 20130141529 | Sathish | Jun 2013 | A1 |
| 20130166639 | Shaffer et al. | Jun 2013 | A1 |
| 20130232251 | Pauley | Sep 2013 | A1 |
| 20130305287 | Wong et al. | Nov 2013 | A1 |
| 20140047335 | Lewis et al. | Feb 2014 | A1 |
| 20140129942 | Rathod | May 2014 | A1 |
| 20150005010 | Zhang et al. | Jan 2015 | A1 |
| 20160070809 | Rathus | Mar 2016 | A1 |
| 20170124385 | Ganong et al. | May 2017 | A1 |
| Number | Date | Country |
|---|---|---|
| 101167066 | Apr 2008 | CN |
| 102265649 | Nov 2011 | CN |
| 2000-067057 | Mar 2000 | JP |
| 2004-222056 | Aug 2004 | JP |
| 2006-101095 | Apr 2006 | JP |
| 2007-249821 | Sep 2007 | JP |
| 2008-077445 | Apr 2008 | JP |
| 2008-146174 | Jun 2008 | JP |
| 2008-146230 | Jun 2008 | JP |
| 2010-237751 | Oct 2010 | JP |
| 2010-263621 | Nov 2010 | JP |
| 2011-526013 | Sep 2011 | JP |
| 2012-514276 | Jun 2012 | JP |
| 2009009204 | Jan 2015 | JP |
| 10-2009-0080063 | Jul 2009 | KR |
| 10-2010-0120282 | Nov 2010 | KR |
| 10-2012-0029861 | Mar 2012 | KR |
| 10-2012-0034600 | Apr 2012 | KR |
| 2007113462 | Oct 2007 | WO |
| 2007135871 | Nov 2007 | WO |
| 2007135971 | Nov 2007 | WO |
| 2008045701 | Apr 2008 | WO |
| 2009076555 | Jun 2009 | WO |
| 2009082814 | Jul 2009 | WO |
| WO 2010075049 | Jul 2010 | WO |
| 2010108157 | Sep 2010 | WO |
| 2011001587 | Jan 2011 | WO |
| 2011149961 | Dec 2011 | WO |
| 2011149961 | Dec 2011 | WO |
| WO 2011149961 | Dec 2011 | WO |
| Entry |
|---|
| Unpublished U.S. Appl. No. 13/346,385, filed Jan. 9, 2012. |
| Non-Final Rejection received for U.S. Appl. No. 13/219,835, filed Aug. 29, 2011, dated Apr. 2, 2013. |
| International Search Report and Written Opinion for corresponding PCT application No. PCT/US2013/066257, dated Apr. 25, 2014. |
| Notice of Allowance received for U.S. Appl. No. 13/050,195, filed Mar. 17, 2011, dated Aug. 15, 2014. |
| Written Opinion from ISA received for related PCT Application No. PCT/US2012/024874, dated Aug. 18, 2013. |
| Japanese Office Action received in Japanese Application No. 2013-554527, dated Aug. 18, 2015. |
| International Preliminary Report on Patentability received for related PCT Application No. PCT/US2012/024874, dated Aug. 21, 2013. |
| International Search Report received for related PCT Application No. PCT/US2012/024874, dated Aug. 23, 2012. |
| International Search Report and Written Opinion From ISA received for related PCT Application No. PCT/US2013/048318, dated Dec. 9, 2014. |
| Final Rejection received for U.S. Appl. No. 13/346,385, filed Jan. 9, 2012, dated Feb. 14, 2014. |
| Notice of Allowance received for U.S. Appl. No. 13/050,195, filed Mar. 17, 2011, dated Jan. 24, 2014. |
| Office Action received for U.S. Appl. No. 13/050,195, filed Mar. 17, 2011, dated Jan. 29, 2013. |
| International Preliminary Report on Patentability received for related PCT Application No. PCT/US2013/048318, dated Jan. 8, 2015. |
| Non-Final Rejection received for U.S. Appl. No. 13/929,498, filed Jun. 27, 2013, dated Jul. 15, 2015. |
| Non-Final Rejection for U.S. Appl. No. 13/658,490, filed Oct. 23, 2012, dated Jul. 2, 2015. |
| Non-Final Rejection received for U.S. Appl. No. 13/346,385, filed Jan. 9, 2012, dated Jul. 22, 2015. |
| First Action Interview Office Action for U.S. Appl. No. 13/658,490, filed Oct. 23, 2012, dated Jul. 24, 2014. |
| Notice of Allowance received for U.S. Appl. No. 13/534,555, filed Jun. 27, 2012, dated Jun. 11, 2015. |
| Final Rejection for U.S. Appl. No. 13/219,835, filed Aug. 29, 2011, dated Jun. 4, 2015. |
| Pre-Interview First Office Action received for U.S. Appl. No. 13/658,490, filed Oct. 23, 2012, dated Mar. 18, 2014. |
| Non-Final Rejection received for U.S. Appl. No. 13/050,195, filed Mar. 17, 2011, dated Mar. 28, 2014. |
| Office Action received for U.S. Appl. No. 13/050,195, filed Mar. 17, 2011, dated Mar. 28, 2014. |
| Final Rejection for U.S. Appl. No. 13/542,379, filed Jul. 5, 2012, dated May 12, 2015. |
| Examination Report received for Australian Patent Application No. 2012217920, dated Nov. 25, 2014. |
| Final Rejection received for U.S. Appl. No. 13/658,490, filed Oct. 23, 2012, dated Nov. 7, 2014. |
| Non-Final Rejection received for U.S. Appl. No. 13/219,835, filed Aug. 29, 2011, dated Oct. 24, 2014. |
| Non-Final Rejection for U.S. Appl. No. 13/542,379, filed Jul. 5, 2012, dated Oct. 3, 2014. |
| Final Rejection received for U.S. Appl. No. 13/219,835, filed Aug. 29, 2011, dated Oct. 9, 2013. |
| Office Action received for U.S. Appl. No. 13/050,195, filed Mar. 17, 2011, dated Sep. 10, 2013. |
| Non-Final Rejection received for U.S. Appl. No. 13/346,385, filed Jan. 9, 2012, dated Sep. 20, 2013. |
| Non-Final Rejection received for U.S. Appl. No. 13/534,555, filed Jun. 27, 2012, dated Sep. 30, 2014. |
| Notice of Allowance for U.S. Appl. No. 13/661,425, dated Dec. 8, 2015, 14 pages. |
| Korean Notice of Preliminary Rejection received in Korean Patent Application No. 10-2015-7002159 (with Translation), dated Jan. 26, 2016, 15 pages. |
| Final Rejection for U.S. Appl. No. 13/661,425, dated Mar. 10, 2015, 17 pages. |
| Australian Patent Office. Patent Examination Report No. 1 for Australian Patent Application No. 2013334606, dated Nov. 25, 2015, 2 pages. |
| Non-Final Rejection received for U.S. Appl. No. 14/571,213, dated Sep. 24, 2015, 21 pages. |
| Non-Final Rejection for U.S. Appl. No. 13/661,425, dated Sep. 18, 2014, 24 pages. |
| Notice of Allowance received for U.S. Appl. No. 13/050,195, filed Mar. 17, 2011 dated Aug. 15, 2014, dated Aug. 15, 2014, 27 pages. |
| Notice of Allowance received for U.S. Appl. No. 13/534,555, filed Jun. 27, 2012, dated Oct. 23, 2015, 31 pages. |
| Non-Final Rejection for U.S. Appl. No. 13/542,379, filed Jul. 5, 2012, dated Nov. 19, 2015, 42 pages. |
| JPO Notice of Allowance for Japanese Application No. 2013-554527, dated Nov. 20, 2015, 6 pages. |
| PCT. International Search Report and the Written Opinion of the International Search Authority received in related PCT Application No. PCT/US2013/066414, dated Jan. 29, 2014, 8 pages. |
| Notice of Reason for Rejection for Japanese Application No. 2013-554527, dated Aug. 18, 2015, 9 pages. |
| “Extended European Search Report”, EP Application No. 13737075.5, dated Jan. 13, 2016, 8 Pages. |
| “JPO Notice of Reasons for Rejection”, Japanese Application No. 2015-539724, dated Apr. 4, 2016. |
| “KIPO Notice of Preliminary Rejection”, Korean Patent Application No. 10-2015-7013481, dated Feb. 5, 2016, 11 Pages. |
| “Notice of Reasons for Rejection”, Japanese Application No. 2015-520534, dated Mar. 7, 2016, 9 pages. |
| “USPTO”, Notice of Allowance in U.S. Appl. No. 13/534,555, dated Mar. 8, 2016. |
| “USPTO”, Notice of Allowance for U.S. Appl. No. 13/658,490, dated Feb. 19, 2016, 40 Pages. |
| “USPTO,”, Final Rejection in U.S. Appl. No. 13/929,498, dated Feb. 25, 2016. |
| Alt, et al., “Location-based Crowdsourcing: Extending Crowdsourcing to the Real World”, Proceedings NordiCHI, Oct. 16-20, 2010, pp. 13-22. |
| Hayakawa, “Remarkable Software at the Cutting Edge”, MacPeople. ASCII Media Works Inc., vol. 15, No. 8, Aug. 1, 2009, p. 176. |
| KIPO, Notice of Preliminary Rejection (with English translation) for Korean Patent Application No. 10-2015-7010982, dated Mar. 8, 2016, 9 pages. |
| USPTO, Final Office Action for U.S. Appl. No. 13/346,385, dated Feb. 5, 2016, 11 pages. |
| USPTO, Non-Final Office Action for U.S. Appl. No. 14/162,708, dated Dec. 8, 2017, 15 pages. |
| SIPO, Notification for Patent Registration Formalities and Notification of the Grant of Patent Right for Invention (with English translation) for Chinese Patent Application No. 2013800559264, dated Dec. 1, 2017, 4 pages. |
| USPTO, Notice of Allowance for U.S. Appl. No. 15/205,602, dated Dec. 12, 2017, 8 pages. |
| SIPO, Second Office Action (with English translation) for Chinese Patent Application No. 201380055474.X, dated Nov. 3, 2017, 20 pages. |
| USPTO, Non-final Office Action for U.S. Appl. No. 13/219,835, dated Dec. 15, 2017, 31 pages. |
| Australia IP, Second Examination Report for Australian Patent Application No. 2016219660, dated Sep. 25, 2017, 3 pages. |
| EPO, Communication under Rule 71(3) for European Patent Application No. 13737075.5, dated Mar. 20, 2018, 37 pages. |
| SIPO, Notice of Allowance for Chinese Patent Application No. 201380055474.X, dated Feb. 24, 2018, 3 pages. |
| Supplementary Search Report in European Application No. 13786816.2, dated Jun. 3, 2016, 7 pp. |
| USPTO, Non-final Office Action for U.S. Appl. No. 13/346,385, dated Jun. 30, 2016, 12 pages. |
| USPTO, Non-final Office Action for U.S. Appl. No. 13/219,835, dated Jul. 15, 2016, 49 pages. |
| “Australian Patent Examination Report No. 3”, in Australian Application No. 2012217920, dated Aug. 19, 2016. |
| “KIPO Notice of Last Preliminary Rejection”, in Korean Patent Application No. 10-2015-7002159, dated Aug. 10, 2016. |
| USPTO, Final Office Action for U.S. Appl. No. 13/346,385, dated Oct. 20, 2016, 15 pages. |
| USPTO, Non-final Office Action for U.S. Appl. No. 14/162,708, dated Sep. 27, 2016, 14 pages. |
| Notice of Final Rejection in Korean Application No. 10-2015-7002159. |
| EPO, “Communication pursuant to Article 94(3) EPC in European Application No. 12711045.0”, dated Feb. 10, 2017, 6 pp. |
| SIPO, “Notification of First Office Action mailed in Chinese application No. 201380055474.X”, dated Feb. 27, 2017, 25 pages. |
| USPTO, “Final Rejection in U.S. Appl. No. 13/219,835” dated Feb. 27, 2017. |
| USPTO, “Non-Final Office Action in U.S. Appl. No. 13/929,498”, dated Feb. 13, 2017, 32 pages. |
| USPTO, “Final Office Action in U.S. Appl. No. 14/162,708”, dated Mar. 22, 2017. |
| Non-Final Office Action for U.S. Appl. No. 13/346,385, dated Apr. 7, 2017, 18 pages. |
| Received for Canadian Application No. 2,885,504 filed Oct. 23, 2013, dated May 2, 2017, 1 page. |
| Notice of Allowance in Korean Application No. 10-2015-7002159. |
| Office Action mailed in EP Application No. 13737075.5. |
| First Examination Report mailed in Australian Application No. 2016219660, dated May 12, 2017. |
| Notice of Allowance in Canadian Application No. 2885504, dated May 2, 2017. |
| First Office Action in Chinese Application No. 201380055926.4, dated May 31, 2017. |
| U.S. Office Action mailed in U.S. Appl. No. 13/929,498, dated Jul. 13, 2017. |
| U.S. Office Action mailed in U.S. Appl. No. 15/205,602, dated Jul. 13, 2017. |
| Notification of Grant in Chinese Application No. 201380040026.2, dated Jul. 27, 2017. |
| Decision of Rejection and Decision to Reject the Amendments in Japanese Application No. 2015-520534, dated Aug. 7, 2017, 6 pages. |
| Notice of Reasons for Rejection in Japanese Application No. 2015-520534, dated Jan. 16, 2017, 8 pages. |
| Notice of Acceptance for Application No. 2013334606, dated Nov. 18, 2016, 2 pages. |
| JPO, Notice of Allowance for U.S. Appl. No. 2015-539724, dated Dec. 12, 2016, 3 pages. |
| JPO, Notice of Allowance for Japanese Patent Application No. 2015-538160, Dec. 5, 2016, 3 Pages. |
| KIPO, Notice of Allowance in Korean Patent Application No. 10-2015-7010982. |
| USPTO, Non-Final Rejection in U.S. Appl. No. 13/929,498, dated Feb. 13, 2017, 32 Pages. |
| KIPO, Notice of Allowance mailed in Korean Patent Application No. 10-2015-7013481. |
| SIPO, Notification of First Office Action mailed in Chinese application No. 201380040026.2, dated Nov. 28, 2016, 25 pages. |
| USPTO, Final Office Action for U.S. Appl. No. 13/219,835, dated May 10, 2018, 15 pages. |
| USPTO, Notice of Allowance in U.S. Appl. No. 14/162,708, dated Jun. 28, 2018, 22 pages. |
| USPTO, Non-Final Office Action in U.S. Appl. No. 15/960,394, dated Jul. 27, 2018, 6 pages. |
| KIPO, Notice of Preliminary Rejection for Korean Patent Application No. 10-2017-7026395, dated Jul. 10, 2018, 3 pages. |
| Number | Date | Country | |
|---|---|---|---|
| 20160267500 A1 | Sep 2016 | US |
| Number | Date | Country | |
|---|---|---|---|
| Parent | 13658490 | Oct 2012 | US |
| Child | 15159188 | US |