The present invention relates generally to providing user representations in computing environments. The present invention further relates to managing a particular user's representation in avatar form in computing environments.
Environments that use avatars to represent users typically provide their own avatar creation tools. An avatar created by a user in one online environment is usually confined to that environment, so for each new environment, the user typically must create a separate, different avatar. To update characteristics of multiple avatars, the user must change each avatar separately within each environment, which can be time consuming.
Despite the apparent inefficiency of such a system, having a multiplicity of avatars may serve a practical purpose. Just as in real life, digital users exist in multiple contexts and may require different identities in different environments; for example, one identity for work, another for family and friends, another for video games, others for interests and hobbies, and so on. A different avatar in each situation allows the user to present a contextually relevant appearance.
Nevertheless, as the number of digital environments grows, the user is compelled to create and manage an ever-increasing number of avatars, which creates a disincentive to design a separate avatar for each new environment. This reduces the value of avatars generally, and for environments that use avatars, adds a barrier to adoption.
There are mechanisms that attempt to solve this problem by enabling users to use the same avatar in multiple environments, such as one disclosed by Mason et al. in U.S. patent application Ser. No. 12/279,643 (published as US 2010/0011422 A1).
However, such mechanisms require an avatar to be rendered identically in each environment, and therefore fail to provide the avatar's fundamental benefits, which include giving the user a contextually relevant identity, and each environment a consistent look and feel.
Therefore, what is needed is a solution to address at least some of these limitations.
In accordance with an aspect of the present invention, the present disclosure relates to a system and method for providing avatars adaptable to multiple environments. Avatar means any representation of a user, which may be manifested visually in an environment. For example, an avatar may be manifested as a character in a video game, a user profile picture in a social networking website, or an emoticon in a messaging application, etc. Environment in this disclosure broadly means any environment where an avatar may be manifested. For example, an environment may be an avatar creation application, video game, social networking website, messaging application, smartphone address book, or any other application where a user may want to have a representation.
In accordance with an aspect of the present invention, more particularly, the system and method provides data about an avatar's components, such as discrete facial and bodily features, and the relative positions of the features to another, for the purpose of rendering different versions of a user's avatar which may be visually adapted to suit multiple environments.
In accordance with an aspect of the present invention, the system and method enables data about a single avatar to be adapted and re-rendered in multiple environments in a virtually unlimited number of ways. This allows service providers (such as operators of a service environment) to reap the benefits of providing avatars designed to visually match their environments, while relieving the user of the need to create a new avatar specifically for each service. The user can instead create an avatar in one environment, and link the avatar data to a user account in a second environment. The avatar can then be automatically re-rendered with the second environment's art set using avatar data from the first environment. The second environment can therefore choose not to offer an avatar creation tool at all, yet still give the user the full experience of interacting in its environment with a relevant and personalized avatar.
In accordance with an aspect of the present invention, there is provided a method comprising: a first computing device maintaining avatar data describing at least one avatar component of a first avatar designed for a first service environment, the avatar data associated with an identifier; the first computing device receiving a request from a second service environment, the request including identification data and at least one requested avatar component; based at least partly on a correspondence between the respective identifier and the received identification data, the first computing device transmitting to the second service environment the avatar data describing the at least one avatar component corresponding to the at least one requested avatar component type; and the second service environment generating a second avatar for a second service environment based at least partly on the avatar data received from the first computing device.
In accordance with an aspect of the present invention, in this respect, before explaining at least one embodiment of the invention in detail, it is to be understood that the invention is not limited in its application to the details of construction and to the arrangements of the components set forth in the following description or the examples provided therein, or illustrated in the drawings. The invention is capable of other embodiments and of being practiced and carried out in various ways. Also, it is to be understood that the phraseology and terminology employed herein are for the purpose of description and should not be regarded as limiting.
In the drawings, embodiments of the invention are illustrated by way of example. It is to be expressly understood that the description and drawings are only for the purpose of illustration and as an aid to understanding, and are not intended as a definition of the limits of the invention.
As noted above, the present disclosure relates to a system and method for providing avatars adaptable to multiple environments.
In an aspect, the system and method enables data about a single avatar to be adapted and re-rendered in a virtually unlimited number of ways. This is achieved by generating data about an avatar's discrete facial and bodily features, and each feature's relative position to another, for the purpose of rendering different versions of that avatar which are visually adapted to suit multiple service environments.
This allows a service provider to reap the benefits of providing users with avatars that visually match its service environment, while relieving the user of the requirement to create a separate avatar specifically for that service.
Avatars are typically designed and intended for use in a bounded environment, such as a web portal, branded application or website, or video game. Avatars in each service environment may have a distinct look and feel, and services that use avatars may not want to allow sharing of their avatars for use in other environments.
To address the need for each environment to maintain its own look and feel, and simultaneously eliminate the time and effort required of users to create a new avatar in each environment, the present system and method reduces an avatar to a data set of components, which may be used to re-render a different version of that avatar matched to the look and feel of another environment. This effectively adapts the avatar into something new in each environment, while the use of a common data set of components provides a consistency between different versions of the avatar in different environments.
An adaptable avatar designed to be location and appearance-agnostic benefits an environment by automatically customizing an avatar that is appropriate to its look and feel, and provides the user with control over a consistent set of components for personalized avatars across multiple environments.
Various aspects of the present system and method will now be described in more detail with reference to the figures.
Referring to
In this way, the avatar (14) is rendered to the second environment B in a responsive and contextually relevant way, without the need to deploy substantial infrastructure on at the server farm (12), i.e. only the API (17) is required.
Avatars may be created from avatar data saved or maintained by a computing device 100.
Avatar components are combined to form an avatar using an avatar composition utility. Operation of the avatar composition utility is illustrated (
Each of the service environments may operate on a respective computing device (e.g. a computer server) or be associated with a respective computing device. Each service environment may also be associated with or operate on a single computing device, in the case of a videogame console or personal computer operating multiple games or applications locally. Each service environment may define its own avatar definitions and requirements. For example, while Environment A may require avatar components related to the avatar's face and body, Environment E may only require avatar components related to the avatar's face.
Avatars may be created locally on a particular computing device for a service environment, and then transmitted to a computer server configured with operating or maintaining some aspects of the service environment. The computer server may maintain each avatar dataset created in this way, and make avatar datasets, or data derived therefrom, available to other service environments operating either on the same computer server, or on a different computing device.
Now referring to
The avatar composition utility may embody one or more user workflows for guiding a user through the steps involved in creating an avatar, or modifying an existing avatar. The avatar composition utility may also access one or more avatar templates or “types”, where each template or type constitutes a collection of components which may be organized hierarchically into component categories and component sub-categories. These components may relate to visual components of the avatar, as described above.
A representative workflow in accordance with the method of the present invention is shown in
In another embodiment, the avatar composition utility defines a series of avatar composition rules that may include, for example, minimum requirements for composing an avatar, so as to ensure that minimum rules for establishing an avatar data set to render an avatar in multiple environments are also met.
The user may save the avatar based on the selected components. If the user is saving the avatar for the first time, the system generates and saves a unique alphanumeric avatar ID and an “avatar profile”. An avatar profile includes avatar component data corresponding to each selected component as shown in
Additional avatar components may be defined by the computing device 100 and saved as avatar data by defining a corresponding avatar component. In particular, non-visual data or meta data associated with a respective avatar may also be stored in the avatar dataset in this way, including data related to an emotion, experience, activity, or other non-visual characteristic. Accordingly, not all avatar components in an avatar dataset are intended to be displayed or will be displayed when rendering the avatar in a service environment. However, avatar components not rendered may be used by the service environment to modify other aspects of a display, or to modify the display of other avatar components. For example, an avatar component storing avatar component data representing that the user has participated in an activity, witnessed an event, completed a task, or used the avatar for a particular number of hours, may cause the service environment to display a representative icon, a title, or change the appearance of the avatar to appear stronger, tired, or in other ways.
The avatar ID generated by the system may be displayed to the user, or the user may create a unique username that corresponds to the ID. The avatar profile may further include an identification of a user associated with the avatar, either in the avatar ID itself or separately.
Avatar profiles are stored in a server database and may be made accessible to Environment E through an API. As the avatar designed by the user for Environment A may by stored in a proprietary data format or may include proprietary avatar components, Environment A may be configured to derive and store a “universal” dataset for each avatar based on the avatar dataset. The universal dataset may be stored in a format usable by Environment E for generating a subsequent avatar. Alternatively, the API may provide for any required format conversions on demand upon receiving a request from Environment E for a particular avatar dataset, or avatar component associated with a particular user.
An Environment E may access Environment A's avatar dataset through Environment A's API for the purpose of generating and mapping Environment E's avatar components with Environment A's. Environment E can ask Environment A for its entire dataset, including every individual component and corresponding avatar component data, and then generate any or all components in its own style. Environment E can ask Environment A only for particular subsets of information, such as all avatar components and avatar component data with gender equal to female, for the purpose of, for example, generating its own female-only avatars; or Environment E can ask Environment A only for all subcategories of all categories, for the purpose of, for example, generating one simplified hairstyle to correspond to each subcategory. Environment E may also map each avatar component it generates to corresponding avatar component data in Environment A by adding relevant identifying tags to each component.
In one embodiment of the present invention, the API may be accessed through an avatar data mapping utility (not shown). Such a utility may include a graphical user interface which displays available avatar components and avatar data generated by Environment A. Environment E can generate its own corresponding avatar components in its own style, and then upload, or drag-and-drop, each component beside Environment A's corresponding avatar components. Each matched component automatically generates a relevant identifying tag, which Environment E can export and then use to generate adapted avatars when users connect their avatars from Environment A to Environment E. As above, Environment E may generate and map components to each of Environment A's components, only some components, or only subsets of components. Such a visual utility simplifies and speeds the process of generating and mapping avatar components and data between environments.
In instances where there is no corresponding component in Environment A, a component may be mapped to an arbitrary parallel component. For example, if an avatar in Environment E has the appearance of a cartoon hedgehog, avatar skin color data in Environment A may be mapped to fur color, and avatar clothing data may be discarded. Instead of requesting the entire avatar dataset for a particular identifier, Environment E may request only avatar component data supported by Environment E. Environment E may also request only one or more avatar components from the API. Accordingly, any request from Environment E may specify the avatar component data requested in addition to including identification data. The identification data may identify a particular avatar by avatar identifier, or may include other data or identifiers associated with a particular user, such as name or email address.
A user may enter Environment E and see an option to connect avatar saved in Environment A to Environment E. User enters unique avatar ID or username and password created in Environment A. Environment E sends avatar ID and password to Environment A via API. Environment A then authenticates identity of user. Environment E requests user's avatar component data from Environment A via API, and Environment A sends requested avatar component data. Environment E can ask Environment A for a user's complete avatar profile, or only for avatar component data from that profile that is relevant to Environment E's avatars. For example, Environment E may only require a user's avatar body shape, skin color and hair color, or only require a user's avatar eye color and emotion data. This way, Environment E can choose to automatically modify, adapt, replace or remove any of the user's avatar components. After Environment E receives the requested avatar component data, it generates an adapted version of the user's avatar consisting of Environment E's relevant avatar component data and presents this adapted avatar to the user.
Environment E may offer the user further avatar customization options using its own avatar composition utility, allowing the user to modify existing avatar components, or add new avatar components not present in Environment A, such as accessories, clothing, weapons, etc. The user may save the adapted avatar in Environment E, and Environment E may generate a new avatar profile of the adapted avatar. The avatar profile of the adapted avatar in Environment E may then be sent to Environment A through the API, and Environment A may then make relevant updates to the user's avatar profile in Environment A, such as modifications to hairstyle or additions of new avatar components.
In an embodiment, the present system and method includes an application programming interface or API which may be integrated with a third party service provider's online environment. For example, in one possible implementation, the API may embody programming required to enable the use of the avatar profile to populate the third party online environment with the avatar, or to generate an avatar derivative based on the base avatar profile, and then populate the third party online environment with the resulting derivative avatar. As another illustrative example, the API may include a translator that is operable to translate the avatar profile into a set of instructions that may be processed by the systems associated with the third party online environment to populate the avatar or derivative avatar into the third party online environment.
The creation of avatars that are adaptive and responsive to different online environments and can move seamlessly from one online environment to another online environment provides the possibility of creating a stronger connection between the user and his/her avatar, because the avatar with a common set of components becomes strongly identifiable by the user. In an embodiment, this connection with the user can be leveraged for the purpose of a variety of user engagement scenarios that may bring value to an avatar platform as well as its various participating partners.
By adapting an existing avatar to multiple service environments, it may be possible to reduce the amount of time spent by a user creating or modifying avatars in each environment, which may be referred to as “onboarding” time. New digital applications are being created or modified regularly, each with their own onboarding tasks, including creating user profiles or avatars. While often necessary, creating multiple user profiles across multiple applications is very time consuming. A single application to create, update and manage an adaptable avatar across multiple other applications and platforms would save users time versus creating and updating different avatars for each environment. For example, a user may register for a banking application that shows the user's avatar's mood based on each month's personal spending, a music application that shows the user's avatar dancing whenever music plays, and a video game where the user's avatar has the appearance of a robot. Creating an adaptable avatar once and then connecting to it from each of the three applications would be 75% faster than creating avatars separately in each service; connecting to 10 such applications would be over 90% faster, and so on. For application providers, giving users the option to connect to their adaptable avatars results in faster and simpler onboarding.
Adaptable avatars may assist in establishing a unifying user experience across disparate applications. Each computer software application or service environment may serve a different purpose and may include some necessary differences in features and user interface. Nevertheless, the user of such environments remains a single person, and so the experience of switching between environments can feel disjointed and lack continuity. A personalized and recognizable avatar that may be adapted to fit other contexts helps to establish consistency between otherwise disparate environments. The resulting user experience may feel more harmonious and unified. For example, switching between a location-sharing application and a health-monitoring application while seeing essentially the same avatar gives these very different applications a source of continuity, making the user's overall digital experience feel more seamless. As applications evolve and new ones are introduced, an adaptable and easily updatable avatar also provides users continuity through time.
The present invention promotes for example environments that are based on collaboration between different online operators who may collaborate to provide a single online environment. The system and method provides the ability for an avatar to be rendered seamlessly across multiple associated environments.
Adaptable avatars may increase the relevance of an application to a user. An application may be more useful when the user experience is personalized. An application that does not use avatars, or that uses non-personalized avatars, may be missing a visual cue for making the application and its features personally relevant to users. Allowing users to connect un-adapted portable avatars might add some relevance to the user experience, but it risks creating an unacceptably inconsistent experience because the avatars are not properly contextualized. For example, a weather forecasting application might tell the user how to dress on a given day by showing a generic avatar wearing appropriate clothing, on a warm and rainy day, the avatar might appear in the application wearing a T-shirt and shorts and holding an umbrella. An improvement might be to give the user an option to connect and import an existing un-adapted avatar from another application, but this would make the user experience visually fragmented. Still another improvement might be to give the user the option to create a new avatar from scratch within the application, but this would add a time-consuming customization task and give the user an additional avatar to manage. If such an application integrated with an adaptable avatar system, the user would only need to connect a previously made avatar to the application to see a contextualized version of him- or herself appear in the weather forecast automatically, thereby making it much more personally relevant. For the application provider, each adaptable avatar becomes an immediate and familiar-looking common connection between itself and the user.
Adaptable avatars may increase user engagement in a particular application. Two challenges faced by most applications are initially grabbing users' interest, and then establishing a pattern of behavior so that they become engaged, habitual users. Adaptable avatars provide an ideal mechanism for establishing such user interest and engagement. Users build an affinity for their adaptable avatars through repeated and varied use in multiple applications. Consequently, serving a pre-existing, yet adapted, avatar of a user within the context of a new application may make that application more interesting to the user than applications where either the avatar needs to be created from scratch and therefore has no history with the user, or is previously designed by the user in another application but not adapted. For example, a video game set in a different galaxy might allow a user to connect a pre-existing adaptable avatar, and then automatically generate a new adapted alien version that captures the original avatar's essence by retaining appropriate components, such as eye color, skin tone, and face shape, etc. Presenting such an adapted avatar at the start of the game provides a vehicle for instantly capturing the user's attention and immersing the user in the game. All applications that use an adaptive avatar system become collectively more engaging by giving users a single, unifying visual identity that transcends any one application, thus making the entire digital experience personally meaningful and satisfying as a whole.
Referring to
Use Cases
As will be appreciated, the present system and method may be adapted for various use cases. For example, the present system and method may be used to link a user's avatar ID and avatar profile to websites and have their unique versions of a user appear automatically. A user's avatar ID and avatar profile may also be linked to a video game to play as a re-rendered version of the user's avatar in that game's world. Alternatively, the system of the present invention may be linked to a social networking platform, and sign in to the social networking using social networking login information.
The system of the present invention may be configured to extract information relevant to a user to add context to the rendering of the avatar. For example, the system may connect to social media channels in order to extract information regarding the interests of a user, topics or themes of current interest to the user, and so on. This information may be used to adapt content including avatars, as explained herein.
In another use case, the system and method may automatically render different versions of avatars for different occasions or branding purposes. For example, the avatar can automatically turn into a zombie on Halloween, or be modified by clothing or features to represent a popular character in a show.
In another use case, the system and method may be used to improve digital advertising and web technical support by presenting complementary avatars based on a user's avatar component data. The system and method can also improve user profiling by compiling a user's physiological avatar data, such as height and body shape, behavioral avatar data, such as past and current body poses and postures, and emotional history based on past and current avatar emotion data, across multiple environments.
In another use case, the system and method may be used to improve digital advertising by presenting the user's avatar in online advertisements. The system and method may also be used to improve advertising targeting by using avatar component data to glean physiological information, such as height and body shape, or emotional state based on current or historical avatar emotion data, to serve more relevant ads.
In another use case, the system and method can improve content targeting by using avatar component data to glean physiological information, such as height and body shape, or emotional state based on current or historical avatar emotion data, to serve more relevant news, stories, games and videos.
In another use case, users' avatar data may be categorized in aggregate to ascertain distributions in user physiological, behavioural, emotion and other data within and between populations. Aggregate data may be used to correlate avatar data with other digital user activities, such as online shopping and content consumption. Historical and current aggregate data may also be used to build predictive models for real-world activities, such as economic and sociological trends.
In another use case, the system and method can improve personalized content and storytelling by presenting different versions of a user's avatar in custom comics and animation.
In another use case, the system and method may provide branded personalized emoticon packs featuring a user's avatar to represent a particular theme or particular well-known character.
In another use case, the system and method can present custom versions of a user's avatar in e-greeting cards.
In another use case, the present system and method can present custom versions of a user's avatar in different outfits based on body data.
These various use cases may be adapted to virtually any type of media or digital environment.
While illustrative embodiments of the invention have been described above, it will be appreciated that various changes and modifications may be made without departing from the scope of the present invention.
In further aspects, the disclosure provides systems, devices, methods, and computer programming products, including non-transient machine-readable instruction sets, for use in implementing such methods and enabling the functionality described previously.
Although the disclosure has been described and illustrated in exemplary forms with a certain degree of particularity, it is noted that the description and illustrations have been made by way of example only. Numerous changes in the details of construction and combination and arrangement of parts and steps may be made. Accordingly, such changes are intended to be included in the invention, the scope of which is defined by the claims.
Except to the extent explicitly stated or inherent within the processes described, including any optional steps or components thereof, no required order, sequence, or combination is intended or implied. As will be will be understood by those skilled in the relevant arts, with respect to both processes and any systems, devices, etc., described herein, a wide range of variations is possible, and even advantageous, in various circumstances, without departing from the scope of the invention, which is to be limited only by the claims.
This application is a continuation of, and claims priority to, U.S. patent application Ser. No. 13/979,974, filed May 8, 2013 and entitled “SYSTEM AND METHOD FOR ADAPTABLE AVATARS,” which is a U.S. National Stage Application filed Under § 371 from International Application Serial No. PCT/CA2013/000454, filed May 8, 2013, entitled “SYSTEM AND METHOD FOR GENERATING AND DISPLAYING AVATARS,” which in turn claims priority to U.S. Provisional Patent Application Ser. No. 61/644,057, filed May 2012, and entitled “SYSTEM AND METHOD FOR ADAPTABLE AVATARS,” the contents of which are incorporated herein by reference in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
5880731 | Liles et al. | Mar 1999 | A |
6023270 | Brush, II et al. | Feb 2000 | A |
6038295 | Mattes | Mar 2000 | A |
6223165 | Lauffer | Apr 2001 | B1 |
6772195 | Hatlelid et al. | Aug 2004 | B1 |
6842779 | Nishizawa | Jan 2005 | B1 |
6980909 | Root et al. | Dec 2005 | B2 |
7173651 | Knowles | Feb 2007 | B1 |
7342587 | Danzig et al. | Mar 2008 | B2 |
7411493 | Smith | Aug 2008 | B2 |
7468729 | Levinson | Dec 2008 | B1 |
7535890 | Rojas | May 2009 | B2 |
7636755 | Blattner et al. | Dec 2009 | B2 |
7639251 | Gu et al. | Dec 2009 | B2 |
7775885 | Van Luchene et al. | Aug 2010 | B2 |
7859551 | Bulman et al. | Dec 2010 | B2 |
7885931 | Seo et al. | Feb 2011 | B2 |
7925703 | Dinan et al. | Apr 2011 | B2 |
8088044 | Tchao et al. | Jan 2012 | B2 |
8095878 | Bates et al. | Jan 2012 | B2 |
8108774 | Finn et al. | Jan 2012 | B2 |
8117281 | Robinson et al. | Feb 2012 | B2 |
8130219 | Fleury et al. | Mar 2012 | B2 |
8131597 | Hudetz | Mar 2012 | B2 |
8146005 | Jones et al. | Mar 2012 | B2 |
8151191 | Nicol | Apr 2012 | B2 |
8199747 | Rojas et al. | Jun 2012 | B2 |
8332475 | Rosen et al. | Dec 2012 | B2 |
8384719 | Reville et al. | Feb 2013 | B2 |
RE44054 | Kim | Mar 2013 | E |
8396708 | Park et al. | Mar 2013 | B2 |
8425322 | Gillo et al. | Apr 2013 | B2 |
8458601 | Castelli et al. | Jun 2013 | B2 |
8462198 | Lin et al. | Jun 2013 | B2 |
8484158 | Deluca et al. | Jul 2013 | B2 |
8495503 | Brown et al. | Jul 2013 | B2 |
8495505 | Smith et al. | Jul 2013 | B2 |
8504926 | Wolf | Aug 2013 | B2 |
8559980 | Pujol | Oct 2013 | B2 |
8564621 | Branson et al. | Oct 2013 | B2 |
8564710 | Nonaka et al. | Oct 2013 | B2 |
8581911 | Becker et al. | Nov 2013 | B2 |
8597121 | Andres Del | Dec 2013 | B2 |
8601051 | Wang | Dec 2013 | B2 |
8601379 | Marks et al. | Dec 2013 | B2 |
8632408 | Gillo et al. | Jan 2014 | B2 |
8648865 | Dawson et al. | Feb 2014 | B2 |
8659548 | Hildreth | Feb 2014 | B2 |
8683354 | Khandelwal et al. | Mar 2014 | B2 |
8692830 | Nelson et al. | Apr 2014 | B2 |
8718333 | Wolf et al. | May 2014 | B2 |
8724622 | Rojas | May 2014 | B2 |
8810513 | Ptucha et al. | Aug 2014 | B2 |
8812171 | Filev et al. | Aug 2014 | B2 |
8832201 | Wall | Sep 2014 | B2 |
8832552 | Arrasvuori et al. | Sep 2014 | B2 |
8839327 | Amento et al. | Sep 2014 | B2 |
8874677 | Rosen et al. | Oct 2014 | B2 |
8890926 | Tandon et al. | Nov 2014 | B2 |
8892999 | Nims et al. | Nov 2014 | B2 |
8909679 | Roote et al. | Dec 2014 | B2 |
8924250 | Bates et al. | Dec 2014 | B2 |
8935656 | Dandia et al. | Jan 2015 | B2 |
8963926 | Brown et al. | Feb 2015 | B2 |
8989786 | Feghali | Mar 2015 | B2 |
8995433 | Rojas | Mar 2015 | B2 |
9040574 | Wang et al. | May 2015 | B2 |
9055416 | Rosen et al. | Jun 2015 | B2 |
9086776 | Ye et al. | Jul 2015 | B2 |
9100806 | Rosen et al. | Aug 2015 | B2 |
9100807 | Rosen et al. | Aug 2015 | B2 |
9105014 | Collet et al. | Aug 2015 | B2 |
9191776 | Root et al. | Nov 2015 | B2 |
9204252 | Root | Dec 2015 | B2 |
9241184 | Weerasinghe | Jan 2016 | B2 |
9256860 | Herger et al. | Feb 2016 | B2 |
9298257 | Hwang et al. | Mar 2016 | B2 |
9314692 | Konoplev et al. | Apr 2016 | B2 |
9330483 | Du et al. | May 2016 | B2 |
9357174 | Li et al. | May 2016 | B2 |
9361510 | Yao et al. | Jun 2016 | B2 |
9378576 | Bouaziz et al. | Jun 2016 | B2 |
9402057 | Kaytaz et al. | Jul 2016 | B2 |
9412192 | Mandel et al. | Aug 2016 | B2 |
9443227 | Evans et al. | Sep 2016 | B2 |
9460541 | Li et al. | Oct 2016 | B2 |
9480924 | Haslam | Nov 2016 | B2 |
9489661 | Evans et al. | Nov 2016 | B2 |
9489760 | Li et al. | Nov 2016 | B2 |
9491134 | Rosen et al. | Nov 2016 | B2 |
9503845 | Vincent | Nov 2016 | B2 |
9508197 | Quinn et al. | Nov 2016 | B2 |
9544257 | Ogundokun et al. | Jan 2017 | B2 |
9576400 | Van Os et al. | Feb 2017 | B2 |
9589357 | Li et al. | Mar 2017 | B2 |
9592449 | Barbalet et al. | Mar 2017 | B2 |
9641870 | Cormie et al. | May 2017 | B1 |
9648376 | Chang et al. | May 2017 | B2 |
9697635 | Quinn et al. | Jul 2017 | B2 |
9706040 | Kadirvel et al. | Jul 2017 | B2 |
9744466 | Fujioka | Aug 2017 | B2 |
9746990 | Anderson et al. | Aug 2017 | B2 |
9749270 | Collet et al. | Aug 2017 | B2 |
9792714 | Li et al. | Oct 2017 | B2 |
9839844 | Dunstan et al. | Dec 2017 | B2 |
9883838 | Kaleal, III et al. | Feb 2018 | B2 |
9898849 | Du et al. | Feb 2018 | B2 |
9911073 | Spiegel et al. | Mar 2018 | B1 |
9936165 | Li et al. | Apr 2018 | B2 |
9959037 | Chaudhri et al. | May 2018 | B2 |
9980100 | Charlton et al. | May 2018 | B1 |
9990373 | Fortkort | Jun 2018 | B2 |
10039988 | Lobb et al. | Aug 2018 | B2 |
10097492 | Tsuda et al. | Oct 2018 | B2 |
10116598 | Tucker et al. | Oct 2018 | B2 |
10155168 | Blackstock et al. | Dec 2018 | B2 |
10242477 | Charlton et al. | Mar 2019 | B1 |
10242503 | McPhee et al. | Mar 2019 | B2 |
10262250 | Spiegel et al. | Apr 2019 | B1 |
10362219 | Wilson et al. | Jul 2019 | B2 |
10432559 | Baldwin et al. | Oct 2019 | B2 |
10475225 | Park et al. | Nov 2019 | B2 |
10504266 | Blattner et al. | Dec 2019 | B2 |
10573048 | Ni et al. | Feb 2020 | B2 |
10657701 | Osman et al. | May 2020 | B2 |
10880246 | Baldwin et al. | Dec 2020 | B2 |
10938758 | Allen et al. | Mar 2021 | B2 |
20020067362 | Agostino Nocera et al. | Jun 2002 | A1 |
20020169644 | Greene | Nov 2002 | A1 |
20050162419 | Kim et al. | Jul 2005 | A1 |
20050206610 | Cordelli | Sep 2005 | A1 |
20050280660 | Seo et al. | Dec 2005 | A1 |
20060294465 | Ronen et al. | Dec 2006 | A1 |
20070113181 | Blattner et al. | May 2007 | A1 |
20070168863 | Blattner et al. | Jul 2007 | A1 |
20070176921 | Iwasaki et al. | Aug 2007 | A1 |
20070218987 | Luchene et al. | Sep 2007 | A1 |
20070260984 | Marks | Nov 2007 | A1 |
20080158222 | Li et al. | Jul 2008 | A1 |
20080201638 | Nair | Aug 2008 | A1 |
20090016617 | Bregman-amitai et al. | Jan 2009 | A1 |
20090055484 | Vuong et al. | Feb 2009 | A1 |
20090070688 | Gyorfi et al. | Mar 2009 | A1 |
20090099925 | Mehta et al. | Apr 2009 | A1 |
20090106672 | Burstrom | Apr 2009 | A1 |
20090144639 | Nims et al. | Jun 2009 | A1 |
20090150778 | Nicol | Jun 2009 | A1 |
20090158170 | Narayanan et al. | Jun 2009 | A1 |
20090177976 | Bokor et al. | Jul 2009 | A1 |
20090202114 | Morin et al. | Aug 2009 | A1 |
20090254859 | Arrasvuori | Oct 2009 | A1 |
20090265604 | Howard et al. | Oct 2009 | A1 |
20090300525 | Jolliff et al. | Dec 2009 | A1 |
20090303984 | Clark et al. | Dec 2009 | A1 |
20090319178 | Khosravy et al. | Dec 2009 | A1 |
20090328122 | Amento et al. | Dec 2009 | A1 |
20100011422 | Mason et al. | Jan 2010 | A1 |
20100023885 | Reville | Jan 2010 | A1 |
20100083140 | Dawson et al. | Apr 2010 | A1 |
20100100828 | Khandelwal et al. | Apr 2010 | A1 |
20100115426 | Liu et al. | May 2010 | A1 |
20100121915 | Wang | May 2010 | A1 |
20100162149 | Sheleheda et al. | Jun 2010 | A1 |
20100197396 | Fujii | Aug 2010 | A1 |
20100203968 | Gill et al. | Aug 2010 | A1 |
20100227682 | Reville et al. | Sep 2010 | A1 |
20100290756 | Karaoguz et al. | Nov 2010 | A1 |
20100332980 | Sun et al. | Dec 2010 | A1 |
20110093780 | Dunn | Apr 2011 | A1 |
20110115798 | Nayar et al. | May 2011 | A1 |
20110148864 | Lee et al. | Jun 2011 | A1 |
20110202598 | Evans et al. | Aug 2011 | A1 |
20110239136 | Goldman et al. | Sep 2011 | A1 |
20110239143 | Ye et al. | Sep 2011 | A1 |
20110248992 | van Os | Oct 2011 | A1 |
20120013770 | Stafford et al. | Jan 2012 | A1 |
20120113106 | Choi et al. | May 2012 | A1 |
20120124458 | Cruzada | May 2012 | A1 |
20120130717 | Xu et al. | May 2012 | A1 |
20120209924 | Evans et al. | Aug 2012 | A1 |
20120215879 | Bozo | Aug 2012 | A1 |
20120223940 | Dunstan | Sep 2012 | A1 |
20130103760 | Golding et al. | Apr 2013 | A1 |
20130152000 | Liu et al. | Jun 2013 | A1 |
20130155169 | Hoover et al. | Jun 2013 | A1 |
20130201187 | Tong et al. | Aug 2013 | A1 |
20130249948 | Reitan | Sep 2013 | A1 |
20130257877 | Davis | Oct 2013 | A1 |
20140011576 | Barbalet et al. | Jan 2014 | A1 |
20140040066 | Fujioka | Feb 2014 | A1 |
20140043329 | Wang et al. | Feb 2014 | A1 |
20140055554 | Du et al. | Feb 2014 | A1 |
20140125678 | Wang et al. | May 2014 | A1 |
20140128166 | Tam | May 2014 | A1 |
20140129343 | Finster et al. | May 2014 | A1 |
20140160149 | Blackstock et al. | Jun 2014 | A1 |
20150121251 | Kadirvel | Apr 2015 | A1 |
20150206349 | Rosenthal et al. | Jul 2015 | A1 |
20160045834 | Burns | Feb 2016 | A1 |
20160134840 | Mcculloch | May 2016 | A1 |
20160234149 | Tsuda et al. | Aug 2016 | A1 |
20170006322 | Dury et al. | Jan 2017 | A1 |
20170080346 | Abbas | Mar 2017 | A1 |
20170087473 | Siegel et al. | Mar 2017 | A1 |
20170118145 | Aittoniemi et al. | Apr 2017 | A1 |
20170199855 | Fishbeck | Jul 2017 | A1 |
20170235848 | Van Deusen et al. | Aug 2017 | A1 |
20170310934 | Du et al. | Oct 2017 | A1 |
20170312634 | Ledoux et al. | Nov 2017 | A1 |
20180047200 | O'hara et al. | Feb 2018 | A1 |
20180113587 | Allen et al. | Apr 2018 | A1 |
20180115503 | Baldwin et al. | Apr 2018 | A1 |
20180315076 | Andreou | Nov 2018 | A1 |
20180315133 | Brody et al. | Nov 2018 | A1 |
20180315134 | Amitay et al. | Nov 2018 | A1 |
20190001223 | Blackstock | Jan 2019 | A1 |
20190057616 | Cohen | Feb 2019 | A1 |
20190188920 | Mcphee et al. | Jun 2019 | A1 |
20190386941 | Baldwin et al. | Dec 2019 | A1 |
20210152505 | Baldwin et al. | May 2021 | A1 |
20210266277 | Allen et al. | Aug 2021 | A1 |
Number | Date | Country |
---|---|---|
2887596 | Jul 2015 | CA |
109863532 | Jun 2019 | CN |
110023985 | Jul 2019 | CN |
110168478 | Aug 2019 | CN |
2184092 | May 2010 | EP |
2001230801 | Aug 2001 | JP |
2014006881 | Jan 2014 | JP |
5497931 | Mar 2014 | JP |
1020050036963 | Apr 2005 | KR |
20070008417 | Jan 2007 | KR |
101445263 | Sep 2014 | KR |
20160016532 | Feb 2016 | KR |
20160028028 | Mar 2016 | KR |
20160051536 | May 2016 | KR |
102233700 | Mar 2021 | KR |
WO-2003094072 | Nov 2003 | WO |
WO-2004095308 | Nov 2004 | WO |
WO-2006107182 | Oct 2006 | WO |
WO-2007134402 | Nov 2007 | WO |
WO-2012139276 | Oct 2012 | WO |
WO-2013027893 | Feb 2013 | WO |
WO-2013152454 | Oct 2013 | WO |
WO-2013166588 | Nov 2013 | WO |
WO-2014031899 | Feb 2014 | WO |
WO-2014194439 | Dec 2014 | WO |
WO-2016090605 | Jun 2016 | WO |
WO-2018081013 | May 2018 | WO |
WO-2018102562 | Jun 2018 | WO |
WO-2018129531 | Jul 2018 | WO |
WO-2019089613 | May 2019 | WO |
Entry |
---|
“A Guide to Advertising on Campus With Snapchat Geofilters”, College Marketing Group, [Online] Retrieved from the Internet: <URL: https://collegemarketinggroup.com/blog/a-guide-toadvertising-on-campus-with-snapchat-geofilters/>, (Jul. 25, 2016), 5 pgs. |
“U.S. Appl. No. 15/369,499, Final Office Action dated Oct. 1, 2019”, 17 pgs. |
“U.S. Appl. No. 15/369,499, Non Final Office Action dated Jun. 17, 2019”, 17 pgs. |
“U.S. Appl. No. 15/369,499, Response filed Sep. 10, 2019 to Non-Final Office Action dated Jun. 17, 2019”, 9 pgs. |
“U.S. Appl. No. 16/115,259, Non Final Office Action dated Jul. 30, 2019”, 21 pgs. |
“U.S. Appl. No. 16/115,259, Response filed Oct. 30, 2019 to Non Final Office Action dated Jul. 30, 2019”, 9 pgs. |
“European Application Serial No. 17876226.6, Extended European Search Report dated Sep. 5, 2019”, 10 pgs. |
“International Application Serial No. PCT/US2017/063981, International Preliminary Report on Patentability dated Jun. 13, 2019”, 10 pgs. |
Alex, Heath, “What do Snapchat's emojis mean?—Understanding these emojis will turn you into a Snapchat pro”, Business Insider, [Online] Retrieved from the Internet: <URL: https://www.businessinsider.com/what-do-snapchats-emojismean-2016-5?international=true&r=US&IR=T>, (May 28, 2016), 1 pg. |
Karen, Tumbokon, “Snapchat Update: How to Add Bitmoji to customizable Geofilters”, International Business Times, [Online] Retrieved from the Internet: <URL: https://www.ibtimes.com/snapchat-update-how-add-bitmojicustornizable-geofilters-2448152>, (Nov. 18, 2016), 6 pgs. |
Sophia, Bernazzani, “A Brief History of Snapchat”, Hubspot, [Online] Retrieved from the Internet: <URL: https://blog.hubspot.com/marketing/history-of-snapchat>, (Feb. 10, 2017), 12 pgs. |
“U.S. Appl. No. 15/369,499, Response filed Feb. 3, 2020 to Final Office Action dated Oct. 1, 2019”, 10 pgs. |
“U.S. Appl. No. 16/115,259, Final Office Action dated Dec. 16, 2019”, 23 pgs. |
“U.S. Appl. No. 16/126,869, Non Final Office Action dated Feb. 5, 2020”, 7 pgs. |
“U.S. Appl. No. 15/369,499, Non Final Office Action dated Mar. 2, 2020”, 17 pgs. |
“U.S. Appl. No. 16/115,259, Response filed Mar. 13, 2020 to Final Office Action dated Dec. 16, 2019”. |
“U.S. Appl. No. 15/369,499, Final Office Action dated Jun. 15, 2020”, 17 pgs. |
“U.S. Appl. No. 15/369,499, Response filed Jun. 2, 2020 to Non Final Office Action dated Mar. 2, 2020”, 9 pgs. |
“U.S. Appl. No. 16/115,259, Final Office Action dated Jul. 22, 2020”, 20 pgs. |
“U.S. Appl. No. 16/115,259, Non Final Office Action dated Apr. 9, 2020”, 18 pgs. |
“U.S. Appl. No. 16/115,259, Response filed Jul. 9, 2020 to Non Final Office Action dated Apr. 9, 2020”. |
“U.S. Appl. No. 16/126,869, Final Office Action dated Jul. 7, 2020”, 8 pgs. |
“U.S. Appl. No. 16/126,869, Response filed May 5, 2020 to Non Final Office Action dated Feb. 5, 2020”, 8 pgs. |
“European Application Serial No. 17876226.6, Communication Pursuant to Article 94(3) EPC dated May 29, 2020”, 5 pgs. |
“European Application Serial No. 17876226.6, Response filed Mar. 30, 2020 to Extended European Search Report dated Sep. 5, 2019”, 22 pgs. |
“Korean Application Serial No. 10-2019-7018501, Notice of Preliminary Rejection dated Apr. 16, 2020”, w/ English Translation, 20 pgs. |
“Korean Application Serial No. 10-2019-7018501, Response filed Jun. 16, 2020 to Notice of Preliminary Rejection dated Apr. 16, 2020”, w/ English Claims, 17 pgs. |
“U.S. Appl. No. 15/369,499, Examiner Interview Summary dated Sep. 21, 2020”, 3 pgs. |
“U.S. Appl. No. 15/369,499, Examiner Interview Summary dated Oct. 9, 2020”, 2 pgs. |
“U.S. Appl. No. 15/369,499, Notice of Allowance dated Oct. 26, 2020”, 17 pgs. |
“U.S. Appl. No. 15/369,499, Response filed Sep. 15, 2020 to Final Office Action dated Jun. 15, 2020”, 10 pgs. |
“U.S. Appl. No. 16/115,259, Non Final Office Action dated Jan. 11, 2021”, 17 pgs. |
“U.S. Appl. No. 16/115,259, Response filed Oct. 22, 2020 to Final Office Action dated Jul. 22, 2020”, 10 pgs. |
“U.S. Appl. No. 16/126,869, Non Final Office Action dated Oct. 30, 2020”, 9 pgs. |
“U.S. Appl. No. 16/126,869, Response filed Oct. 7, 2020 to Final Office Action dated Jul. 7, 2020”, 10 pgs. |
“U.S. Appl. No. 16/552,003, Notice of Allowance datd Aug. 27, 2020”, 15 pgs. |
“European Application Serial No. 17876226.6, Response filed Oct. 2, 2020 to Communication Pursuant to Article 94(3) EPC dated May 29, 2020”, 22 pgs. |
“Korean Application Serial No. 10-2019-7018501, Final Office Action dated Sep. 8, 2020”, w/ English translation, 9 pgs. |
“Korean Application Serial No. 10-2019-7018501, Response filed Dec. 7, 2020 to Final Office Action dated Sep. 8, 2020”, w/ English Claims, 26 pgs. |
“U.S. Appl. No. 16/126,869, Response filed May 10, 2021 to Final Office Action dated Feb. 8, 2021 ”, 10 pgs. |
“U.S. Appl. No. 16/115,259, Response filed May 11, 2021 to Non Final Office Action dated Jan. 11, 2021”, 14 pgs. |
U.S. Appl. No. 16/552,003, filed Aug. 27, 2019, Generating and Displaying Customized Avatars in Electronic Messages. |
“Korean Application Serial No. 10-2020-7035136, Notice of Preliminary Rejection dated Feb. 25, 2021”, w/ English Translation, 5 pgs. |
“U.S. Appl. No. 16/115,259, Preliminary Amendment filed Oct. 18, 2018 t”, 6 pgs. |
“U.S. Appl. No. 15/369,499, Response filed Nov. 15, 2018 to Non Final Office Action dated Aug. 15, 2018”, 10 pgs. |
“U.S. Appl. No. 13/979,974, Corrected Notice of Allowability dated Nov. 19, 2018”, 2 pgs. |
“U.S. Appl. No. 15/365,046, Non Final Office Action dated Dec. 20, 2018”, 36 pgs. |
U.S. Appl. No. 16/126,869, filed Sep. 10, 2018, System and Method for Generating and Displaying Avatars. |
U.S. Appl. No. 16/115,259, filed Aug. 28, 2018, Generating and Displaying Customized Avatars in Media Overlays. |
“U.S. Appl. No. 15/365,046, Notice of Allowance dated May 21, 2019”, 14 pgs. |
“U.S. Appl. No. 15/369,499, Final Office Action dated Jan. 31, 2019”, 22 pgs. |
“U.S. Appl. No. 15/369,499, Response filed Mar. 14, 2019 to Final Office Action dated Jan. 31, 2019”, 12 pgs. |
“U.S. Appl. No. 15/365,046, Response filed Mar. 20, 2019 to Non Final Office Action dated Dec. 20, 2018”, 20 pgs. |
U.S. Appl. No. 13/979,974, filed Jul. 16, 2013, System and Method for Adaptable Avatars. |
U.S. Appl. No. 15/365,046, filed Nov. 30, 2016, Generating and Displaying Customized Avatars in Electronic Messages. |
U.S. Appl. No. 15/369,499, filed Dec. 5, 2016, Generating and Displaying Customized Avatars in Media Overlays. |
“U.S. Appl. No. 13/979,974, Final Office Action dated Apr. 25, 2018”, 18 pgs. |
“U.S. Appl. No. 13/979,974, Notice of Allowance dated Aug. 10, 2018”, 9 pgs. |
“U.S. Appl. No. 13/979,974, Response filed Jan. 30, 2018 to Non Final Office Action dated Oct. 3, 2017”, 8 pgs. |
“U.S. Appl. No. 13/979,974, Response filed Jul. 25, 2018 to Final Office Action dated Apr. 25, 2018”, 10 pgs. |
“U.S. Appl. No. 15/369,499, Non Final Office Action dated Aug. 15, 2018”, 22 pgs. |
“International Application Serial No. PCT/US2017/063981, International Search Report dated Mar. 22, 2018”, 3 pgs. |
“International Application Serial No. PCT/US2017/063981, Written Opinion dated Mar. 22, 2018”, 8 pgs. |
“U.S. Appl. No. 13/979,974, Examiner Interview Summary dated Jun. 29, 2017”, 3 pgs. |
“U.S. Appl. No. 13/979,974, Examiner Interview Summary dated Sep. 15, 2017”, 3 pgs. |
“U.S. Appl. No. 13/979,974, Final Office Action dated Jun. 9, 2017”, 20 pgs. |
“U.S. Appl. No. 13/979,974, Final Office Action dated Oct. 12, 2016”, 13 pgs. |
“U.S. Appl. No. 13/979,974, Non Final Office Action dated Feb. 22, 2017”, 17 pgs. |
“U.S. Appl. No. 13/979,974, Non Final Office Action dated Apr. 27, 2016”, 16 pgs. |
“U.S. Appl. No. 13/979,974, Non Final Office Action dated Oct. 3, 2017”, 17 pgs. |
“U.S. Appl. No. 13/979,974, Response filed May 22, 2017 to Non Final Office Action dated Feb. 22, 2017”, 10 pgs. |
“U.S. Appl. No. 13/979,974, Response filed Jul. 26, 2016 to Non Final Office Action dated Apr. 27, 2016”, 8 pgs. |
“U.S. Appl. No. 13/979,974, Response filed Sep. 11, 2017 to Final Office Action dated Jun. 9, 2017”, 8 pgs. |
“U.S. Appl. No. 13/979,974, Response filed Jan. 12, 2017 to Non Final Office Action dated Apr. 27, 2016”, 8 pgs. |
“International Application Serial No. PCT/CA2013/000454, International Preliminary Report on Patentability dated Nov. 20, 2014”, 9 pgs. |
“International Application Serial No. PCT/CA2013/000454, International Search Report dated Aug. 20, 2013”, 3 pgs. |
“International Application Serial No. PCT/CA2013/000454, Written Opinion dated Aug. 20, 2013”, 7 pgs. |
Leyden, John, “This SMS will self-destruct in 40 seconds”, [Online]. Retrieved from the Internet: <URL: http://www.theregister.co.uk/2005/12/12/stealthtext/, (Dec. 12, 2005), 1 pg. |
“U.S. Appl. No. 15/369,499, Corrected Notice of Allowability dated Jan. 28, 2021”, 3 pgs. |
“U.S. Appl. No. 16/126,869, Final Office Action dated Feb. 8, 2021”, 8 pgs. |
“U.S. Appl. No. 16/126,869, Response filed Feb. 1, 2021 to Non Final Office Action dated Oct. 30, 2020”, 9 pgs. |
“U.S. Appl. No. 17/247,169, Preliminary Amendment filed Feb. 2, 2021”, 7 pgs. |
“U.S. Appl. No. 16/126,869, Non Final Office Action dated May 19, 2021”, 8 pgs. |
“U.S. Appl. No. 16/115,259, Final Office Action dated Jul. 13, 2021”, 18 pgs. |
“U.S. Appl. No. 16/126,869, Response filed Aug. 19, 21 to Non Final Office Action dated May 19, 2021”, 10 pgs. |
“U.S. Appl. No. 17/247,169, Corrected Notice of Allowability dated Jul. 9, 2021”, 2 pgs. |
“U.S. Appl. No. 17/247,169, Notice of Allowance dated Jul. 2, 2021”, 14 pgs. |
“U.S. Appl. No. 16/115,259, Non-Final Office Action dated Nov. 8, 2021”, 17 pgs. |
“U.S. Appl. No. 16/115,259, Response filed Oct. 13, 2021 to Final Office Action dated Jul. 13, 2021”, 10 pgs. |
“U.S. Appl. No. 16/126,869, Final Office Action dated Nov. 29, 2021”, 9 pgs. |
“U.S. Appl. No. 17/247,169, Corrected Notice of Allowability dated Oct. 15, 2021”, 7 pgs. |
“U.S. Appl. No. 17/247,169, Corrected Notice of Allowability dated Nov. 30, 2021”, 2 pgs. |
Number | Date | Country | |
---|---|---|---|
20170113140 A1 | Apr 2017 | US |
Number | Date | Country | |
---|---|---|---|
61644057 | May 2012 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13979974 | US | |
Child | 15401926 | US |