Conditional modification of augmented reality object

Information

  • Patent Grant
  • 10789749
  • Patent Number
    10,789,749
  • Date Filed
    Wednesday, May 6, 2020
    4 years ago
  • Date Issued
    Tuesday, September 29, 2020
    3 years ago
Abstract
A server machine modifies an augmented reality (AR) object in response to fulfillment of a condition. The machine provides, to a user device, object data that defines the AR object. The object data specifies a physical geolocation of the AR object, a presentation attribute of the AR object, a conditional modification program, and a trigger condition for execution of the conditional modification program. The object data causes the user device to present the AR object with a first appearance, located at the physical geolocation. The machine detects fulfillment of the trigger condition, and in response, the machine executes the conditional modification program. This modifies the object data by modifying the presentation attribute. The machine provides, to the user device, the modified object data, which causes the user device to present the AR object with a second appearance based on the modified presentation attribute.
Description
TECHNICAL FIELD

The subject matter disclosed herein generally relates to the technical field of special-purpose machines that facilitate generation and presentation of computer graphics, including software-configured computerized variants of such special-purpose machines and improvements to such variants, and to the technologies by which such special-purpose machines become improved compared to other special-purpose machines that facilitate generation and presentation of computer graphics. Specifically, the present disclosure addresses systems and methods to facilitate presentation of an augmented reality (AR) object.


BACKGROUND

A machine (e.g., a user's device) may be configured to interact with one or more users by generating computer graphics and causing presentation thereof in an AR environment or a virtual reality (VR) environment. For example, a user's device in the example form of a smartphone may generate AR content (e.g., one or more AR objects) and cause the AR content to be displayed to the user (e.g., on a display screen of the smartphone or a head-mounted display (HMD) controlled by the smart phone).





BRIEF DESCRIPTION OF THE DRAWINGS

Some embodiments are illustrated by way of example and not limitation in the figures of the accompanying drawings.



FIG. 1 is a network diagram illustrating a network environment suitable for facilitating conditional modification of an AR object, according to some example embodiments.



FIG. 2 is a block diagram illustrating components of a server machine suitable for facilitating conditional modification of an AR object, according to some example embodiments.



FIG. 3 is a block diagram illustrating components of a database suitable for facilitating conditional modification of an AR object, according to some example embodiments.



FIGS. 4-9 are flowcharts illustrating operations of the server machine in performing a method of conditionally modifying of an AR object, according to some example embodiments.



FIG. 10 is a block diagram illustrating components of a machine, according to some example embodiments, able to read instructions from a machine-readable medium and perform any one or more of the methodologies discussed herein.



FIG. 11 and FIG. 12 are diagrams illustrating screenshots of an AR object before and after conditional modification of the AR object, according to some example embodiments.





DETAILED DESCRIPTION

Example methods (e.g., algorithms) facilitate conditional modification of one or more AR objects, and example systems (e.g., special-purpose machines configured by special-purpose software) are configured to facilitate conditional modification of one or more AR objects. Examples merely typify possible variations. Unless explicitly stated otherwise, structures (e.g., structural components, such as modules) are optional and may be combined or subdivided, and operations (e.g., in a procedure, algorithm, or other function) may vary in sequence or be combined or subdivided. In the following description, for purposes of explanation, numerous specific details are set forth to provide a thorough understanding of various example embodiments. It will be evident to one skilled in the art, however, that the present subject matter may be practiced without these specific details.


A machine (e.g., a server machine) is configured (e.g., by suitable software, hardware, or both) to provide a consistent set of one or more AR objects to one or more user devices (e.g., one or more smartphones, game controllers, HMDs, or combinations thereof), and the machine is configured to modify one or more of these AR objects in response to detected fulfillment of one or more conditions. For clarity and brevity, the discussion herein focuses on conditional modification of a single AR object, presented with an appearance of being present in the real world, though the example systems and methods discussed herein are applicable to conditional modification of multiple AR objects similarly presented.


Configured in accordance with the systems and methods discussed herein, the machine provides, to a user device, object data that defines an AR object. The object data of the AR object, in some example embodiments, specifies four things: (1) a physical geolocation of the AR object, (2) a presentation attribute of the AR object, (3) a conditional modification program of the AR object, and (4) a trigger condition for execution of the conditional modification program. The provided object data causes the user device, as well as any number of other user devices in the area, to present the AR object with a consistent first appearance and consistently located at the physical geolocation. The machine then detects fulfillment of the trigger condition specified by the object data, and based on the detected fulfillment of the trigger condition, the machine executes the conditional modification program of the AR object. The executing of the conditional modification program modifies the object data of the AR object by modifying the presentation attribute of the AR object. The machine next provides, to the user device, as well as any other user devices present, the modified object data of the AR object. The modified object data causes the user device, as well as any other user device in the area, to present the AR object with a consistent second appearance that is based on the modified presentation attribute of the AR object.


By causing the AR object to change from its first appearance to its second appearance, in accordance with the methodologies described herein, the machine provides interactivity between the AR object and one or more users. For example, the detected trigger condition may be or include a user or her device intersecting the AR object, and in response to this intersection, the AR object may change color (e.g., from gray to blue, darker hue to lighter hue, or both) and emit animated sparkles. As another example, the detected trigger condition may be or include an increase in ambient audio loudness, and in response to this increase, the AR object may change color and begin pulsating or flashing. As yet another example, the detected trigger condition may be or include a pattern of sound (e.g., a candidate pattern that matches a reference pattern, which may represent clapping, stomping, or chanting), and in response to this pattern of sound, the AR object may change size (e.g., appear to inflate).


As a further example, the detected trigger condition may be or include a user's device selecting the AR object for editing, and in response to this selection, the AR object may be displayed with a cursor or other visual highlight that indicates which portion of the AR object is ready to be edited. This may have the effect of assisting the user in planning or making an edit to the AR object. As a still further example, the detected trigger condition may be or include an interaction count transgressing a threshold value (e.g., a minimum or maximum interaction count), and in response to this transgression, the AR object may bounce vertically in space and emit animated fireworks. This may have the effect of indicating the attaining of a milestone in the AR object's user interactivity. As a yet further example, the detected trigger condition may be or include an interaction time transgressing a threshold value (e.g., minimum or maximum interaction time), and in response to this transgression, the AR object may change its size, its physical geolocation in the real world, or both. This may have the effect of altering the attractiveness or availability of the AR object for further interactions with users.


According to various example embodiments, the conditional modification program may be provided by the machine (e.g., as part of generating the object data of the AR object) or provided by a user's device (e.g., as an edit to the object data the AR object). In general, the object data of the AR object may contain one or more user-provided conditional modification programs, which may be a subset of all conditional modification programs specified for the AR object. In some example embodiments, a user-provided conditional modification program is incorporated by the machine into the object data for a specified period of time (e.g., in exchange for a fee payment or a task performance from the user's device). After the specified time period is over, the machine may remove the user-provided conditional modification program, prompt the user to renew the conditional modification program for an additional period of time, or both.



FIG. 1 is a network diagram illustrating a network environment 100 suitable for conditional modification of an AR object, according to some example embodiments. The network environment 100 includes a server machine 110, a database 115, and devices 130 and 150, all communicatively coupled to each other via a network 190. The server machine 110, with or without the database 115, may form all or part of a cloud 118 (e.g., a geographically distributed set of multiple machines configured to function as a single server), which may form all or part of a network-based system 105 (e.g., a cloud-based server system configured to provide one or more network-based services, such as AR services, to the devices 130 and 150). The server machine 110, the database 115, and the devices 130 and 150 may each be implemented in a special-purpose (e.g., specialized) computer system, in whole or in part, as described below with respect to FIG. 10.


Also shown in FIG. 1 are users 132 and 152. One or both of the users 132 and 152 may be a human user (e.g., a human being), a machine user (e.g., a computer configured by a software program to interact with the device 130 or 150), or any suitable combination thereof (e.g., a human assisted by a machine or a machine supervised by a human). The user 132 is associated with the device 130 and may be a user of the device 130. For example, the device 130 may be a desktop computer, a vehicle computer, a home media system (e.g., a home theater system or other home entertainment system), a tablet computer, a navigational device, a portable media device, a smartphone (e.g., with a touchscreen configured to detect one or more touch gestures), or a wearable device (e.g., an HMD, such as an AR headset, a smart watch, smart glasses, smart clothing, or smart jewelry) belonging to the user 132. Likewise, the user 152 is associated with the device 150 and may be a user of the device 150. As an example, the device 150 may be a desktop computer, a vehicle computer, a home media system (e.g., a home theater system or other home entertainment system), a tablet computer, a navigational device, a portable media device, a smartphone (e.g., with a touchscreen configured to detect one or more touch gestures), or a wearable device (e.g., an HMD, such as an AR headset, a smart watch, smart glasses, smart clothing, or smart jewelry) belonging to the user 152.


Any of the systems or machines (e.g., databases and devices) shown in FIG. 1 may be, include, or otherwise be implemented in a special-purpose (e.g., specialized or otherwise non-conventional and non-generic) computer that has been modified to perform one or more of the functions described herein for that system or machine (e.g., configured or programmed by special-purpose software, such as one or more software modules of a special-purpose application, operating system, firmware, middleware, or other software program). For example, a special-purpose computer system able to implement any one or more of the methodologies described herein is discussed below with respect to FIG. 10, and such a special-purpose computer may accordingly be a means for performing any one or more of the methodologies discussed herein. Within the technical field of such special-purpose computers, a special-purpose computer that has been specially modified (e.g., configured by special-purpose software) by the structures discussed herein to perform the functions discussed herein is technically improved compared to other special-purpose computers that lack the structures discussed herein or are otherwise unable to perform the functions discussed herein. Accordingly, a special-purpose machine configured according to the systems and methods discussed herein provides an improvement to the technology of similar special-purpose machines.


As used herein, a “database” is a data storage resource and may store data structured as a text file, a table, a spreadsheet, a relational database (e.g., an object-relational database), a triple store, a hierarchical data store, or any suitable combination thereof. Moreover, any two or more of the systems or machines illustrated in FIG. 1 may be combined into a single system or machine, and the functions described herein for any single system or machine may be subdivided among multiple systems or machines.


The network 190 may be any network that enables communication between or among systems, machines, databases, and devices (e.g., between the machine 110 and the device 130). Accordingly, the network 190 may be a wired network, a wireless network (e.g., a mobile or cellular network), or any suitable combination thereof. The network 190 may include one or more portions that constitute a private network, a public network (e.g., the Internet), or any suitable combination thereof. Accordingly, the network 190 may include one or more portions that incorporate a local area network (LAN), a wide area network (WAN), the Internet, a mobile telephone network (e.g., a cellular network), a wired telephone network (e.g., a plain old telephone service (POTS) network), a wireless data network (e.g., a WiFi network or WiMax network), or any suitable combination thereof. Any one or more portions of the network 190 may communicate information via a transmission medium. As used herein, “transmission medium” refers to any intangible (e.g., transitory) medium that is capable of communicating (e.g., transmitting) instructions for execution by a machine (e.g., by one or more processors of such a machine), and includes digital or analog communication signals or other intangible media to facilitate communication of such software.



FIG. 2 is a block diagram illustrating components of the server machine 110, according to some example embodiments. The server machine 110 is shown as including a client interface 210, a trigger condition detector 220, and an object modifier 230, all configured to communicate with each other (e.g., via a bus, shared memory, or a switch). The client interface 210 may be or include a client interface module or similarly suitable software code configured to interface with one or more clients of the server machine 110 (e.g., one or more of the devices 130 and 150). The trigger condition detector 220 may be or include a trigger detection module or similarly suitable software code configured to detect one or more trigger conditions (e.g., a trigger condition for execution of a conditional modification program for an AR object). The object modifier 230 may be or include a program execution module or similarly suitable software code configured to execute a program (e.g., a conditional modification program for an AR object).


As shown in FIG. 2, the client interface 210, the trigger condition detector 220, the object modifier 230, or any suitable combination thereof, may form all or part of an application 200 (e.g., a server-side app or other server-executed program) that is stored (e.g., installed) on the server machine 110 (e.g., responsive to or otherwise as a result of data being received via the network 190). Furthermore, one or more processors 299 (e.g., hardware processors, digital processors, or any suitable combination thereof) may be included (e.g., temporarily or permanently) in the application 200, the client interface 210, the trigger condition detector 220, the object modifier 230, or any suitable combination thereof.


Any one or more of the components (e.g., modules) described herein may be implemented using hardware alone (e.g., one or more of the processors 299) or a combination of hardware and software. For example, any component described herein may physically include an arrangement of one or more of the processors 299 (e.g., a subset of or among the processors 299) configured to perform the operations described herein for that component. As another example, any component described herein may include software, hardware, or both, that configure an arrangement of one or more of the processors 299 to perform the operations described herein for that component. Accordingly, different components described herein may include and configure different arrangements of the processors 299 at different points in time or a single arrangement of the processors 299 at different points in time. Each component (e.g., module) described herein is an example of a means for performing the operations described herein for that component. Moreover, any two or more components described herein may be combined into a single component, and the functions described herein for a single component may be subdivided among multiple components. Furthermore, according to various example embodiments, components described herein as being implemented within a single system or machine (e.g., a single device) may be distributed across multiple systems or machines (e.g., multiple devices).



FIG. 3 is a block diagram illustrating components of the database 115, according to some example embodiments. The database 115 is shown as storing object data 300 of an AR object. The object data 300 corresponds to the AR object, defines the AR object, and specifies how the AR object is to be presented (e.g., visually, acoustically, or both) in an AR environment. Accordingly, one or more portions of the object data 300 may be considered as metadata of the AR object to which the object data 300 corresponds.


As shown in FIG. 3, the object data 300 specifies (e.g., by inclusion or by reference) one or more of various characteristics or other descriptors of the AR object. For example, the object data 300 is shown as specifying (e.g., by inclusion) a physical geolocation 310 of the AR object in the real world. In some example embodiments, the physical geolocation 310 is specified by global geolocation coordinates relative to the earth (e.g., a triplet of latitude, longitude, and altitude). In other example embodiments, the physical geolocation 310 is specified by relative geolocation coordinates relative to a room, a building, a landmark, or other suitable frame of reference. In some example embodiments, the physical geolocation 310 is treated as a type of presentation attribute for presenting the AR object.


The object data 300 may specify physical dimensions 312 of the AR object in the real world. The physical dimensions 312 may fully or partially define the size of the AR object, the shape of the AR object, the orientation of the AR object, or any suitable combination thereof. For example, the physical dimensions 312 may specify one or more linear dimensions (e.g., length, width, and height) of the AR object. In some example embodiments, the physical dimensions 312 are treated as presentation attributes for presenting the AR object.


The object data 300 may specify one or more presentation attributes 320 and 322 (e.g., additional presentation attributes beyond the physical geolocation 310 and the physical dimensions 312) of the AR object. A presentation attribute specifies how the AR object is to be presented (e.g., visually, acoustically, haptically, or any suitable combination thereof). Examples of a presentation attribute include a color of the AR object, a color pattern (e.g., chroma texture map) applied to the AR object, a brightness of the AR object, a brightness pattern (e.g., a luma texture map) applied to the AR object, a flash indicator of the AR object (e.g., to indicate whether the AR object should be displayed in a flashing state), a movement indicator of the AR object (e.g., to indicate whether the AR object should be displayed in a vertically bouncing state, horizontally jiggling state, or both), an audio file of the AR object (e.g., to be played when one or more users is physically within a threshold distance of the physical geolocation 310 of the AR object), an audio volume of the AR object (e.g., to provide a basis for audio playback of the audio file of the AR object), or any suitable combination thereof.


The object data 300 may specify one or more pairs of conditional modification programs of the AR object, each with one or more corresponding trigger conditions for execution of that conditional modification program. FIG. 3 illustrates a conditional modification program 330 paired with a corresponding trigger condition 332, as well as another conditional modification program 340 paired with a corresponding trigger condition 342. Accordingly, if the trigger condition detector 220 (e.g., in the server machine 110) detects fulfillment of the trigger condition 332, the corresponding conditional modification program 330 is executed (e.g., by the object modifier 230 in the server machine 110). Similarly, if the trigger condition detector 220 detects fulfillment of the trigger condition 342, the corresponding conditional modification program 340 is executed.


According to some example embodiments, the object data 300 specifies a total interaction count 350, which may quantify (e.g., by enumerating) a total number of discrete user interactions experienced by the AR object. The total interaction count 350 may be paired with one or more corresponding demographic interaction counts. FIG. 3 illustrates the total interaction count 350 paired with a single corresponding demographic interaction count 352, though additional corresponding demographic interaction counts may also be included in the object data 300 and associated with (e.g., linked, mapped, or otherwise assigned to) the total interaction count 350. The demographic interaction count 352 may quantify a total number of discrete user interactions experienced by the AR object with users from a particular demographic group (e.g., defined by age, gender, education, occupation, favorite ice cream flavor, any other demographic descriptor, or any suitable combination thereof).


According to certain example embodiments, the object data 300 specifies a total interaction time 360, which may quantify (e.g., by valuing) a total amount of time during which the AR object experienced user interactions. The total interaction time 360 may be paired with one more corresponding demographic interaction times. FIG. 3 illustrates the total interaction time 360 being paired with a single corresponding demographic interaction time 362, though additional corresponding demographic interaction times may also be included in the object data 300 and associated with (e.g., linked, mapped, or otherwise assigned to) the total interaction time 360. The demographic interaction time 362 may quantify a total amount of time during which the object experienced user interactions with users from a particular demographic group (e.g., defined by age, gender, education, occupation, favorite ice cream flavor, any other demographic descriptor, or any suitable combination thereof).



FIGS. 4-9 are flowcharts illustrating operations in performing a method 400 of conditionally modifying of an AR object, according to some example embodiments. For clarity and brevity, operations in the method 400 are presently described as being performed by the server machine 110, using components (e.g., modules) described above with respect to FIG. 2, using one or more processors (e.g., microprocessors or other hardware processors), or using any suitable combination thereof. However, in various example embodiments, the method 400 may be performed by the device 130, alone or in conjunction with one or more other machines (e.g., the server machine 110, the database 115, or the device 150). As shown in FIG. 4, the method 400 includes operations 410, 420, 430, and 440.


In operation 410, the client interface 210 provides the object data 300 to the device 130. For example, the client interface 210 may retrieve the object data 300 from the database 115 and provide the object data 300 to the device 130 via the network 190. As another example, the client interface 210 may command or otherwise cause the database 115 to provide the object data 300 to the device 130. As noted above, the object data 300 corresponds to an AR object and specifies, for example, the physical geolocation 310, the presentation attribute 320 (e.g., color or size), the conditional modification program 330, and the corresponding trigger condition 332. The object data 300 causes the device 130 to present (e.g., via a display screen or an AR headset) the AR object defined by the object data 300. In accordance with the provided object data 300, the device 130 presents the AR object at the physical geolocation 310 and with a first appearance (e.g., visual appearance, acoustic appearance, or both) specified by the object data 300 (e.g., specified by the presentation attribute 320).


In operation 420, the trigger condition detector 220 detects that the trigger condition 332 has been fulfilled. This may be performed by receiving an indicator or other signal from the client interface 210 (e.g., as a result of a status message being received by the client interface 210 via the network 190 from the device 130). As noted above, the trigger condition 332 corresponds to the conditional modification program 330, and the object data 300 of the AR object specifies that the trigger condition 332 is a trigger for executing the corresponding conditional modification program 330 to modify the AR object.


To illustrate an example, suppose that the trigger condition 332 for executing the conditional modification program 330 specifies reception of an indication that a portion of the AR object is being edited. In this situation, the detecting in operation 420 that the trigger condition 332 has been fulfilled may include receiving such an indication (e.g., from the device 130 via the network 190) that a portion of the AR object is being edited.


To illustrate another example, suppose that the trigger condition 332 for executing the conditional modification program 330 specifies a threshold level (e.g., a reference threshold level) of ambient audio loudness. In such a situation, the detecting in operation 420 that the trigger condition 332 has been fulfilled may include receiving an indication that a microphone of the device 130 has detected an actual level of ambient audio loudness that transgresses (e.g., meets or exceeds) the threshold level of ambient audio loudness.


To illustrate a further example, suppose that the trigger condition 332 specifies a reference pattern of sounds (e.g., clapping or a song, such as “The Star-Spangled Banner”). In such a situation, the detecting in operation 420 that the trigger condition 332 has been fulfilled may include receiving an indication that the microphone of the device 130 has captured an actual pattern of sounds that matches the reference pattern of sounds.


In operation 430, the object modifier 230 executes the conditional modification program 330 that corresponds to the trigger condition 332 that was detected to be fulfilled in operation 420. This is performed in response to the detected fulfillment of the trigger condition 332. Accordingly, the execution of the conditional modification program 330 modifies the object data 300 of the AR object and consequently modifies the AR object. For example, execution of the conditional modification program 330 may modify the object data 300 by modifying the presentation attribute 320 (e.g., color or size) within the object data 300. The modification of the object data 300 may occur within the database 115, within the server machine 110 (e.g., within a memory of the server machine 110), or within any suitable combination thereof.


To illustrate an example, suppose that that the trigger condition 332 for executing the conditional modification program 330 specifies reception of an indication that a portion of the AR object is being edited. In this situation, in response to a detection in operation 420 that the trigger condition 332 has been fulfilled, the executing of the conditional modification program 330 may add an AR cursor to the AR object in its second appearance, and the added AR cursor may indicate that the portion of the AR object is being edited. For example, the AR cursor may indicate a location of a non-committed edit to the AR object (e.g., an edit that is not yet saved or otherwise persistently incorporated into the object data 300 of the AR object).


In operation 440, the client interface 210 provides at least the modified portion of the object data 300 to the device 130. For example, the entirety of the modified object data 300 may be communicated to the device 130 via the network 190, or the modified presentation attribute 320 (e.g., size or color of the AR object) may be communicated to the device 130 via the network 190. In some example embodiments, the modified object data 300 is stored in the database 115, and the client interface 210 commands or otherwise causes the device 130 to retrieve at least the modified portion (e.g., the modified presentation attribute 320) of the object data 300 via the network 190. The modified object data 300 causes the device 130 to present (e.g., via a display screen or an AR headset) the AR object in modified form, as modified in accordance with a modified object data 300. Thus, in accordance with the modified object data 300, the device 130 presents the AR object with a second appearance specified by the modified object data 300 (e.g., specified by the modified presentation attribute 320).


For example, the presentation attribute modified in operation 430 (e.g., presentation attribute 320 or physical dimensions 312) may specify a physical dimension (e.g., length, width, height, depth, thickness, or any suitable combination thereof) of the AR object, and that physical dimension may fully or partially determine the real-world size of the AR object. Accordingly, the executing of the conditional modification program 330 in operation 430 may modify this physical dimension of the AR object and consequently alter the real-world size of the AR object, as presented with its second appearance.


As another example, the presentation attribute modified in operation 430 (e.g., presentation attribute 320) may specify a color (e.g., hue, saturation, brightness, or any suitable combination thereof) of at least a portion of the AR object. Accordingly, the executing of the conditional modification program 330 in operation 430 may modify this color of the AR object and consequently alter the color of at least a portion of the AR object, as presented with its second appearance.


As a further example, the presentation attribute modified in operation 430 (e.g., presentation attribute 320 or physical geolocation 310) may specify the physical geolocation of the AR object in the real world, and that physical geolocation may fully or partially determine the real-world geolocation of the AR object. Accordingly, the executing of the conditional modification program 330 in operation 430 may modify this real-world geolocation of the AR object and consequently alter where the AR object is presented relative to the real world, in accordance with its second appearance.


As shown in FIG. 5, in addition to any one or more of the operations previously described, the method 400 may include one or more of operations 510, 512, 520, 530, 532, 534, 540, and 542. According to some example embodiments, operations 510 and in 512 are performed prior to operation 420, in which the trigger condition detector 220 detects fulfillment of the trigger condition 332.


In operation 510, the client interface 210 receives an indication that the device 130 submitted an edit to the AR object (e.g., an edit in the form of a newly added portion, removal of an existing portion, or modification to an existing portion of the AR object). For example, the client interface 210 may receive a request that an edit be performed on the AR object (e.g., stored in the object data 300 or otherwise persistently incorporated into the definition of the AR object), and such a request may be received by the network 190 from the device 130. In some example embodiments, the received indication also refers to a user profile (e.g., stored by the device 130, the server machine 110, the database 115, or any suitable combination thereof) that describes the user 132 and indicates that the user 132 belongs to a particular demographic group (e.g., single women between ages 25 and 40 whose favorite ice cream flavor is chocolate).


In operation 512, the client interface 210 increments an interaction count that corresponds to the AR object (e.g., the total interaction count 350, the demographic interaction count 352, or both), in response to the indication received in operation 510. For example, the client interface 210 may increment the total interaction count 350 that corresponds to the AR object. As another example, if the received indication in operation 510 indicates that the user 132 is a member of a particular demographic group (e.g., single women between 25 and 40 whose favorite ice cream flavor is chocolate), the demographic interaction count for that particular demographic group (e.g., demographic interaction count 352) may be incremented by the client interface 210. In such a situation, the total interaction count 350 may also be incremented as well (e.g., to tally user interactions of multiple kinds, from multiple demographic groups, or both).


In example embodiments that include operation 512, operation 520 may be performed as part (e.g., a precursor task, a subroutine, or a portion) of operation 420, in which the trigger condition detector 220 detects fulfillment of the trigger condition 332. In operation 520, the trigger condition detector 220 performs a comparison of the interaction count (e.g., the total interaction count 350 or the demographic interaction count 352) incremented in operation 512 to a corresponding threshold interaction count (e.g., a predetermined total interaction count or a predetermined demographic interaction count, either of which may be a maximum interaction count or a minimum interaction count).


Accordingly, the detection of the fulfillment of the trigger condition 332 in operation 420 may be based on the comparison performed in operation 520. For example, if the total interaction count 350 was incremented in operation 512, the trigger condition detector 220 may compare the total interaction count 350 to a reference total interaction count (e.g., a maximum interaction count of all types and from all users). As another example, if the demographic interaction count 352 was incremented in operation 512, the trigger condition detector 220 may compare the demographic interaction count 352 to a reference demographic interaction count (e.g., a maximum interaction count or a minimum interaction count for that demographic group).


As shown in FIG. 5, one or more of operations 530, 532, and 534 may be performed as part of operation 430, in which the object modifier 230 executes the conditional modification program 330 to modify the object data 300 of the AR object. In operation 530, the object modifier 230 modifies a physical dimension (e.g., as specified by the presentation attribute 320 or the physical dimensions 312) of the AR object by executing the conditional modification program 330. As noted above, this may have the effect of altering the real-world size of the AR object, as presented with its second appearance.


In operation 532, the object modifier 230 modifies the physical geolocation (e.g., specified by the presentation attribute 320 or the physical geolocation 310) of the AR object by executing the conditional modification program 330. As noted above, this may have the effect of altering where the AR object is presented relative to the real world, according to its second appearance.


In operation 534, the object modifier 230 modifies a presentation attribute (e.g., specified by the presentation attribute 320, which may, for example, specify color) of the AR object by executing the conditional modification program 330. As noted above, this may have the effect of altering the appearance of the AR object from its first appearance to its second appearance.


As shown in FIG. 5, one or more of operations 540 and 542 may be performed after operation 430, in which the object modifier 230 executes the conditional modification program 330 to modify the object data 300 of the AR object. In operation 540, the client interface 210 generates a geographical map (e.g., a map that illustrates roads and landmarks within a geographical area, such as a city). The geographical map includes a depiction or other representation (e.g., symbolic representation) of the AR object and indicates the physical geolocation 310 of the AR object. In addition, the depiction or other representation may have a color that indicates a frequency (e.g., averaging five times a day, twice a week, or once a month) or range of frequencies at which the AR object is edited (e.g., by user interactions). For example, a gray depiction may indicate that the AR object is edited by users less than once a week, while a blue depiction may indicate that the AR object is edited by users more than once a week.


In operation 542, the client interface 210 provides the geographical map generated in operation 540 to one or more devices. For example, the geographical map may be provided to the device 150 (e.g., as a guide for the user 152 to locate the AR object for further interaction, such as editing). Accordingly, the color of the depiction of the AR object within the provided geographical map can notify the user 152 of how often the AR object is edited.


As shown in FIG. 6, in addition to any one or more of the operations previously described, the method 400 may include one or more of operations 610, 612, and 620. According to some example embodiments, operation 610 and 612 are performed prior to operation 420, in which the trigger condition detector 220 detects fulfillment of the trigger condition 332.


In operation 610, the client interface 210 receives an indication that the device 130 has been editing the AR object or has edited the AR object for a period of time (e.g., a period of time during which the AR object has been kept in an editing state or other edit-accepting state). For example, the client interface 210 may receive a request that the AR object be unlocked for editing (e.g., with write-permissions for the object data 300 that defines the AR object), and such a request may be received by the network 190 from the device 130. In some example embodiments, the received indication also refers to a user profile (e.g., stored by the device 130, the server machine 110, the database 115, or any suitable combination thereof) that describes the user 132 and indicates that the user 132 belongs to a particular demographic group (e.g., married men between ages 40 and 55 whose favorite ice cream flavor is vanilla).


In operation 612, the client interface 210 updates an interaction time that corresponds to the AR object (e.g., the total interaction time 360, the demographic interaction time 362, or both), in response to the indication received in operation 610. For example, the client interface 210 may update the total interaction time 360 that corresponds to the AR object. As another example, if the received indication in operation 610 indicates that the user 132 is a member of a particular demographic group (e.g., married men between 40 and 55 whose favorite ice cream flavor is vanilla), the demographic interaction time for that particular demographic group (e.g., demographic interaction time 362) may be updated by the client interface 210. In such a situation, the total interaction time 360 may also be incremented as well (e.g., to track cumulative user interaction times from multiple demographic groups).


In example embodiments that include operation 612, operation 620 may be performed as part of operation 420, in which the trigger condition detector 220 detects fulfillment of the trigger condition 332. In operation 620, the trigger condition detector 220 performs a comparison of the interaction time (e.g., the total interaction time 360 or the demographic interaction time 362) updated in operation 512 to a corresponding threshold interaction time (e.g., a predetermined total interaction time or a predetermined demographic interaction time, either of which may be a maximum interaction time or a minimum interaction time).


Accordingly, the detection of the fulfillment of the trigger condition 332 in operation 420 may be based on the comparison performed in operation 620. For example, if the total interaction time 360 was updated in operation 612, the trigger condition detector 220 may compare the total interaction time 360 to a reference total interaction time (e.g., a maximum interaction time for all users). As another example, if the demographic interaction time 362 was updated in operation 612, the trigger condition detector 220 may compare the demographic interaction time 362 to a reference demographic interaction time (e.g., a maximum interaction time or a minimum interaction time for that demographic group).


As shown in FIG. 7, in addition to any one or more of the operations previously described, the method 400 may include one or more of operations 710, 512, and 520, the latter two of which have been previously described. According to some example embodiments, operations 710 and in 512 are performed prior to operation 420, in which the trigger condition detector 220 detects fulfillment of the trigger condition 332.


In operation 710, the client interface 210 receives an indication that the device 130 has intersected the AR object (e.g., an indication that the geolocation of the device 130 has coincided with the physical geolocation 310 of the AR object or has been encompassed by an outer surface of the AR object, as defined by the physical dimensions 312 of the AR object). For example, the client interface 210 may receive such an indication in the form of geolocation data of the device 130, which geolocation data may be received by the network 190 from the device 130. In some example embodiments, the received indication also refers to a user profile (e.g., stored by the device 130, the server machine 110, the database 115, or any suitable combination thereof) that describes the user 132 and indicates that the user 132 belongs to a particular demographic group (e.g., married women between ages 55 and 70 whose favorite ice cream flavor is coffee).


As noted above, in operation 512, the client interface 210 increments an interaction count that corresponds to the AR object (e.g., the total interaction count 350, the demographic interaction count 352, or both). This may be performed in response to the indication received in operation 710. For example, the client interface 210 may increment the total interaction count 350 that corresponds to the AR object. As another example, if the received indication in operation 710 indicates that the user 132 is a member of a particular demographic group (e.g., married women between 55 and 70 whose favorite ice cream flavor is coffee), the demographic interaction count for that particular demographic group (e.g., demographic interaction count 352) may be incremented by the client interface 210. In such a situation, the total interaction count 350 may also be incremented as well (e.g., to tally user interactions of multiple kinds, from multiple demographic groups, or both).


As noted above, in example embodiments that include operation 512, operation 520 may be performed as part (e.g., a precursor task, a subroutine, or a portion) of operation 420, in which the trigger condition detector 220 detects fulfillment of the trigger condition 332. In operation 520, the trigger condition detector 220 performs a comparison of the interaction count (e.g., the total interaction count 350 or the demographic interaction count 352) incremented in operation 512 to a corresponding threshold interaction count (e.g., a predetermined total interaction count or a predetermined demographic interaction count, either of which may be a maximum interaction count or a minimum interaction count).


Accordingly, the detection of the fulfillment of the trigger condition 332 in operation 420 may be based on the comparison performed in operation 520. For example, if the total interaction count 350 was incremented in operation 512, the trigger condition detector 220 may compare the total interaction count 350 to a reference total interaction count (e.g., a maximum interaction count of all types and from all users). As another example, if the demographic interaction count 352 was incremented in operation 512, the trigger condition detector 220 may compare the demographic interaction count 352 to a reference demographic interaction count (e.g., a maximum interaction count or a minimum interaction count for that demographic group).


As shown in FIG. 8, in addition to any one or more of the operations previously described, the method 400 may include one or more of operations 810, 612, and 620, the latter two of which have been previously described. According to some example embodiments, operation 810 and 612 are performed prior to operation 420, in which the trigger condition detector 220 detects fulfillment of the trigger condition 332.


In operation 810, the client interface 210 receives an indication that the device 130 has been intersecting the AR object or has intersected the AR object for a period of time (e.g., a period of time during which the geolocation of the device 130 has coincided with the physical geolocation 310 of the AR object or has been encompassed by an outer surface of the AR object, as defined by the physical dimensions 312 of the AR object). For example, the client interface 210 may receive a status message indicating a running tally of intersection time, and such a status message may be received by the network 190 from the device 130. In some example embodiments, the received indication also refers to a user profile (e.g., stored by the device 130, the server machine 110, the database 115, or any suitable combination thereof) that describes the user 132 and indicates that the user 132 belongs to a particular demographic group (e.g., married men between ages 40 and 55 whose favorite ice cream flavor is vanilla).


As noted above, in operation 612, the client interface 210 updates an interaction time that corresponds to the AR object (e.g., the total interaction time 360, the demographic interaction time 362, or both). This may be performed in response to the indication received in operation 810. For example, the client interface 210 may update the total interaction time 360 that corresponds to the AR object. As another example, if the received indication in operation 810 indicates that the user 132 is a member of a particular demographic group (e.g., married women between 55 and 70 whose favorite ice cream flavor is coffee), the demographic interaction time for that particular demographic group (e.g., demographic interaction time 362) may be updated by the client interface 210. In such a situation, the total interaction time 360 may also be incremented as well (e.g., to track cumulative user interaction times from multiple demographic groups).


As noted above, in example embodiments that include operation 612, operation 620 may be performed as part of operation 420, in which the trigger condition detector 220 detects fulfillment of the trigger condition 332. In operation 620, the trigger condition detector 220 performs a comparison of the interaction time (e.g., the total interaction time 360 or the demographic interaction time 362) updated in operation 512 to a corresponding threshold interaction time (e.g., a predetermined total interaction time or a predetermined demographic interaction time, either of which may be a maximum interaction time or a minimum interaction time).


Accordingly, the detection of the fulfillment of the trigger condition 332 in operation 420 may be based on the comparison performed in operation 620. For example, if the total interaction time 360 was updated in operation 612, the trigger condition detector 220 may compare the total interaction time 360 to a reference total interaction time (e.g., a maximum interaction time for all users). As another example, if the demographic interaction time 362 was updated in operation 612, the trigger condition detector 220 may compare the demographic interaction time 362 to a reference demographic interaction time (e.g., a maximum interaction time or a minimum interaction time for that demographic group).


As shown in FIG. 9, in addition to any one or more of the operations previously described, the method 400 may include one or more of operations 900, 902, and 920. FIG. 9 illustrates operations 900 and 902 being performed prior to operation 410, in which the client interface 210 provides the object data 300 of the AR object to the device 130.


In operation 900, the client interface 210 receives a user submission that includes or otherwise identifies (e.g., by reference) the conditional modification program 330. That is, the conditional modification program 330 may be user-submitted (e.g., by the user 132 or by the user 152). Thus, the client interface 210 may receive the user submission (e.g., within a request or other message) via the network 190 from the device 130 or from the device 150. The user submission may also include the corresponding trigger condition 332 for execution of the conditional modification program 330. In some example embodiments, the user submission further includes a corresponding value accorded to the conditional modification program 330 (e.g., a value that indicates a bid or fee paid for incorporating the submitted conditional modification program 330 into the object data 300 of the AR object).


In operation 902, the client interface 210 determines a period of time during which the submitted conditional modification program 330 is to be specified (e.g., by inclusion or by reference) by the object data 300 that defines the AR object. In some example embodiments, this may be performed based on (e.g., in response to) the value accorded to the conditional modification program 330. For example, the user 152 may have paid a fee to have the submitted conditional modification program 330 temporarily incorporated into the object data 300 of the AR object, and this fee may be reflected in the submission received in operation 900. Accordingly, the client interface 210 may determine (e.g., via a lookup table stored by the database 115) a time period that corresponds to the amount of the fee paid, such that the client interface 210 can add the conditional modification program 330 and its trigger condition 332 to the object data 300 during the time period, and then remove the conditional modification program 330 and its trigger condition 332 from the object data 300 after expiration of the time period.


In example embodiments that include operation 902, operation 920 may be performed as part of operation 420, in which the trigger condition detector 220 to detect fulfillment of the trigger condition 332. In operation 920, the trigger condition detector 220 detects the fulfillment of the trigger condition 332 during the period of time that was determined in operation 902 for the corresponding conditional modification program 330.


According to various example embodiments, one or more of the methodologies described herein may facilitate conditional modification of an AR object. Moreover, one or more of the methodologies described herein may facilitate dynamic behavior of the AR object in response to one or more of various user interactions, cumulative user interactions, or both. In addition, one or more of the methodologies described herein are operable to improve the variety of interactive experiences provided by an AR object, including support of user-submitted conditional modification programs and corresponding trigger conditions. Hence, one or more of the methodologies described herein may facilitate provision of crowd-sourced creative experiences with one or more AR objects, as well as regularly refreshed or updated AR content that is more resistant to loss of user interest or user engagement, compared to capabilities of pre-existing systems and methods.


When these effects are considered in aggregate, one or more of the methodologies described herein may obviate a need for certain efforts or resources that otherwise would be involved in conditional modification of an AR object. Efforts expended by a user in configuring sophisticated conditional modifications to an AR object may be reduced by use of (e.g., reliance upon) a special-purpose machine that implements one or more of the methodologies described herein. Computing resources used by one or more systems or machines (e.g., within the network environment 100) may similarly be reduced (e.g., compared to systems or machines that lack the structures discussed herein or are otherwise unable to perform the functions discussed herein). Examples of such computing resources include processor cycles, network traffic, computational capacity, main memory usage, graphics rendering capacity, graphics memory usage, data storage capacity, power consumption, and cooling capacity.



FIG. 10 is a block diagram illustrating components of a machine 1000, according to some example embodiments, able to read instructions 1024 from a machine-readable medium 1022 (e.g., a non-transitory machine-readable medium, a machine-readable storage medium, a computer-readable storage medium, or any suitable combination thereof) and perform any one or more of the methodologies discussed herein, in whole or in part. Specifically, FIG. 10 shows the machine 1000 in the example form of a computer system (e.g., a computer) within which the instructions 1024 (e.g., software, a program, an application, an applet, an app, or other executable code) for causing the machine 1000 to perform any one or more of the methodologies discussed herein may be executed, in whole or in part.


In alternative embodiments, the machine 1000 operates as a standalone device or may be communicatively coupled (e.g., networked) to other machines. In a networked deployment, the machine 1000 may operate in the capacity of a server machine or a client machine in a server-client network environment, or as a peer machine in a distributed (e.g., peer-to-peer) network environment. The machine 1000 may be a server computer, a client computer, a personal computer (PC), a tablet computer, a laptop computer, a netbook, a cellular telephone, a smart phone, a set-top box (STB), a personal digital assistant (PDA), a web appliance, a network router, a network switch, a network bridge, or any machine capable of executing the instructions 1024, sequentially or otherwise, that specify actions to be taken by that machine. Further, while only a single machine is illustrated, the term “machine” shall also be taken to include any collection of machines that individually or jointly execute the instructions 1024 to perform all or part of any one or more of the methodologies discussed herein.


The machine 1000 includes a processor 1002 (e.g., one or more central processing units (CPUs), one or more graphics processing units (GPUs), one or more digital signal processors (DSPs), one or more application specific integrated circuits (ASICs), one or more radio-frequency integrated circuits (RFICs), or any suitable combination thereof), a main memory 1004, and a static memory 1006, which are configured to communicate with each other via a bus 1008. The processor 1002 contains solid-state digital microcircuits (e.g., electronic, optical, or both) that are configurable, temporarily or permanently, by some or all of the instructions 1024 such that the processor 1002 is configurable to perform any one or more of the methodologies described herein, in whole or in part. For example, a set of one or more microcircuits of the processor 1002 may be configurable to execute one or more modules (e.g., software modules) described herein. In some example embodiments, the processor 1002 is a multicore CPU (e.g., a dual-core CPU, a quad-core CPU, an 8-core CPU, or a 128-core CPU) within which each of multiple cores behaves as a separate processor that is able to perform any one or more of the methodologies discussed herein, in whole or in part. Although the beneficial effects described herein may be provided by the machine 1000 with at least the processor 1002, these same beneficial effects may be provided by a different kind of machine that contains no processors (e.g., a purely mechanical system, a purely hydraulic system, or a hybrid mechanical-hydraulic system), if such a processor-less machine is configured to perform one or more of the methodologies described herein.


The machine 1000 may further include a graphics display 1010 (e.g., a plasma display panel (PDP), a light emitting diode (LED) display, a liquid crystal display (LCD), a projector, a cathode ray tube (CRT), or any other display capable of displaying graphics or video). The machine 1000 may also include an alphanumeric input device 1012 (e.g., a keyboard or keypad), a pointer input device 1014 (e.g., a mouse, a touchpad, a touchscreen, a trackball, a joystick, a stylus, a motion sensor, an eye tracking device, a data glove, or other pointing instrument), a data storage 1016, an audio generation device 1018 (e.g., a sound card, an amplifier, a speaker, a headphone jack, or any suitable combination thereof), and a network interface device 1020.


The data storage 1016 (e.g., a data storage device) includes the machine-readable medium 1022 (e.g., a tangible and non-transitory machine-readable storage medium) on which are stored the instructions 1024 embodying any one or more of the methodologies or functions described herein. The instructions 1024 may also reside, completely or at least partially, within the main memory 1004, within the static memory 1006, within the processor 1002 (e.g., within the processor's cache memory), or any suitable combination thereof, before or during execution thereof by the machine 1000. Accordingly, the main memory 1004, the static memory 1006, and the processor 1002 may be considered machine-readable media (e.g., tangible and non-transitory machine-readable media). The instructions 1024 may be transmitted or received over the network 190 via the network interface device 1020. For example, the network interface device 1020 may communicate the instructions 1024 using any one or more transfer protocols (e.g., hypertext transfer protocol (HTTP)).


In some example embodiments, the machine 1000 may be a portable computing device (e.g., a smart phone, a tablet computer, or a wearable device) and may have one or more additional input components 1030 (e.g., sensors or gauges). Examples of such input components 1030 include an image input component (e.g., one or more cameras), an audio input component (e.g., one or more microphones), a direction input component (e.g., a compass), a location input component (e.g., a global positioning system (GPS) receiver), an orientation component (e.g., a gyroscope), a motion detection component (e.g., one or more accelerometers), an altitude detection component (e.g., an altimeter), a temperature input component (e.g., a thermometer), and a gas detection component (e.g., a gas sensor). Input data gathered by any one or more of these input components 1030 may be accessible and available for use by any of the modules described herein (e.g., with suitable privacy notifications and protections, such as opt-in consent or opt-out consent, implemented in accordance with user preference, applicable regulations, or any suitable combination thereof).


As used herein, the term “memory” refers to a machine-readable medium able to store data temporarily or permanently and may be taken to include, but not be limited to, random-access memory (RAM), read-only memory (ROM), buffer memory, flash memory, and cache memory. While the machine-readable medium 1022 is shown in an example embodiment to be a single medium, the term “machine-readable medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, or associated caches and servers) able to store instructions. The term “machine-readable medium” shall also be taken to include any medium, or combination of multiple media, that is capable of carrying (e.g., storing or communicating) the instructions 1024 for execution by the machine 1000, such that the instructions 1024, when executed by one or more processors of the machine 1000 (e.g., processor 1002), cause the machine 1000 to perform any one or more of the methodologies described herein, in whole or in part. Accordingly, a “machine-readable medium” refers to a single storage apparatus or device, as well as cloud-based storage systems or storage networks that include multiple storage apparatus or devices. The term “machine-readable medium” shall accordingly be taken to include, but not be limited to, one or more tangible and non-transitory data repositories (e.g., data volumes) in the example form of a solid-state memory chip, an optical disc, a magnetic disc, or any suitable combination thereof.


A “non-transitory” machine-readable medium, as used herein, specifically excludes propagating signals per se. According to various example embodiments, the instructions 1024 for execution by the machine 1000 can be communicated via a carrier medium (e.g., a machine-readable carrier medium). Examples of such a carrier medium include a non-transient carrier medium (e.g., a non-transitory machine-readable storage medium, such as a solid-state memory that is physically movable from one place to another place) and a transient carrier medium (e.g., a carrier wave or other propagating signal that communicates the instructions 1024).


Certain example embodiments are described herein as including modules. Modules may constitute software modules (e.g., code stored or otherwise embodied in a machine-readable medium or in a transmission medium), hardware modules, or any suitable combination thereof. A “hardware module” is a tangible (e.g., non-transitory) physical component (e.g., a set of one or more processors) capable of performing certain operations and may be configured or arranged in a certain physical manner. In various example embodiments, one or more computer systems or one or more hardware modules thereof may be configured by software (e.g., an application or portion thereof) as a hardware module that operates to perform operations described herein for that module.


In some example embodiments, a hardware module may be implemented mechanically, electronically, hydraulically, or any suitable combination thereof. For example, a hardware module may include dedicated circuitry or logic that is permanently configured to perform certain operations. A hardware module may be or include a special-purpose processor, such as a field programmable gate array (FPGA) or an ASIC. A hardware module may also include programmable logic or circuitry that is temporarily configured by software to perform certain operations. As an example, a hardware module may include software encompassed within a CPU or other programmable processor. It will be appreciated that the decision to implement a hardware module mechanically, hydraulically, in dedicated and permanently configured circuitry, or in temporarily configured circuitry (e.g., configured by software) may be driven by cost and time considerations.


Accordingly, the phrase “hardware module” should be understood to encompass a tangible entity that may be physically constructed, permanently configured (e.g., hardwired), or temporarily configured (e.g., programmed) to operate in a certain manner or to perform certain operations described herein. Furthermore, as used herein, the phrase “hardware-implemented module” refers to a hardware module. Considering example embodiments in which hardware modules are temporarily configured (e.g., programmed), each of the hardware modules need not be configured or instantiated at any one instance in time. For example, where a hardware module includes a CPU configured by software to become a special-purpose processor, the CPU may be configured as respectively different special-purpose processors (e.g., each included in a different hardware module) at different times. Software (e.g., a software module) may accordingly configure one or more processors, for example, to become or otherwise constitute a particular hardware module at one instance of time and to become or otherwise constitute a different hardware module at a different instance of time.


Hardware modules can provide information to, and receive information from, other hardware modules. Accordingly, the described hardware modules may be regarded as being communicatively coupled. Where multiple hardware modules exist contemporaneously, communications may be achieved through signal transmission (e.g., over circuits and buses) between or among two or more of the hardware modules. In embodiments in which multiple hardware modules are configured or instantiated at different times, communications between such hardware modules may be achieved, for example, through the storage and retrieval of information in memory structures to which the multiple hardware modules have access. For example, one hardware module may perform an operation and store the output of that operation in a memory (e.g., a memory device) to which it is communicatively coupled. A further hardware module may then, at a later time, access the memory to retrieve and process the stored output. Hardware modules may also initiate communications with input or output devices, and can operate on a resource (e.g., a collection of information from a computing resource).


The various operations of example methods described herein may be performed, at least partially, by one or more processors that are temporarily configured (e.g., by software) or permanently configured to perform the relevant operations. Whether temporarily or permanently configured, such processors may constitute processor-implemented modules that operate to perform one or more operations or functions described herein. As used herein, “processor-implemented module” refers to a hardware module in which the hardware includes one or more processors. Accordingly, the operations described herein may be at least partially processor-implemented, hardware-implemented, or both, since a processor is an example of hardware, and at least some operations within any one or more of the methods discussed herein may be performed by one or more processor-implemented modules, hardware-implemented modules, or any suitable combination thereof.


Moreover, such one or more processors may perform operations in a “cloud computing” environment or as a service (e.g., within a “software as a service” (SaaS) implementation). For example, at least some operations within any one or more of the methods discussed herein may be performed by a group of computers (e.g., as examples of machines that include processors), with these operations being accessible via a network (e.g., the Internet) and via one or more appropriate interfaces (e.g., an application program interface (API)). The performance of certain operations may be distributed among the one or more processors, whether residing only within a single machine or deployed across a number of machines. In some example embodiments, the one or more processors or hardware modules (e.g., processor-implemented modules) may be located in a single geographic location (e.g., within a home environment, an office environment, or a server farm). In other example embodiments, the one or more processors or hardware modules may be distributed across a number of geographic locations.



FIG. 11 and FIG. 12 are diagrams illustrating screenshots of an AR object 1112 before and after conditional modification of the AR object 1112, according to some example embodiments. As shown in FIG. 11, the device 130 is being used by the user 132 in a real-world scene 1100 and a camera of the device 130 is capturing a view of the real world scene 1100. In addition, a display screen of the device 130 depicts an AR scene 1110 that combines the real world scene 1100 with the AR object 1112. Specifically, the AR object 1112 is depicted with a pre-modification (e.g., first) appearance (e.g., a miniature hippopotamus), in accordance with its defining object data (e.g., object data 300) prior to the modification process described above.


As shown in FIG. 12, the display screen of the device 130 continues to detect the AR scene 1110, and the AR scene 1110 continues to combine the real world scene 1100 with the AR object 1112. However, the AR object 1112 is now depicted with a post-modification (e.g., second) appearance (e.g., a miniature hippopotamus with elephant ears and elephant tusks) in accordance with its defining object data (e.g., object data 300) after the modification process described above.


Throughout this specification, plural instances may implement components, operations, or structures described as a single instance. Although individual operations of one or more methods are illustrated and described as separate operations, one or more of the individual operations may be performed concurrently, and nothing requires that the operations be performed in the order illustrated. Structures and their functionality presented as separate components and functions in example configurations may be implemented as a combined structure or component with combined functions. Similarly, structures and functionality presented as a single component may be implemented as separate components and functions. These and other variations, modifications, additions, and improvements fall within the scope of the subject matter herein.


Some portions of the subject matter discussed herein may be presented in terms of algorithms or symbolic representations of operations on data stored as bits or binary digital signals within a memory (e.g., a computer memory or other machine memory). Such algorithms or symbolic representations are examples of techniques used by those of ordinary skill in the data processing arts to convey the substance of their work to others skilled in the art. As used herein, an “algorithm” is a self-consistent sequence of operations or similar processing leading to a desired result. In this context, algorithms and operations involve physical manipulation of physical quantities. Typically, but not necessarily, such quantities may take the form of electrical, magnetic, or optical signals capable of being stored, accessed, transferred, combined, compared, or otherwise manipulated by a machine. It is convenient at times, principally for reasons of common usage, to refer to such signals using words such as “data,” “content,” “bits,” “values,” “elements,” “symbols,” “characters,” “terms,” “numbers,” “numerals,” or the like. These words, however, are merely convenient labels and are to be associated with appropriate physical quantities.


Unless specifically stated otherwise, discussions herein using words such as “accessing,” “processing,” “detecting,” “computing,” “calculating,” “determining,” “generating,” “presenting,” “displaying,” or the like refer to actions or processes performable by a machine (e.g., a computer) that manipulates or transforms data represented as physical (e.g., electronic, magnetic, or optical) quantities within one or more memories (e.g., volatile memory, non-volatile memory, or any suitable combination thereof), registers, or other machine components that receive, store, transmit, or display information. Furthermore, unless specifically stated otherwise, the terms “a” or “an” are herein used, as is common in patent documents, to include one or more than one instance. Finally, as used herein, the conjunction “or” refers to a non-exclusive “or,” unless specifically stated otherwise.


The following enumerated descriptions describe various examples of methods, machine-readable media, and systems (e.g., machines, devices, or other apparatus) discussed herein.


A first example provides a method comprising:


providing, to a user device and by one or more processors of a machine, object data that defines an augmented reality (AR) object, the object data of the AR object specifying a physical geolocation of the AR object, a presentation attribute of the AR object, a conditional modification program of the AR object, and a trigger condition for execution of the conditional modification program, the provided object data causing the user device to present the AR object with a first appearance and located at the physical geolocation;


detecting, by one or more processors of the machine, fulfillment of the trigger condition specified by the object data for execution of the conditional modification program of the AR object,


based on the detected fulfillment of the trigger condition and by one or more processors of the machine, executing the conditional modification program of the AR object, the executing of the conditional modification program modifying the object data of the AR object by modifying the presentation attribute of the AR object; and


providing, to the user device and by one or more processors of the machine, the modified object data of the AR object, the modified object data causing the user device to present the AR object with a second appearance based on the modified presentation attribute of the AR object.


A second example provides a method according to the first example, wherein:


the presentation attribute of the AR object specifies a physical dimension of the AR object; and


the executing of the conditional modification program modifies the physical dimension of the AR object, the second appearance of the AR object differing in size from the first appearance of the AR object.


A third example provides a method according to the first example or the second example, wherein:


the presentation attribute of the AR object specifies a color of at least a portion of the AR object; and


the executing of the conditional modification program modifies a color of the portion of the AR object, the second appearance of the AR object at least partially differing in color from the first appearance of the AR object.


A fourth example provides a method according to any of the first through third examples, wherein:


the presentation attribute of the AR object specifies the physical geolocation of the AR object; and


the executing of the conditional modification program modifies the physical geolocation of the AR object, the second appearance of the AR object differing in physical geolocation from the first appearance of the AR object.


A fifth example provides a method according to any of the first through fourth examples, further comprising:


receiving an indication that the user device submitted an edit to the AR object; and


in response to the received indication, incrementing an interaction count that corresponds to the AR object; and wherein:


the detecting of the fulfillment of the trigger condition is based on a comparison of the incremented interaction count to a threshold interaction count.


A sixth example provides a method according to any of the first through fifth examples, further comprising:


receiving an indication that the user device edited the AR object for a period of time; and


in response to the received indication, updating a total interaction time that corresponds to the AR object based on the period of time; and wherein:


the detecting of the fulfillment of the trigger condition is based on a comparison of the updated total interaction time to a threshold total interaction time.


A seventh example provides a method according to any of the first through sixth examples, further comprising:


receiving an indication that a device geolocation of the user device intersected the AR object; and


in response to the received indication, incrementing an interaction count that corresponds to the AR object; and wherein:


the detecting of the fulfillment of the trigger condition is based on a comparison of the incremented interaction count to a threshold interaction count.


An eighth example provides a method according to any of the first through seventh examples, further comprising:


receiving an indication that the user device edited the AR object and corresponds to a user profile that specifies a demographic descriptor of a user of the user device; and


in response to the received indication, incrementing an interaction count that corresponds to the AR object and corresponds to the demographic descriptor; and wherein:


the detecting of the fulfillment of the trigger condition is based on a comparison of the incremented interaction count to a threshold interaction count.


A ninth example provides a method according to any of the first through eighth examples, further comprising:


receiving an indication that the user device edited the AR object; and


in response to the received indication, incrementing an interaction count that corresponds to the AR object; and


generating a geographical map that includes a depiction of the edited AR object, the depiction having a color that indicates a frequency at which the AR object is edited.


A tenth example provides a method according to the ninth example, wherein:


the user device is a first user device; and


the method further comprises:


providing the generated geographical map to a second user device, the provided geographical map including the depiction whose color indicates the frequency at which the AR object is edited.


An eleventh example provides a method according to any of the first through tenth examples, further comprising:


receiving a user submission that includes the conditional modification program, the trigger condition for execution of the conditional modification program, and a corresponding value accorded to the conditional modification program; and


based on the value accorded to the conditional modification program, determining a period of time during which the conditional modification program is to be specified by the object data that defines the AR object; and wherein:


the detecting of the fulfillment of the trigger condition occurs within the determined period of time.


A twelfth example provides a method according to any of the first through eleventh examples, wherein:


the trigger condition for executing the conditional modification program specifies a threshold level of ambient audio loudness; and


the detecting of the fulfillment of the trigger condition includes receiving an indication that a microphone of the user device detected an actual level of ambient audio loudness that transgresses the threshold level of ambient audio loudness.


A thirteenth example provides a method according to any of the first through twelfth examples, wherein:


the trigger condition for executing the conditional modification program specifies a reference pattern of sounds; and


the detecting of the fulfillment of the trigger condition includes receiving an indication that a microphone of the user device captured the reference pattern of sounds.


A fourteenth example provides a method according to any of the first through thirteenth examples, further comprising:


receiving an indication that the user device is editing the AR object; and wherein:


the detecting of the fulfillment of the trigger condition is based on the received indication that the user device is editing the AR object; and


the executing of the conditional modification program adds an AR cursor to the second appearance of AR object, the added AR cursor indicating a location of a non-committed edit to the AR object.


A fifteenth example provides a machine-readable medium (e.g., a non-transitory machine-readable storage medium) comprising instructions that, when executed by one or more processors of a machine, cause the machine to perform operations comprising:


providing, to a user device, object data that defines an augmented reality (AR) object, the object data of the AR object specifying a physical geolocation of the AR object, a presentation attribute of the AR object, a conditional modification program of the AR object, and a trigger condition for execution of the conditional modification program, the provided object data causing the user device to present the AR object with a first appearance and located at the physical geolocation;


detecting, by one or more processors of the machine, fulfillment of the trigger condition specified by the object data for execution of the conditional modification program of the AR object;


based on the detected fulfillment of the trigger condition, executing the conditional modification program of the AR object, the executing of the conditional modification program modifying the object data of the AR object by modifying the presentation attribute of the AR object; and


providing, to the user device, the modified object data of the AR object, the modified object data causing the user device to present the AR object with a second appearance based on the modified presentation attribute of the AR object.


A sixteenth example provides a machine-readable medium according to the fifteenth example, wherein the operations further comprise:


receiving an indication that the user device submitted an edit to the AR object; and


in response to the received indication, incrementing an interaction count that corresponds to the AR object; and wherein:


the detecting of the fulfillment of the trigger condition is based on a comparison of the incremented interaction count to a threshold interaction count.


A seventeenth example provides a machine-readable medium according to the fifteenth example or the sixteenth example, wherein the operations further comprise:


receiving an indication that a device geolocation of the user device intersected the AR object; and


in response to the received indication, incrementing an interaction count that corresponds to the AR object; and wherein:


the detecting of the fulfillment of the trigger condition is based on a comparison of the incremented interaction count to a threshold interaction count.


An eighteenth example provides a system comprising:


one or more processors; and


a memory storing instructions that, when executed by at least one processor among the one or more processors, cause the system to perform operations comprising:


providing, to a user device, object data that defines an augmented reality (AR) object, the object data of the AR object specifying a physical geolocation of the AR object, a presentation attribute of the AR object, a conditional modification program of the AR object, and a trigger condition for execution of the conditional modification program, the provided object data causing the user device to present the AR object with a first appearance and located at the physical geolocation;


detecting, by one or more processors of the machine, fulfillment of the trigger condition specified by the object data for execution of the conditional modification program of the AR object;


based on the detected fulfillment of the trigger condition, executing the conditional modification program of the AR object, the executing of the conditional modification program modifying the object data of the AR object by modifying the presentation attribute of the AR object; and


providing, to the user device, the modified object data of the AR object, the modified object data causing the user device to present the AR object with a second appearance based on the modified presentation attribute of the AR object.


A nineteenth example provides a system according to the eighteenth example, wherein the operations further comprise:


receiving an indication that the user device edited the AR object for a period of time; and


in response to the received indication, updating a total interaction time that corresponds to the AR object based on the period of time; and wherein:


the detecting of the fulfillment of the trigger condition is based on a comparison of the updated total interaction time to a threshold total interaction time.


A twentieth example provides a system according to the eighteenth example or the nineteenth example, wherein the operations further comprise:


receiving an indication that the user device edited the AR object; and


in response to the received indication, incrementing an interaction count that corresponds to the AR object; and


generating a geographical map that includes a depiction of the edited AR object, the depiction having a color that indicates a frequency at which the AR object is edited.


A twenty-first example provides a carrier medium carrying machine-readable instructions for controlling a machine to carry out the operations (e.g., method operations) performed in any one of the previously described examples.

Claims
  • 1. A method comprising: providing, to a user device and by one or more processors, object data that defines an augmented reality (AR) object, the object data specifying a physical geolocation of the AR object, a presentation attribute of the AR object, a conditional modification program of the AR object, and a trigger condition for execution of the conditional modification program, the provided object data causing the user device to present the AR object with a first appearance and located at the physical geolocation;receiving, by the one or more processors, an indication that the AR object is being edited via the user device;detecting, by the one or more processors, fulfillment of the trigger condition specified by the object data, the detecting being based on the received indication that the user device is editing the AR object;based on the detected fulfillment of the trigger condition and by the one or more processors, executing the conditional modification program, the executed conditional modification program modifying the object data of the AR object by modifying the presentation attribute of the AR object, the modified presentation attribute adding an AR indicator to the second appearance of AR object, the added AR indicator indicating a location of a non-committed edit to the AR object; andproviding, to the user device and by the one or more processors, the modified object data of the AR object, the modified object data causing the user device to present the AR object with a second appearance based on the modified presentation attribute.
  • 2. The method of claim 1, wherein: the presentation attribute of the AR object specifies a physical dimension of the AR object; andthe executing of the conditional modification program modifies the physical dimension of the AR object, the second appearance of the AR object differing in size from the first appearance of the AR object.
  • 3. The method of claim 1, wherein: the presentation attribute of the AR object specifies a color of at least a portion of the AR object; andthe executing of the conditional modification program modifies the color of the portion of the AR object, the second appearance of the AR object at least partially differing in color from the first appearance of the AR object.
  • 4. The method of claim 1, wherein: the presentation attribute of the AR object specifies the physical geolocation of the AR object; andthe executing of the conditional modification program modifies the physical geolocation of the AR object, the second appearance of the AR object differing in physical geolocation from the first appearance of the AR object.
  • 5. The method of claim 1, further comprising: receiving an indication that the user device submitted an edit to the AR object; andin response to the received indication, incrementing an interaction count that corresponds to the AR object; and wherein:the detecting of the fulfillment of the trigger condition is based on a comparison of the incremented interaction count to a threshold interaction count.
  • 6. The method of claim 1, further comprising: receiving an indication that the user device edited the AR object for a period of time; andin response to the received indication, updating a total interaction time that corresponds to the AR object based on the period of time; and wherein:the detecting of the fulfillment of the trigger condition is based on a comparison of the updated total interaction time to a threshold total interaction time.
  • 7. The method of claim 1, further comprising: receiving an indication that a device geolocation of the user device intersected the AR object; andin response to the received indication, incrementing an interaction count that corresponds to the AR object; and wherein:the detecting of the fulfillment of the trigger condition is based on a comparison of the incremented interaction count to a threshold interaction count.
  • 8. The method of claim 1, further comprising: receiving an indication that the user device edited the AR object and corresponds to a user profile that specifies a demographic descriptor of a user of the user device; andin response to the received indication, incrementing an interaction count that corresponds to the AR object and corresponds to the demographic descriptor; and wherein:the detecting of the fulfillment of the trigger condition is based on a comparison of the incremented interaction count to a threshold interaction count.
  • 9. The method of claim 1, further comprising: receiving an indication that the user device edited the AR object;in response to the received indication, incrementing an interaction count that corresponds to the AR object; andgenerating a geographical map that includes a depiction of the edited AR object, the depiction having a color that indicates a frequency at which the AR object is edited.
  • 10. The method of claim 9, wherein: the user device is a first user device; and
  • 11. The method of claim 1, further comprising: receiving a user submission that includes the conditional modification program, the trigger condition for execution of the conditional modification program, and a corresponding value accorded to the conditional modification program; andbased on the value accorded to the conditional modification program, determining a period of time during which the conditional modification program is to be specified by the object data that defines the AR object, and wherein:the detecting of the fulfillment of the trigger condition occurs within the determined period of time.
  • 12. The method of claim 1, wherein: the trigger condition for executing the conditional modification program specifies a threshold level of ambient audio loudness; andthe detecting of the fulfillment of the trigger condition includes receiving an indication that a microphone of the user device detected an actual level of ambient audio loudness that transgresses the threshold level of ambient audio loudness.
  • 13. The method of claim 1, wherein: the trigger condition for executing the conditional modification program specifies a reference pattern of sounds; andthe detecting of the fulfillment of the trigger condition includes receiving an indication that a microphone of the user device captured the reference pattern of sounds.
  • 14. A non-transitory machine-readable storage medium comprising instructions that, when executed by one or more processors of a machine, cause the machine to perform operations comprising: providing, to a user device, object data that defines an augmented reality (AR) object, the object data specifying a physical geolocation of the AR object, a presentation attribute of the AR object, a conditional modification program of the AR object, and a trigger condition for execution of the conditional modification program, the provided object data causing the user device to present the AR object with a first appearance and located at the physical geolocation;receiving an indication that the AR object is being edited via the user device;detecting fulfillment of the trigger condition specified by the object data, the detecting being based on the received indication that the user device is editing the AR object;based on the detected fulfillment of the trigger condition, executing the conditional modification program, the executed conditional modification program modifying the object data of the AR object by modifying the presentation attribute of the AR object, the modified presentation attribute adding an AR indicator to the second appearance of AR object, the added AR indicator indicating a location of a non-committed edit to the AR object; andproviding, to the user device, the modified object data of the AR object, the modified object data causing the user device to present the AR object with a second appearance based on the modified presentation attribute.
  • 15. The non-transitory machine-readable storage medium of claim 14, wherein the operations further comprise: receiving an indication that the user device submitted an edit to the AR object; andin response to the received indication, incrementing an interaction count that corresponds to the AR object; and wherein:the detecting of the fulfillment of the trigger condition is based on a comparison of the incremented interaction count to a threshold interaction count.
  • 16. The non-transitory machine-readable storage medium of claim 14, wherein: receiving an indication that a device geolocation of the user device intersected the AR object; andin response to the received indication, incrementing an interaction count that corresponds to the AR object; and wherein:the detecting of the fulfillment of the trigger condition is based on a comparison of the incremented interaction count to a threshold interaction count.
  • 17. The non-transitory machine-readable storage medium of claim 14, wherein the operations further comprise: receiving a user submission that includes the conditional modification program, the trigger condition for execution of the conditional modification program, and a corresponding value accorded to the conditional modification program; andbased on the value accorded to the conditional modification program, determining a period of time during which the conditional modification program is to be specified by the object data that defines the AR object, and wherein:the detecting of the fulfillment of the trigger condition occurs within the determined period of time.
  • 18. A system comprising: one or more processors; anda memory storing instructions that, when executed by at least one processor among the one or more processors, cause the system to perform operations comprising:providing, to a user device, object data that defines an augmented reality (AR) object, the object data specifying a physical geolocation of the AR object, a presentation attribute of the AR object, a conditional modification program of the AR object, and a trigger condition for execution of the conditional modification program, the provided object data causing the user device to present the AR object with a first appearance and located at the physical geolocation;receiving an indication that the AR object is being edited via the user device;detecting fulfillment of the trigger condition specified by the object data, the detecting being based on the received indication that the user device is editing the AR object;based on the detected fulfillment of the trigger condition, executing the conditional modification program, the executed conditional modification program modifying the object data of the AR object by modifying the presentation attribute of the AR object, the modified presentation attribute adding an AR indicator to the second appearance of AR object, the added AR indicator indicating a location of a non-committed edit to the AR object; andproviding, to the user device, the modified object data of the AR object, the modified object data causing the user device to present the AR object with a second appearance based on the modified presentation attribute.
  • 19. The system of claim 18, wherein the operations further comprise: receiving an indication that the user device edited the AR object for a period of time; andin response to the received indication, updating a total interaction time that corresponds to the AR object based on the period of time; and wherein:the detecting of the fulfillment of the trigger condition is based on a comparison of the updated total interaction time to a threshold total interaction time.
  • 20. The system of claim 18, wherein the operations further comprise: receiving an indication that the user device edited the AR object;in response to the received indication, incrementing an interaction count that corresponds to the AR object; andgenerating a geographical map that includes a depiction of the edited AR object, the depiction having a color that indicates a frequency at which the AR object is edited.
CLAIM OF PRIORITY

This application is a continuation of U.S. patent application Ser. No. 16/043,995, filed on Jul. 24, 2018, which is incorporated herein by reference in its entirety.

US Referenced Citations (621)
Number Name Date Kind
666223 Shedlock Jan 1901 A
4581634 Williams Apr 1986 A
4975690 Torres Dec 1990 A
5072412 Henderson, Jr. et al. Dec 1991 A
5493692 Theimer et al. Feb 1996 A
5713073 Warsta Jan 1998 A
5754939 Herz et al. May 1998 A
5855008 Goldhaber et al. Dec 1998 A
5883639 Walton et al. Mar 1999 A
5999932 Paul Dec 1999 A
6012098 Bayeh et al. Jan 2000 A
6014090 Rosen et al. Jan 2000 A
6029141 Bezos et al. Feb 2000 A
6038295 Mattes Mar 2000 A
6049711 Ben-Yehezkel et al. Apr 2000 A
6154764 Nitta et al. Nov 2000 A
6167435 Druckenmiller et al. Dec 2000 A
6204840 Petelycky et al. Mar 2001 B1
6205432 Gabbard et al. Mar 2001 B1
6216141 Straub et al. Apr 2001 B1
6285381 Sawano et al. Sep 2001 B1
6285987 Roth et al. Sep 2001 B1
6310694 Okimoto et al. Oct 2001 B1
6317789 Rakavy et al. Nov 2001 B1
6334149 Davis, Jr. et al. Dec 2001 B1
6349203 Asaoka et al. Feb 2002 B1
6353170 Eyzaguirre et al. Mar 2002 B1
6446004 Cao et al. Sep 2002 B1
6449657 Stanbach, et al. Sep 2002 B2
6456852 Bar et al. Sep 2002 B2
6484196 Maurille Nov 2002 B1
6487601 Hubacher et al. Nov 2002 B1
6523008 Avrunin Feb 2003 B1
6542749 Tanaka et al. Apr 2003 B2
6549768 Fraccaroli Apr 2003 B1
6618593 Drutman et al. Sep 2003 B1
6622174 Ukita et al. Sep 2003 B1
6631463 Floyd et al. Oct 2003 B1
6636247 Hamzy et al. Oct 2003 B1
6636855 Holloway et al. Oct 2003 B2
6643684 Malkin et al. Nov 2003 B1
6658095 Yoakum et al. Dec 2003 B1
6665531 Soderbacka et al. Dec 2003 B1
6668173 Greene Dec 2003 B2
6684238 Dutta Jan 2004 B1
6684257 Camut et al. Jan 2004 B1
6698020 Zigmond et al. Feb 2004 B1
6700506 Winkler Mar 2004 B1
6720860 Narayanaswami Apr 2004 B1
6724403 Santoro et al. Apr 2004 B1
6757713 Ogilvie et al. Jun 2004 B1
6832222 Zimowski Dec 2004 B1
6834195 Brandenberg et al. Dec 2004 B2
6836792 Chen Dec 2004 B1
6898626 Ohashi May 2005 B2
6959324 Kubik et al. Oct 2005 B1
6970088 Kovach Nov 2005 B2
6970907 Ullmann et al. Nov 2005 B1
6980909 Root et al. Dec 2005 B2
6981040 Konig et al. Dec 2005 B1
7020494 Spriestersbach et al. Mar 2006 B2
7027124 Foote et al. Apr 2006 B2
7072963 Anderson et al. Jul 2006 B2
7085571 Kalhan et al. Aug 2006 B2
7110744 Freeny, Jr. Sep 2006 B2
7124164 Chemtob Oct 2006 B1
7149893 Leonard et al. Dec 2006 B1
7173651 Knowles Feb 2007 B1
7188143 Szeto Mar 2007 B2
7203380 Chiu et al. Apr 2007 B2
7206568 Sudit Apr 2007 B2
7227937 Yoakum et al. Jun 2007 B1
7237002 Estrada et al. Jun 2007 B1
7240089 Boudreau Jul 2007 B2
7269426 Kokkonen et al. Sep 2007 B2
7280658 Amini et al. Oct 2007 B2
7315823 Brondrup Jan 2008 B2
7349768 Bruce et al. Mar 2008 B2
7356564 Hartselle et al. Apr 2008 B2
7394345 Ehlinger et al. Jul 2008 B1
7411493 Smith Aug 2008 B2
7423580 Markhovsky et al. Sep 2008 B2
7454442 Cobleigh et al. Nov 2008 B2
7508419 Toyama et al. Mar 2009 B2
7512649 Faybishenko et al. Mar 2009 B2
7519670 Hagale et al. Apr 2009 B2
7535890 Rojas May 2009 B2
7546554 Chiu et al. Jun 2009 B2
7607096 Oreizy et al. Oct 2009 B2
7639943 Kalajan Dec 2009 B1
7650231 Gadler Jan 2010 B2
7668537 DeVries Feb 2010 B2
7770137 Forbes et al. Aug 2010 B2
7778973 Choi Aug 2010 B2
7779444 Glad Aug 2010 B2
7787886 Markhovsky et al. Aug 2010 B2
7796946 Eisenbach Sep 2010 B2
7801954 Cadiz et al. Sep 2010 B2
7856360 Kramer et al. Dec 2010 B2
8001204 Burtner et al. Aug 2011 B2
8032586 Challenger et al. Oct 2011 B2
8082255 Carlson, Jr. et al. Dec 2011 B1
8090351 Klein Jan 2012 B2
8098904 Ioffe et al. Jan 2012 B2
8099109 Altman et al. Jan 2012 B2
8112716 Kobayashi Feb 2012 B2
8131597 Hudetz et al. Mar 2012 B2
8135166 Rhoads Mar 2012 B2
8136028 Loeb et al. Mar 2012 B1
8146001 Reese Mar 2012 B1
8161115 Yamamoto Apr 2012 B2
8161417 Lee Apr 2012 B1
8195203 Tseng Jun 2012 B1
8199747 Rojas et al. Jun 2012 B2
8208943 Petersen Jun 2012 B2
8214443 Hamburg Jul 2012 B2
8234350 Gu et al. Jul 2012 B1
8276092 Narayanan et al. Sep 2012 B1
8279319 Date Oct 2012 B2
8280406 Ziskind et al. Oct 2012 B2
8285199 Hsu et al. Oct 2012 B2
8287380 Nguyen et al. Oct 2012 B2
8301159 Hamynen et al. Oct 2012 B2
8306922 Kunal et al. Nov 2012 B1
8312086 Velusamy et al. Nov 2012 B2
8312097 Siegel et al. Nov 2012 B1
8326315 Phillips et al. Dec 2012 B2
8326327 Hymel et al. Dec 2012 B2
8332475 Rosen et al. Dec 2012 B2
8352546 Dollard Jan 2013 B1
8379130 Forutanpour et al. Feb 2013 B2
8385950 Wagner et al. Feb 2013 B1
8402097 Szeto Mar 2013 B2
8405773 Hayashi et al. Mar 2013 B2
8418067 Cheng et al. Apr 2013 B2
8423409 Rao Apr 2013 B2
8471914 Sakiyama et al. Jun 2013 B2
8472935 Fujisaki Jun 2013 B1
8510383 Hurley et al. Aug 2013 B2
8527345 Rothschild et al. Sep 2013 B2
8554627 Svendsen et al. Oct 2013 B2
8560612 Kilmer et al. Oct 2013 B2
8594680 Ledlie et al. Nov 2013 B2
8613089 Holloway et al. Dec 2013 B1
8660358 Bergboer et al. Feb 2014 B1
8660369 Llano et al. Feb 2014 B2
8660793 Ngo et al. Feb 2014 B2
8682350 Altman et al. Mar 2014 B2
8718333 Wolf et al. May 2014 B2
8724622 Rojas May 2014 B2
8732168 Johnson May 2014 B2
8744523 Fan et al. Jun 2014 B2
8745132 Obradovich Jun 2014 B2
8761800 Kuwahara Jun 2014 B2
8768876 Shim et al. Jul 2014 B2
8775972 Spiegel Jul 2014 B2
8788680 Naik Jul 2014 B1
8790187 Walker et al. Jul 2014 B2
8797415 Arnold Aug 2014 B2
8798646 Wang et al. Aug 2014 B1
8856349 Jain et al. Oct 2014 B2
8874677 Rosen et al. Oct 2014 B2
8886227 Schmidt et al. Nov 2014 B2
8909679 Root et al. Dec 2014 B2
8909725 Sehn Dec 2014 B1
8972357 Shim et al. Mar 2015 B2
8995433 Rojas Mar 2015 B2
9015285 Ebsen et al. Apr 2015 B1
9020745 Johnston et al. Apr 2015 B2
9040574 Wang et al. May 2015 B2
9055416 Rosen et al. Jun 2015 B2
9094137 Sehn et al. Jul 2015 B1
9100806 Rosen et al. Aug 2015 B2
9100807 Rosen et al. Aug 2015 B2
9113301 Spiegel et al. Aug 2015 B1
9119027 Sharon et al. Aug 2015 B2
9123074 Jacobs et al. Sep 2015 B2
9143382 Bhogal et al. Sep 2015 B2
9143681 Ebsen et al. Sep 2015 B1
9152477 Campbell et al. Oct 2015 B1
9191776 Root et al. Nov 2015 B2
9204252 Root et al. Dec 2015 B2
9225897 Sehn et al. Dec 2015 B1
9258459 Hartley Feb 2016 B2
9344606 Hartley et al. May 2016 B2
9385983 Sehn Jul 2016 B1
9396354 Murphy et al. Jul 2016 B1
9407712 Sehn Aug 2016 B1
9407816 Sehn Aug 2016 B1
9430783 Sehn Aug 2016 B1
9439041 Parvizi et al. Sep 2016 B2
9443227 Evans et al. Sep 2016 B2
9450907 Pridmore et al. Sep 2016 B2
9459778 Hogeg et al. Oct 2016 B2
9489661 Evans et al. Nov 2016 B2
9491134 Rosen et al. Nov 2016 B2
9532171 Allen et al. Dec 2016 B2
9537811 Allen et al. Jan 2017 B2
9628950 Noeth et al. Apr 2017 B1
9710821 Heath Jul 2017 B2
9854219 Sehn Dec 2017 B2
10318569 Funk et al. Jun 2019 B1
10679393 Canberk et al. Jun 2020 B2
20020047868 Miyazawa Apr 2002 A1
20020078456 Hudson et al. Jun 2002 A1
20020087631 Sharma Jul 2002 A1
20020097257 Miller et al. Jul 2002 A1
20020122659 Mcgrath et al. Sep 2002 A1
20020128047 Gates Sep 2002 A1
20020144154 Tomkow Oct 2002 A1
20030001846 Davis et al. Jan 2003 A1
20030016247 Lai et al. Jan 2003 A1
20030017823 Mager et al. Jan 2003 A1
20030020623 Cao et al. Jan 2003 A1
20030023874 Prokupets et al. Jan 2003 A1
20030037124 Yamaura et al. Feb 2003 A1
20030052925 Daimon et al. Mar 2003 A1
20030101230 Benschoter et al. May 2003 A1
20030110503 Perkes Jun 2003 A1
20030126215 Udell Jul 2003 A1
20030148773 Spriestersbach et al. Aug 2003 A1
20030164856 Prager et al. Sep 2003 A1
20030229607 Zellweger et al. Dec 2003 A1
20040027371 Jaeger Feb 2004 A1
20040064429 Hirstius et al. Apr 2004 A1
20040078367 Anderson et al. Apr 2004 A1
20040111467 Willis Jun 2004 A1
20040158739 Wakai et al. Aug 2004 A1
20040189465 Capobianco et al. Sep 2004 A1
20040203959 Coombes Oct 2004 A1
20040215625 Svendsen et al. Oct 2004 A1
20040243531 Dean Dec 2004 A1
20040243688 Wugofski Dec 2004 A1
20050021444 Bauer et al. Jan 2005 A1
20050022211 Veselov et al. Jan 2005 A1
20050048989 Jung Mar 2005 A1
20050056997 Wakitani et al. Mar 2005 A1
20050078804 Yomoda Apr 2005 A1
20050097176 Schatz et al. May 2005 A1
20050102381 Jiang et al. May 2005 A1
20050104976 Currans May 2005 A1
20050114783 Szeto May 2005 A1
20050119936 Buchanan et al. Jun 2005 A1
20050122405 Voss et al. Jun 2005 A1
20050193340 Amburgey et al. Sep 2005 A1
20050193345 Klassen et al. Sep 2005 A1
20050198128 Anderson Sep 2005 A1
20050223066 Buchheit et al. Oct 2005 A1
20050288954 McCarthy et al. Dec 2005 A1
20060026067 Nicholas et al. Feb 2006 A1
20060107297 Toyama et al. May 2006 A1
20060114338 Rothschild Jun 2006 A1
20060119882 Harris et al. Jun 2006 A1
20060242239 Morishima et al. Oct 2006 A1
20060252438 Ansamaa et al. Nov 2006 A1
20060265417 Amato et al. Nov 2006 A1
20060270419 Crowley et al. Nov 2006 A1
20060287878 Wadhwa et al. Dec 2006 A1
20070004426 Pfleging et al. Jan 2007 A1
20070038715 Collins et al. Feb 2007 A1
20070040931 Nishizawa Feb 2007 A1
20070073517 Panje Mar 2007 A1
20070073823 Cohen et al. Mar 2007 A1
20070075898 Markhovsky et al. Apr 2007 A1
20070082707 Flynt et al. Apr 2007 A1
20070136228 Petersen Jun 2007 A1
20070192128 Celestini Aug 2007 A1
20070198340 Lucovsky et al. Aug 2007 A1
20070198495 Buron et al. Aug 2007 A1
20070208751 Cowan et al. Sep 2007 A1
20070210936 Nicholson Sep 2007 A1
20070214180 Crawford Sep 2007 A1
20070214216 Carrer et al. Sep 2007 A1
20070233556 Koningstein Oct 2007 A1
20070233801 Eren et al. Oct 2007 A1
20070233859 Zhao et al. Oct 2007 A1
20070243887 Bandhole et al. Oct 2007 A1
20070244750 Grannan et al. Oct 2007 A1
20070255456 Funayama Nov 2007 A1
20070281690 Altman et al. Dec 2007 A1
20080022329 Glad Jan 2008 A1
20080025701 Ikeda Jan 2008 A1
20080032703 Krumm et al. Feb 2008 A1
20080033930 Warren Feb 2008 A1
20080043041 Hedenstroem et al. Feb 2008 A2
20080049704 Witteman et al. Feb 2008 A1
20080062141 Chandhri Mar 2008 A1
20080076505 Ngyen et al. Mar 2008 A1
20080092233 Tian et al. Apr 2008 A1
20080094387 Chen Apr 2008 A1
20080104503 Beall et al. May 2008 A1
20080109844 Baldeschwieler et al. May 2008 A1
20080120409 Sun et al. May 2008 A1
20080147730 Lee et al. Jun 2008 A1
20080148150 Mall Jun 2008 A1
20080158230 Sharma et al. Jul 2008 A1
20080168033 Ott et al. Jul 2008 A1
20080168489 Schraga Jul 2008 A1
20080189177 Anderton et al. Aug 2008 A1
20080207176 Brackbill et al. Aug 2008 A1
20080208692 Garaventi et al. Aug 2008 A1
20080214210 Rasanen et al. Sep 2008 A1
20080222545 Lemay Sep 2008 A1
20080255976 Altberg et al. Oct 2008 A1
20080256446 Yamamoto Oct 2008 A1
20080256577 Funaki et al. Oct 2008 A1
20080266421 Takahata et al. Oct 2008 A1
20080270938 Carlson Oct 2008 A1
20080288338 Wiseman et al. Nov 2008 A1
20080306826 Kramer et al. Dec 2008 A1
20080313329 Wang et al. Dec 2008 A1
20080313346 Kujawa et al. Dec 2008 A1
20080318616 Chipalkatti et al. Dec 2008 A1
20090006191 Arankalle et al. Jan 2009 A1
20090006565 Velusamy et al. Jan 2009 A1
20090015703 Kim et al. Jan 2009 A1
20090024956 Kobayashi Jan 2009 A1
20090030774 Rothschild et al. Jan 2009 A1
20090030999 Gatzke et al. Jan 2009 A1
20090040324 Nonaka Feb 2009 A1
20090042588 Lottin et al. Feb 2009 A1
20090058822 Chaudhri Mar 2009 A1
20090079846 Chou Mar 2009 A1
20090089678 Sacco et al. Apr 2009 A1
20090089710 Wood et al. Apr 2009 A1
20090093261 Ziskind Apr 2009 A1
20090132341 Klinger May 2009 A1
20090132453 Hangartner et al. May 2009 A1
20090132665 Thomsen et al. May 2009 A1
20090148045 Lee et al. Jun 2009 A1
20090153492 Popp Jun 2009 A1
20090157450 Athsani et al. Jun 2009 A1
20090157752 Gonzalez Jun 2009 A1
20090160970 Fredlund et al. Jun 2009 A1
20090163182 Gatti et al. Jun 2009 A1
20090177299 Van De Sluis Jul 2009 A1
20090192900 Collison Jul 2009 A1
20090199242 Johnson et al. Aug 2009 A1
20090215469 Fisher et al. Aug 2009 A1
20090232354 Camp, Jr. et al. Sep 2009 A1
20090234815 Boerries et al. Sep 2009 A1
20090239552 Churchill et al. Sep 2009 A1
20090249222 Schmidt et al. Oct 2009 A1
20090249244 Robinson et al. Oct 2009 A1
20090265647 Martin et al. Oct 2009 A1
20090288022 Almstrand et al. Nov 2009 A1
20090291672 Treves et al. Nov 2009 A1
20090292608 Polachek Nov 2009 A1
20090319607 Belz et al. Dec 2009 A1
20090327073 Li Dec 2009 A1
20100062794 Han Mar 2010 A1
20100082427 Burgener et al. Apr 2010 A1
20100082693 Hugg et al. Apr 2010 A1
20100100568 Papin et al. Apr 2010 A1
20100113065 Narayan et al. May 2010 A1
20100130233 Lansing May 2010 A1
20100131880 Lee et al. May 2010 A1
20100131895 Wohlert May 2010 A1
20100153144 Miller et al. Jun 2010 A1
20100159944 Pascal et al. Jun 2010 A1
20100161658 Hamynen et al. Jun 2010 A1
20100161831 Haas et al. Jun 2010 A1
20100162149 Sheleheda et al. Jun 2010 A1
20100183280 Beauregard et al. Jul 2010 A1
20100185552 Deluca et al. Jul 2010 A1
20100185665 Horn et al. Jul 2010 A1
20100191631 Weidmann Jul 2010 A1
20100197318 Petersen et al. Aug 2010 A1
20100197319 Petersen et al. Aug 2010 A1
20100198683 Aarabi Aug 2010 A1
20100198694 Muthukrishnan Aug 2010 A1
20100198826 Petersen et al. Aug 2010 A1
20100198828 Petersen et al. Aug 2010 A1
20100198862 Jennings et al. Aug 2010 A1
20100198870 Petersen et al. Aug 2010 A1
20100198917 Petersen et al. Aug 2010 A1
20100201482 Robertson et al. Aug 2010 A1
20100201536 Robertson et al. Aug 2010 A1
20100214436 Kim et al. Aug 2010 A1
20100223128 Dukellis et al. Sep 2010 A1
20100223343 Bosan et al. Sep 2010 A1
20100250109 Johnston et al. Sep 2010 A1
20100257196 Waters et al. Oct 2010 A1
20100259386 Holley et al. Oct 2010 A1
20100273509 Sweeney et al. Oct 2010 A1
20100281045 Dean Nov 2010 A1
20100306669 Della Pasqua Dec 2010 A1
20110004071 Faiola et al. Jan 2011 A1
20110010205 Richards Jan 2011 A1
20110029512 Folgner et al. Feb 2011 A1
20110040783 Uemichi et al. Feb 2011 A1
20110040804 Peirce et al. Feb 2011 A1
20110050909 Ellenby et al. Mar 2011 A1
20110050915 Wang et al. Mar 2011 A1
20110064388 Brown et al. Mar 2011 A1
20110066743 Hurley et al. Mar 2011 A1
20110083101 Sharon et al. Apr 2011 A1
20110102630 Rukes May 2011 A1
20110119133 Igelman et al. May 2011 A1
20110137881 Cheng et al. Jun 2011 A1
20110145564 Moshir et al. Jun 2011 A1
20110159890 Fortescue et al. Jun 2011 A1
20110164163 Bilbrey et al. Jul 2011 A1
20110187743 Hwang et al. Aug 2011 A1
20110197194 D'Angelo et al. Aug 2011 A1
20110202598 Evans et al. Aug 2011 A1
20110202968 Nurmi Aug 2011 A1
20110211534 Schmidt et al. Sep 2011 A1
20110213845 Logan et al. Sep 2011 A1
20110215966 Kim et al. Sep 2011 A1
20110225048 Nair Sep 2011 A1
20110238763 Shin et al. Sep 2011 A1
20110242134 Miller et al. Oct 2011 A1
20110255736 Thompson et al. Oct 2011 A1
20110273575 Lee Nov 2011 A1
20110282799 Huston Nov 2011 A1
20110283188 Farrenkopf Nov 2011 A1
20110314419 Dunn et al. Dec 2011 A1
20110320373 Lee et al. Dec 2011 A1
20120028659 Whitney et al. Feb 2012 A1
20120033718 Kauffman et al. Feb 2012 A1
20120036015 Sheikh Feb 2012 A1
20120036443 Ohmori et al. Feb 2012 A1
20120054797 Skog et al. Mar 2012 A1
20120059722 Rao Mar 2012 A1
20120062805 Candelore Mar 2012 A1
20120075285 Oyagi et al. Mar 2012 A1
20120084731 Filman et al. Apr 2012 A1
20120084835 Thomas et al. Apr 2012 A1
20120099800 Llano et al. Apr 2012 A1
20120108293 Law et al. May 2012 A1
20120110096 Smarr et al. May 2012 A1
20120113143 Adhikari et al. May 2012 A1
20120113272 Hata May 2012 A1
20120123830 Svendsen et al. May 2012 A1
20120123871 Svendsen et al. May 2012 A1
20120123875 Svendsen et al. May 2012 A1
20120124126 Alcazar et al. May 2012 A1
20120124176 Curtis et al. May 2012 A1
20120124458 Cruzada May 2012 A1
20120131507 Sparandara et al. May 2012 A1
20120131512 Takeuchi et al. May 2012 A1
20120143760 Abulafia et al. Jun 2012 A1
20120150978 Monaco Jun 2012 A1
20120165100 Lalancette et al. Jun 2012 A1
20120166971 Sachson et al. Jun 2012 A1
20120169855 Oh Jul 2012 A1
20120172062 Altman et al. Jul 2012 A1
20120173991 Roberts et al. Jul 2012 A1
20120176401 Hayward et al. Jul 2012 A1
20120184248 Speede Jul 2012 A1
20120197724 Kendall Aug 2012 A1
20120200743 Blanchflower et al. Aug 2012 A1
20120209924 Evans et al. Aug 2012 A1
20120210244 De Francisco et al. Aug 2012 A1
20120212632 Mate et al. Aug 2012 A1
20120218300 Hayakawa Aug 2012 A1
20120220264 Kawabata Aug 2012 A1
20120226748 Bosworth et al. Sep 2012 A1
20120233000 Fisher et al. Sep 2012 A1
20120236162 Imamura Sep 2012 A1
20120239761 Linner et al. Sep 2012 A1
20120250951 Chen Oct 2012 A1
20120252418 Kandekar et al. Oct 2012 A1
20120254325 Majeti et al. Oct 2012 A1
20120278387 Garcia et al. Nov 2012 A1
20120278692 Shi Nov 2012 A1
20120290637 Perantatos et al. Nov 2012 A1
20120299954 Wada et al. Nov 2012 A1
20120304052 Tanaka et al. Nov 2012 A1
20120304080 Wormald et al. Nov 2012 A1
20120307096 Ford et al. Dec 2012 A1
20120307112 Kunishige et al. Dec 2012 A1
20120319904 Lee et al. Dec 2012 A1
20120323933 He et al. Dec 2012 A1
20120324018 Metcalf et al. Dec 2012 A1
20130006759 Srivastava et al. Jan 2013 A1
20130024757 Doll et al. Jan 2013 A1
20130036364 Johnson Feb 2013 A1
20130045753 Obermeyer et al. Feb 2013 A1
20130050260 Reitan Feb 2013 A1
20130055083 Fino Feb 2013 A1
20130057587 Leonard et al. Mar 2013 A1
20130059607 Herz et al. Mar 2013 A1
20130060690 Oskolkov et al. Mar 2013 A1
20130063369 Malhotra et al. Mar 2013 A1
20130067027 Song et al. Mar 2013 A1
20130071093 Hanks et al. Mar 2013 A1
20130080254 Thramann Mar 2013 A1
20130085790 Palmer et al. Apr 2013 A1
20130086072 Peng et al. Apr 2013 A1
20130090171 Holton et al. Apr 2013 A1
20130095857 Garcia et al. Apr 2013 A1
20130100165 Komiyama Apr 2013 A1
20130104053 Thornton et al. Apr 2013 A1
20130110885 Brundrett, III May 2013 A1
20130111514 Slavin et al. May 2013 A1
20130128059 Kristensson May 2013 A1
20130129252 Lauper May 2013 A1
20130132477 Bosworth et al. May 2013 A1
20130145286 Feng et al. Jun 2013 A1
20130147837 Stroila Jun 2013 A1
20130159110 Rajaram et al. Jun 2013 A1
20130159919 Leydon Jun 2013 A1
20130169822 Zhu et al. Jul 2013 A1
20130173729 Starenky et al. Jul 2013 A1
20130182133 Tanabe Jul 2013 A1
20130185131 Sinha et al. Jul 2013 A1
20130191198 Carlson et al. Jul 2013 A1
20130194301 Robbins et al. Aug 2013 A1
20130198176 Kim Aug 2013 A1
20130218965 Abrol et al. Aug 2013 A1
20130218968 Mcevilly et al. Aug 2013 A1
20130222323 Mckenzie Aug 2013 A1
20130227476 Frey Aug 2013 A1
20130232194 Knapp et al. Sep 2013 A1
20130263031 Oshiro et al. Oct 2013 A1
20130265450 Barnes, Jr. Oct 2013 A1
20130267253 Case et al. Oct 2013 A1
20130275505 Gauglitz et al. Oct 2013 A1
20130290443 Collins et al. Oct 2013 A1
20130304646 De Geer Nov 2013 A1
20130311255 Cummins et al. Nov 2013 A1
20130325964 Berberat Dec 2013 A1
20130344896 Kirmse et al. Dec 2013 A1
20130346869 Asver et al. Dec 2013 A1
20130346877 Borovoy et al. Dec 2013 A1
20140006129 Heath Jan 2014 A1
20140011538 Mulcahy et al. Jan 2014 A1
20140019264 Wachman et al. Jan 2014 A1
20140028712 Keating et al. Jan 2014 A1
20140032682 Prado et al. Jan 2014 A1
20140043204 Basnayake et al. Feb 2014 A1
20140045530 Gordon et al. Feb 2014 A1
20140047016 Rao Feb 2014 A1
20140047045 Baldwin et al. Feb 2014 A1
20140047335 Lewis et al. Feb 2014 A1
20140049652 Moon et al. Feb 2014 A1
20140052485 Shidfar Feb 2014 A1
20140052633 Gandhi Feb 2014 A1
20140055491 Malamud et al. Feb 2014 A1
20140057660 Wager Feb 2014 A1
20140063062 Fateh Mar 2014 A1
20140082651 Sharifi Mar 2014 A1
20140092130 Anderson et al. Apr 2014 A1
20140096029 Schultz Apr 2014 A1
20140114565 Aziz et al. Apr 2014 A1
20140122658 Haeger et al. May 2014 A1
20140122787 Shalvi et al. May 2014 A1
20140125701 Hayakawa May 2014 A1
20140129953 Spiegel May 2014 A1
20140143143 Fasoli et al. May 2014 A1
20140149519 Redfern et al. May 2014 A1
20140155102 Cooper et al. Jun 2014 A1
20140173424 Hogeg et al. Jun 2014 A1
20140173457 Wang et al. Jun 2014 A1
20140189592 Benchenaa et al. Jul 2014 A1
20140207679 Cho Jul 2014 A1
20140210859 Ihara et al. Jul 2014 A1
20140214471 Schreiner, III Jul 2014 A1
20140222564 Kranendonk et al. Aug 2014 A1
20140258405 Perkin Sep 2014 A1
20140265359 Cheng et al. Sep 2014 A1
20140266703 Dalley, Jr. et al. Sep 2014 A1
20140279061 Elimeliah et al. Sep 2014 A1
20140279436 Dorsey et al. Sep 2014 A1
20140279540 Jackson Sep 2014 A1
20140280537 Pridmore et al. Sep 2014 A1
20140282096 Rubinstein et al. Sep 2014 A1
20140287779 O'keefe et al. Sep 2014 A1
20140289833 Briceno et al. Sep 2014 A1
20140306986 Gottesman et al. Oct 2014 A1
20140317302 Naik Oct 2014 A1
20140324627 Haver et al. Oct 2014 A1
20140324629 Jacobs Oct 2014 A1
20140325383 Brown et al. Oct 2014 A1
20140368542 Tawara et al. Dec 2014 A1
20150006276 Cohen Jan 2015 A1
20150020086 Chen et al. Jan 2015 A1
20150029220 Hittel et al. Jan 2015 A1
20150046278 Pei et al. Feb 2015 A1
20150071619 Brough Mar 2015 A1
20150087263 Branscomb et al. Mar 2015 A1
20150088622 Ganschow et al. Mar 2015 A1
20150095020 Leydon Apr 2015 A1
20150096042 Mizrachi Apr 2015 A1
20150116529 Wu et al. Apr 2015 A1
20150169827 Laborde Jun 2015 A1
20150172534 Miyakawa et al. Jun 2015 A1
20150178260 Brunson Jun 2015 A1
20150222814 Li et al. Aug 2015 A1
20150261917 Smith Sep 2015 A1
20150262197 Chao et al. Sep 2015 A1
20150312184 Langholz et al. Oct 2015 A1
20150350136 Flynn, III et al. Dec 2015 A1
20150365795 Allen et al. Dec 2015 A1
20150378502 Hu et al. Dec 2015 A1
20160004320 Lundberg et al. Jan 2016 A1
20160006927 Sehn Jan 2016 A1
20160014063 Hogeg et al. Jan 2016 A1
20160085773 Chang et al. Mar 2016 A1
20160085863 Allen et al. Mar 2016 A1
20160093107 Yamamoto et al. Mar 2016 A1
20160099901 Allen et al. Apr 2016 A1
20160148430 Lin et al. May 2016 A1
20160180887 Sehn Jun 2016 A1
20160182422 Sehn et al. Jun 2016 A1
20160182875 Sehn Jun 2016 A1
20160239248 Sehn Aug 2016 A1
20160277419 Allen et al. Sep 2016 A1
20160321708 Sehn Nov 2016 A1
20170006094 Abou Mahmoud et al. Jan 2017 A1
20170061308 Chen et al. Mar 2017 A1
20170199855 Fishbeck Jul 2017 A1
20170287006 Azmoodeh et al. Oct 2017 A1
20180005443 Poulos et al. Jan 2018 A1
20180114365 Egri et al. Apr 2018 A1
20180136465 Chi et al. May 2018 A1
20190011703 Robaina et al. Jan 2019 A1
20190130436 Ma et al. May 2019 A1
20200035003 Canberk et al. Jan 2020 A1
Foreign Referenced Citations (32)
Number Date Country
2887596 Jul 2015 CA
2051480 Apr 2009 EP
2151797 Feb 2010 EP
2399928 Sep 2004 GB
19990073076 Oct 1999 KR
20010078417 Aug 2001 KR
WO-1996024213 Aug 1996 WO
WO-1999063453 Dec 1999 WO
WO-2000058882 Oct 2000 WO
WO-2001029642 Apr 2001 WO
WO-2001050703 Jul 2001 WO
WO-2006118755 Nov 2006 WO
WO-2007092668 Aug 2007 WO
WO-2009043020 Apr 2009 WO
WO-2011040821 Apr 2011 WO
WO-2011119407 Sep 2011 WO
WO-2013008238 Jan 2013 WO
WO-2013045753 Apr 2013 WO
WO-2014006129 Jan 2014 WO
WO-2014068573 May 2014 WO
WO-2014115136 Jul 2014 WO
WO-2014194262 Dec 2014 WO
WO-2015192026 Dec 2015 WO
WO-2016044424 Mar 2016 WO
WO-2016054562 Apr 2016 WO
WO-2016065131 Apr 2016 WO
WO-2016100318 Jun 2016 WO
WO-2016100318 Jun 2016 WO
WO-2016100342 Jun 2016 WO
WO-2016149594 Sep 2016 WO
WO-2016179166 Nov 2016 WO
2020023264 Jan 2020 WO
Non-Patent Literature Citations (29)
Entry
U.S. Appl. No. 16/043,995, filed Jul. 24, 2018, Conditional Modification of Augmented Reality Object.
“A Whole New Story”, Snap, Inc., [Online] Retrieved from the Internet: <URL: https://www.snap.com/en-US/news/>, (2017), 13 pgs.
“Adding photos to your listing”, eBay, [Online] Retrieved from the Internet: <URL: http://pages.ebay.com/help/sell/pictures.html>, (accessed May 24, 2017), 4 pgs.
“U.S. Appl. No. 16/043,995, Non Final Office Action dated Oct. 31, 2019”, 9 pgs.
“U.S. Appl. No. 16/043,995, Notice of Allowance dated Feb. 7, 2020”, 8 pgs.
“U.S. Appl. No. 16/043,995, Response filed Jan. 20, 2020 to Non Final Office Action dated Oct. 31, 2019”, 12 pgs.
“BlogStomp”, StompSoftware, [Online] Retrieved from the Internet: <URL: http://stompsoftware.com/blogstomp>, (accessed May 24, 2017), 12 pgs.
“Cup Magic Starbucks Holiday Red Cups come to life with AR app”, Blast Radius, [Online] Retrieved from the Internet: <URL: https://web.archive.org/web/20160711202454/http://www.blastradius.com/work/cup-magic>, (2016), 7 pgs.
“Daily App: InstaPlace (iOS/Android): Give Pictures a Sense of Place”, TechPP, [Online] Retrieved from the Internet: <URL: http://techpp.com/2013/02/15/instaplace-app-review>, (2013), 13 pgs.
“InstaPlace Photo App Tell The Whole Story”, [Online] Retrieved from the Internet: <URL: youtu.beiuF_gFkg1hBM>, (Nov. 8, 2013), 113 pgs., 1:02 min.
“International Application Serial No. PCT/US2015/037251, International Search Report dated Sep. 29, 2015”, 2 pgs.
“International Application Serial No. PCT/US2019/042183, International Search Report dated Oct. 30, 2019”, 4 pgs.
“International Application Serial No. PCT/US2019/042183, Written Opinion dated Oct. 30, 2019”, 5 pgs.
“Introducing Snapchat Stories”, [Online] Retrieved from the Internet: <URL: https://web.archive.org/web/20131026084921/https://www.youtube.com/watch?v=88Cu3yN-LIM>, (Oct. 3, 2013), 92 pgs.; 00:47 min.
“Macy's Believe-o-Magic”, [Online] Retrieved from the Internet: <URL: https://web.archive.org/web/20190422101854/https://www.youtube.com/watch?v=xvzRXy3J0zo&feature=youtu.be>, (Nov. 7, 2011), 102 pgs.; 00:51 min.
“Macy's Introduces Augmented Reality Experience in Stores across Country as Part of Its 2011 Believe Campaign”, Business Wire, [Online] Retrieved from the Internet: <URL: https://www.businesswire.com/news/home/20111102006759/en/Macys-Introduces-Augmented-Reality-Experience-Stores-Country>, (Nov. 2, 2011), 6 pgs.
“Starbucks Cup Magic”, [Online] Retrieved from the Internet: <URL: https://www.youtube.com/watch?v=RWNQXi9RG0w>, (Nov. 8, 2011), 87 pgs.; 00:47 min.
“Starbucks Cup Magic for Valentine's Day”, [Online] Retrieved from the Internet: <URL: https://www.youtube.com/watch?v=8nvgOzjg10w>, (Feb. 6, 2012), 88 pgs. 00:45 min.
“Starbucks Holiday Red Cups Come to Life, Signaling the Return of the Merriest Season”, Business Wire, [Online] Retrieved from the Internet: <URL: http://www.businesswire.com/news/home/20111115005744/en/2479513/Starbucks-Holiday-Red-Cups-Life-Signaling-Return>, (Nov. 15, 2011), 5 pgs.
Carthy, Roi, “Dear All Photo Apps: Mobli Just Won Filters”, TechCrunch, [Online] Retrieved from the Internet: <URL: https://techcrunch.com/2011109/08/mobli-filters>, (Sep. 8, 2011), 10 pgs.
Fuste, Anna, “HoloBits: Creativity and Fun in Mixed Reality”, URL: https://www.media.mit.edu/projects/holobits/overview/, (accessed May 11, 2018), 4 pgs.
Janthong, Isaranu, “Instaplace ready on Android Google Play store”, Android App Review Thailand, [Online] Retrieved from the Internet: <URL: http://www.android-free-app-review.com/2013/01/instaplace-android-google-play-store.html>, (Jan. 23, 2013), 9 pgs.
MacLeod, Duncan, “Macys Believe-o-Magic App”, [Online] Retrieved from the Internet: <URL: http://theinspirationroom.com/daily/2011/macys-believe-o-magic-app>, (Nov. 14, 2011), 10 pgs.
MacLeod, Duncan, “Starbucks Cup Magic Lets Merry”, [Online] Retrieved from the Internet: <URL: http://theinspirationroom.com/daily/2011/starbucks-cup-magic>, (Nov. 12, 2011), 8 pgs.
Notopoulos, Katie, “A Guide to the New Snapchat Filters And Big Fonts”, [Online] Retrieved from the Internet: <URL: https://www.buzzfeed.com/katienotopoulos/a-guide-to-the-new-snapchat-filters-and-big-fonts?utm_term=.bkQ9qVZWe#.nv58YXpkV>, (Dec. 22, 2013), 13 pgs.
Panzarino, Matthew, “Snapchat Adds Filters, A Replay Function and for Whatever Reason, Time, Temperature and Speed Overlays”, TechCrunch, [Online] Retrieved form the Internet: <URL: https://techcrunch.com/2013/12/20/snapchat-adds-filters-new-font-and-for-some-reason-time-temperature-and-speed-overlays/>, (Dec. 20, 2013), 12 pgs.
Placenote, “Introduction to Placenote SDK”, URL: https://placenote.com/documentation/, (accessed May 17, 2018), 3 pgs.
Tripathi, Rohit, “Watermark Images in PHP and Save File on Server”, [Online] Retrieved from the Internet: <URL: http://code.rohitink.com/2012/12/28/watermark-images-in-php-and-save-file-on-server>, (Dec. 28, 2012), 4 pgs.
Voropaev, Pavel, “AR Synchronization System by Prisma”, URL: https://blog.prismalabs.ai/ar-synchronization-system-by-prisma-3c41ff55a2a7, (accessed May 11, 2018), 6 pgs.
Related Publications (1)
Number Date Country
20200265626 A1 Aug 2020 US
Continuations (1)
Number Date Country
Parent 16043995 Jul 2018 US
Child 16868315 US