Multichannel system

Information

  • Patent Grant
  • 11249617
  • Patent Number
    11,249,617
  • Date Filed
    Thursday, August 1, 2019
    5 years ago
  • Date Issued
    Tuesday, February 15, 2022
    2 years ago
Abstract
Systems and methods for presenting content that depicts one or more stories are provided. Content that depicts one or more stories is received from a first client device. The content that depicts the one or more stories is grouped into an edition being used to represent the one or more stories. A discovery page that enables selection of the edition is generated. Presentation of the discovery page is caused on a second client device.
Description
TECHNICAL FIELD

Embodiments of the present disclosure relate generally to data processing and, more particularly, but not by way of limitation, to discovery of content grouped into an edition.


BACKGROUND

In recent years, use of mobile devices, wearable devices, smart devices, and the like have each increased among consumers. In fact, these devices pervade nearly every aspect of modern life. Further, the use of mobile devices for messaging and consumption of information has also increased.





BRIEF DESCRIPTION OF THE DRAWINGS

Various ones of the appended drawings merely illustrate example embodiments of the present disclosure and should not be considered as limiting its scope.



FIG. 1 is a block diagram illustrating a networked system, according to some example embodiments.



FIG. 2 is a block diagram illustrating components of a discovery system, according to some example embodiments.



FIG. 3 is a user interface of a discovery page, according to some example embodiments.



FIGS. 4-7 are user interfaces of a story from an edition, according to some example embodiments.



FIG. 8 is a user interface of annotated content, according to some example embodiments.



FIG. 9 is a user interface of an upload page, according some example embodiments.



FIGS. 10-12 are flowcharts illustrating operations of a discovery system in performing a method of presentation of content that depicts one or more stories, according to some example embodiments.



FIG. 13 is a user interface diagram depicting an example mobile device and mobile operating system interface, according to some example embodiments.



FIG. 14 is a block diagram illustrating an example of a software architecture that may be installed on a machine, according to some example embodiments,



FIG. 15 is a block diagram presenting a diagrammatic representation of a machine in the form of a computer system within which a set of instructions may be executed for causing the machine to perform any of the methodologies discussed herein, according to an example embodiment.





The headings provided herein are merely for convenience and do not necessarily affect the scope or meaning of the terms used.


DETAILED DESCRIPTION

The description that follows includes systems, methods, techniques, instruction sequences, and computing machine program products that embody illustrative embodiments of the disclosure. In the following description, for the purposes of explanation, numerous specific details are set forth in order to provide an understanding of various embodiments of the inventive subject matter. It will be evident, however, to those skilled in the art, that embodiments of the inventive subject matter may be practiced without these specific details. In general, well-known instruction instances, protocols, structures, and techniques are not necessarily shown in detail.


In various example embodiments, a system makes an edition available for a user to view in a user interface of a messaging application on a client device. In some instances, the edition is a group of stories that are bundled together. Moreover, the system may generate a discovery page featuring a variety of channels. The user may access the edition from one of the channels featured in the discovery page. In other words, the channel featured on the discovery page serves as a link to the edition. As stated, the edition can include one or more stories. For example, each story can include an article, multimedia items, and the like. Moreover, each story may correspond to a specific topic. As soon as the user selects the edition from the discovery page, the user may view the one or more stories from the edition. Also, the content within each story may include more than one layer. For example, the content within a story may include a first layer, also known as a top-snap, and a second layer, also known as longform content. In some embodiments, the top-snap for each story may be viewed by tapping or swiping horizontally across the screen of the client device and the longform content may be viewed by selection of the top-snap. For instance, the user may swipe vertically on the top-snap in order to view the longform content.


In various example embodiments, upon viewing the edition, the user may share the content from the edition with another user. The user may send a request to the system to share the content with the other user. In some instances, the user may also annotate the content prior to sharing it with other users. Once shared, the other users may view the annotated content from their client devices. In some cases, the annotated content is sent as an ephemeral message and is only available for a predefined duration of time.


In various example embodiments, the system includes a publication application that enables editions to be uploaded to the discovery page from a content publisher or an uploading party. The system may present an upload page to the uploading party. The upload page of the publication application may assist the uploading party in uploading the content for the edition to a database server. Once the content for the edition is uploaded at the database server, the content is made available to view from the discovery page. Accordingly, the system generates a channel on the discovery page which provides access to the one or more editions. In some embodiments, channels may be implemented to correspond to uploading parties. Although the discovery page provides access to the one or more editions, in some instances, only one edition may be viewed at a time on the client device.



FIG. 1 is a network diagram depicting a network system 100 having a client-server architecture configured for exchanging data over a network, according to one embodiment. For example, the network system 100 may be a messaging system where clients communicate and exchange data within the network system 100. The data may pertain to various functions (e.g., sending and receiving text and media communication, determining geolocation, etc.) and aspects (e.g., publication of media content, management of media content, etc.) associated with the network system 100 and its users. Although illustrated herein as client-server architecture, other embodiments may include other network architectures, such as peer-to-peer or distributed network environments.


As shown in FIG. 1, the network system 100 includes a social messaging system 130. The social messaging system 130 is generally based on a three-tiered architecture, consisting of an interface layer 124, an application logic layer 126, and a data layer 128. As is understood by skilled artisans in the relevant computer and Internet-related arts, each module or engine shown in FIG. 1 represents a set of executable software instructions and the corresponding hardware (e.g., memory and processor) for executing the instructions. To avoid obscuring the inventive subject matter with unnecessary detail, various functional modules and engines that are not germane to conveying an understanding of the inventive subject matter have been omitted from FIG. 1. Of course, additional functional modules and engines may be used with a social messaging system, such as that illustrated in FIG. 1, to facilitate additional functionality that is not specifically described herein. Furthermore, the various functional modules and engines depicted in FIG. 1 may reside on a single server computer, or may be distributed across several server computers in various arrangements. Moreover, although the social messaging system 130 is depicted in FIG. 1 as a three-tiered architecture, the inventive subject matter is by no means limited to such an architecture.


As shown in FIG. 1, the interface layer 124 consists of interface module(s) (e.g., a web server) 140, which receives requests from various client-computing devices and servers, such as client device(s) 110 executing client application(s) 112, and third party server(s) 120 executing third party application(s) 122. In response to received requests, the interface module(s) 140 communicates appropriate responses to requesting devices via a network 104. For example, the interface module(s) 140 can receive requests such as Hypertext Transfer Protocol (HTTP) requests, or other web-based, Application Programming Interface (API) requests.


The client device(s) 110 can execute conventional web browser applications or applications (also referred to as “apps”) that have been developed for a specific platform to include any of a wide variety of mobile computing devices and mobile-specific operating systems (e.g., IOS™, ANDROID™, WINDOWS® PHONE). In an example, the client device(s) 110 are executing the client application(s) 112. The client application(s) 112 can provide functionality to present information to a user 106 and communicate via the network 104 to exchange information with the social messaging system 130. Each of the client device(s) 110 can comprise a computing device that includes at least a display and communication capabilities with the network 104 to access the social messaging system 130. The client device(s) 110 comprise, but are not limited to, remote devices, work stations, computers, general purpose computers, Internet appliances, hand-held devices, wireless devices, portable devices, wearable computers, cellular or mobile phones, personal digital assistants (PDAs), smart phones, tablets, ultrabooks, netbooks, laptops, desktops, multi-processor systems, microprocessor-based or programmable consumer electronics, game consoles, set-top boxes, network PCs, mini-computers, and the like. User(s) 106 can be a person, a machine, or other means of interacting with the client device(s) 110. In some embodiments, the user(s) 106 interact with the social messaging system 130 via the client device(s) 110.


As shown in FIG. 1, the data layer 128 has database server(s) 132 that facilitate access to information storage repositories or database(s) 134. The database(s) 134 are storage devices that store data such as member profile data, social graph data (e.g., relationships between members of the social messaging system 130), and other user data.


An individual can register with the social messaging system 130 to become a member of the social messaging system 130. Once registered, a member can form social network relationships (e.g., friends, followers, or contacts) the social messaging system 130 and interact with a broad range of applications provided by the social messaging system 130.


The application logic layer 126 includes various application logic module(s) 150, which, in conjunction with the interface module(s) 140, generate various user interfaces with data retrieved from various data sources or data services in the data layer 128. Individual application logic module(s) 150 may be used to implement the functionality associated with various applications, services, and features of the social messaging system 130. For instance, a social messaging application can be implemented with one or more of the application logic module(s) 150. The social messaging application provides a messaging mechanism for users of the client device(s) 110 to send and receive messages that include text and media content such as pictures and video. The client device(s) 110 may access and view the messages from the social messaging application for a specified period of time (e.g., limited or unlimited). In an example, a particular message is accessible to a message recipient for a predefined duration (e.g., specified by a message sender) that begins when the particular message is first accessed. After the predefined duration elapses, the message is deleted and is no longer accessible to the message recipient. Of course, other applications and services may be separately embodied in their own application server module(s) 150. As another example, a publication application can be implemented with one or more of the application logic module(s) 150. The publication application provides a messaging mechanism for users (e.g., content publishers) of the client device(s) 110 to upload media content to the social messaging system 130.


As illustrated in FIG. 1, the social messaging system 130 includes a discovery system 160. In various embodiments, the discovery system 160 can be implemented as a standalone system and is not necessarily included in the social messaging system 130. In some embodiments, the client device(s) 110 includes a portion of the discovery system 160 (e.g., a portion of the discovery system may be included in the client application(s) 112). In these embodiments, the client device (110 can work alone or in conjunction with the portion of the discovery system 160 included in a particular application server or the social messaging system 130.


The communication module 250 provides various communications functionality. For example, the communication module 250 may be configured to exchange network communications with database server(s) 132, clients device(s) 110, as well as any third party server(s) 120. The types of information and data exchanged may include configuration data for media content, such as pictograph data (e.g., an image file); geographic indications (e.g., a particular latitude and longitude); data associated with the user (e.g., member profile data from an online account or social network service data); or other data to facilitate the functionality described herein.


The presentation module 240 provides various presentation and user interface functionalities and allows the transmission of information to and from the user (e.g., visually displaying information on a screen, acoustic output, haptic feedback). For instance, the presentation module 240 may present content depicting one or more stories on a user interface or in a display of a user device. The user may provide input to interact with the user interface in many possible manners, such as alphanumeric, point based (e.g., cursor), tactile, or other input (e.g., touch screen, tactile sensor, light sensor, infrared sensor, biometric sensor, microphone, gyroscope, accelerometer, or other sensors), and the like.



FIG. 2 is a block diagram illustrating components of a discovery system 160, according to some example embodiments. The discovery system 160 is shown as including a reception module 210, a group module 220, a generation module 230, and a presentation module 240, all configured to communicate with each other (e.g., via a bus, network coupling, shared memory, a switch, and the like). Any one or more of the modules described herein may be implemented using hardware (e.g., one or more processors of a machine) or a combination of hardware and software. For example, any module described herein may configure a processor (e.g., among one or more processors of a machine) to perform the operations described herein for that module. Moreover, multiple modules may be combined into a single module, and the functions described herein for a single module may be subdivided among multiple modules. Furthermore, according to various example embodiments, modules described herein as being implemented within a single machine, database, or device may be distributed across multiple machines, databases, or devices. Other modules not pertinent to example embodiments can also be included, but are not shown.


In various example embodiments, the reception module 210 is configured to receive content that depicts one or more stories. The reception module 210 may receive the content from a first client device which is operated by a first user. The first user may be an individual, a group of users or any other entity. The content may include a video clip, an image, a description, and the like. Each story may revolve around a specific topic or idea. For instance, a story may include information about a particular travel destination. Accordingly, the content for the travel destination may include pictures and descriptions of the travel destination. In some instances, the content may include multiple layers (e.g., a first layer, a second layer, etc.). A user a may view the first layer, also known as a top-snap, prior to the second layer, also known as longform content. In some instances, the first layer is selectable to cause display of the second layer. For example, the first layer content for a music video may be a ten second preview of the entire music video. The second layer content for the music video may be a full length clip of the music video. As another example, the first layer content for the travel destination may be a picture of the travel destination. The second layer content for the travel destination may be descriptions about things to do at the travel destination.


In various example embodiments, the reception module 210 is further configured to receive a request from the first client device to upload the content into the edition. The first device may have client application software (e.g., a messaging application) installed which includes a feature to send the request to the reception module 210. In response, the presentation module 240 is configured to display an upload page on the first client device, as further explained below.


In various example embodiments, the group module 220 is configured to group the content that depicts the one or more stories into an edition. In example embodiments, these editions may be presented as part of the messaging application installed on a client device. Each story of the edition may occupy an entire screen when viewed on the client device. Moreover, a user may scroll through each story of the edition using a tapping or swiping vesture or motion. Specifically, the first layer (e.g., top-snap) for each story may be viewed by tapping or swiping horizontally across the screen of the client device. Similarly, the second layer (e.g., longform content) may be viewed by selection of the top-snap. In various example embodiments, the reception module 210 is further configured to store the received content in a database, such as database(s) 134. Once stored, the group module 220 can retrieve the content from the database and group the content into the edition.


The group module 220 may be further configured to identify that the content is received from the first client device within a specific predefined window of time. For instance, all of the stories grouped into the edition may be identified by the group module 220 as being received within a span of a week. Therefore, these stories within the edition may remain current or up to date.


In various example embodiments, the generation module 230 is configured to generate a discovery page that allows a user to select an edition. In some embodiments, the discovery page is a landing page that includes a link or channel to an edition. For instance, each link/channel may be generated by generation module 230. In order to select an edition for viewing, a user may activate or select the desired link on the discovery page. In some embodiments, the channels may correspond to (or serve to identify) the source of the content in the edition.


In various example embodiments, the presentation module 240 may present the discovery page on a second client device, operated by a second user. Thus, the presentation module 240 is further configured to present the icon of the channel as part of the discovery page. The channel may also serve as a link to other editions that are received from the same source. In various example embodiments, the presentation module 240 may present the content on the second client device based on the selection of the channel. Therefore, the presentation of the content is performed based on selection of the edition from the discovery page. The presentation module 240 may also present the multiple layers of the content (e.g., the first layer and the second layer) on the second client device.


In various example embodiments, the presentation module 240 is further configured to present the upload page on the first client device. The presentation of the upload page may be performed in response to the received request to upload content. The upload page enables a user (e.g., the first user operating the first client device) to upload content that depicts the one or more stories. For instance, the upload page may include a section where the user can drag content to indicate that the content is to be uploaded into the edition.


In various example embodiments, the reception module 210 is further configured to receive an annotated version of the content that depicts the one or more stories. The reception module 210 may also receive an indication to share the annotated version of the content with another or third client device. In this regard, the annotated version of the content may be received from the second client device. The annotated version of the content may include additional markings added by the second user to the original content that was displayed on the second client device. Specifically, upon viewing the content on the second client device, the second user may edit or alter the content in order to generate an annotated version of the content. Annotations may include: adding a further description, a drawing, an animation or a picture, adding a layer on top of original content (including e.g., any number of photofilters), deleting a portion of the original content, and the like. Accordingly, the presentation module 240 is configured to cause presentation of the annotated version of the content on the third client device for a predefined duration of time. Therefore, the third user on the third client device will be able to view the annotated version of the content sent from the second user on the second client device. In some instances, the annotated version of the content is an ephemeral message that is available for the predefined duration of time. After the predefined duration of time elapses, the message may be deleted, and no longer able to be viewed, on the third client device.


In various example embodiments, the reception module 210 is further configured to receive a scheduled time from the first client device. The scheduled time is used to indicate when to present the edition on the second client device. Therefore, in some instances, the presentation module 240 is further configured to present the edition on the second client device during the scheduled time. For example, the edition is made available on the discovery page during the scheduled time. Thus, when the scheduled time elapses, the second user is no longer able to select the edition for viewing from the discovery page.



FIG. 3 is a user interface 300 of a discovery page, according to some example embodiments. The discovery page is displayed on a client device belonging to a user. As shown, the user interface 300 includes a variety of channels. Each channel represents an icon and can be used to access an edition. In some instances, each channel corresponds to a source for the content or editions included in the channel. As shown, the user interface 300 includes a first channel 310 and a second channel 320. The first channel 310 may be a link that is selectable by the user to view an edition from the first channel 310. Likewise, the second channel 320 is a link that is selectable by the user to view an edition from the second channel 320. Selection of a channel may include a touch or a gesture over the portion of the discovery page that displays the channel.



FIG. 4 is a user interface 400 of a story from an edition, according to some example embodiments. The user interface 400 is displayed on a client device belonging to a user. Further, the user interface 400 may be displayed after selection of a channel from the discovery page of FIG. 3. As shown, the user interface 400 of the story displays content that includes a description 410 and a video clip 420. For example, the story may be a rock music video, as indicated by the description 410. Further included in the user interface 400 is a video clip 420 of the rock music video. In some instances, the video clip 420 is a preview of the entire rock music video and therefore a shortened version of the rock music video. Also included in the content displayed in the user interface 400 is a description 405 to indicate that the story for the rock music video belongs to the edition.



FIG. 5 is a user interface 500 of a story from an edition, depicting a second layer of a story, according to some example embodiments. The user interface 500 is displayed on a client device belonging to a user. Further, the user interface 500 may be displayed after selection of the content displayed in the user interface 400 shown in FIG. 4. For example, the user may swipe vertically on the user interface 400 in order to view the user interface 500, As shown, the content displayed in the user interface 500 is the second layer of the story for the rock music video. Moreover, the content displayed in the user interface 500 includes a description 510 and a video clip 520. The video clip 520 may be the entire version of the rock music video, as indicated by the description 510. Also included in the content displayed in the user interface 500 is a description 505 to indicate that the story for the rock music video belongs to the edition. The video clip 520 may also be included with an image 525. Specifically, embodiments of the invention contemplate allowing portable network graphics images (“PNG images”) or the like to be overlaid on top of video. Thus, the image 525 may appear in front of the video clip 520 as the video clip 520 is playing. Because the PNG image is able to present text, icons, colors, etc. in a more detailed manner than a video counterpart, the content appears more tightly integrated than other methods.



FIG. 6 is a user interface 600 of a story from an edition, according to some example embodiments. The user interface 600 is displayed on a client device belonging to a user and displays a story for a travel destination (e.g., Chicago). The content displayed in the user interface 600 includes a description 610 and an image 620. In this example, the description 610 indicates that the story is about visiting Chicago. Further, the image 620 is used to depict the Chicago skyline. Also included in the content displayed in the user interface 600 is a description 605 to indicate that the story for the travel destination belongs to the edition.



FIG. 7 is a user interface 700 of a story from an edition, according to some example embodiments. The user interface 700 is displayed on a client device belonging to a user. Further, the user interface 700 may be displayed after selection of the user interface 600 shown in FIG. 6. For example, the user may swipe vertically on the content displayed in the user interface 600 in order to view the user interface 700. As shown, the content displayed in the user interface 700 is the second layer of the story for the travel destination. Moreover, the content displayed in the user interface 700 includes a first description 710, a second description 720, and a third description 730. The first description indicates that the story is about visiting Chicago. The second description 720 provides information about events in Chicago. The third description 730 provides information about food in Chicago. Also included in the content displayed in the user interface 700 is a description 705 to indicate that the story for the travel destination belongs to the edition.



FIG. 8 is a user interface 800 of annotated content, according to some example embodiments. The user interface 800 is displayed on a client device belonging to a recipient. Further, the user interface 800 includes content that was intended to be shared with the recipient and received from a sender. As shown, the content displayed in the user interface 800 includes a first description 810, a second description. 820, and a third description 830. In some instances, the first description 810, the second description 820, and the third description 830 are respectively identical to the first description 710, the second description 720, and the third description 730 of FIG. 7. Also shown in FIG. 8 is a fourth description 850. The fourth description 850 may be additional content that is added by the sender. For instance, the sender may edit the content displayed in the user interface 700 of FIG. 7 by adding the fourth description 850 (e.g., by typing using a keyboard or handwriting via the touch sensitive display screen). As shown, the fourth description 850 may include text, as well as graphics, animations, emojis, photo filters, etc. Thereafter, the sender will send the annotated content to the recipient to be displayed on the client device belonging to the recipient. Also shown in FIG. 8 is a timer 840 to indicate an amount of time remaining before the annotated content disappears. Therefore, the annotated content may appear on the screen on the client device for the length of time indicated by the timer 840. Also shown in FIG. 8 is a link 860 in the form of an arrow. In some instances, the link is selectable to cause navigation to the discovery page. For example, upon selection of the link, an animation may appear on the screen of the client device which will eventually display the discovery page.



FIG. 9 is a user interface 900 of an upload page, according some example embodiments. The user interface 900 is displayed on a device belonging to a user, such as a desktop computer or a mobile device. For example, as shown in FIG. 9, the user interface 900 of the upload page is displayed on a desktop computer. Moreover, the user is able to upload original content by interacting with the user interface 900. Specifically, the user interface 900 includes a portion 910 where the user is able to upload specific content. In particular, the user is able to upload content for one or more stories by moving files over to the portion 910 of the user interface. For example, the user may upload a video clip by moving the video clip to the portion 910 of the user interface. As another example, the user may upload an image by moving the image to the portion 910 of the user interface. The user interface 900 further includes a scheduling window 920 that enables the user to input a scheduled time. The scheduled time corresponds to when the edition is made available on the discovery page. Although not shown in FIG. 9, in some instances, the upload page may include editing tools, such as a text or video editor, to facilitate the creation and/or editing of content.



FIG. 10-12 are flowcharts illustrating operations of the discovery system 160 in performing a method 1000 of presentation of content that depicts one or more stories, according to some example embodiments. Operations in the method 1000 may be performed by the discovery system 160, using modules described above with respect to FIG. 2. As shown in FIG. 10, the method 1000 includes operations 1010, 1020, 1030, 1040, and 1050.


At operation 1010, the reception module 210 receives content that depicts one or more stories. The reception module 210 receives the content from a first client device which is operated by a first user. As an example, the reception module 210 receives first content that depicts a first story and second content that depicts a second story. The first user may upload the content from an upload page that is displayed on the first client device. Further, the content may include a video clip, an image, a description, and the like.


At operation 1020, the group module 220 groups the content into an edition used to represent the one or more stories. As an example, the group module 220 groups the first content and the second content received at operation 1020 into the edition used to represent the one or more stories.


At operation 1030, the generation module 230 generates a discovery page that enables selection of the edition. A second user may select the link on the discovery page in order to view the edition. The link to the edition may be a channel. As shown in NG. 3, the discovery page includes a variety of channels for the second user to choose from.


At operation 1040, the presentation module 240 causes presentation of the discovery page. The presentation of the discovery page may be on a second client device belonging to the second user (e.g., the discovery page may be broadcast to multiple recipient or target devices). In some instances, the discovery page is a feature of a messaging application installed on the second client device.


At operation 1050, the presentation module 240 causes presentation of the content. In some cases, the presentation module 240 is further configured to cause presentation of the multiple layers of the content (e.g., the first layer and the second layer) on the screen of the second client device belonging to the second user. Once displayed on the screen of the second client device, the second user may select the first layer to cause display of the second layer of the content. For instance, the second user may swipe vertically on the first layer displayed on the screen of the second client device. Alternatively, the second user may tap on the first layer displayed on the screen of the second client device. In some instances, the first layer indicates an area where the second user may tap in order to cause display of the second layer.


In some instances, the presentation module 240 simultaneously displays a video clip and one or more images. For example, the presentation module 240 may cause the one or more images (e.g., the PNG images described above) to be displayed over the video clip as the video clip is being played. Referring back to FIG. 5, the video clip 520 of the rock music video includes an image (e.g., image 525) that is overlaid on top of the video clip 520.


As shown in FIG. 11, the method 1000 may include one or more of sub-operations 1110, 1120, 1130, and 1140. Operations 1110 and 1120 may be performed as part of operation 1020. Further, operation 1130 may be performed as part of operation 1030 and operation 1140 may be performed prior to operation 1040.


At operation 1110, the group module 220 may be used to associate a time setting to the content. For example, a user may be able to identify a time span, after which the content may no longer be ‘current’ or accessible. Thus, the content included within the edition may correspond to a certain day of the week. As shown in FIGS. 4-8, each piece of content displayed in the user interfaces is indicated as being part of an edition that corresponds to a day of the week (e.g., Monday).


At operation 1120, the group module 220 generates a collection of stories depicted by the content, which may be displayed on a client device. As an example, the pieces of content displayed in FIGS. 3-8 may represent a story within the collection of stories. Further, a user may be able to navigate through each story by performing a tapping or swiping gesture.


At operation 1130, the generation module 230 generates an icon of a channel selectable to result in the presentation of the edition. Moreover, the icon of the channel may include a thumbnail image related to the source of the content included in the edition. Also, as shown in FIG. 3, the generated icon may be displayed on a discovery page. Although FIG. 3 does not specifically show thumbnail images, each of the icons may include a thumbnail image that is related to the source of the content for the edition. In some cases, when all the editions from a channel have been viewed on a client device, the icon of the channel will change appearance (e.g., an animation that modifies a color of the icon).


At operation 1140, the reception module 210 receives a scheduled time, which may be used to indicate when the edition is available on the discovery page. In some instances, the scheduled time is indicated by a user from the upload page. For example, as shown in FIG. 9, the upload page includes a section where a user may input the scheduled time.


As shown in FIG. 12, the method 1000 may include one or more of sub-operations 1210, 1220, 1230, and 1240. Operations 1210 and 1220 may be performed prior to operation 1010. Further, operations 1230 and 1240 may be performed after operation 1050.


At operation 1210, the reception module 210 receives a request to upload content to the edition. The request may be received from the first client device. For instance, client application software installed on the first client device may include a feature which sends a request to the reception module 210. Moreover, the operation 1020 may be performed based on the request received at operation 1210.


At operation 1220, the presentation module 240 causes presentation of an upload page. The upload page enables the user of the first client device to upload the content that depicts the one or more stories. For instance, the upload page may include a section where a user can drag content to indicate that the content is to be uploaded into the edition. From the upload page, the user is also able, to upload multiple layers (e.g., a first layer and a second layer) of content that depicts one or more stories which will be grouped into the edition. For example, the user may upload each layer by dragging the content (e.g., a file that includes the content) into the appropriate section of the upload page. Further, the user may continue dragging multiple files until the multiple layers of the content have been uploaded.


At operation 1230, the reception module 210 receives an annotated version of the content that depicts the one or more stories. For instance, the reception module 210 receives the annotated version of the content from the second client device. The second user of the second device, upon viewing the content displayed on the second client device, may alter or edit the content. Subsequently, the second user may send the annotated content from the second client device to the reception module 210. As an example, as shown in FIG. 8, the annotated version of the content may include an additional description that is added to the original content. Further, the reception module 210 receives an indication to share the annotated version of the content with a third client device. For example, the second user of the second client device may double tap on the screen of the second device in order to share the annotated version of the content.


At operation 1240, the presentation module 240 causes presentation of the annotated version of the content. The presentation module 240 may cause presentation of the annotated version of the content on the third client device as indicated at operation 1230. Further, the annotated version of the content is viewable on the third client device for a predefined duration of time. This is because the annotated version of the content can be an ephemeral message that is available for the predefined duration of time. For example, as shown in FIG. 8, the annotated content includes a timer which indicates an amount of time remaining before the annotated content disappears.


Modules, Components, and Logic


Certain embodiments are described herein as including logic or a number of components, modules, or mechanisms. Modules can constitute either software modules (e.g., code embodied on a machine-readable medium or in a transmission signal) or hardware modules. A “hardware module” is a tangible unit capable of performing certain operations and can be configured or arranged in a certain physical manner. In various example embodiments, one or more computer systems (e.g., a standalone computer system, a client computer system, or a server computer system) or one or more hardware modules of a computer system (e.g., a processor or a group of processors) is configured by software (e.g., an application or application portion) as a hardware module that operates to perform certain operations as described herein.


In some embodiments, a hardware module is implemented mechanically, electronically, or any suitable combination thereof. For example, a hardware module can include dedicated circuitry or logic that is permanently configured to perform certain operations. For example, a hardware module can be a special-purpose processor, such as a Field-Programmable Gate Array (FPGA) or an Application Specific Integrated Circuit (ASIC). A hardware module may also include programmable logic or circuitry that is temporarily configured by software to perform certain operations. For example, a hardware module can include software encompassed within a general-purpose processor or other programmable processor. It will be appreciated that the decision to implement a hardware module mechanically, in dedicated and permanently configured circuitry, or in temporarily configured circuitry (e.g., configured by software) can be driven by cost and time considerations.


Accordingly, the phrase “hardware module” should be understood to encompass a tangible entity, be that an entity that is physically constructed, permanently configured (e.g., hardwired), or temporarily configured (e.g., programmed) to operate in a certain manner or to perform certain operations described herein. As used herein, “hardware-implemented module” refers to a hardware module. Considering embodiments in which hardware modules are temporarily configured (e.g., programmed), each of the hardware modules need not be configured or instantiated at any one instance in time. For example, where a hardware module comprises a general-purpose processor configured by software to become a special-purpose processor, the general-purpose processor may be configured as respectively different special-purpose processors (e.g., comprising different hardware modules) at different times. Software can accordingly configure a particular processor or processors, for example, to constitute a particular hardware module, at one instance of time and to constitute a different hardware module at a different instance of time.


Hardware modules can provide information to, and receive information from, other hardware modules. Accordingly, the described hardware modules can be regarded as being communicatively coupled. Where multiple hardware modules exist contemporaneously, communications can be achieved through signal transmission (e.g., over appropriate circuits and buses) between or among two or more of the hardware modules. In embodiments in which multiple hardware modules are configured or instantiated at different times, communications between such hardware modules may be achieved, for example, through the storage and retrieval of information in memory structures to which the multiple hardware modules have access. For example, one hardware module performs an operation and stores the output of that operation in a memory device to which it is communicatively coupled. A further hardware module can then, at a later time, access the memory device to retrieve and process the stored output. Hardware modules can also initiate communications with input or output devices, and can operate on a resource (e.g., a collection of information).


The various operations of example methods described herein can be performed, at least partially, by one or more processors that are temporarily configured (e.g., by software) or permanently configured to perform the relevant operations. Whether temporarily or permanently configured, such processors constitute processor-implemented modules that operate to perform one or more operations or functions described herein. As used herein, “processor-implemented module” refers to a hardware module implemented using one or more processors.


Similarly, the methods described herein can be at least partially processor-implemented, with a particular processor or processors being an example of hardware. For example, at least some of the operations of a method can be performed by one or more processors or processor-implemented modules. Moreover, the one or more processors may also operate to support performance of the relevant operations in a “cloud computing” environment or as a “software as a service” (SaaS). For example, at least some of the operations may be performed by a group of computers (as examples of machines including processors), with these operations being accessible via a network (e.g., the Internet) and via one or more appropriate interfaces (e.g., an Application Program Interface (API)).


The performance of certain of the operations may be distributed among the processors, not only residing within a single machine, but deployed across a number of machines. In some example embodiments, the processors or processor-implemented modules are located in a single geographic location (e.g., within a home environment, an office environment, or a server farm). In other example embodiments, the processors or processor-implemented modules are distributed across a number of geographic locations.


Applications



FIG. 13 illustrates an example mobile device 1300 executing a mobile operating system (e.g., IOS™, ANDROID™, WINDOWS® Phone, or other mobile operating systems), consistent with some embodiments. In one embodiment, the mobile device 1300 includes a touch screen operable to receive tactile data from a user 1302. For instance, the user 1302 may physically touch 1304 the mobile device 1300, and in response to the touch 1304, the mobile device 1300 may determine tactile data such as touch location, touch force, or gesture motion. In various example embodiments, the mobile device 1300 displays a home screen 1306 (e.g., Springboard on IOS™) operable to launch applications or otherwise manage various aspects of the mobile device 1300. In some example embodiments, the home screen 1306 provides status information such as battery life, connectivity, or other hardware statuses. The user 1302 can activate user interface elements by touching an area occupied by a respective user interface element. In this manner, the user 1302 interacts with the applications of the mobile device 1300. For example, touching the area occupied by a particular icon included in the home screen 1306 causes launching of an application corresponding to the particular icon.


Many varieties of applications (also referred to as “apps”) can be executing on the mobile device 1300, such as native applications (e.g., applications programmed in Objective-C, Swift, or another suitable language running on IOS™, or applications programmed in Java running on ANDROID™), mobile web applications (e.g., applications written in Hypertext Markup Language-5 (HTML5)), or hybrid applications (e.g., a native shell application that launches an HTML5 session). For example, the mobile device 1300 includes a messaging app, an audio recording app, a camera app, a book reader app, a media app, a fitness app, a file management app, a location app, a browser app, a settings app, a contacts app, a telephone call app, or other apps (e.g., gaming apps, social networking apps, biometric monitoring apps). In another example, the mobile device 1300 includes a social messaging app 1308 such as SNAPCHAT® that, consistent with some embodiments, allows users to exchange ephemeral messages that include media content. In this example, the social messaging app 1308 can incorporate aspects of embodiments described herein.


Software Architecture



FIG. 14 is a block diagram 1400 illustrating an architecture of software 1402, which can be installed on any one or more of the devices described above. FIG. 14 is merely a non-limiting example of a software architecture, and it will be appreciated that many other architectures can be implemented to facilitate the functionality described herein. In various embodiments, the software 1402 is implemented by hardware such as machine a 1500 of FIG. 15 that includes processors 1510, memory 1530, and I/O components 1550. In this example architecture, the software 1402 can be conceptualized as a stack of layers where each layer may provide a particular functionality. For example, the software 1402 includes layers such as an operating system 1404, libraries 1406, frameworks 1408, and applications 1410. Operationally, the applications 1410 invoke application programming interface (API) calls 1412 through the software stack and receive messages 1414 in response to the API calls 1412, consistent with some embodiments.


In various implementations, the operating system 1404 manages hardware resources and provides common services. The operating system 1404 includes, for example, a kernel 1420, services 1422, and drivers 1424. The kernel 1420 acts as an abstraction layer between the hardware and the other software layers consistent with some embodiments. For example, the kernel 1420 provides memory, management, processor management (e.g., scheduling), component management, networking, and security settings, among other functionality. The services 1422 can provide other common services for the other software layers. The drivers 1424 are responsible for controlling or interfacing with the underlying hardware, according to some embodiments. For instance, the drivers 1424 can include display drivers, camera drivers, BLUETOOTH® drivers, flash memory drivers, serial communication drivers (e.g., Universal Serial Bus (USB) drivers), WI-FI® drivers, audio drivers, power management drivers, and so forth.


In some embodiments, the libraries 1406 provide a low-level common infrastructure utilized by the applications 1410. The libraries 1406 can include system libraries 1430 (e.g., C standard library) that can provide functions such as memory allocation functions, string manipulation functions, mathematic functions, and the like. In addition, the libraries 1406 can include API libraries 1432 such as media libraries (e.g., libraries to support presentation and manipulation of various media formats such as Moving Picture Experts Group-4 (MPEG4), Advanced Video Coding (H.264 or AVC), Moving Picture Experts Group Layer-3 (MP3), Advanced Audio Coding (AAC), Adaptive Multi-Rate (AMR) audio codec, Joint Photographic Experts Group (JPEG or JPG), or Portable Network Graphics (PNG)), graphics libraries (e.g., an OpenGL framework used to render in two dimensions (2D) and three dimensions (3D) in a graphic content on a display), database libraries (e.g., SQLite to provide various relational database functions), web libraries (e.g., WebKit to provide web browsing functionality), and the like. The libraries 1406 can also include a wide variety of other libraries 1434 to provide many other APIs to the applications 1410.


The frameworks 1408 provide a high-level common infrastructure that can be utilized by the applications 1410, according to some embodiments. For example, the frameworks 1408 provide various graphic user interface (GUI) functions, high-level resource management, high-level location services, and so forth. The frameworks 1408 can provide a broad spectrum of other APIs that can be utilized by the applications 1410, some of which may be specific to a particular operating system or platform.


In an example embodiment, the applications 1410 include a home application 1450, a contacts application 1452, a browser application 1454, a book reader application 1456, a location application 1458, a media application 1460, a messaging application 1462, a game application 1464, and a broad assortment of other applications such as a third party application 1466. According to some embodiments, the applications 1410 are programs that execute functions defined in the programs. Various programming languages can be employed to create one or more of the applications 1410, structured in a variety of manners, such as object-oriented programming languages (e.g., Objective-C, Java, or C++) or procedural programming languages (e.g., C or assembly language). In a specific example, the third party application 1466 (e.g., an application developed using the ANDROID™ or IOS™ software development kit (SDK) by an entity other than the vendor of the particular platform) may be mobile software running on a mobile operating system such as IOS™, ANDROID™, WINDOWS® PHONE, or another mobile operating systems. In this example, the third party application 1466 can invoke the API calls 1412 provided by the operating system 1404 to facilitate functionality described herein.


Example Machine Architecture and Machine-Readable Medium



FIG. 15 is a block diagram illustrating components of a machine 1500, according to some embodiments, able to read instructions from a machine-readable medium (e.g., a machine-readable storage medium) and perform any one or more of the methodologies discussed herein. Specifically, FIG. 15 shows a diagrammatic representation of the machine 1500 in the example form of a computer system, within which instructions 1516 (e.g., software, a program, an application, an applet, an app, or other executable code) for causing the machine 1500 to perform any one or more of the methodologies discussed herein can be executed. In alternative embodiments, the machine 1500 operates as a standalone device or can be coupled (e.g., networked) to other machines. In a networked deployment, the machine 1500 may operate in the capacity of a server machine or a client machine in a server-client network environment, or as a peer machine in a peer-to-peer (or distributed) network environment. The machine 1500 can comprise, but not be limited to, a server computer, a client computer, a personal computer (PC), a tablet computer, a laptop computer, a netbook, a set-top box (STB), a personal digital assistant (PDA), an entertainment media system, a cellular telephone, a smart phone, a mobile device, a wearable device (e.g., a smart watch), a smart home device (e.g., a smart appliance), other smart devices, a web appliance, a network router, a network switch, a network bridge, or any machine capable of executing the instructions 1516, sequentially or otherwise, that specify actions to be taken by the machine 1500. Further, while only a single machine 1500 is illustrated, the term “machine” shall also be taken to include a collection of machines 1500 that individually or jointly execute the instructions 1516 to perform any one or more of the methodologies discussed herein.


In various embodiments, the machine 1500 comprises processors 1510, memory 1530, and I/O components 1550, which can be configured to communicate with each other via a bus 1502. In an example embodiment, the processors 1510 (e.g., a Central Processing Unit (CPU), a Reduced Instruction Set Computing (RISC) processor, a Complex Instruction Set Computing (CISC) processor, a Graphics Processing Unit (GPU), a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Radio-Frequency Integrated Circuit (RTIC), another processor, or any suitable combination thereof) includes, for example, a processor 1512 and a processor 1514 that may execute the instructions 1516. The term “processor” is intended to include multi-core processors that may comprise two or more independent processors (also referred to as “cores”) that can execute instructions contemporaneously. Although FIG. 15 shows multiple processors, the machine 1500 may include a single processor with a single core, a single processor with multiple cores (e.g., a multi-core processor), multiple processors with a single core, multiple processors with multiples cores, or any combination thereof.


The memory 1530 comprises a main memory 1532, a static memory 1534, and a storage unit 1536 accessible to the processors 1510 via the bus 1502, according to some embodiments. The storage unit 1536 can include a machine-readable medium 1538 on which are stored the instructions 1516 embodying any one or more of the methodologies or functions described herein. The instructions 1516 can also reside, completely or at least partially, within the main memory 1532, within the static memory 1534, within at least one of the processors 1510 (e.g., within the processor's cache memory), or any suitable combination thereof, during execution thereof by the machine 1500. Accordingly, in various embodiments, the main memory 1532, the static memory 1534, and the processors 1510 are considered machine-readable media 1538.


As used herein, the term “memory” refers to a machine-readable medium 1538 able to store data temporarily or permanently and may be taken to include, but not be limited to, random-access memory (RAM), read-only memory (ROM), buffer memory, flash memory, and cache memory. While the machine-readable medium 1538 is shown in an example embodiment to be a single medium, the term “machine-readable medium” should be taken to include a single medium or multiple media (e.g., a centralized or distributed database, or associated caches and servers) able to store the instructions 1516. The term “machine-readable medium” shall also be taken to include any medium, or combination of multiple media, that is capable of storing instructions (e.g., instructions 1516) for execution by a machine (e.g., machine 1500), such that the instructions, when executed by one or more processors of the machine 1500 (e.g., processors 1510), cause the machine 1500 to perform any one or more of the methodologies described herein. Accordingly, a “machine-readable medium” refers to a single storage apparatus or device, as well as “cloud-based” storage systems or storage networks that include multiple storage apparatus or devices. The term “machine-readable medium” shall accordingly be taken to include, but not be limited to, one or more data repositories in the form of a solid-state memory (e.g., flash memory), an optical medium, a magnetic medium, other non-volatile memory (e.g., Erasable Programmable Read-Only Memory (EPROM)), or any suitable combination thereof. The term “machine-readable medium” specifically excludes non-statutory signals per se.


The I/O components 1550 include a wide variety of components to receive input, provide output, produce output, transmit information, exchange information, capture measurements, and so on. In general, it will be appreciated that the I/O components 1550 can include many other components that are not shown in FIG. 15. The I/O components 1550 are grouped according to functionality merely for simplifying the following discussion, and the grouping is in no way limiting. In various example embodiments, the I/O components 1550 include output components 1552 and input components 1554. The output components 1552 include visual components (e.g., a display such as a plasma display panel (PDP), a light emitting diode (LED) display, a liquid crystal display (LCD), a projector, or a cathode ray tube (CRT)), acoustic components (e.g., speakers), haptic components (e.g., a vibratory motor), other signal generators, and so forth. The input components 1554 include alphanumeric input components (e.g., a keyboard, a touch screen configured to receive alphanumeric input, a photo-optical keyboard, or other alphanumeric input components), point based input components (e.g., a mouse, a touchpad, a trackball, a joystick, a motion sensor, or other pointing instruments), tactile input components (e.g., a physical button, a touch screen that provides location and force of touches or touch gestures, or other tactile input components), audio input components (e.g., a microphone), and the like.


In some further example embodiments, the I/O components 1550 include biometric components 1556, motion components 1558, environmental components 1560, or position components 1562, among a wide array of other components. For example, the biometric components 1556 include components to detect expressions (e.g., hand expressions, facial expressions, vocal expressions, body gestures, or eye tracking), measure biosignals (e.g., blood pressure, heart rate, body temperature, perspiration, or brain waves), identify a person (e.g., voice identification, retinal identification, facial identification, fingerprint identification, or electroencephalogram based identification), and the like. The motion components 1558 include acceleration sensor components (e.g., accelerometer), gravitation sensor components, rotation sensor components (e.g., gyroscope), and so forth. The environmental components 1560 include, for example, illumination sensor components (e.g., photometer), temperature sensor components (e.g., one or more thermometers that detect ambient temperature), humidity sensor components, pressure sensor components (e.g., barometer), acoustic sensor components (e.g., one or more microphones that detect background noise), proximity sensor components (e.g., infrared sensors that detect nearby objects), gas sensor components (e.g., machine olfaction detection sensors, gas detection sensors to detect concentrations of hazardous gases for safety or to measure pollutants in the atmosphere), or other components that may provide indications, measurements, or signals corresponding to a surrounding physical environment. The position components 1562 include location sensor components (e.g., a Global Positioning System (GPS) receiver component), altitude sensor components (e.g., altimeters or barometers that detect air pressure from which altitude may be derived), orientation sensor components (e.g., magnetometers), and the lite.


Communication can be implemented using a wide variety of technologies. The I/O components 1550 may include communication components 1564 operable to couple the machine 1500 to a network 1580 or devices 1570 via a coupling 1582 and a coupling 1572, respectively. For example, the communication components 1564 include a network interface component or another suitable device to interface with the network 1580. In further examples, communication components 1564 include wired communication components, wireless communication components, cellular communication components, Near Field Communication (NFC) components, BLUETOOTH® components (e.g., BLUETOOTH® Low Energy), WI-FI® components, and other communication components to provide communication via other modalities. The devices 1570 may be another machine or any of a wide variety of peripheral devices (e.g., a peripheral device coupled via a Universal Serial Bus (USB)).


Moreover, in some embodiments, the communication components 1564 detect identifiers or include components operable to detect identifiers. For example, the communication components 1564 include Radio Frequency Identification (RFID) tag reader components, NFC smart tag detection components, optical reader components (e.g., an optical sensor to detect a one-dimensional bar codes such as a Universal Product Code (UPC) bar code, multi-dimensional bar codes such as a Quick Response (QR) code, Aztec Code, Data Matrix, Dataglyph, MaxiCode, PDF417, Ultra Code, Uniform Commercial Code Reduced Space Symbology (UCC RSS)-2D bar codes, and other optical codes), acoustic detection components (e.g., microphones to identify tagged audio signals), or any suitable combination thereof. In addition, a variety of information can be derived via the communication components 1564, such as location via Internet Protocol (IP) geolocation, location via WI-FI® signal triangulation, location via detecting a. BLUETOOTH® or NEC beacon signal that may indicate a particular location, and so forth.


Transmission Medium


In various example embodiments, one or more portions of the network 1580 can be an ad hoc network, an intranet, an extranet, a virtual private network (VPN), a local area network (LAN), a wireless LAN (WLAN), a wide area network (WAN), a wireless WAN (WWAN), a metropolitan area network (MAN), the Internet, a portion of the Internet, a portion of the Public Switched Telephone Network (PSTN), a plain old telephone service (POTS) network, a cellular telephone network, a wireless network, a WI-FI® network, another type of network, or a combination of two or more such networks. For example, the network 1580 or a portion of the network 1580 may include a wireless or cellular network, and the coupling 1582 may be a Code Division Multiple Access (CDMA) connection, a. Global System for Mobile communications (GSM) connection, or another type of cellular or wireless coupling. In this example, the coupling 1582 can implement any of a variety of types of data transfer technology, such as Single Carrier Radio Transmission Technology (1×RTT), Evolution-Data Optimized (EVDO) technology, General Packet Radio Service ((CPRS) technology, Enhanced Data rates for GSM Evolution (EDGE) technology, third Generation Partnership Project (3GPP) including 3G, fourth generation wireless (4G) networks, Universal Mobile Telecommunications System (UMTS), High Speed Packet Access (HSPA), Worldwide Interoperability for Microwave Access (WiMAX), Long Term Evolution (LTE) standard, others defined by various standard-setting organizations, other long range protocols, or other data transfer technology.


In example embodiments, the instructions 1516 are transmitted or received over the network 1580 using a transmission medium via a network interface device (e.g., a network interface component included in the communication components 1564) and utilizing any one of a number of well-known transfer protocols (e.g., Hypertext Transfer Protocol (HTTP)). Similarly, in other example embodiments, the instructions 1516 are transmitted or received using a transmission medium via the coupling 1572 (e.g., a peer-to-peer coupling) to the devices 1570. The term “transmission medium” shall be taken to include any intangible medium that is capable of storing, encoding, or carrying the instructions 1516 for execution by the machine 1500, and includes digital or analog communications signals or other intangible media to facilitate communication of such software.


Furthermore, the machine-readable medium 1538 is non-transitory (in other words, not having any transitory signals) in that it does not embody a propagating signal. However, labeling the machine-readable medium 1538 “non-transitory” should not be construed to mean that the medium is incapable of movement; the medium should be considered as being transportable from one physical location to another. Additionally, since the machine-readable medium 1538 is tangible, the medium may be considered to be a machine-readable device.


Language


Throughout this specification, plural instances may implement components, operations, or structures described as a single instance. Although individual operations of one or more methods are illustrated and described as separate operations, one or more of the individual operations may be performed concurrently, and nothing requires that the operations be performed in the order illustrated. Structures and functionality presented as separate components in example configurations may be implemented as a combined structure or component. Similarly, structures and functionality presented as a single component may be implemented as separate components. These and other variations, modifications, additions, and improvements fall within the scope of the subject matter herein.


Although an overview of the inventive subject matter has been described with reference to specific example embodiments, various modifications and changes may be made to these embodiments without departing from the broader scope of embodiments of the present disclosure. Such embodiments of the inventive subject matter may be referred to herein, individually or collectively, by the term “invention” merely for convenience and without intending to voluntarily, limit the scope of this application to any single disclosure or inventive concept if more than one is, in fact disclosed.


The embodiments illustrated herein are described in sufficient detail to enable those skilled in the art to practice the teachings disclosed. Other embodiments may be used and derived therefrom, such that structural and logical substitutions and changes may be made without departing from the scope of this disclosure. The Detailed Description, therefore, is not to be taken in a limiting sense, and the scope of various embodiments is defined only by the appended claims, along with the full range of equivalents to which such claims are entitled.


As used herein, the term “or” may be construed in either an inclusive or exclusive sense. Moreover, plural instances may be provided for resources, operations, or structures described herein as a single instance. Additionally, boundaries between various resources, operations, modules, engines, and data stores are somewhat arbitrary, and particular operations are illustrated in a context of specific illustrative configurations. Other allocations of functionality are envisioned and may fall within a scope of various embodiments of the present disclosure. In general, structures and functionality presented as separate resources in the example configurations may be implemented as a combined structure or resource. Similarly, structures and functionality presented as a single resource may be implemented as separate resources. These and other variations, modifications, additions, and improvements fall within a scope of embodiments of the present disclosure as represented by the appended claims. The specification and drawings are, accordingly, to be regarded in an illustrative rather than a restrictive sense.

Claims
  • 1. A method comprising: receiving a first plurality of content items from a first user account of a social networking system, the receiving comprising receiving multiple files each corresponding to a different layer of content, a first file of the multiple files received from the first user account comprising a first content item of the first plurality of content items being a first video including video content, and a second file of the multiple files received from the first user account comprising shortened video content of the first video, the second file being received separately from the first file from the first user account;generating a first layer of content that includes the second video and a second layer of content that includes the first video, the first layer of content being configured to display the second layer of content in response to receiving a particular user input;causing display of the first layer of content on a client device;causing display of an indication of an area of the display that is configured to receive touch input to cause display of the second layer;receiving, from the client device, a user interaction with the first layer of content, the user interaction comprising touch input at the area of the display;responsive to the receiving of the user interaction, causing display of the second layer of content; andpresenting, together with the second layer of content, a visual indication of an amount of time remaining before an annotation associated with the second layer of content is deleted.
  • 2. The method of claim 1, wherein the second video includes less than all of the video content of the first video, further comprising: receiving a second plurality of content items from a second user account of the social networking system;generating a first graphic in response to the reception of the first plurality of content items received from the first user account;generating a second graphic in response to the reception of the second plurality of content items received from the second user account;generating a graphical user interface to include the first graphic;configuring the graphical user interface to display content from the first user account upon selection of the first graphic based, in part, on the first plurality of content items being received from the first user account;further generating the graphical user interface to include the second graphic;configuring the graphical user interface to display content from the second user account upon selection of the second graphic based, in part, on the second plurality of content items being received from the second user account; andcausing presentation, via a server of the social networking system, of the graphical user interface in a social networking client application of a third user account.
  • 3. The method of claim 1, further comprising: receiving, from a first user of the client device, the annotation to the second layer of content displayed; andmodifying the second layer of content based on the received annotation, wherein the annotated second layer of content is presented to a second user of another client device.
  • 4. The method of claim 1, further comprising: receiving a plurality of content items from a plurality of additional user accounts of the social networking system;generating a corresponding plurality of graphics corresponding to each of the additional user accounts; andgenerating a graphical user interface to display each of the plurality of graphics, each of the plurality of graphics configured to display content from the corresponding user account in response to a selection of the graphic.
  • 5. The method of claim 1, further comprising: grouping the first plurality of content items based on the first plurality of content items being received within a predetermined window of elapsed time.
  • 6. The method of claim 1, further comprising: receiving additional content from a device of a provider; andgenerating a link in a generated home page that corresponds to the additional content.
  • 7. The method of claim 6, further comprising: generating a thumbnail image in the generated home page that corresponds to the provider, an appearance of the thumbnail image being subject to change based on display of the additional content.
  • 8. The method of claim 1, wherein: the first layer of content is displayed on a screen of the client device; andthe user interaction includes a swipe across the first layer of content displayed on the screen of the client device.
  • 9. The method of claim 1, further comprising receiving third and fourth files of the multiple files from the first user account, the third file comprising an image or video associated with a travel destination, and the fourth file comprising descriptive information about the travel destination.
  • 10. The method of claim 1, wherein the second video comprises a video clip corresponding to the first video, further comprising overlaying one or more images on top of the video clip as the video clip is being played in the first layer of content.
  • 11. The method of claim 1, further comprising: causing display of a plurality of indicators of different subsets of content, wherein the first layer of content corresponding to a first subset of the different subsets of content is displayed in response to receiving input that selects a given indicator of the plurality of indicators that is associated with the first video, and wherein the given indicator is animated to change an appearance of the given indicator in response to determining that all content associated with the first subset has been.
  • 12. The method of claim 11, wherein the annotated second layer of content is available for presentation to the second user for a predefined duration of time.
  • 13. The method of claim 9, further comprising: generating a third layer of content that includes the image or video associated with the travel destination,receiving, from the client device, a user interaction with the third layer of content; andresponsive to the receiving of the user interaction with the third layer of content, causing display of a fourth layer of content that includes a plurality of description comprising the descriptive information about the travel destination.
  • 14. The method of claim 1, further comprising: causing an upload page to be presented to a user associated with the first user account;receiving, via the upload page, the first file in response receiving first input from the user that drags the first file into a first section of the upload page corresponding to the second layer, andreceiving, via the upload page, the second file in response receiving second input from the user that drags the second file into a second section of the upload page corresponding to the first layer.
  • 15. The method of claim 13, wherein the plurality of descriptions in the fourth layer of content comprising a story about the travel destination, one or more events at the travel destination, and food associated with the travel destination.
  • 16. The method of claim 1, further comprising: receiving a scheduled time for presenting the first layer of content from a first user associated with the first user account; andcausing presentation of the first layer of content to a second user, associated with a second user account, during the scheduled time.
  • 17. The method of claim 1, further comprising: causing display of a plurality of indicators of different subsets of content, wherein the first layer of content is displayed in response to receiving input that selects a given indicator of the plurality of indicators that is associated with the first video; andcausing display of a selectable link in the second layer of content together with the first video, wherein the plurality of indicators are displayed in response detecting activation of the selectable link, and wherein an animation is displayed in response to detecting activation of the selectable link.
  • 18. The method of claim 17, wherein each of the subsets of content is associated with a different day of the week wherein the visual indication comprises a number that indicates the amount of time remaining.
  • 19. A system comprising: one or more hardware processors and executable instructions accessible on a non-transitory computer-readable medium that, when executed, configure the one or more hardware processors to perform operations comprising:receiving a first plurality of content items from a first user account of a social networking system, the receiving comprising receiving multiple files each corresponding to a different layer of content, a first file of the multiple files received from the first user account comprising a first content item of the first plurality of content items being a first video including video content, and a second file of the multiple files received from the first user account comprising shortened video content of the first video, the second file being received separately from the first file from the first user account;generating a first layer of content that includes the second video and a second layer of content that includes the first video, the first layer of content being configured to display the second layer of content in response to receiving a particular user input;causing display of the first layer of content on a client device;causing display of an indication of an area of the display that is configured to receive touch input to cause display of the second layer;receiving, from the client device, a user interaction with the first layer of content, the user interaction comprising touch input at the area of the display; andresponsive to the receiving of the user interaction, causing display of the second layer of content; andpresenting, together with the second layer of content, a visual indication of an amount of time remaining before an annotation associated with the second layer of content is deleted.
  • 20. A non-transitory machine-readable medium storing instructions that, when executed by one or more processors of a machine, cause the machine to perform operations comprising: receiving a first plurality of content items from a first user account of a social networking system, the receiving comprising receiving multiple files each corresponding to a different layer of content, a first file of the multiple files received from the first user account comprising a first content item of the first plurality of content items being a first video including video content, and a second file of the multiple files received from the first user account comprising a second video corresponding to shortened video content of the first video, the second file being received separately from the first file from the first user account;generating a first layer of content that includes the second video and a second layer of content that includes the first video, the first layer of content being configured to display the second layer of content in response to receiving a particular user input;causing display of the first layer of content on a client device;causing display of an indication of an area of the display that is configured to receive touch input to cause display of the second layer;receiving, from the client device, a user interaction with the first layer of content, the user interaction comprising touch input at the area of the display; andresponsive to the receiving of the user interaction, causing display of the second layer of content; andpresenting, together with the second layer of content, a visual indication of an amount of time remaining before an annotation associated with the second layer of content is deleted.
RELATED APPLICATION

This application is a continuation of and claims priority of U.S. patent application Ser. No. 15/152,975, filed on May 12, 2016, which is a continuation of and claims priority to U.S. patent application Ser. No. 14/612,692, filed Feb. 3, 2015, which claims the priority benefit of U.S. Provisional Application Ser. No. 62/105,199, filed Jan. 19, 2015, each of which are hereby incorporated herein by reference in their entirety.

US Referenced Citations (933)
Number Name Date Kind
666223 Shedlock Jan 1901 A
4581634 Williams Apr 1986 A
4975690 Torres Dec 1990 A
5072412 Henderson, Jr. et al. Dec 1991 A
5493692 Theimer et al. Feb 1996 A
5713073 Warsta Jan 1998 A
5754939 Herz et al. May 1998 A
5855008 Goldhaber et al. Dec 1998 A
5883639 Walton et al. Mar 1999 A
5999932 Paul Dec 1999 A
6012098 Bayeh et al. Jan 2000 A
6014090 Rosen et al. Jan 2000 A
6029141 Bezos et al. Feb 2000 A
6038295 Mattes Mar 2000 A
6049711 Yehezkel et al. Apr 2000 A
6075535 Fitzhugh et al. Jun 2000 A
6154764 Nitta et al. Nov 2000 A
6158044 Tibbetts Dec 2000 A
6167435 Druckenmiller et al. Dec 2000 A
6204840 Petelycky et al. Mar 2001 B1
6205432 Gabbard et al. Mar 2001 B1
6216141 Straub et al. Apr 2001 B1
6285381 Sawano et al. Sep 2001 B1
6285987 Roth et al. Sep 2001 B1
6290504 Benitz et al. Sep 2001 B1
6310694 Okimoto et al. Oct 2001 B1
6317789 Rakavy et al. Nov 2001 B1
6334149 Davis, Jr. et al. Dec 2001 B1
6349203 Asaoka et al. Feb 2002 B1
6353170 Eyzaguirre et al. Mar 2002 B1
6363380 Dimitrova Mar 2002 B1
6446004 Cao et al. Sep 2002 B1
6449657 Stanbach et al. Sep 2002 B2
6456852 Bar et al. Sep 2002 B2
6484196 Maurille Nov 2002 B1
6487586 Ogilvie et al. Nov 2002 B2
6487601 Hubacher et al. Nov 2002 B1
6499016 Anderson Dec 2002 B1
6523008 Avrunin Feb 2003 B1
6542749 Tanaka et al. Apr 2003 B2
6549768 Fraccaroli Apr 2003 B1
6618593 Drutman et al. Sep 2003 B1
6622174 Ukita et al. Sep 2003 B1
6631463 Floyd et al. Oct 2003 B1
6636247 Hamzy et al. Oct 2003 B1
6636855 Holloway et al. Oct 2003 B2
6643684 Malkin et al. Nov 2003 B1
6658095 Yoakum et al. Dec 2003 B1
6665531 Soderbacka et al. Dec 2003 B1
6668173 Greene Dec 2003 B2
6684238 Dutta Jan 2004 B1
6684257 Camut et al. Jan 2004 B1
6698020 Zigmond et al. Feb 2004 B1
6700506 Winkler Mar 2004 B1
6701347 Ogilvie Mar 2004 B1
6711608 Ogilvie Mar 2004 B1
6720860 Narayanaswami Apr 2004 B1
6724403 Santoro et al. Apr 2004 B1
6757713 Ogilvie et al. Jun 2004 B1
6832222 Zimowski Dec 2004 B1
6834195 Brandenberg et al. Dec 2004 B2
6836792 Chen Dec 2004 B1
6898626 Ohashi May 2005 B2
6959324 Kubik et al. Oct 2005 B1
6970088 Kovach Nov 2005 B2
6970907 Ullmann et al. Nov 2005 B1
6980909 Root et al. Dec 2005 B2
6981040 Konig et al. Dec 2005 B1
7004394 Kim Feb 2006 B2
7020494 Spriestersbach et al. Mar 2006 B2
7027124 Foote et al. Apr 2006 B2
7072963 Anderson et al. Jul 2006 B2
7085571 Kalhan et al. Aug 2006 B2
7110744 Freeny, Jr. Sep 2006 B2
7124164 Chemtob Oct 2006 B1
7142823 Logue et al. Nov 2006 B1
7149893 Leonard et al. Dec 2006 B1
7173651 Knowles Feb 2007 B1
7188143 Szeto Mar 2007 B2
7203380 Chiu et al. Apr 2007 B2
7206568 Sudit Apr 2007 B2
7227937 Yoakum et al. Jun 2007 B1
7237002 Estrada et al. Jun 2007 B1
7240089 Boudreau Jul 2007 B2
7243163 Friend et al. Jul 2007 B1
7254585 Frieden et al. Aug 2007 B2
7269426 Kokkonen et al. Sep 2007 B2
7278168 Chaudhury et al. Oct 2007 B1
7280658 Amini et al. Oct 2007 B2
7315823 Brondrup Jan 2008 B2
7349768 Bruce et al. Mar 2008 B2
7356564 Hartselle et al. Apr 2008 B2
7376715 Cunningham et al. May 2008 B2
7394345 Ehlinger et al. Jul 2008 B1
7411493 Smith Aug 2008 B2
7423580 Markhovsky et al. Sep 2008 B2
7454442 Cobleigh et al. Nov 2008 B2
7478402 Christensen et al. Jan 2009 B2
7496347 Puranik Feb 2009 B2
7508419 Toyama et al. Mar 2009 B2
7512649 Faybishenko et al. Mar 2009 B2
7519670 Hagale et al. Apr 2009 B2
7535890 Rojas May 2009 B2
7546554 Chiu Jun 2009 B2
7607096 Oreizy et al. Oct 2009 B2
7639943 Kalajan Dec 2009 B1
7650231 Gadler Jan 2010 B2
7668537 DeVries Feb 2010 B2
7703140 Nath et al. Apr 2010 B2
7770137 Forbes et al. Aug 2010 B2
7778973 Choi Aug 2010 B2
7779444 Glad Aug 2010 B2
7787886 Markhovsky et al. Aug 2010 B2
7796946 Eisenbach Sep 2010 B2
7801954 Cadiz et al. Sep 2010 B2
7856360 Kramer et al. Dec 2010 B2
7856449 Martino et al. Dec 2010 B1
7912896 Wolovitz et al. Mar 2011 B2
7934156 Forstall et al. Apr 2011 B2
8001204 Burtner et al. Aug 2011 B2
8032586 Challenger et al. Oct 2011 B2
8063797 Sonnabend et al. Nov 2011 B1
8082255 Carlson, Jr. et al. Dec 2011 B1
8090351 Klein Jan 2012 B2
8098904 Ioffe et al. Jan 2012 B2
8099109 Altman et al. Jan 2012 B2
8112716 Kobayashi Feb 2012 B2
8127035 Hood et al. Feb 2012 B1
8131597 Hudetz Mar 2012 B2
8135166 Rhoads et al. Mar 2012 B2
8136028 Loeb et al. Mar 2012 B1
8146001 Reese Mar 2012 B1
8161115 Yamamoto Apr 2012 B2
8161417 Lee Apr 2012 B1
8170957 Richard May 2012 B2
8195203 Tseng Jun 2012 B1
8199747 Rojas et al. Jun 2012 B2
8208943 Petersen Jun 2012 B2
8214443 Hamburg Jul 2012 B2
8234350 Gu et al. Jul 2012 B1
8238947 Lottin et al. Aug 2012 B2
8244593 Klinger et al. Aug 2012 B2
8276092 Narayanan et al. Sep 2012 B1
8279319 Date Oct 2012 B2
8280406 Ziskind et al. Oct 2012 B2
8285199 Hsu et al. Oct 2012 B2
8287380 Nguyen et al. Oct 2012 B2
8301159 Hamynen et al. Oct 2012 B2
8306922 Kunal et al. Nov 2012 B1
8312086 Velusamy et al. Nov 2012 B2
8312097 Siegel et al. Nov 2012 B1
8326315 Phillips et al. Dec 2012 B2
8326327 Hymel et al. Dec 2012 B2
8332475 Rosen et al. Dec 2012 B2
8352494 Badoiu Jan 2013 B1
8352546 Dollard Jan 2013 B1
8369866 Ashley, Jr. et al. Feb 2013 B2
8379130 Forutanpour et al. Feb 2013 B2
8385950 Wagner et al. Feb 2013 B1
8402097 Szeto Mar 2013 B2
8405773 Hayashi et al. Mar 2013 B2
8418067 Cheng et al. Apr 2013 B2
8423409 Rao Apr 2013 B2
8428453 Spiegel et al. Apr 2013 B1
8471914 Sakiyama et al. Jun 2013 B2
8472935 Fujisaki Jun 2013 B1
8494481 Bacco et al. Jul 2013 B1
8510383 Hurley et al. Aug 2013 B2
8527345 Rothschild et al. Sep 2013 B2
8542685 Forbes, Jr. et al. Sep 2013 B2
8554627 Svendsen et al. Oct 2013 B2
8560612 Kilmer et al. Oct 2013 B2
8570907 Garcia, Jr. et al. Oct 2013 B2
8594680 Ledlie et al. Nov 2013 B2
8613089 Holloway et al. Dec 2013 B1
8639648 Koponen et al. Jan 2014 B2
8660358 Bergboer et al. Feb 2014 B1
8660369 Llano et al. Feb 2014 B2
8660793 Ngo et al. Feb 2014 B2
8666152 Ramanathan et al. Mar 2014 B1
8682350 Altman et al. Mar 2014 B2
8686962 Christie Apr 2014 B2
8687021 Bathiche et al. Apr 2014 B2
8718333 Wolf et al. May 2014 B2
8724622 Rojas May 2014 B2
8732168 Johnson May 2014 B2
8744523 Fan et al. Jun 2014 B2
8745132 Obradovich Jun 2014 B2
8761800 Kuwahara Jun 2014 B2
8768876 Shim et al. Jul 2014 B2
8775401 Zhou et al. Jul 2014 B2
8775972 Spiegel Jul 2014 B2
8788680 Naik Jul 2014 B1
8788947 Putz et al. Jul 2014 B2
8790187 Walker et al. Jul 2014 B2
8797415 Arnold Aug 2014 B2
8798646 Wang et al. Aug 2014 B1
8856349 Jain et al. Oct 2014 B2
8874677 Rosen et al. Oct 2014 B2
8886227 Schmidt et al. Nov 2014 B2
8909679 Root et al. Dec 2014 B2
8909714 Agarwal et al. Dec 2014 B2
8909725 Sehn Dec 2014 B1
8914752 Spiegel Dec 2014 B1
8925106 Steiner et al. Dec 2014 B1
8943140 Kothari Jan 2015 B1
8972357 Shim et al. Mar 2015 B2
8995433 Rojas Mar 2015 B2
9015285 Ebsen et al. Apr 2015 B1
9020745 Johnston et al. Apr 2015 B2
9026943 Spiegel May 2015 B1
9037577 Saylor et al. May 2015 B1
9040574 Wang et al. May 2015 B2
9055416 Rosen et al. Jun 2015 B2
9063638 Schrock et al. Jun 2015 B1
9083770 Drose et al. Jul 2015 B1
9094137 Sehn Jul 2015 B1
9098832 Scardino Aug 2015 B1
9100806 Rosen et al. Aug 2015 B2
9100807 Rosen et al. Aug 2015 B2
9113301 Spiegel et al. Aug 2015 B1
9119027 Sharon et al. Aug 2015 B2
9123074 Jacobs et al. Sep 2015 B2
9143382 Bhogal et al. Sep 2015 B2
9143681 Ebsen et al. Sep 2015 B1
9148424 Yang Sep 2015 B1
9152477 Campbell et al. Oct 2015 B1
9159364 Matias et al. Oct 2015 B1
9191776 Root et al. Nov 2015 B2
9204252 Root Dec 2015 B2
9225805 Kujawa et al. Dec 2015 B2
9225897 Sehn et al. Dec 2015 B1
9237202 Sehn Jan 2016 B1
9258459 Hartley Feb 2016 B2
9264463 Rubinstein et al. Feb 2016 B2
9276886 Samaranayake Mar 2016 B1
9294425 Son Mar 2016 B1
9319472 Cathcart et al. Apr 2016 B2
9344606 Hartley et al. May 2016 B2
9385983 Sehn Jul 2016 B1
9396354 Murphy et al. Jul 2016 B1
9407712 Sehn Aug 2016 B1
9407816 Sehn Aug 2016 B1
9417754 Smith Aug 2016 B2
9430783 Sehn Aug 2016 B1
9439041 Parvizi et al. Sep 2016 B2
9443227 Evans et al. Sep 2016 B2
9450907 Pridmore et al. Sep 2016 B2
9459778 Hogeg et al. Oct 2016 B2
9477391 Flynn, III et al. Oct 2016 B2
9482882 Hanover et al. Nov 2016 B1
9482883 Meisenholder Nov 2016 B1
9489661 Evans et al. Nov 2016 B2
9491134 Rosen et al. Nov 2016 B2
9532171 Allen et al. Dec 2016 B2
9537811 Allen et al. Jan 2017 B2
9542422 Duggal et al. Jan 2017 B2
9560006 Prado et al. Jan 2017 B2
9628950 Noeth et al. Apr 2017 B1
9641572 Yeskel et al. May 2017 B1
9652896 Jurgenson et al. May 2017 B1
9659244 Anderton et al. May 2017 B2
9672538 Vaynblat et al. Jun 2017 B1
9674660 Vaynblat et al. Jun 2017 B1
9693191 Sehn Jun 2017 B2
9705831 Spiegel Jul 2017 B2
9710821 Heath Jul 2017 B2
9742713 Spiegel et al. Aug 2017 B2
9785796 Murphy et al. Oct 2017 B1
9802121 Ackley et al. Oct 2017 B2
9825898 Sehn Nov 2017 B2
9843720 Ebsen et al. Dec 2017 B1
9854219 Sehn Dec 2017 B2
9866999 Noeth Jan 2018 B1
9881094 Pavlovskaia Jan 2018 B2
9961520 Brooks et al. May 2018 B2
10133705 Allen et al. Nov 2018 B1
10182311 Sehn Jan 2019 B2
10200813 Allen et al. Feb 2019 B1
10284508 Allen et al. May 2019 B1
10311916 Sehn Jun 2019 B2
10416845 Allen et al. Sep 2019 B1
10438094 Ko et al. Oct 2019 B1
10448201 Sehn et al. Oct 2019 B1
10476830 Allen et al. Nov 2019 B2
10514876 Sehn Dec 2019 B2
10524087 Allen et al. Dec 2019 B1
10572681 Murphy et al. Feb 2020 B1
10580458 Sehn Mar 2020 B2
10616239 Allen et al. Apr 2020 B2
10623891 Sehn et al. Apr 2020 B2
10659914 Allen et al. May 2020 B1
10708210 Allen et al. Jul 2020 B1
10779113 Sehn et al. Sep 2020 B2
10811053 Sehn Oct 2020 B2
10893055 Allen et al. Jan 2021 B2
10944710 Allen et al. Mar 2021 B1
10958608 Allen et al. Mar 2021 B1
10990697 Murphy et al. Apr 2021 B2
11012398 Allen et al. May 2021 B1
11038829 Allen et al. Jun 2021 B1
20010025316 Oh Sep 2001 A1
20010028787 Nomura et al. Oct 2001 A1
20020023101 Kurihara et al. Feb 2002 A1
20020047858 Bayliss et al. Apr 2002 A1
20020047868 Miyazawa Apr 2002 A1
20020078456 Hudson Jun 2002 A1
20020087631 Sharma Jul 2002 A1
20020097257 Miller et al. Jul 2002 A1
20020098850 Akhteruzzaman et al. Jul 2002 A1
20020122659 Mcgrath et al. Sep 2002 A1
20020128047 Gates Sep 2002 A1
20020141378 Bays et al. Oct 2002 A1
20020144154 Tomkow Oct 2002 A1
20020146103 Holt et al. Oct 2002 A1
20020171669 Meron et al. Nov 2002 A1
20030001846 Davis et al. Jan 2003 A1
20030016247 Lai et al. Jan 2003 A1
20030016253 Aoki et al. Jan 2003 A1
20030017823 Mager et al. Jan 2003 A1
20030020623 Cao et al. Jan 2003 A1
20030023874 Prokupets et al. Jan 2003 A1
20030037124 Yamaura et al. Feb 2003 A1
20030052925 Daimon et al. Mar 2003 A1
20030074404 Parker et al. Apr 2003 A1
20030101230 Benschoter et al. May 2003 A1
20030110503 Perkes Jun 2003 A1
20030126215 Udell Jul 2003 A1
20030131362 Jasinschi et al. Jul 2003 A1
20030148773 Spriestersbach et al. Aug 2003 A1
20030163370 Chen et al. Aug 2003 A1
20030164856 Prager et al. Sep 2003 A1
20030210280 Baker et al. Nov 2003 A1
20030217106 Adar et al. Nov 2003 A1
20030217118 Kobayashi et al. Nov 2003 A1
20030229607 Zellweger et al. Dec 2003 A1
20030236823 Patzer et al. Dec 2003 A1
20040027371 Jaeger Feb 2004 A1
20040064429 Hirstius et al. Apr 2004 A1
20040078367 Anderson et al. Apr 2004 A1
20040111467 Willis Jun 2004 A1
20040158739 Wakai et al. Aug 2004 A1
20040185877 Asthana et al. Sep 2004 A1
20040189465 Capobianco et al. Sep 2004 A1
20040199402 Walker et al. Oct 2004 A1
20040203959 Coombes Oct 2004 A1
20040205480 Moore Oct 2004 A1
20040205514 Sommerer et al. Oct 2004 A1
20040215625 Svendsen et al. Oct 2004 A1
20040243531 Dean Dec 2004 A1
20040243688 Wugofski Dec 2004 A1
20050019014 Yoo et al. Jan 2005 A1
20050021444 Bauer et al. Jan 2005 A1
20050022211 Veselov et al. Jan 2005 A1
20050048989 Jung Mar 2005 A1
20050071435 Karstens Mar 2005 A1
20050078804 Yomoda Apr 2005 A1
20050097176 Schatz et al. May 2005 A1
20050102381 Jiang et al. May 2005 A1
20050104976 Currans May 2005 A1
20050114783 Szeto May 2005 A1
20050119936 Buchanan et al. Jun 2005 A1
20050122405 Voss et al. Jun 2005 A1
20050193340 Amburgey et al. Sep 2005 A1
20050193345 Klassen et al. Sep 2005 A1
20050198128 Anderson et al. Sep 2005 A1
20050223066 Buchheit et al. Oct 2005 A1
20050288954 McCarthy et al. Dec 2005 A1
20060004630 Criddle et al. Jan 2006 A1
20060026067 Nicholas et al. Feb 2006 A1
20060107297 Toyama et al. May 2006 A1
20060109238 Lau et al. May 2006 A1
20060114338 Rothschild Jun 2006 A1
20060119882 Harris et al. Jun 2006 A1
20060127054 Matsuyama Jun 2006 A1
20060242239 Morishima et al. Oct 2006 A1
20060242554 Gerace et al. Oct 2006 A1
20060252438 Ansamaa et al. Nov 2006 A1
20060265417 Amato Nov 2006 A1
20060270419 Crowley et al. Nov 2006 A1
20060282819 Graham Dec 2006 A1
20060287878 Wadhwa et al. Dec 2006 A1
20070003221 Hamada et al. Jan 2007 A1
20070004426 Pfleging et al. Jan 2007 A1
20070028183 Ostojic et al. Feb 2007 A1
20070038715 Collins et al. Feb 2007 A1
20070040931 Nishizawa Feb 2007 A1
20070064899 Boss et al. Mar 2007 A1
20070073517 Panje Mar 2007 A1
20070073823 Cohen et al. Mar 2007 A1
20070075898 Markhovsky et al. Apr 2007 A1
20070082707 Flynt et al. Apr 2007 A1
20070106706 Ahrens et al. May 2007 A1
20070136228 Petersen Jun 2007 A1
20070192128 Celestini Aug 2007 A1
20070198340 Lucovsky et al. Aug 2007 A1
20070198495 Buron et al. Aug 2007 A1
20070208751 Cowan et al. Sep 2007 A1
20070210936 Nicholson Sep 2007 A1
20070214180 Crawford Sep 2007 A1
20070214216 Carrer et al. Sep 2007 A1
20070233556 Koningstein Oct 2007 A1
20070233801 Eren et al. Oct 2007 A1
20070233859 Zhao et al. Oct 2007 A1
20070243887 Bandhole et al. Oct 2007 A1
20070244750 Grannan et al. Oct 2007 A1
20070250791 Halliday et al. Oct 2007 A1
20070255456 Funayama Nov 2007 A1
20070268988 Hedayat et al. Nov 2007 A1
20070281690 Altman et al. Dec 2007 A1
20070294735 Luo Dec 2007 A1
20070299807 Lea et al. Dec 2007 A1
20080005240 Knighton et al. Jan 2008 A1
20080012987 Hirata et al. Jan 2008 A1
20080022329 Glad Jan 2008 A1
20080025701 Ikeda Jan 2008 A1
20080032703 Krumm et al. Feb 2008 A1
20080033930 Warren Feb 2008 A1
20080043041 Hedenstroem et al. Feb 2008 A2
20080046476 Anderson et al. Feb 2008 A1
20080046831 Imai et al. Feb 2008 A1
20080049704 Witteman et al. Feb 2008 A1
20080055269 Lemay et al. Mar 2008 A1
20080062141 Chandhri Mar 2008 A1
20080076505 Nguyen et al. Mar 2008 A1
20080091723 Zuckerberg et al. Apr 2008 A1
20080092233 Tian et al. Apr 2008 A1
20080094387 Chen Apr 2008 A1
20080104503 Beall et al. May 2008 A1
20080109844 Baldeschwieler et al. May 2008 A1
20080120409 Sun et al. May 2008 A1
20080147730 Lee et al. Jun 2008 A1
20080148150 Mall Jun 2008 A1
20080158230 Sharma et al. Jul 2008 A1
20080168033 Ott et al. Jul 2008 A1
20080168489 Schraga Jul 2008 A1
20080172413 Chiu Jul 2008 A1
20080189177 Anderton et al. Aug 2008 A1
20080193101 Agnihotri et al. Aug 2008 A1
20080207176 Brackbill et al. Aug 2008 A1
20080208692 Garaventi et al. Aug 2008 A1
20080214210 Rasanen et al. Sep 2008 A1
20080222158 Saika Sep 2008 A1
20080222545 Lemay Sep 2008 A1
20080244438 Peters et al. Oct 2008 A1
20080255976 Altberg et al. Oct 2008 A1
20080256430 Gold Oct 2008 A1
20080256446 Yamamoto Oct 2008 A1
20080256450 Takakura et al. Oct 2008 A1
20080256577 Funaki et al. Oct 2008 A1
20080263103 Mcgregor et al. Oct 2008 A1
20080266421 Takahata et al. Oct 2008 A1
20080270938 Carlson Oct 2008 A1
20080284587 Saigh et al. Nov 2008 A1
20080288338 Wiseman et al. Nov 2008 A1
20080306826 Kramer et al. Dec 2008 A1
20080313329 Wang et al. Dec 2008 A1
20080313346 Kujawa et al. Dec 2008 A1
20080316181 Nurmi Dec 2008 A1
20080318616 Chipalkatti et al. Dec 2008 A1
20090006191 Arankalle et al. Jan 2009 A1
20090006565 Velusamy et al. Jan 2009 A1
20090015703 Kim et al. Jan 2009 A1
20090024956 Kobayashi Jan 2009 A1
20090030774 Rothschild et al. Jan 2009 A1
20090030999 Gatzke et al. Jan 2009 A1
20090040324 Nonaka Feb 2009 A1
20090042588 Lottin et al. Feb 2009 A1
20090058822 Chaudhri Mar 2009 A1
20090079846 Chou Mar 2009 A1
20090087161 Roberts et al. Apr 2009 A1
20090089378 Maresh Apr 2009 A1
20090089678 Sacco et al. Apr 2009 A1
20090089710 Wood et al. Apr 2009 A1
20090093261 Ziskind et al. Apr 2009 A1
20090132341 Klinger May 2009 A1
20090132453 Hangartner et al. May 2009 A1
20090132665 Thomsen et al. May 2009 A1
20090148045 Lee et al. Jun 2009 A1
20090153492 Popp Jun 2009 A1
20090157450 Athsani et al. Jun 2009 A1
20090157752 Gonzalez Jun 2009 A1
20090158183 Mccurdy et al. Jun 2009 A1
20090160970 Fredlund et al. Jun 2009 A1
20090163182 Gatti et al. Jun 2009 A1
20090169062 Cheung et al. Jul 2009 A1
20090177299 Van De Sluis Jul 2009 A1
20090177588 Marchese Jul 2009 A1
20090177730 Annamalai et al. Jul 2009 A1
20090187825 Sandquist et al. Jul 2009 A1
20090192900 Collision Jul 2009 A1
20090199242 Johnson et al. Aug 2009 A1
20090215469 Fisher et al. Aug 2009 A1
20090232354 Camp, Jr. et al. Sep 2009 A1
20090234815 Boerries et al. Sep 2009 A1
20090235155 Ueda Sep 2009 A1
20090239552 Churchill et al. Sep 2009 A1
20090249222 Schmidt Oct 2009 A1
20090249244 Robinson et al. Oct 2009 A1
20090260010 Burkhart et al. Oct 2009 A1
20090265647 Martin et al. Oct 2009 A1
20090284658 Cho Nov 2009 A1
20090288022 Almstrand et al. Nov 2009 A1
20090291665 Gaskarth et al. Nov 2009 A1
20090291672 Treves et al. Nov 2009 A1
20090292608 Polachek Nov 2009 A1
20090319607 Belz et al. Dec 2009 A1
20090327073 Li Dec 2009 A1
20100001980 Kim et al. Jan 2010 A1
20100011316 Sar et al. Jan 2010 A1
20100014833 Pjanovic et al. Jan 2010 A1
20100039505 Inoue et al. Feb 2010 A1
20100041378 Aceves et al. Feb 2010 A1
20100062794 Han Mar 2010 A1
20100073509 Shioji Mar 2010 A1
20100082427 Burgener et al. Apr 2010 A1
20100082693 Hugg et al. Apr 2010 A1
20100100568 Papin et al. Apr 2010 A1
20100100729 Read et al. Apr 2010 A1
20100113065 Narayan et al. May 2010 A1
20100115281 Camenisch et al. May 2010 A1
20100130233 Lansing May 2010 A1
20100131880 Lee et al. May 2010 A1
20100131895 Wohlert May 2010 A1
20100153144 Miller et al. Jun 2010 A1
20100156933 Jones et al. Jun 2010 A1
20100159944 Pascal et al. Jun 2010 A1
20100161635 Dey Jun 2010 A1
20100161658 Hamynen et al. Jun 2010 A1
20100161831 Haas et al. Jun 2010 A1
20100162149 Sheleheda et al. Jun 2010 A1
20100183280 Beauregard et al. Jul 2010 A1
20100185552 Deluca et al. Jul 2010 A1
20100185665 Horn et al. Jul 2010 A1
20100185987 Yang et al. Jul 2010 A1
20100191631 Weidmann Jul 2010 A1
20100197318 Petersen et al. Aug 2010 A1
20100197319 Petersen et al. Aug 2010 A1
20100198683 Aarabi Aug 2010 A1
20100198694 Muthukrishnan Aug 2010 A1
20100198826 Petersen et al. Aug 2010 A1
20100198828 Petersen et al. Aug 2010 A1
20100198862 Jennings et al. Aug 2010 A1
20100198870 Petersen et al. Aug 2010 A1
20100198917 Petersen et al. Aug 2010 A1
20100199166 Fisk, III Aug 2010 A1
20100199227 Xiao et al. Aug 2010 A1
20100201482 Robertson et al. Aug 2010 A1
20100201536 Robertson et al. Aug 2010 A1
20100214436 Kim et al. Aug 2010 A1
20100223128 Dukellis et al. Sep 2010 A1
20100223343 Bosan et al. Sep 2010 A1
20100247064 Yeh et al. Sep 2010 A1
20100250109 Johnston et al. Sep 2010 A1
20100251143 Thomas et al. Sep 2010 A1
20100257196 Waters et al. Oct 2010 A1
20100259386 Holley et al. Oct 2010 A1
20100273463 Bonnefoy Oct 2010 A1
20100273509 Sweeney et al. Oct 2010 A1
20100281045 Dean Nov 2010 A1
20100293105 Blinn Nov 2010 A1
20100306669 Della Pasqua Dec 2010 A1
20100332958 Weinberger et al. Dec 2010 A1
20110004071 Faiola et al. Jan 2011 A1
20110010205 Richards Jan 2011 A1
20110029512 Folgner et al. Feb 2011 A1
20110040783 Uemichi et al. Feb 2011 A1
20110040804 Peirce et al. Feb 2011 A1
20110050909 Ellenby et al. Mar 2011 A1
20110050915 Wang et al. Mar 2011 A1
20110064388 Brown et al. Mar 2011 A1
20110066743 Hurley et al. Mar 2011 A1
20110083101 Sharon et al. Apr 2011 A1
20110085059 Noh Apr 2011 A1
20110099507 Nesladek et al. Apr 2011 A1
20110099519 Ma et al. Apr 2011 A1
20110102630 Rukes May 2011 A1
20110106882 Takakura et al. May 2011 A1
20110119133 Igelman et al. May 2011 A1
20110131633 Macaskill et al. Jun 2011 A1
20110137881 Cheng et al. Jun 2011 A1
20110141025 Tsai Jun 2011 A1
20110145564 Moshir et al. Jun 2011 A1
20110159890 Fortescue et al. Jun 2011 A1
20110164163 Bilbrey et al. Jul 2011 A1
20110184980 Jeong et al. Jul 2011 A1
20110191368 Muzatko Aug 2011 A1
20110197194 D'Angelo et al. Aug 2011 A1
20110202598 Evans et al. Aug 2011 A1
20110202968 Nurmi Aug 2011 A1
20110211534 Schmidt et al. Sep 2011 A1
20110213845 Logan et al. Sep 2011 A1
20110215966 Kim et al. Sep 2011 A1
20110225048 Nair Sep 2011 A1
20110238762 Soni et al. Sep 2011 A1
20110238763 Shin et al. Sep 2011 A1
20110249551 Rollins Oct 2011 A1
20110255736 Thompson et al. Oct 2011 A1
20110273575 Lee Nov 2011 A1
20110282799 Huston Nov 2011 A1
20110283188 Farrenkopf Nov 2011 A1
20110286586 Saylor et al. Nov 2011 A1
20110294541 Zheng et al. Dec 2011 A1
20110296474 Babic Dec 2011 A1
20110302525 Jeon Dec 2011 A1
20110306387 Moon Dec 2011 A1
20110314084 Saretto et al. Dec 2011 A1
20110314419 Dunn et al. Dec 2011 A1
20110320373 Lee et al. Dec 2011 A1
20120004956 Hustan Jan 2012 A1
20120028659 Whitney et al. Feb 2012 A1
20120033718 Kauffman et al. Feb 2012 A1
20120036015 Sheikh Feb 2012 A1
20120036443 Ohmori et al. Feb 2012 A1
20120054797 Skog et al. Mar 2012 A1
20120054811 Spears Mar 2012 A1
20120059722 Rao Mar 2012 A1
20120062805 Candelore Mar 2012 A1
20120070045 Vesper et al. Mar 2012 A1
20120084731 Filman et al. Apr 2012 A1
20120084835 Thomas et al. Apr 2012 A1
20120099800 Llano et al. Apr 2012 A1
20120108293 Law et al. May 2012 A1
20120110096 Smarr et al. May 2012 A1
20120113143 Adhikari et al. May 2012 A1
20120113272 Hata May 2012 A1
20120117456 Koskimies May 2012 A1
20120123830 Svendsen et al. May 2012 A1
20120123871 Svendsen et al. May 2012 A1
20120123875 Svendsen et al. May 2012 A1
20120124126 Alcazar et al. May 2012 A1
20120124147 Hamlin et al. May 2012 A1
20120124176 Curtis et al. May 2012 A1
20120124458 Cruzada May 2012 A1
20120127196 Landry May 2012 A1
20120131507 Sparandara et al. May 2012 A1
20120131512 Takeuchi et al. May 2012 A1
20120136998 Hough et al. May 2012 A1
20120143760 Abulafia et al. Jun 2012 A1
20120150978 Monaco et al. Jun 2012 A1
20120158532 Fitzsimmons Jun 2012 A1
20120163664 Zhu Jun 2012 A1
20120165100 Lalancette et al. Jun 2012 A1
20120166462 Pathak et al. Jun 2012 A1
20120166971 Sachson et al. Jun 2012 A1
20120169855 Oh Jul 2012 A1
20120172062 Altman et al. Jul 2012 A1
20120173991 Roberts et al. Jul 2012 A1
20120176401 Hayward et al. Jul 2012 A1
20120184248 Speede Jul 2012 A1
20120197724 Kendall Aug 2012 A1
20120200743 Blanchflower et al. Aug 2012 A1
20120201362 Crossan et al. Aug 2012 A1
20120203849 Collins et al. Aug 2012 A1
20120208564 Clark et al. Aug 2012 A1
20120209892 Macaskill et al. Aug 2012 A1
20120209921 Adafin et al. Aug 2012 A1
20120209924 Evans et al. Aug 2012 A1
20120210244 De Francisco et al. Aug 2012 A1
20120212632 Mate et al. Aug 2012 A1
20120220264 Kawabata Aug 2012 A1
20120226748 Bosworth et al. Sep 2012 A1
20120233000 Fisher et al. Sep 2012 A1
20120236162 Imamura Sep 2012 A1
20120239761 Linner et al. Sep 2012 A1
20120250951 Chen Oct 2012 A1
20120252418 Kandekar et al. Oct 2012 A1
20120254324 Majeti et al. Oct 2012 A1
20120254325 Majeti et al. Oct 2012 A1
20120259815 Olson Oct 2012 A1
20120278387 Garcia et al. Nov 2012 A1
20120278692 Shi Nov 2012 A1
20120281129 Wang et al. Nov 2012 A1
20120288147 Fujitani Nov 2012 A1
20120290637 Perantatos et al. Nov 2012 A1
20120299954 Wada et al. Nov 2012 A1
20120304052 Tanaka et al. Nov 2012 A1
20120304080 Wormald et al. Nov 2012 A1
20120307096 Ford et al. Dec 2012 A1
20120307112 Kunishige et al. Dec 2012 A1
20120311465 Nealer et al. Dec 2012 A1
20120311623 Davis et al. Dec 2012 A1
20120319904 Lee et al. Dec 2012 A1
20120323933 He et al. Dec 2012 A1
20120324018 Metcalf et al. Dec 2012 A1
20130006759 Srivastava et al. Jan 2013 A1
20130018960 Knysz et al. Jan 2013 A1
20130024292 David Jan 2013 A1
20130024757 Doll et al. Jan 2013 A1
20130036364 Johnson Feb 2013 A1
20130045753 Obermeyer et al. Feb 2013 A1
20130050260 Reitan Feb 2013 A1
20130055083 Fino Feb 2013 A1
20130057489 Morton Mar 2013 A1
20130057587 Leonard et al. Mar 2013 A1
20130059607 Herz et al. Mar 2013 A1
20130060690 Oskolkov et al. Mar 2013 A1
20130063369 Malhotra et al. Mar 2013 A1
20130067027 Song et al. Mar 2013 A1
20130071093 Hanks et al. Mar 2013 A1
20130076758 Li et al. Mar 2013 A1
20130080254 Thramann Mar 2013 A1
20130082959 Shimazu et al. Apr 2013 A1
20130085790 Palmer et al. Apr 2013 A1
20130086072 Peng et al. Apr 2013 A1
20130090171 Holton et al. Apr 2013 A1
20130095857 Garcia et al. Apr 2013 A1
20130104053 Thornton et al. Apr 2013 A1
20130110885 Brundrett, III May 2013 A1
20130111514 Slavin et al. May 2013 A1
20130115872 Huang et al. May 2013 A1
20130122862 Horn et al. May 2013 A1
20130122929 Al-mufti et al. May 2013 A1
20130128059 Kristensson May 2013 A1
20130129252 Lauper May 2013 A1
20130132477 Bosworth et al. May 2013 A1
20130132908 Lee et al. May 2013 A1
20130144979 Kansal et al. Jun 2013 A1
20130145286 Feng et al. Jun 2013 A1
20130159110 Rajaram et al. Jun 2013 A1
20130159919 Leydon Jun 2013 A1
20130169822 Zhu et al. Jul 2013 A1
20130173729 Starenky et al. Jul 2013 A1
20130182133 Tanabe Jul 2013 A1
20130185131 Sinha et al. Jul 2013 A1
20130191198 Carlson et al. Jul 2013 A1
20130194301 Robbins et al. Aug 2013 A1
20130198176 Kim Aug 2013 A1
20130210518 Barclay et al. Aug 2013 A1
20130218965 Abrol et al. Aug 2013 A1
20130218968 Mcevilly et al. Aug 2013 A1
20130222323 Mckenzie Aug 2013 A1
20130227476 Frey Aug 2013 A1
20130232194 Knapp et al. Sep 2013 A1
20130263031 Oshiro et al. Oct 2013 A1
20130265450 Barnes, Jr. Oct 2013 A1
20130267253 Case et al. Oct 2013 A1
20130275505 Gauglitz et al. Oct 2013 A1
20130283167 Xu Oct 2013 A1
20130290337 Lansford et al. Oct 2013 A1
20130290443 Collins et al. Oct 2013 A1
20130304646 De Geer Nov 2013 A1
20130311255 Cummins et al. Nov 2013 A1
20130325964 Berberat Dec 2013 A1
20130344896 Kirmse et al. Dec 2013 A1
20130346869 Asver et al. Dec 2013 A1
20130346877 Borovoy et al. Dec 2013 A1
20140003739 S V et al. Jan 2014 A1
20140006129 Heath Jan 2014 A1
20140011538 Mulcahy et al. Jan 2014 A1
20140019264 Wachman et al. Jan 2014 A1
20140028589 Reilly Jan 2014 A1
20140029034 Toriyama Jan 2014 A1
20140032682 Prado et al. Jan 2014 A1
20140040712 Chang et al. Feb 2014 A1
20140043204 Basnayake et al. Feb 2014 A1
20140043355 Kim et al. Feb 2014 A1
20140045530 Gordon et al. Feb 2014 A1
20140047016 Rao Feb 2014 A1
20140047045 Baldwin et al. Feb 2014 A1
20140047335 Lewis et al. Feb 2014 A1
20140049652 Moon et al. Feb 2014 A1
20140052281 Eronen et al. Feb 2014 A1
20140052485 Shidfar Feb 2014 A1
20140052633 Gandhi Feb 2014 A1
20140057660 Wager Feb 2014 A1
20140059479 Hamburg et al. Feb 2014 A1
20140068692 Archibong et al. Mar 2014 A1
20140082651 Sharifi Mar 2014 A1
20140089264 Talagala et al. Mar 2014 A1
20140089314 Iizuka et al. Mar 2014 A1
20140092130 Anderson et al. Apr 2014 A1
20140096029 Schultz Apr 2014 A1
20140114565 Aziz et al. Apr 2014 A1
20140122658 Haeger et al. May 2014 A1
20140122787 Shalvi et al. May 2014 A1
20140129627 Baldwin et al. May 2014 A1
20140129953 Spiegel May 2014 A1
20140136985 Albir et al. May 2014 A1
20140143143 Fasoli et al. May 2014 A1
20140149519 Redfern et al. May 2014 A1
20140155102 Cooper et al. Jun 2014 A1
20140164979 Deeter et al. Jun 2014 A1
20140172877 Rubinstein et al. Jun 2014 A1
20140173424 Hogeg et al. Jun 2014 A1
20140173457 Wang et al. Jun 2014 A1
20140188815 Mentz Jul 2014 A1
20140189592 Benchenaa et al. Jul 2014 A1
20140201527 Krivorot Jul 2014 A1
20140207679 Cho Jul 2014 A1
20140207860 Wang et al. Jul 2014 A1
20140214471 Schreiner, III Jul 2014 A1
20140222564 Kranendonk et al. Aug 2014 A1
20140222913 Cathcart et al. Aug 2014 A1
20140250465 Mulholland et al. Sep 2014 A1
20140258405 Perkin Sep 2014 A1
20140265359 Cheng et al. Sep 2014 A1
20140266703 Dalley, Jr. et al. Sep 2014 A1
20140279061 Elimeliah et al. Sep 2014 A1
20140279128 Sagebin Sep 2014 A1
20140279436 Dorsey et al. Sep 2014 A1
20140279540 Jackson Sep 2014 A1
20140280140 Ling et al. Sep 2014 A1
20140280537 Pridmore et al. Sep 2014 A1
20140282096 Rubinstein Sep 2014 A1
20140286566 Rhoads Sep 2014 A1
20140287779 O'keefe et al. Sep 2014 A1
20140289157 Kenna, III et al. Sep 2014 A1
20140289833 Briceno Sep 2014 A1
20140298210 Park et al. Oct 2014 A1
20140304622 Jorasch et al. Oct 2014 A1
20140306986 Gottesman et al. Oct 2014 A1
20140317302 Naik Oct 2014 A1
20140324627 Haver et al. Oct 2014 A1
20140324629 Jacobs Oct 2014 A1
20140325383 Brown et al. Oct 2014 A1
20140331188 Sandstrom et al. Nov 2014 A1
20140359024 Spiegel Dec 2014 A1
20140359032 Spiegel et al. Dec 2014 A1
20140372844 Zumkhawala Dec 2014 A1
20140372850 Campbell et al. Dec 2014 A1
20140379683 Bazaz Dec 2014 A1
20150012603 Saito Jan 2015 A1
20150013016 Kanter et al. Jan 2015 A1
20150020086 Chen et al. Jan 2015 A1
20150040011 Chun Feb 2015 A1
20150042572 Lombardi Feb 2015 A1
20150043033 Sugimoto Feb 2015 A1
20150046278 Pei et al. Feb 2015 A1
20150055197 Romanoff et al. Feb 2015 A1
20150058957 Halliday et al. Feb 2015 A1
20150063724 Ikeda et al. Mar 2015 A1
20150071619 Brough Mar 2015 A1
20150081630 Linsalata et al. Mar 2015 A1
20150087263 Branscomb et al. Mar 2015 A1
20150088622 Ganschow et al. Mar 2015 A1
20150094106 Grossman et al. Apr 2015 A1
20150095020 Leydon Apr 2015 A1
20150096042 Mizrachi Apr 2015 A1
20150116529 Wu et al. Apr 2015 A1
20150127754 Clark et al. May 2015 A1
20150130178 Clements May 2015 A1
20150134318 Cuthbert et al. May 2015 A1
20150142753 Soon-shiong May 2015 A1
20150161178 B Doiu Jun 2015 A1
20150169827 Laborde Jun 2015 A1
20150172534 Miyakawa et al. Jun 2015 A1
20150177937 Poletto et al. Jun 2015 A1
20150178260 Brunson Jun 2015 A1
20150188869 Gilad et al. Jul 2015 A1
20150193685 Srinivasan et al. Jul 2015 A1
20150199082 Scholler et al. Jul 2015 A1
20150206349 Rosenthal et al. Jul 2015 A1
20150222814 Li et al. Aug 2015 A1
20150227602 Ramu et al. Aug 2015 A1
20150248683 Walkingshaw Sep 2015 A1
20150261917 Smith Sep 2015 A1
20150271779 Alavudin Sep 2015 A1
20150312184 Langholz et al. Oct 2015 A1
20150325268 Berger et al. Nov 2015 A1
20150326510 Tomlinson et al. Nov 2015 A1
20150334347 Kang Nov 2015 A1
20150350136 Flynn, III et al. Dec 2015 A1
20150358806 Salqvist Dec 2015 A1
20150365795 Allen et al. Dec 2015 A1
20150367233 Hicks et al. Dec 2015 A1
20150378502 Hu et al. Dec 2015 A1
20150381682 Rao et al. Dec 2015 A1
20150381688 Jenkins et al. Dec 2015 A1
20160006927 Sehn Jan 2016 A1
20160014063 Hogeg et al. Jan 2016 A1
20160034253 Bang et al. Feb 2016 A1
20160048369 Zenoff Feb 2016 A1
20160085773 Chang et al. Mar 2016 A1
20160085863 Allen et al. Mar 2016 A1
20160085994 Pereira Mar 2016 A1
20160086670 Gross et al. Mar 2016 A1
20160092561 Liu et al. Mar 2016 A1
20160092962 Wasserman et al. Mar 2016 A1
20160099901 Allen et al. Apr 2016 A1
20160105387 Jackson Apr 2016 A1
20160119272 Rubinstein et al. Apr 2016 A1
20160134941 Selvaraj May 2016 A1
20160139748 Iwaizumi et al. May 2016 A1
20160149843 Spicer et al. May 2016 A1
20160180887 Sehn Jun 2016 A1
20160182422 Sehn et al. Jun 2016 A1
20160182875 Sehn Jun 2016 A1
20160234023 Mozer et al. Aug 2016 A1
20160239248 Sehn Aug 2016 A1
20160253833 Lew Sep 2016 A1
20160253912 Heilman et al. Sep 2016 A1
20160274705 Kapadia et al. Sep 2016 A1
20160277419 Allen et al. Sep 2016 A1
20160321708 Sehn Nov 2016 A1
20160352659 Krishnamoorth Dec 2016 A1
20160359957 Laliberte Dec 2016 A1
20160359987 Laliberte Dec 2016 A1
20170006094 Abou Mahmoud et al. Jan 2017 A1
20170061308 Chen et al. Mar 2017 A1
20170111617 Kuwahara et al. Apr 2017 A1
20170134821 D'amelio et al. May 2017 A1
20170149717 Sehn May 2017 A1
20170161382 Ouimet et al. Jun 2017 A1
20170185256 Bennett Jun 2017 A1
20170230315 Zubas et al. Aug 2017 A1
20170263029 Yan et al. Sep 2017 A1
20170287006 Azmoodeh et al. Oct 2017 A1
20170295250 Samaranayake et al. Oct 2017 A1
20170310888 Wright et al. Oct 2017 A1
20170329481 Stoop et al. Nov 2017 A1
20170374003 Allen et al. Dec 2017 A1
20170374508 Davis et al. Dec 2017 A1
20180103002 Sehn Apr 2018 A1
20180278562 Tang Sep 2018 A1
20180279016 Tang Sep 2018 A1
20180316575 Son et al. Nov 2018 A1
20190097812 Toth Mar 2019 A1
20190237106 Sehn Aug 2019 A1
20190342699 Sehn et al. Nov 2019 A1
20190372991 Allen et al. Dec 2019 A1
20200057590 Sehn Feb 2020 A1
20200105304 Sehn Apr 2020 A1
20200193053 Murphy et al. Jun 2020 A1
20200213804 Sehn et al. Jul 2020 A1
20200288270 Allen et al. Sep 2020 A1
20200329336 Sehn et al. Oct 2020 A1
20200411058 Sehn Dec 2020 A1
20210006526 Allen et al. Jan 2021 A1
20210006527 Allen et al. Jan 2021 A1
20210006528 Allen et al. Jan 2021 A1
20210014238 Allen et al. Jan 2021 A1
20210273903 Allen et al. Sep 2021 A1
Foreign Referenced Citations (66)
Number Date Country
2887596 Jul 2015 CA
2894332 Aug 2018 CA
2910158 Jun 2019 CA
101635763 Jan 2010 CN
102238107 Nov 2011 CN
103095768 May 2013 CN
106663264 May 2017 CN
107004225 Aug 2017 CN
107111828 Aug 2017 CN
107251006 Oct 2017 CN
107637099 Jan 2018 CN
107710772 Feb 2018 CN
2051480 Apr 2009 EP
2151797 Feb 2010 EP
2399928 Sep 2004 GB
19990073076 Oct 1999 KR
20010078417 Aug 2001 KR
1020060038872 May 2006 KR
1020080006729 Jan 2008 KR
1020080017854 Feb 2008 KR
1020120140404 Dec 2012 KR
20130061724 Jun 2013 KR
1020140066795 Jun 2014 KR
1020140115413 Oct 2014 KR
10-1822920 Jan 2018 KR
101869473 Jun 2018 KR
WO-1996024213 Aug 1996 WO
WO-1999063453 Dec 1999 WO
WO-2000058882 Oct 2000 WO
WO-2001029642 Apr 2001 WO
WO-2001050703 Jul 2001 WO
WO-2006118755 Nov 2006 WO
WO-2007092668 Aug 2007 WO
WO-2009043020 Apr 2009 WO
WO-2011040821 Apr 2011 WO
WO-2011119407 Sep 2011 WO
WO-2012000107 Jan 2012 WO
WO-2013006584 Jan 2013 WO
WO-2013008238 Jan 2013 WO
WO-2013008251 Jan 2013 WO
WO-2013045753 Apr 2013 WO
WO-2013058897 Apr 2013 WO
WO-2013126784 Aug 2013 WO
WO-2014006129 Jan 2014 WO
2014031899 Feb 2014 WO
WO-2014068573 May 2014 WO
WO-2014093668 Jun 2014 WO
WO-2014115136 Jul 2014 WO
WO-2014194262 Dec 2014 WO
WO-2015192026 Dec 2015 WO
WO-2016007285 Jan 2016 WO
WO-2016044424 Mar 2016 WO
WO-2016054562 Apr 2016 WO
WO-2016065131 Apr 2016 WO
WO-2016100318 Jun 2016 WO
WO-2016100318 Jun 2016 WO
WO-2016100342 Jun 2016 WO
WO-2016112299 Jul 2016 WO
WO-2016149594 Sep 2016 WO
WO-2016179166 Nov 2016 WO
WO-2016179235 Nov 2016 WO
WO-2016202890 Dec 2016 WO
WO-2017106529 Jun 2017 WO
WO-2017176739 Oct 2017 WO
WO-2017176992 Oct 2017 WO
WO-2018005644 Jan 2018 WO
Non-Patent Literature Citations (553)
Entry
US 10,075,404 B1, 09/2018, Allen et al. (withdrawn)
US 10,425,370 B2, 09/2019, Allen et al. (withdrawn)
US 10,484,394 B2, 11/2019, Allen et al. (withdrawn)
US 10,503,924 B1, 12/2019, Murphy et al. (withdrawn)
“Android Getting Started Guide”, Voxer Business, [Online] Retrieved from the Internet: <URL: https://voxer.com/assets/AndroidGuide.pdf>, (Feb. 1, 2014), 18 pgs.
“U.S. Appl. No. 14/304,855, Corrected Notice of Allowance dated Jun. 26, 2015”, 8 pgs.
“U.S. Appl. No. 14/304,855, Final Office Action dated Feb. 18, 2015”, 10 pgs.
“U.S. Appl. No. 14/304,855, Non Final Office Action dated Mar. 18, 2015”, 9 pgs.
“U.S. Appl. No. 14/304,855, Non Final Office Action dated Oct. 22, 2014”, 11 pgs.
“U.S. Appl. No. 14/304,855, Notice of Allowance dated Jun. 1, 2015”, 11 pgs.
“U.S. Appl. No. 14/304,855, Response filed Feb. 25, 2015 to Final Office Action dated Feb. 18, 2015”, 5 pgs.
“U.S. Appl. No. 14/304,855, Response filed Apr. 1, 2015 to Non Final Office Action dated Mar. 18, 2015”, 4 pgs.
“U.S. Appl. No. 14/304,855, Response filed Nov. 7, 2014 to Non Final Office Action dated Oct. 22, 2014”, 5 pgs.
“U.S. Appl. No. 14/505,478, Advisory Action dated Apr. 14, 2015”, 3 pgs.
“U.S. Appl. No. 14/505,478, Corrected Notice of Allowance dated May 18, 2016”, 2 pgs.
“U.S. Appl. No. 14/505,478, Corrected Notice of Allowance dated Jul. 22, 2016”, 2 pgs.
“U.S. Appl. No. 14/505,478, Final Office Action dated Mar. 17, 2015”, 16 pgs.
“U.S. Appl. No. 14/505,478, Non Final Office Action dated Jan. 27, 2015”, 13 pgs.
“U.S. Appl. No. 14/505,478, Non Final Office Action dated Sep. 4, 2015”, 19 pgs.
“U.S. Appl. No. 14/505,478, Notice of Allowance dated Apr. 28, 2016”, 11 pgs.
“U.S. Appl. No. 14/505,478, Notice of Allowance dated Aug. 26, 2016”, 11 pgs.
“U.S. Appl. No. 14/505,478, Response filed Jan. 30, 2015 to Non Final Office Action dated Jan. 27, 2015”, 10 pgs.
“U.S. Appl. No. 14/505,478, Response filed Mar. 4, 2016 to Non Final Office Action dated Sep. 4, 2015”, 12 pgs.
“U.S. Appl. No. 14/505,478, Response filed Apr. 1, 2015 to Final Office Action dated Mar. 17, 2015”, 6 pgs.
“U.S. Appl. No. 14/506,478, Response filed Aug. 17, 2015 to Advisory Action dated Apr. 14, 2015”, 10 pgs.
“U.S. Appl. No. 14/523,728, Non Final Office Action dated Dec. 12, 2014”, 10 pgs.
“U.S. Appl. No. 14/523,728, Notice of Allowance dated Mar. 24, 2015”, 8 pgs.
“U.S. Appl. No. 14/523,728, Notice of Allowance dated Apr. 15, 2015”, 8 pgs.
“U.S. Appl. No. 14/523,728, Notice of Allowance dated Jun. 5, 2015”, 8 pgs.
“U.S. Appl. No. 14/523,728, Response filed Aug. 25, 2014 to Non Final Office Action dated Jan. 16, 2015”, 5 pgs.
“U.S. Appl. No. 14/529,064, Non Final Office Action dated Mar. 12, 2015”, 20 pgs.
“U.S. Appl. No. 14/529,064, Response filed Feb. 5, 2015 to Restriction Requirement dated Feb. 2, 2015”, 6 pgs.
“U.S. Appl. No. 14/529,064, Response filed Mar. 26, 2015 to Non Final Office Action dated Mar. 12, 2015”, 8 pgs.
“U.S. Appl. No. 14/529,064, Restriction Requirement dated Feb. 2, 2015”, 5 pgs.
“U.S. Appl. No. 14/578,258, Examiner Interview Summary dated Nov. 25, 2015”, 3 pgs.
“U.S. Appl. No. 14/578,258, Non Final Office Action dated Jun. 10, 2015”, 12 pgs.
“U.S. Appl. No. 14/578,258, Notice of Allowance dated Feb. 26, 2016”, 5 pgs.
“U.S. Appl. No. 14/578,258, Response filed Dec. 10, 2015 to Non Final Office Action dated Jun. 10, 2015”, 11 pgs.
“U.S. Appl. No. 14/578,271, Corrected Notice of Allowance dated Oct. 30, 2017”, 2 pgs.
“U.S. Appl. No. 14/578,271, Final Office Action dated Dec. 3, 2015”, 15 pgs.
“U.S. Appl. No. 14/578,271, Non Final Office Action dated Aug. 7, 2015”, 12 pgs.
“U.S. Appl. No. 14/578,271, Notice of Allowability dated Oct. 29, 2017”, 3 pgs.
“U.S. Appl. No. 14/578,271, Notice of Allowance dated Aug. 1, 2017”, 5 pgs.
“U.S. Appl. No. 14/578,271, Notice of Allowance dated Oct. 7, 2016”, 7 pgs.
“U.S. Appl. No. 14/578,271, Response filed Feb. 9, 2016 to Final Office Action dated Dec. 2, 2015”, 10 pgs.
“U.S. Appl. No. 14/578,271, Response filed Jun. 19, 2015 to Restriction Requirement dated Apr. 23, 2015”, 6 pgs.
“U.S. Appl. No. 14/578,271, Response filed Oct. 28, 2015 to Non Final Office Action dated Aug. 7, 2015”, 9 pgs.
“U.S. Appl. No. 14/578,271, Restriction Requirement dated Apr. 23, 2015”, 8 pgs.
“U.S. Appl. No. 14/612,692, Examiner Interview Summary dated Jan. 29, 2016”, 5 pgs.
“U.S. Appl. No. 14/612,692, Examiner Interview Summary dated May 14, 2018”, 3 pgs.
“U.S. Appl. No. 14/612,692, Examiner Interview Summary dated Jul. 6, 2016”, 4 pgs.
“U.S. Appl. No. 14/612,692, Examiner Interview Summary dated Aug. 14, 2015”, 3 pgs.
“U.S. Appl. No. 14/612,692, Examiner Interview Summary dated Sep. 8, 2016”, 3 pgs.
“U.S. Appl. No. 14/612,692, Examiner Interview Summary dated Nov. 13, 2017”, 13 pgs.
“U.S. Appl. No. 14/612,692, Final Office Action dated Aug. 15, 2016”, 18 pgs.
“U.S. Appl. No. 14/612,692, Final Office Action dated Aug. 25, 2017”, 18 pgs.
“U.S. Appl. No. 14/612,692, Final Office Action dated Nov. 23, 2015”, 15 pgs.
“U.S. Appl. No. 14/612,692, Non Final Office Action dated Jan. 3, 2017”, 17 pgs.
“U.S. Appl. No. 14/612,692, Non Final Office Action dated Jan. 9, 2018”, 19 pgs.
“U.S. Appl. No. 14/612,692, Non Final Office Action dated Mar. 28, 2016”, 15 pgs.
“U.S. Appl. No. 14/612,692, Non Final Office Action dated Jul. 20, 2015”, 25 pgs.
“U.S. Appl. No. 14/612,692, Notice of Allowance dated Jul. 5, 2018”, 11 pgs.
“U.S. Appl. No. 14/612,692, Response filed Feb. 23, 2016 to Final Office Action dated Nov. 23, 2015”, 10 pgs.
“U.S. Appl. No. 14/612,692, Response filed May 3, 2017 to Non Final Office Action dated Jan. 3, 2017”, 18 pgs.
“U.S. Appl. No. 14/612,692, Response filed May 9, 2018 to Non Final Office Action dated Jan. 9, 2018”, 15 pgs.
“U.S. Appl. No. 14/612,692, Response filed Nov. 14, 2016 to Final Office Action dated Aug. 15, 2016”, 15 pgs.
“U.S. Appl. No. 14/612,692, Response filed Nov. 22, 2017 to Final Office Action dated Aug. 25, 2017”, 11 pgs.
“U.S. Appl. No. 14/612,692, Response filed Jun. 28, 2016 to Non Final Office Action dated Mar. 28, 2016”, 14 pgs.
“U.S. Appl. No. 14/612,692, Response filed Oct. 19, 2015 to Non Final Office Action dated Jul. 20, 2015”, 11 pgs.
“U.S. Appl. No. 14/634,417, Advisory Action dated Mar. 14, 2017”, 3 pgs.
“U.S. Appl. No. 14/634,417, Examiner Interview Summary dated Aug. 7, 2017”, 3 pgs.
“U.S. Appl. No. 14/634,417, Final Office Action dated Jan. 31, 2017”, 27 pgs.
“U.S. Appl. No. 14/634,417, Non Final Office Action dated Jun. 8, 2017”, 17 pgs.
“U.S. Appl. No. 14/634,417, Non Final Office Action dated Aug. 30, 2016”, 23 pgs.
“U.S. Appl. No. 14/634,417, Notice of Allowance dated May 22, 2018”, 9 pgs.
“U.S. Appl. No. 14/634,417, Notice of Allowance dated Oct. 25, 2017”, 9 pgs.
“U.S. Appl. No. 14/634,417, Response filed Mar. 2, 2017 to Final Office Action dated Jan. 31, 2017”, 23 pgs.
“U.S. Appl. No. 14/634,417, Response filed Sep. 21, 2017 to Non Final Office Action dated Jun. 8, 2017”, 16 pgs.
“U.S. Appl. No. 14/634,417, Response filed Nov. 30, 2016 to Non Final Office Action dated Aug. 30, 2016”, 18 pgs.
“U.S. Appl. No. 14/723,400, Final Office Action dated Jan. 4, 2016”, 14 pgs.
“U.S. Appl. No. 14/723,400, Non Final Office Action dated Jul. 20, 2015”, 14 pgs.
“U.S. Appl. No. 14/723,400, Notice of Allowance dated Mar. 28, 2016”, 12 pgs.
“U.S. Appl. No. 14/723,400, Notice of Non Compliant Amendment dated Sep. 21, 2015”, 2 pgs.
“U.S. Appl. No. 14/723,400, Notice of Non Compliant Amendment dated Nov. 10, 2015”, 2 pgs.
“U.S. Appl. No. 14/723,400, Response filed Jan. 29, 2016 to Final Office Action dated Jan. 4, 2016”, 8 pgs.
“U.S. Appl. No. 14/723,400, Response filed Aug. 13, 2015 to Non Final Office Action dated Jul. 20, 2015”, 7 pgs.
“U.S. Appl. No. 14/723,400, Response filed Sep. 23, 2015 to Notice of Non Compliant Amendment dated Sep. 21, 2015”, 5 pgs.
“U.S. Appl. No. 14/723,400, Response filed Nov. 19, 2015 to Notice of Non Compliant Amendment dated Nov. 10, 2015”, 5 pgs.
“U.S. Appl. No. 14/738,069, Non Final Office Action dated Mar. 21, 2016”, 12 pgs.
“U.S. Appl. No. 14/738,069, Notice of Allowance dated Aug. 17, 2016”, 6 pgs.
“U.S. Appl. No. 14/738,069, Response filed Jun. 10, 2016 to Non Final Office Action dated Mar. 21, 2016”, 10 pgs.
“U.S. Appl. No. 14/808,283, Notice of Allowance dated Apr. 12, 2016”, 9 pgs.
“U.S. Appl. No. 14/808,283, Notice of Allowance dated Jul. 14, 2016”, 8 pgs.
“U.S. Appl. No. 14/808,283, Preliminary Amendment filed Jul. 24, 2015”, 8 pgs.
“U.S. Appl. No. 14/967,472, Final Office Action dated Mar. 10, 2017”, 15 pgs.
“U.S. Appl. No. 14/967,472, Final Office Action dated Jun. 25, 2018”, 14 pgs.
“U.S. Appl. No. 14/967,472, Non Final Office Action dated Jan. 12, 2018”, 17 pgs.
“U.S. Appl. No. 14/967,472, Non Final Office Action dated Sep. 8, 2016”, 11 pgs.
“U.S. Appl. No. 14/967,472, Preliminary Amendment filed Dec. 15, 2015”, 6 pgs.
“U.S. Appl. No. 14/967,472, Response filed Mar. 16, 2018 Non Final Office Action dated Jan. 12, 2018”, 13 pgs.
“U.S. Appl. No. 14/967,472, Response filed Jun. 7, 2017 to Final Office Action dated Mar. 10, 2017”, 12 pgs.
“U.S. Appl. No. 14/967,472, Response filed Dec. 5, 2016 to Non Final Office Action dated Sep. 8, 2016”, 11 pgs.
“U.S. Appl. No. 15/074,029, Non Final Office Action dated Nov. 30, 2017”, 16 pgs.
“U.S. Appl. No. 15/074,029, Response filed Feb. 28, 2018 to Non Final Office Action dated Nov. 30, 2017”, 12 pgs.
“U.S. Appl. No. 15/137,608, Preliminary Amendment filed Apr. 26, 2016”, 6 pgs.
“U.S. Appl. No. 15/152,975, Examiner Interview Summary dated May 14, 2018”, 3 pgs.
“U.S. Appl. No. 15/152,975, Examiner Interview Summary dated Nov. 13, 2017”, 13 pgs.
“U.S. Appl. No. 15/152,975, Final Office Action dated Jun. 30, 2017”, 17 pgs.
“U.S. Appl. No. 15/152,975, Non Final Office Action dated Jan. 10, 2018”, 18 pgs.
“U.S. Appl. No. 15/152,975, Non Final Office Action dated Jan. 12, 2017”, 36 pgs.
“U.S. Appl. No. 15/152,975, Preliminary Amendment filed May 19, 2016”, 8 pgs.
“U.S. Appl. No. 15/152,975, Response filed May 10, 2018 to Non Final Office Action dated Jan. 10, 2018”, 13 pgs.
“U.S. Appl. No. 15/152,975, Response filed Jun. 12, 2017 to Non Final Office Action dated Jan. 12, 2017”, 13 pgs.
“U.S. Appl. No. 15/152,975, Response filed Nov. 30, 2017 to Final Office Action dated Jun. 17, 2017”, 9 pgs.
“U.S. Appl. No. 15/208,460, Notice of Allowance dated Feb. 27, 2017”, 8 pgs.
“U.S. Appl. No. 15/208,460, Notice of Allowance dated Dec. 30, 2016”, 9 pgs.
“U.S. Appl. No. 15/208,460, Supplemental Preliminary Amendment filed Jul. 18, 2016”, 8 pgs.
“U.S. Appl. No. 15/212,095, Final Office Action dated Mar. 14, 2017”, 9 pgs.
“U.S. Appl. No. 15/212,095, Non Final Office Action dated Feb. 2, 2017”, 8 pgs.
“U.S. Appl. No. 15/212,095, Notice of Allowance dated Jun. 1, 2017”, 8 pgs.
“U.S. Appl. No. 15/212,095, Notice of Allowance dated Sep. 8, 2017”, 2 pgs.
“U.S. Appl. No. 15/212,095, Response filed Feb. 28, 2017 to Non Final Office Action dated Feb. 2, 2017”, 2 pgs.
“U.S. Appl. No. 15/212,095, Response filed May 5, 2017 to Final Office Action dated Mar. 14, 2017”, 2 pgs.
“U.S. Appl. No. 15/224,312, Final Office Action dated Apr. 20, 2018”, 22 pgs.
“U.S. Appl. No. 15/224,312, Non Final Office Action dated Oct. 11, 2017”, 29 pgs.
“U.S. Appl. No. 15/224,312, Preliminary Amendment filed Feb. 1, 2017”, 11 pgs.
“U.S. Appl. No. 15/224,343, Final Office Action dated Apr. 19, 2018”, 20 pgs.
“U.S. Appl. No. 15/224,343, Non Final Office Action dated Oct. 4, 2017”, 26 pgs.
“U.S. Appl. No. 15/224,343, Preliminary Amendment filed Jan. 31, 2017”, 10 pgs.
“U.S. Appl. No. 15/224,343, Response filed Mar. 5, 2018 to Non Final Office Action dated Oct. 4, 2017”, 23 pgs.
“U.S. Appl. No. 15/224,355, Examiner Interview Summary dated Oct. 25, 2017”, 3 pgs.
“U.S. Appl. No. 15/224,355, Final Office Action dated Apr. 24, 2018”, 20 pgs.
“U.S. Appl. No. 15/224,355, Non Final Office Action dated Sep. 6, 2017”, 30 pgs.
“U.S. Appl. No. 15/224,355, Preliminary Amendment filed Apr. 3, 2017”, 12 pgs.
“U.S. Appl. No. 15/224,355, Response filed Mar. 6, 2018 to Non Final Office Action dated Sep. 6, 2017”, 25 pgs.
“U.S. Appl. No. 15/224,359, Final Office Action dated Apr. 2, 2018”, 18 pgs.
“U.S. Appl. No. 15/224,359, Non Final Office Action dated Jul. 20, 2017”, 33 pgs.
“U.S. Appl. No. 15/224,359, Response filed Jan. 22, 2018 to Non Final Office Action dated Jul. 20, 2017”, 13 pgs.
“U.S. Appl. No. 15/224,365, Final Office Action dated Apr. 2, 2018”, 19 pgs.
“U.S. Appl. No. 15/224,365, Non Final Office Action dated Aug. 8, 2017”, 41 pgs.
“U.S. Appl. No. 15/224,365, Response filed Feb. 2, 2018 to Non Final Office Action dated Aug. 8, 2017”, 14 pgs.
“U.S. Appl. No. 15/224,372, Final Office Action dated Apr. 3, 2018”, 18 pgs.
“U.S. Appl. No. 15/224,372, Non Final Office Action dated Aug. 7, 2017”, 40 pgs.
“U.S. Appl. No. 15/224,372, Preliminary Amendment filed May 5, 2017”, 10 pgs.
“U.S. Appl. No. 15/224,372, Response filed Jan. 8, 2017 to Non Final Office Action dated Aug. 7, 2017”, 22 pgs.
“U.S. Appl. No. 15/224,377, Final Office Action dated Jan. 2, 2018”, 29 pgs.
“U.S. Appl. No. 15/224,377, Non Final Office Action dated Jun. 15, 2018”, 19 pgs.
“U.S. Appl. No. 15/224,377, Non Final Office Action dated Aug. 4, 2017”, 41 pgs.
“U.S. Appl. No. 15/224,377, Response filed Dec. 6, 2017 to Non Final Office Action dated Aug. 4, 2017”, 22 pgs.
“U.S. Appl. No. 15/224,383, Examiner Interview Summary dated Oct. 25, 2017”, 3 pgs.
“U.S. Appl. No. 15/224,383, Final Office Action dated Feb. 14, 2018”, 25 pgs.
“U.S. Appl. No. 15/224,383, Non Final Office Action dated Aug. 30, 2017”, 26 pgs.
“U.S. Appl. No. 15/224,383, Preliminary Amendment filed May 9, 2017”, 13 pgs.
“U.S. Appl. No. 15/224,383, Response filed Jan. 3, 2018 to Non Final Office Action dated Aug. 30, 2017”, 25 pgs.
“U.S. Appl. No. 15/224,383, Response filed Jun. 14, 2018 to Final Office Action dated Feb. 14, 2018”, 14 pgs.
“U.S. Appl. No. 15/224,359, Preliminary Amendment filed Apr. 19, 2017”, 8 pgs.
“U.S. Appl. No. 15/298,806, Advisory Action dated Jan. 29, 2018”, 4 pgs.
“U.S. Appl. No. 15/298,806, Examiner Interview Summary dated Jan. 12, 2018”, 3 pgs.
“U.S. Appl. No. 15/298,806, Final Office Action dated Oct. 24, 2017”, 15 pgs.
“U.S. Appl. No. 15/298,806, Non Final Office Action dated May 17, 2018”, 16 pgs.
“U.S. Appl. No. 15/298,806, Non Final Office Action dated Jun. 12, 2017”, 26 pgs.
“U.S. Appl. No. 15/298,806, Preliminary Amendment filed Oct. 21, 2016”, 8 pgs.
“U.S. Appl. No. 15/298,806, Response filed Jan. 9, 2018 to Final Office Action dated Oct. 24, 2017”, 17 pgs.
“U.S. Appl. No. 15/298,806, Response filed Sep. 12, 2017 to Non Final Office Action dated Jun. 12, 2017”, 12 pgs.
“U.S. Appl. No. 15/416,846, Notice of Allowance dated Jul. 19, 2017”, 9 pgs.
“U.S. Appl. No. 15/416,846, Preliminary Amendment filed Feb. 18, 2017”, 10 pgs.
“U.S. Appl. No. 15/702,511, Preliminary Amendment filed Sep. 15, 2017”, 13 pgs.
“U.S. Appl. No. 15/729,582, Non Final Office Action dated May 25, 2018”, 14 pgs.
“U.S. Appl. No. 15/787,467, Non Final Office Action dated Apr. 18, 2018”, 17 pgs.
“U.S. Appl. No. 15/787,467, Preliminary Amendment filed Oct. 26, 2017”, 11 pgs.
“U.S. Appl. No. 16/000,657, Preliminary Amendment filed Jun. 6, 2018”, 8 pgs.
“Canadian Application Serial No. 2,894,332 Response filed Jan. 24, 2017 to Office Action dated Aug. 16, 2016-16”, 15 pgs.
“Canadian Application Serial No. 2,894,332, Office Action dated Aug. 16, 2016”, 4 pgs.
“Canadian Application Serial No. 2,894,332, Request for Reinstatement filed Jun. 11, 2011-18”, w/Amended Claims, 17 pgs.
“Canadian Application Serial No. 2,894,332, Response filed Jan. 24, 2017 to Office Action dated Aug. 16, 2016”, 15 pgs.
“Canadian Application Serial No. 2,910,158, Office Action dated Jun. 6, 2018”, 5 pgs.
“Canadian Application Serial No. 2,910,158, Office Action dated Dec. 15, 2016”, 5 pgs.
“Canadian Application Serial No. 2,910,158, Response filed Apr. 11, 2017 to Office Action dated Dec. 15, 2016”, 21 pgs.
“European Application Serial No. 14804343.3, Extended European Search Report dated Sep. 29, 2016”, 12 pgs.
“European Application Serial No. 15870874.3, Extended European Search Report dated Nov. 29, 2017”, 7 pgs.
“How Snaps Are Stored and Deleted”, Snapchat, [Online] Retrieved from the Internet: <URL: https://www.snap.com/en-US/news/post/how-snaps-are-stored-and-deleted/>, (May 9, 2013), 2 pgs.
“International Application Serial No. PCT/EP2008/063682, International Search Report dated Nov. 24, 2008”, 3 pgs.
“International Application Serial No. PCT/US2014/040346, International Search Report dated Mar. 23, 2015”, 2 pgs.
“International Application Serial No. PCT/US2014/040346, Written Opinion I7 Mar. 23, 2015”, 6 pgs.
“International Application Serial No. PCT/US2015/035591, International Preliminary Report on Patentability dated Dec. 22, 2016”, 7 pgs.
“International Application Serial No. PCT/US2015/035591, International Search Report dated Aug. 11, 2015”, 5 pgs.
“International Application Serial No. PCT/US2015/035591, International Written Opinion dated Aug. 11, 2015”, 5 pgs.
“International Application Serial No. PCT/US2015/037251, International Search Report dated Sep. 29, 2015”, 2 pgs.
“International Application Serial No. PCT/US2015/037251, Written Opinion dated Sep. 29, 2015”, 4 pgs.
“International Application Serial No. PCT/US2015/053811, International Preliminary Report on Patentability dated Apr. 13, 2017”, 9 pgs.
“International Application Serial No. PCT/US2015/053811, International Search Report dated Nov. 23, 2015”, 5 pgs.
“International Application Serial No. PCT/US2015/053811, Written Opinion dated Nov. 23, 2015”, 8 pgs.
“International Application Serial No. PCT/US2015/056884, International Preliminary Report on Patentability dated May 4, 2017”, 8 pgs.
“International Application Serial No. PCT/US2015/056884, International Search Report dated Dec. 22, 2015”, 5 pgs.
“International Application Serial No. PCT/US2015/056884, Written Opinion dated Dec. 22, 2015”, 6 pgs.
“International Application Serial No. PCT/US2015/065785, International Preliminary Report on Patentability dated Jun. 29, 2017”, 7 pgs.
“international Application Serial No. PCT/US2015/065785, International Search Report dated Jul. 21, 2016”, 5 pgs.
“International Application Serial No. PCT/US2015/065785, Written Opinion dated Jul. 21, 2016”, 5 pgs.
“International Application Serial No. PCT/US2015/065821, International Preliminary Report on Patentability dated Jun. 29, 2017”, 5 pgs.
“International Application Serial No. PCT/US2015/065821, International Search Report dated Mar. 3, 2016”, 2 pgs.
“International Application Serial No. PCT/US2015/065821, Written Opinion dated Mar. 3, 2016”, 3 pgs.
“International Application Serial No. PCT/US2016/023085, International Preliminary Report on Patentability dated Sep. 28, 2017”, 8 pgs.
“International Application Serial No. PCT/US2016/023085, International Search Report dated Jun. 17, 2016”, 5 pgs.
“International Application Serial No. PCT/US2016/023085, Written Opinion dated Jun. 17, 2016”, 6 pgs.
“IVisit Mobile: Getting Started”, IVISIT, [Online] Retrieved from the Internet: <URL: http://web.archive.org/web/20140830174355/http://ivisit.com/support_mobile>, (Dec. 4, 2013), 16 pgs.
“Pluraleyes by Red Giant”, © 2002-2015 Red Giant LLV, [Online]. Retrieved from the Internet: <URL: http/www.redgiant.com/products/pluraleyes/, (Accessed Nov. 11, 2015), 5 pgs.
Castelluccia, Claude, et al., “EphPub: Toward robust Ephemeral Publishing”, 19th IEEE International Conference on Network Protocols (ICNP), (Oct. 17, 2011), 18 pgs.
Chen, Datong, et al., “Protecting Personal Identification in Video”, Protecting Privacy in Video Surveillance, Springer-Verlag London Ltd., (2009), 115-128.
Clarke, Tangier, “Automatically syncing multiple clips and lots of audio like PluralEyes possible?”, [Online]. Retrieved from the Internet: <URL: https://forums.creativecow.net/thread/344/20553, (May 21, 2013), 8 pgs.
Fajman, “An Extensible Message Format for Message Disposition Notifications”, Request for Comments: 2298, National Institutes of Health, (Mar. 1998), 28 pgs.
Leyden, John, “This SMS will self-destruct in 40 seconds”, [Online] Retrieved from the Internet: <URL: http://www.theregister.co.uk/2005/12/12/stealthtext/>, (Dec. 12, 2005), 1 pg.
Melanson, Mike, “This text message will self destruct in 60 seconds”, [Online] Retrieved from the Internet: <URL: http://readwrite.com/2011/02/11/this_text_message_will_self_destruct_in_60_seconds>, (Feb. 18, 2015), 4 pgs.
Sawers, Paul, “Snapchat for iOS Lets You Send Photos to Friends and Set How long They're Visible for”, [Online] Retrieved from the Internet; <URL: https://thenextweb.com/apps/2012/05/07/snapchat-for-ios-lets-you-send-photos-to-friends-and-set-how-long-theyre-visible-for/>, (May 7, 2012), 5 pgs.
Shein, Esther, “Ephemeral Data”, Communications of the ACM, vol. 56, No. 9, (Sep. 2013), 3 pgs.
Trice, Andrew, “My Favorite New Feature: Multi-Clip Sync in Premiere Pro CC”, [Online]. Retrieved from the Internet: <URL: http://www.tricedesigns.com/2013/06/18/my-favorite-new-feature-multi-cam-synch-in-premiere-pro-cc (Jun. 18, 2013), 5 pgs.
“A Whole New Story”, Snap, Inc., [Online] Retrieved from the Internet: <URL: https://www.snap.com/en-US/news/>, (2017), 13 pgs.
“Adding photos to your listing”, eBay, [Online] Retrieved from the Internet: <URL: http://pages.ebay.com/help/sell/pictures.html>, (accessed May 24, 2017), 4 pgs.
“U.S. Appl. No. 14/494,226, Examiner Interview Summary dated Oct. 27, 2016”, 3 pgs.
“U.S. Appl. No. 14/494,226, Examiner Interview Summary dated Dec. 20, 2017”, 2 pgs.
“U.S. Appl. No. 14/494,226, Final Office Action dated Mar. 7, 2017”, 34 pgs.
“U.S. Appl. No. 14/494,226, Final Office Action dated Jun. 1, 2018”, 33 pgs.
“U.S. Appl. No. 14/494,226, Non Final Office Action dated Sep. 7, 2017”, 36 pgs.
“U.S. Appl. No. 14/494,226, Non Final Office Action dated Sep. 12, 2016”, 32 pgs.
“U.S. Appl. No. 141494,226, Response filed Jan. 8, 2018 to Non Final Office Action dated Sep. 7, 2017”, 15 pgs.
“U.S. Appl. No. 14/494,226, Response filed Jul. 7, 2017 to Final Office Action dated Mar. 7, 2017”, 13 pgs.
“U.S. Appl. No. 14/494,226, Response filed Dec. 12, 2016 to Non Final Office Action dated Sep. 12, 2016”, 16 pgs.
“U.S. Appl. No. 14/510,016, Advisory Action dated Nov. 30, 2017”, 7 pgs.
“U.S. Appl. No. 14/510,016, Final Office Action dated May 22, 2018”, 36 pgs.
“U.S. Appl. No. 14/510,016, Final Office Action dated Sep. 7, 2018”, 34 pgs.
“U.S. Appl. No. 14/510,016, Final Office Action dated Sep. 8, 2017”, 21 pgs.
“U.S. Appl. No. 14/510,016, Non Final Office Action dated Feb. 7, 2018”, 36 pgs.
“U.S. Appl. No. 14/510,016, Non Final Office Action dated Apr. 21, 2017”, 55 pgs.
“U.S. Appl. No. 14/510,016, Response filed Jan. 8, 2017 to Final Office Action dated Sep. 8, 2017”, 22 pgs.
“U.S. Appl. No. 14/510,016, Response filed May 7, 2018 to Non Final Office Action dated Feb. 7, 2018”, 13 pgs.
“U.S. Appl. No. 14/510,016, Response filed Jul. 21, 2017 to Non Final Office Action dated Apr. 21, 2017”, 21 pgs.
“U.S. Appl. No. 14/510,016, Response filed Aug. 23, 2018 to Final Office Action dated May 22, 2018”, 16 pgs.
“U.S. Appl. No. 14/510,016, Response filed Nov. 8, 2017 to Final Office Action dated Sep. 8, 2017”, 24 pgs.
“U.S. Appl. No. 14/529,064, Examiner Interview Summary dated May 23, 2016”, 3 pgs.
“U.S. Appl. No. 14/529,064, Examiner Interview Summary dated Nov. 17, 2016”, 3 pgs.
“U.S. Appl. No. 14/529,064, Final Office Action dated Jan. 25, 2018”, 39 pgs.
“U.S. Appl. No. 14/529,064, Final Office Action dated Aug. 11, 2015”, 23 pgs.
“U.S. Appl. No. 14/529,064, Final Office Action dated Aug. 24, 2016”, 23 pgs.
“U.S. Appl. No. 14/529,064, Non Final Office Action dated Apr. 6, 2017”, 25 pgs.
“U.S. Appl. No. 14/529,064, Non Final Office Action dated Apr. 18, 2016”, 21 pgs.
“U.S. Appl. No. 14/529,064, Non Final Office Action dated Jul. 13, 2018”, 38 pgs.
“U.S. Appl. No. 14/529,064, Response filed May 25, 2018 to Final Office Action dated Jan. 25, 2018”, 20 pgs.
“U.S. Appl. No. 14/529,064, Response filed Jul. 18, 2016 to Non Final Office Action dated Apr. 18, 2016”, 20 pgs.
“U.S. Appl. No. 14/529,064, Response filed Sep. 6,2017 to Non Final Office Action dated Apr. 6, 2017”, 24 pgs.
“U.S. Appl. No. 14/529,064, Response filed Sep. 6, 2017 to Non Final Office Action dated Apr. 6, 2017”, 19 pgs.
“U.S. Appl. No. 14/529,064, Response filed Oct. 12, 2015 to Final Office Action dated Aug. 11, 2015”, 19 pgs.
“U.S. Appl. No. 14/529,064, Response filed Dec. 21, 2016 to Final Office Action dated Aug. 24, 2016”, 17 pgs.
“U.S. Appl. No. 14/539,391, Notice of Allowance dated Mar. 5, 2015”, 17 pgs.
“U.S. Appl. No. 14/548,590, Advisory Action dated Nov. 18, 2016”, 3 pgs.
“U.S. Appl. No. 14/548,590, Final Office Action dated Jul. 5, 2016”, 16 pgs.
“U.S. Appl. No. 14/548,590, Final Office Action dated Jul. 18, 2017”, 20 pgs.
“U.S. Appl. No. 141548,590, Final Office Action dated Sep. 16, 2015”, 15 pgs.
“U.S. Appl. No. 14/548,590, Non Final Office Action dated Jan. 9, 2017”, 14 pgs.
“U.S. Appl. No. 14/548,590, Non Final Office Action dated Feb. 11, 2016”, 16 pgs.
“U.S. Appl. No. 14/548,590, Non Final Office Action dated Apr. 20, 2015”, 14 pgs.
“U.S. Appl. No. 14/548,590, Response filed May 9, 2017 to Non Final Office Action dated Jan. 9, 2017”, 17 pgs.
“U.S. Appl. No. 14/548,590, Response filed May 10, 2016 to Non Final Office Action dated Feb. 11, 2016”, 14 pgs.
“U.S. Appl. No. 14/548,590, Response filed Nov. 7, 2016 to Final Office Action dated Jul. 5, 2016”, 14 pgs.
“U.S. Appl. No. 14/548,590, Response filed Dec. 16, 2015 to Final Office Action dated Sep. 16, 2015”, 13 pgs.
“U.S. Appl. No. 14/548,590, Response filed Jun. 16, 2015 to Non Final Office Action dated Apr. 20, 2015”, 19 pgs.
“U.S. Appl. No. 14/594,410, Non Final Office Action dated Jan. 4, 2016”, 10 pgs.
“U.S. Appl. No. 14/594,410, Notice of Allowance dated Aug. 2, 2016”, 5 pgs.
“U.S. Appl. No. 14/594,410, Notice of Allowance dated Dec. 15, 2016”, 6 pgs.
“U.S. Appl. No. 14/594,410, Response filed Jul. 1, 2016 to Non Final Office Action dated Jan. 4, 2016”, 10 pgs.
“U.S. Appl. No. 14/634,417, Corrected Notice of Allowability dated Mar. 11, 2019”, 3 pgs.
“U.S. Appl. No. 14/634,417, Corrected Notice of Allowability dated Mar. 20, 2019”, 3 pg.s.
“U.S. Appl. No. 14/682,259, Notice of Allowance dated Jul. 27, 2015”, 17 pgs.
“U.S. Appl. No. 14/704,212, Final Office Action dated Jun. 17, 2016”, 12 pgs.
“U.S. Appl. No. 14/704,212, Non Final Office Action dated Mar. 12, 2018”, 7 pgs.
“U.S. Appl. No. 14/704,212, Non Final Office Action dated Jun. 16, 2017”, 13 pgs.
“U.S. Appl. No. 14/704,212, Non Final Office Action dated Nov. 25, 2016”, 13 pgs.
“U.S. Appl. No. 14/704,212, Non Final Office Action dated Dec. 4, 2015”, 17 pgs.
“U.S. Appl. No. 14/704,212, Notice of Allowance dated Jul. 2, 2018”, 7 pgs.
“U.S. Appl. No. 141704,212, Response filed Feb. 27, 2017 to Non Final Office Action dated Nov. 25, 2016”, 14 pgs.
“U.S. Appl. No. 14/704,212, Response filed Mar. 4, 2016 to Non Final Office Action dated Dec. 4, 2015”, 11 pgs.
“U.S. Appl. No. 14/704,212, Response filed Jun. 12, 2018 to Non Final Office Action dated Mar. 12, 2018”, 9 pgs.
“U.S. Appl. No. 14/704,212, Response filed Oct. 17, 2016 to Final Office Action dated Jun. 17, 2016”, 12 pgs.
“U.S. Appl. No. 14/841,987, Notice of Allowance dated Mar. 29, 2017”, 17 pgs.
“U.S. Appl. No. 14/841,987, Notice of Allowance dated Aug. 7, 2017”, 8 pgs.
“U.S. Appl. No. 14/967,472, Corrected Notice of Allowability dated Mar. 18, 2019”, 3 pgs.
“U.S. Appl. No. 14/967,472, Corrected Notice of Allowability dated Apr. 24, 2019”, 3 pgs.
“U.S. Appl. No. 14/967,472, Notice of Allowance dated Jan. 24, 2019”, 6 pgs.
“U.S. Appl. No. 14/967,472, Response filed Sep. 21, 2018 to Final Office Action dated Jun. 25, 2018”, 11 pgs.
“U.S. Appl. No. 14/974,321, Examiner Interview Summary dated Dec. 5, 2017”, 3 pgs.
“U.S. Appl. No. 14/974,321, Final Office Action dated Oct. 26, 2017”, 16 pgs.
“U.S. Appl. No. 14/974,321, Non Final Office Action dated May 31, 2018”, 14 pgs.
“U.S. Appl. No. 14/974,321, Non Final Office Action dated Jun. 29, 2017”, 36 pgs.
“U.S. Appl. No. 14/974,321, Response filed Aug. 30, 2018 to Non Final Office Action dated May 31, 2018”, 14 pgs.
“U.S. Appl. No. 14/974,321, Response filed Sep. 27, 2017 to Non Final Office Action dated Jun. 29, 2017”, 13 pgs.
“U.S. Appl. No. 15/074,029, Advisory Action dated Oct. 11, 2018”, 3 pgs.
“U.S. Appl. No. 15/074,029, Corrected Notice of Allowability dated Aug. 20, 2019”, 10 pgs.
“U.S. Appl. No. 15/074,029, Final Office Action dated Jun. 28, 2018”, 22 pgs.
“U.S. Appl. No. 15/074,029, Non Final Office Action dated Jan. 23, 2019”, 19 pgs.
“U.S. Appl. No. 15/074,029, Notice of Allowance dated Jun. 19, 2019”, 14 pgs.
“U.S. Appl. No. 15/074,029, Response filed Aug. 28, 2018 to Final Office Action dated Jun. 28, 2018”, 21 pgs.
“U.S. Appl. No. 15/074,029, Response filed Apr. 23, 2019 to Non Final Office Action dated Jan. 23, 2019”,15 pgs.
“U.S. Appl. No. 15/137,608, Amendment and Response filed Jan. 25, 2019 to Non Final Office Action dated Nov. 2, 2018”, 13 pgs.
“U.S. Appl. No. 15/137,608, Corrected Notice of Allowability dated Oct. 2, 2019”, 3 pgs.
“U.S. Appl. No. 15/137,608, Final Office Action dated May 13, 2019”, 10 pgs.
“U.S. Appl. No. 15/137,608, Non Final Office Action dated Nov. 2, 2018”, 10 pgs.
“U.S. Appl. No. 15/137,608, Notice of Allowance dated Aug. 8, 2019”, 7 pgs.
“U.S. Appl. No. 15/137,608, Response filed Jul. 12, 2019 to Final Office Action dated May 13, 2019”, 10 pgs.
“U.S. Appl. No. 15/152,975, Examiner Interview Summary dated Feb. 4, 2019”, 7 pgs.
“U.S. Appl. No. 15/152,975, Final Office Action dated Jul. 2, 2018”, 19 pgs.
“U.S. Appl. No. 15/152,975, Non Final Office Action dated Sep. 28, 2018”, 28 pgs.
“U.S. Appl. No. 15/152,975, Notice of Allowance dated May 17, 2019”, 13 pgs.
“U.S. Appl. No. 15/152,975, Response filed Jan. 28, 2019 to Non Final Office Action dated Sep. 28, 2018”, 17 pgs.
“U.S. Appl. No. 15/152,975, Response filed Sep. 19, 2018 to Final Office Action dated Jul. 2, 2018”, 14 pgs.
“U.S. Appl. No. 15/224,312, Advisory Action dated Aug. 27, 2019”, 3 pgs.
“U.S. Appl. No. 15/224,312, Final Office Action dated Apr. 11, 2019”, 15 pgs.
“U.S. Appl. No. 15/224,312, Non Final Office Action dated Oct. 22, 2018”, 15 pgs.
“U.S. Appl. No. 15/224,312, Response filed Feb. 22, 2019 to Non Final Office Action dated Oct. 22, 2018”, 14 pgs.
“U.S. Appl. No. 15/224,312, Response filed Aug. 20, 2018 to Office Action dated Apr. 20, 2018”, 16 pgs.
“U.S. Appl. No. 15/224,312, Response filed Oct. 11, 2019 to Advisory Action dated Aug. 27, 2019”, 17 pgs.
“U.S. Appl. No. 15/224,312, Response filed Aug. 12, 2019 to Final Office Action dated Apr. 11, 2019”, 14 pgs.
“U.S. Appl. No. 15/224,343, Amendment and Response filed Feb. 4, 2019 to Non Final Office Action dated Sep. 4, 2018”, 18 pgs.
“U.S. Appl. No. 15/224,343, Final Office Action dated Mar. 22, 2019”, 17 pgs.
“U.S. Appl. No. 15/224,343, Non Final Office Action dated Sep. 4, 2018”, 20 pgs.
“U.S. Appl. No. 15/224,343, Response filed Jul. 19, 2018 to Final Office Action dated Apr. 19, 2018”, 16 pgs.
“U.S. Appl. No. 15/224,343, Response filed Aug. 22, 2019 to Final Office Action dated Mar. 22, 2019”, 16 pgs.
“U.S. Appl. No. 15/224,355, Final Office Action dated Aug. 9, 2019”, 15 pgs.
“U.S. Appl. No. 15/224,355, Non Final Office Action dated Dec. 20, 2018”, 14 pgs.
“U.S. Appl. No. 15/224,355, Response filed May 20, 2019 to Non Final Office Action dated Dec. 20, 2018”, 13 pgs.
“U.S. Appl. No. 15/224,355, Response filed Sep. 24, 2018 to Final Office Action dated Apr. 24, 2018”, 19 pgs.
“U.S. Appl. No. 15/224,359, Final Office Action dated Apr. 11, 2019”, 15 pgs.
“U.S. Appl. No. 15/224,359, Non Final Office Action dated Sep. 28, 2018”, 15 pgs.
“U.S. Appl. No. 15/224,359, Response filed Feb. 28, 2019 to Non Final Office Action dated Aug. 28, 2018”, 16 pgs.
“U.S. Appl. No. 15/224,359, Response filed Sep. 4, 2018 to Final Office Action dated Apr. 2, 2018”, 14 pgs.
“U.S. Appl. No. 15/224,359, Response filed Sep. 11, 2019 to Final Office Action dated Apr. 11, 2019”, 18 pgs.
“U.S. Appl. No. 15/224,365, Final Office Action dated Aug. 23, 2019”, 12 pgs.
“U.S. Appl. No. 15/224,365, Non Final Office Action dated Jan. 3, 2019”, 11 pgs.
“U.S. Appl. No. 15/224,365, Response filed Oct. 2, 2018 to Final Office Action dated Apr. 2, 2018”, 15 pgs.
“U.S. Appl. No. 15/224,365, Response filed Jun. 3, 2019 to Final Office Action dated Jan. 3, 2019”, 12 pgs.
“U.S. Appl. No. 15/224,372, Final Office Action dated Mar. 6, 2019”, 17 pgs.
“U.S. Appl. No. 15/224,372, Non Final Office Action dated Sep. 14, 2018”, 20 pgs.
“U.S. Appl. No. 15/224,372, Non Final Office Action dated Oct. 16, 2019”, 14 pgs.
“U.S. Appl. No. 15/224,372, Response filed Jan. 16, 2019 to Non Final Office Action dated Sep. 14, 2018”, 18 pgs.
“U.S. Appl. No. 15/224,372, Response filed Jul. 8, 2019 to Final Office Action dated Mar. 6, 2019”, 14 pgs.
“U.S. Appl. No. 15/224,372, Response filed Aug. 3, 2018 to Final Office Action dated Apr. 3, 2018”, 14 pgs.
“U.S. Appl. No. 15/224,377, Examiner Interview Summary dated Mar. 4, 2019”, 5 pgs.
“U.S. Appl. No. 15/224,377, Final Office Action dated Feb. 6, 2019”, 14 pgs.
“U.S. Appl. No. 15/224,377, Non Final Office Action dated Oct. 15, 2019”, 12 pgs.
“U.S. Appl. No. 15/224,377, Response filed Jun. 6, 2019 to Final Office Action dated Feb. 6, 2019”, 10 pgs.
“U.S. Appl. No. 15/224,377, Response filed Dec. 17, 2018 to Non Final Office Action dated Jun. 15, 2018”, 13 pgs.
“U.S. Appl. No. 15/224,383, Examiner Interview Summary dared Aug. 15, 2018”, 4 pgs.
“U.S. Appl. No. 15/224,383, Final Office Action dated Jan. 14, 2019”, 15 pgs.
“U.S. Appl. No. 15/224,383, Non Final Office Action dated Jul. 5, 2018”, 19 pgs.
“U.S. Appl. No. 15/224,383, Non-Final Office Action dated Sep. 23, 2019”, 13 pgs.
“U.S. Appl. No. 15/224,383, Response filed May 14, 2019 to Final Office Action dated Jan. 14, 2019”, 15 pgs.
“U.S. Appl. No. 15/224,383, Response filed Dec. 5, 2018 to Non Final Office Action dated Jul. 5, 2018”, 16 pgs.
“U.S. Appl. No. 15/298,806, Examiner Interview Summary dated Aug. 13, 2018”, 3 pgs.
“U.S. Appl. No. 15/298,806, Notice of Allowance dated Sep. 19, 2018”, 5 pgs.
“U.S. Appl. No. 15/298,806, Response filed Aug. 10, 2018 to Non Final Office Action dated May 17, 2018”, 15 pgs.
“U.S. Appl. No. 15/486,111, Corrected Notice of Allowance dated Sep. 7, 2017”, 3 pgs.
“U.S. Appl. No. 15/486,111, Non Final Office Action dated May 9, 2017”, 17 pgs.
“U.S. Appl. No. 15/486,111, Notice of Allowance dated Aug. 30, 2017”, 5 pgs.
“U.S. Appl. No. 15/486,111, Response filed Aug. 9, 2017 to Non Final Office Action dated May 9, 2017”, 11 pgs.
“U.S. Appl. No. 15/673,137, Non Final Office Action dated Oct. 5, 2018”, 7 pgs.
“U.S. Appl. No. 15/702,511, 312 Amendment filed Jun. 26, 2019”, 11 pgs.
“U.S. Appl. No. 15/702,511, Notice of Allowability dated Sep. 30, 2019”, 2 pgs.
“U.S. Appl. No. 15/702,511, Notice of Allowance dated Mar. 26, 2019”, 7 pgs
“U.S. Appl. No. 15/702,511, Notice of Allowance dated Oct. 26, 2018”, 7 pgs.
“U.S. Appl. No. 15/702,511, PTO Response to Rule 312 Communication dated Aug. 13, 2019”, 2 pgs.
“U.S. Appl. No. 15/729,582, Corrected Notice of Allowability dated Oct. 2, 2019”, 3 pgs.
“U.S. Appl. No. 15/729,582, Final Office Action dated Dec. 13, 2018”, 14 pgs.
“U.S. Appl. No. 15/729,582, Notice of Allowance dated Jul. 22, 2019”, 9 pgs.
“U.S. Appl. No. 15/729,582, Response filed May 13, 2019 to Final Office Action dated Dec. 13, 2018”, 9 pgs.
“U.S. Appl. No. 15/787,467, Corrected Notice of Allowability dated Sep. 24, 2018”, 2 pgs.
“U.S. Appl. No. 15/787,467, Notice of Allowance dated Aug. 31, 2018”, 8 pgs.
“U.S. Appl. No. 15/787,467, Response filed Jul. 18, 2018 to Non Final Office Action dated Apr. 18, 2018”, 12 pgs.
“U.S. Appl. No. 15/835,100, Non Final Office Action dated Jan. 23, 2018”, 18 pgs.
“U.S. Appl. No. 15/946,990, Final Office Action dated May 9, 2019”, 11 pgs.
“U.S. Appl. No. 15/946,990, Non Final Office Action dated Dec. 3, 2018”, 10 pgs.
“U.S. Appl. No. 15/946,990, Notice of Allowance dated Sep. 24, 2019”, 5 pgs.
“U.S. Appl. No. 15/946,990, Response filed Feb. 20, 2019 to Non Final Office Action dated Dec. 3, 2018”, 11 pgs.
“U.S. Appl. No. 15/946,990, Response filed Jul. 9, 2019 to Final Office Action dated May 9, 2019”, 12 pgs.
“U.S. Appl. No. 16/204,886, Corrected Notice of Allowability dated Jul. 15, 2019”, 2 pgs.
“U.S. Appl. No. 16/204,886, Corrected Notice of Allowability dated Aug. 6, 2019”, 2 pgs.
“U.S. Appl. No. 16/204,886, Corrected Notice of Allowability dated Sep. 10, 2019”, 2 pgs.
“U.S. Appl. No. 16/204,886, Non Final Office Action dated Jan. 4, 2019”, 8 pgs.
“U.S. Appl. No. 16/204,886, Notice of Allowance dated May 15, 2019”, 9 pgs.
“U.S. Appl. No. 16/204,886, Response filed Apr. 2, 2019 to Non Final Office Action dated Jan. 4, 2019”, 8 pgs.
“U.S. Appl. No. 16/219,577, Response filed Oct. 3, 2019 to Restriction Requirement dated Aug. 7, 2019”, 6 pgs.
“U.S. Appl. No. 16/219,577, Restriction Requirement dated Aug. 7, 2019”, 6 pgs.
“U.S. Appl. No. 16/376,598, Non Final Office Action dated Jul. 25, 2019”, 7 pgs.
“U.S. Appl. No. 16/376,598, Notice of Allowance dated Oct. 18, 2019”, 5 pgs.
“U.S. Appl. No. 16/376,598, Response filed Oct. 7, 2019 to Non-Final Office Action dated Jul. 25, 2019”, 2 pgs.
“U.S. Appl. No. 16/511,834, Non-Final Office Action dated Aug. 20, 2019”, 11 pgs.
“U.S. Appl. No. 16/511,834, Notice of Allowance dated Oct. 23, 2019”, 8 pgs.
“U.S. Appl. No. 16/511,834, Response filed Oct. 7, 2019 to Non-Final Office Action dated Aug. 20, 2019”, 3 pgs.
“BlogStomp”, StompSoftware, [Online] Retrieved from the Internet: <URL: http://stompsoftware.com/blogstomp, (accessed May 24, 2017), 12 pgs.
“Canadian Application Serial No. 2,910,158, Response filed Dec. 6, 2018 to Office Action dated Jun. 6, 2018”, w/ English Claims, 18 pgs.
“Chinese Application Serial No. 201580031616.8, Office Action dated Jul. 2, 2018”, w/ English translation, 8 pgs.
“Chinese Application Serial No. 201580031616.8, Office Action dated Oct. 9, 2017”, w/ English Translation, 18 pgs.
“Chinese Application Serial No. 201580031616.8, Response filed Feb. 26, 2018 to Office Action dated Oct. 9, 2017”, w/ English Translation, 8 pgs.
“Chinese Application Serial No. 201580076228.1, Office Action dated Jul. 19, 2018”, w/ English translation, 19 pgs.
“Cup Magic Starbucks Holiday Red Cups come to life with AR app”, Blast Radius, [Online] Retrieved from the Internet: <URL: https://web.archive.org/web/20160711202454/http://www.blastradius.com/work/cup-magic>, (2016), 7 pgs.
“Daily App: InstaPlace (iOS/Android): Give Pictures a Sense of Place”, TechPP, [Online] Retrieved from the Internet: <URL: http://techpp.com/2013/02/15/instaplace-app-review>, (2013), 13 pgs.
“European Application Serial No. 15733026.7, Communication Pursuant to Article 94(3) EPC dated Jul. 28, 2017”, 6 pgs.
“European Application Serial No. 15733026.7, Response filed Jan. 30, 2018 to Communication Pursuant to Article 94(3) EPC dated Jul. 28, 2017”, w/ Amended Claims, 37 pgs.
“European Application Serial No. 15782165.3, Communication Pursuant to Article 94(3) EPC dated Sep. 14, 2018”, 7 pgs.
“European Application Serial No. 15782165.3, Response filed Oct. 17, 2017 to Communication pursuant to Rules 161(1) and 162 EPC dated May 10, 2017”, 15 pgs.
“European Application Serial No. 15787854.7, Response filed Dec. 11, 2017 to Communication Pursuant to Rules 161(1) and 162 EPC dated Jun. 11, 2017”, 16 pgs.
“European Application Serial No. 15870861.0, Communication Pursuant to Article 94(3) EPC dated Jul. 12, 2018”, 5 pgs.
“European Application Serial No. 15870861.0, Extended European Search Report dated Jul. 3, 2018”, 4 pgs.
“European Application Serial No. 15870861.0, Response filed Aug. 9, 2017 to Communication Pursuant to Rules 161(2) and 162 EPC dated Aug. 4, 2017”, 10 pgs.
“European Application Serial No. 16716090.2, Response filed May 21, 2018 to Communication pursuant to Rules 161(1) and 162 EPC dated Nov. 10, 2017”, w/ English Claims, 89 pgs.
“InstaPlace Photo App Tell the Whole Story”, [Online] Retrieved from the Internet: <URL: https://youtu.be/uF_gFkg1hBM>, (Nov. 8, 2013), 113 pgs., 1:02 min.
“International Application Serial No. PCT/US2015/050424, International Search Report dated Dec. 4, 2015”, 2 pgs.
“International Application Serial No. PCT/US2015/050424, Written Opinion dated Dec. 4, 2015”, 10 pgs.
“International Application Serial No. PCT/US2016/066976, International Preliminary Report on Patentability dated Jun. 28, 2018”, 9 pgs.
“International Application Serial No. PCT/US2016/066976, International Search Report dated May 17, 2017”, 7 pgs.
“International Application Serial No. PCT/US2016/066976, Invitation to Pay Add'l Fees and Partial Search Rpt dated Mar. 6, 2017”, 8 pgs.
“International Application Serial No. PCT/US2016/066976, Written Opinion dated May 17, 2017”, 7 pgs.
“International Application Serial No. PCT/US2018/024093, International Search Report dated Jul. 19, 2018”, 2 pgs.
“International Application Serial No. PCT/US2018/024093, Written Opinion dated Jul. 19, 2018”, 5 pgs.
“Introducing Snapchat Stories”, [Online] Retrieved from the Internet: <URL: https ://web.archive.org/web/20131026084921/https://www.youtube.com/watch?v =88Cu3yN-LIM>, (Oct. 3, 2013), 92 pgs.; 00:47 min.
“Korean Application Serial No. 10-2017-7001104, Response filed Jul. 25, 2017 to Office Action dated Jun. 26, 2017”, w/ Translation of Claims, 20 pgs.
“Korean Application Serial No. 10-2017-7014135, Notice of Preliminary Rejection dated Jul. 20, 2018”, w/ English Translation, 13 pgs.
“Korean Application Serial No. 10-2017-7020217, Final Office Action dated Jan. 31, 2018”, w/ English Translation, 10 pgs.
“Korean Application Serial No. 10-2017-7020217, Office Action dated Sep. 15, 2017”, w/ English Translation, 11 pgs.
“Korean Application Serial No. 10-2017-7020217, Response filed Feb. 23, 2018 to Final Office Action dated Jan. 31, 2018”, w/ English Translation, 13 pgs.
“Korean Application Serial No. 10-2017-7020217, Response filed Nov. 2, 2017 to Office Action dated Sep. 15, 2017”, w/ English Translation, 17 pgs.
“Korean Application Serial No. 10-2018-7002127, Notice of Preliminary Rejection dated Apr. 10, 2018”, w/ English Translation, 4 pgs.
“Korean Application Serial No. 10-2018-7002127, Response filed May 16, 2018 to Notice of Preliminary Rejection dated Apr. 10, 2018”, w/ English Claims, 14 pgs.
“Korean Office Action Application Serial No. 10-2017-7001104, Office Action dated Jun. 26, 2017”, W/ English Translation, 12 pgs.
“Macy's Believe-o-Magic”, [Online] Retrieved from the Internet: <URL: https://web.archive.org/web/20190422101854/https://www.youtube.com/watch?v=xvzRXy3J0Z0&feature=youtu.be>, (Nov. 7, 2014 102 pgs.; 00:51 min.
“Macy's Introduces Augmented Reality Experience in Stores across Country as Part of Its 2011 Believe Campaign”, Business Wire, [Online] Retrieved from the Internet: <URL: https://www.businesswire.com/news/home/20111102006759/en/Macys-Introduces-Augmented-Reality-Experience-Stores-Country>, (Nov. 2, 2011), 6 pgs.
“Starbucks Cup Magic”, [Online] Retrieved from the internet: <URL: https://www.youtube.com/watch?v=RWwQXi9RG0w>, (Nov. 8, 2011), 87 pgs.; 00:47 min.
“Starbucks Cup Magic for Valentine's Day”, [Online] Retrieved from the Internet: <URL: https://www.youtube.com/watch?v=8nvqOzjq10w>, (Feb. 6, 2012), 88 pgs.; 00:45 min.
“Starbucks Holiday Red Cups Come to Life, Signaling the Return of the Merriest Season”, Business Wire, [Online] Retrieved from the Internet: <URL: http://www.businesswire.com/news/home/20111115005744/en/2479513/Starbucks-Holiday-Red-Cups-Life-Signaling-Return>, (Nov. 15, 2011), 5 pgs.
“To Err is Human. To Self Destruct Messages, There is iDelete for iOS”, The Apple Google, [Online]. Retrieved from the Internet on Mar. 21, 2018: <http://theapplegoogle.com/2013/04/err-human-destruct-messages-idelete-ios/>, (2013), 2 pgs.
Carthy, Roi, “Dear All Photo Apps: Mobli Just Won Filters”, TechCrunch, [Online] Retrieved from the Internet: <URL: https://techcrunch.com/2011/09/08/mobli-filters>, (Sep. 8, 2011), 10 pgs.
Janthong, Isaranu, “Instaplace ready on Android Google Play store”, Android App Review Thailand, [Online] Retrieved from the Internet: <URL: http://www.android-free-app-review.com/2013/01/instaplace-android-google-play-store.html>, (Jan. 23, 2013), 9 pgs.
Macleod, Duncan, “Macys Believe-o-Magic App”, [Online] Retrieved from the Internet: <URL: http://theinspirationroom.com/daily/2011/macys-believe-o-magic-app>, (Jan. 14, 2011), 10 pgs.
Macleod, Duncan, “Starbucks Cup Magic Lets Merry”, [Online] Retrieved from the Internet: <URL: http://theinspirationroom.com/daily/2011/starbucks-cup-magic>, (Nov. 12, 2011), 8 pgs.
Notopoulos, Katie, “A Guide to the New Snapchat Filters and Big Fonts”, [Online] Retrieved from the Internet: <URL: https://www.buzzfeed.com/katienotopoulos/a-guide-to-the-new-snapchat-filters-and-big-fonts?utm_term=.bkQ9qVZWe#.nv58YXpkV>, (Dec. 22, 2013), 13 pgs.
Panzarino, Matthew, “Snapchat Adds Filters, A Replay Function and for Whatever Reason, Time, Temperature and Speed Overlays”, TechCrunch, [Online] Retrieved form the Internet: <URL: https://techcrunch.com/2013/12/20/snapchat-adds-filters-new-font-and-for-some-reason-time-temperature-and-speed-overlays/>, (Dec. 20, 2013), 12 pgs.
Tripathi, Rohit, “Watemark Images in PHP and Save File on Server”, [Online] Retrieved from the Internet: <URL: http://code.rohitink.com/2012/12/28/watermark-images-in-php-and-save-file-on-server>, (Dec. 28, 2012), 4 pags.
Vaas, Lisa, “StealthText, Should You Choose to Accept It”, [Online] Retrieved from the Internet: <URL: http://www.eweek.com/print/c/a/MessagingandCollaboration/StealthTextShouldYouChoosetoAcceptIt>, (Dec. 13, 2005), 2 pgs.
“U.S. Appl. No. 15/074,029, Corrected Notice of Allowability dated Feb. 5, 2020”, 4 pgs.
“U.S. Appl. No. 15/224,312, Final Office Action dated May 1, 2020”, 18 pgs.
“U.S. Appl. No. 15/224,312, Non Final Office Action dated Dec. 16, 2019”, 14 pgs.
“U.S. Appl. No. 15/224,312, Response filed Apr. 16, 2020 to Non Final Office Action dated Dec. 16, 2019”, 14 pgs.
“U.S. Appl. No. 15/224,343, Final Office Action dated Apr. 7, 2020”, 16 pgs.
“U.S. Appl. No. 15/224,343, Non Final Office Action dated Nov. 12, 2019”, 16 pgs.
“U.S. Appl. No. 15/224,343, Response filed Mar. 2, 2020 to Non Final Office Action dated Nov. 12, 2019”, 17 pgs.
“U.S. Appl. No. 15/224,343, Response filed Jun. 30, 2020 to Final Office Action dated Apr. 7, 2020”, 12 pgs.
“U.S. Appl. No. 15/224,355, Final Office Action dated May 1, 2020”, 15 pgs.
“U.S. Appl. No. 15/224,355, Non Final Office Action dated Jan. 22, 2020”, 13 pgs.
“U.S. Appl. No. 15/224,355, Response filed Apr. 22, 2020 to Non Final Office Action dated Jan. 22, 2020”, 13 pgs.
“U.S. Appl. No. 15/224,355, Response filed Nov. 11, 2019 to Final Office Action dated Aug. 9, 2019”, 14 pgs.
“U.S. Appl. No. 15/224,359, Final Office Action dated May 1, 2020”, 13 pgs.
“U.S. Appl. No. 15/224,359, Non Final Office Action dated Dec. 10, 2019”, 12 pgs.
“U.S. Appl. No. 15/224,359, Response filed Apr. 10, 2020 to Non Final Office Action dated Dec. 10, 2019”, 11 pgs.
“U.S. Appl. No. 15/224,365, Final Office Action dated Jul. 2, 2020”, 11 pgs.
“U.S. Appl. No. 15/224,365, Non Final Office Action dated Mar. 13, 2020”, 9 pgs.
“U.S. Appl. No. 15/224,365, Response filed Jan. 23, 2020 to Final Office Action dated Aug. 23, 2019”, 13 pgs.
“U.S. Appl. No. 15/224,365, Response filed Jun. 15, 2020 to Non Final Office Action dated Mar. 13, 2020”, 12 pgs.
“U.S. Appl. No. 15/224,372, Final Office Action dated May 4, 2020”, 15 pgs.
“U.S. Appl. No. 15/224,372, Response filed Apr. 16, 2020 to Non Final Office Action dated Oct. 16, 2019”, 14 pgs.
“U.S. Appl. No. 15/224,377, Final Office Action dated May 5, 2020”, 15 pgs.
“U.S. Appl. No. 15/224,377, Response filed Apr. 15, 2020 to Non Final Office Action dated Oct. 15, 2019”, 13 pgs.
“U.S. Appl. No. 15/224,383, Notice of Allowance dated Feb. 27, 2020”, 7 pgs.
“U.S. Appl. No. 15/224,383, Response filed Jan. 23, 2020 to Non Final Office Action dated Sep. 23, 2019”, 14 pgs.
“U.S. Appl. No. 15/729,582, Corrected Notice of Allowability dated Oct. 30, 2019”, 3 pgs.
“U.S. Appl. No. 15/947,350, Examiner Interview Summary dated Jul. 20, 2020”, 4 pgs.
“U.S. Appl. No. 15/947,350, Final Office Action dated May 4, 2020”, 12 pgs.
“U.S. Appl. No. 15/947,350, Non Final Office Action dated Dec. 13, 2019”, 20 pgs.
“U.S. Appl. No. 15/947,350, Response filed Apr. 13, 2020 to Non Final Office Action dated Dec. 13, 2019”, 12 pgs.
“U.S. Appl. No. 16/000,657, Examiner Interview Summary dated Jun. 12, 2020”, 4 pgs.
“U.S. Appl. No. 16/000,657, Final Office Action dated Jul. 27, 2020”.
“U.S. Appl. No. 16/000,657, Non Final Office Action dated Mar. 6, 2020”, 30 pgs.
“U.S. Appl. No. 16/000,657, Response filed Jul. 6, 2020 to Non Final Office Action dated Mar. 6, 2020”, 13 pgs.
“U.S. Appl. No. 16/219,577, Non Final Office Action dated Oct. 29, 2019”, 7 pgs.
“U.S. Appl. No. 16/219,577, Response filed Dec. 5, 2019 to Non Final Office Action dated Oct. 29, 2019”, 6 pgs.
“U.S. Appl. No. 16/376,598, Notice of Allowability dated Jan. 23, 2020”, 2 pgs.
“U.S. Appl. No. 16/511,834, Corrected Notice of Allowability dated Jan. 27, 2020”, 2 pgs.
“U.S. Appl. No. 16/541,919, Non Final Office Action dated Apr. 14, 2020”, 18 pgs.
“U.S. Appl. No. 16/541,919, Notice of Allowance dated Jun. 30, 2020”, 8 pgs.
“U.S. Appl. No. 16/541,919, Response filed Jun. 12, 2020 to Non Final Office Action dated Apr. 14, 2020”, 8 pgs.
“U.S. Appl. No. 16/662,956, Preliminary Amendment filed Oct. 24, 2019”, 8 pgs.
“U.S. Appl. No. 16/667,814, Preliminary Amendment filed Apr. 20, 2020”, 6 pgs.
“U.S. Appl. No. 16/703,526, Notice of Allowance dated Jun. 19, 2020”, 10 pgs.
“U.S. Appl. No. 16/709,092, Corrected Notice of Allowability dated Jun. 1, 2020”, 2 pgs.
“U.S. Appl. No. 16/709,092, Corrected Notice of Allowability dated Jul. 22, 2020”, 2 pgs.
“U.S. Appl. No. 16/709,092, Notice of Allowance dated Apr. 9, 2020”, 9 pgs.
“Canadian Application Serial No. 3,027,981, Office Action dated Dec. 5, 2019”, 4 pgs.
“Canadian Application Serial No. 3,027,981, Response filed Mar. 31, 2020 to Office Action dated Dec. 5, 2019”, 12 pgs.
“U.S. Appl. No. 15/224,312, Non Final Office Action dated Nov. 9, 2020”, 18 pgs.
“U.S. Appl. No. 15/224,312, Response filed Oct. 1, 2020 to Final Office Action dated May 1, 2020”, 18 pgs.
“U.S. Appl. No. 15/224,343, Notice of Allowance dated Jul. 29, 2020”, 7 pgs.
“U.S. Appl. No. 15/224,343, Notice of Allowance dated Nov. 16, 2020”, 7 pgs.
“U.S. Appl. No. 15/224,355, Response filed Sep. 1, 2020 to Final Office Action dated May 1, 2020”, 16 pgs.
“U.S. Appl. No. 15/224,359, Notice of Allowance dated Nov. 3, 2020”, 15 pgs.
“U.S. Appl. No. 15/224,359, Response filed Sep. 1, 2020 to Final Office Action dated May 1, 2020”, 13 pgs.
“U.S. Appl. No. 15/224,365, Non Final Office Action dated Dec. 10, 2020”, 16 pgs.
“U.S. Appl. No. 15/224,365, Response filed Oct. 2, 2020 to Final Office Action dated Jul. 2, 2020”, 13 pgs.
“U.S. Appl. No. 15/224,372, Notice of Allowance dated Jan. 12, 2021”, 8 pgs.
“U.S. Appl. No. 15/224,372, Response filed Oct. 5, 2020 to Final Office Action dated May 4, 2020”, 17 pgs.
“U.S. Appl. No. 15/224,377, Non Final Office Action dated Oct. 30, 2020”, 15 pgs.
“U.S. Appl. No. 15/224,377, Response filed Sep. 8, 2020 to Final Office Action dated May 5, 2020”, 15 pgs.
“U.S. Appl. No. 15/947,350, Non Final Office Action dated Sep. 28, 2020”, 13 pgs.
“U.S. Appl. No. 15/947,350, Response filed Sep. 4, 2020 to Final Office Action dated May 4, 2020”, 12 pgs.
“U.S. Appl. No. 16/000,657, Advisory Action dated Oct. 19, 2020”, 3 pgs.
“U.S. Appl. No. 16/000,657, Examiner Interview Summary dated Sep. 25, 2020”, 3 pgs.
“U.S. Appl. No. 16/000,657, Response filed Sep. 28, 2020 to Final Office Action dated Jul. 27, 2020”, 12 pgs.
“U.S. Appl. No. 16/541,919, Notice of Allowance dated Oct. 15, 2020”, 8 pgs.
“U.S. Appl. No. 16/662,956, Non Final Office Action dated Oct. 6, 2020”, 13 pgs.
“U.S. Appl. No. 16/662,956, Respolnse filed Dec. 2, 2020 to Non Final Office Action dated Oct. 6, 2020”, 11 pgs.
“U.S. Appl. No. 16/667,814, Corrected Notice of Allowability dated Dec. 23, 2020”, 2 pgs.
“U.S. Appl. No. 16/667,814, Non Final Office Action dated Aug. 17, 2020”, 6 pgs.
“U.S. Appl. No. 16/667,814, Notice of Allowance dated Nov. 23, 2020”, 8 pgs.
“U.S. Appl. No. 16/667,814, Response filed Oct. 29, 2020 to Non Final Office Action dated Aug. 17, 2020”, 7 pgs.
“U.S. Appl. No. 16/703,526, Corrected Notice of Allowability dated Sep. 2, 2020”, 2 pgs.
“U.S. Appl. No. 16/703,526, Supplemental Notice of Allowability dated Aug. 10, 2020”, 2 pgs.
“Canadian Application Serial No. 3,027,981, Office Action dated Oct. 2, 2020”, 5 pgs.
“U.S. Appl. No. 15/224,312, Response filed Apr. 9, 2021 to Non Final Office Action dated Nov. 9, 2020”, 17 pgs.
“U.S. Appl. No. 15/224,377, Final Office Action dated Apr. 14, 2021”, 14 pgs.
“U.S. Appl. No. 15/224,377, Response filed Mar. 30, 2021 to Non Final Office Action dated Oct. 30, 2020”, 14 pgs.
“U.S. Appl. No. 15/947,350, Final Office Action dated Apr. 8, 2021”, 13 pgs.
“U.S. Appl. No. 15/947,350, Response filed Mar. 2, 2021 to Non Final Office Action dated Sep. 28, 2020”, 12 pgs.
“U.S. Appl. No. 16/000,657, Notice of Allowance dated Feb. 4, 2021”, 8 pgs.
“U.S. Appl. No. 16/662,956, Final Office Action dated mailed Mar. 29, 2021”, 17 pgs.
“U.S. Appl. No. 16/667,814, Corrected Notice of Allowability dated Mar. 2, 2021”, 2 pgs.
“U.S. Appl. No. 16/911,854, Non Final Office Action dated Mar. 3, 2021”, 12 pgs.
“U.S. Appl. No. 16/933,205, Non Final Office Action dated Apr. 16, 2021”, 39 pgs.
“U.S. Appl. No. 16/933,279, Non Final Office Action dated Mar. 25, 2021”, 41 pgs.
“U.S. Appl. No. 16/933,366, Non Final Office Action dated Apr. 27, 2021”, 39 pgs.
“Canadian Application Serial No. 3,027,981, Response filed Feb. 2, 2021 to Office Action dated Oct. 2, 2020”, 15 pgs.
“European Application Serial No. 15870874.3, Communication Pursuant to Article 94(3) EPC dated Feb. 22, 2021”, 5 pgs.
“U.S. Appl. No. 16/000,657, 312 Amendment filed Apr. 30, 2021”, 8 pgs.
“U.S. Appl. No. 16/000,657, PTO Response to Rule 312 Communication dated May 11, 2021”, 3 pgs.
“U.S. Appl. No. 15/224,312, Final Office Action dated May 12, 2021”, 21 pgs.
“U.S. Appl. No. 16/841,817, Non Final Office Action dated May 26, 2021”, 7 pgs.
“U.S. Appl. No. 16/911,854, Response filed May 28, 2021 to Non Final Office Action dated Mar. 3, 2021”, 8 pgs.
“U.S. Appl. No. 17/023,175, Non Final Office Action dated Jun. 8, 2021”, 8 pgs.
“U.S. Appl. No. 16/911,854, Notice of Allowance dated Jun. 17, 2021”, 8 pgs.
“U.S. Appl. No. 15/224,355, Notice of Allowance dated Jul. 13, 2021”, 16 pgs.
“U.S. Appl. No. 15/224,377, Response filed Sep. 14, 2021 to Final Office Action dated Apr. 14, 2021”, 13 pages.
“U.S. Appl. No. 16/662,956, Non Final Office Action dated Jul. 21, 2021”, 12 pgs.
“U.S. Appl. No. 16/662,956, Response filed Jun. 24, 2021 to Final Office Action dated Mar. 29, 2021”, 10 pgs.
“U.S. Appl. No. 16/841,817, Corrected Notice of Allowability dated Sep. 16, 2021”, 2 pgs.
“U.S. Appl. No. 16/841,817, Notice of Allowance dated Sep. 3, 2021”, 7 pgs.
“U.S. Appl. No. 16/841,817, Response filed Aug. 26, 2021 to Non Final Office Action dated May 26, 2021”, 6 pgs.
“U.S. Appl. No. 16/911,854, Corrected Notice of Allowability dated Sep. 16, 2021”, 2 pgs.
“U.S. Appl. No. 16/933,366, Response filed Aug. 25, 2021 to Non Final Office Action dated Mar. 25, 2021”, 14 pgs.
“U.S. Appl. No. 16/933,366, filed Aug. 27, 2021 Non Final Office Action dated Apr. 27, 2021”, 16 pgs.
“U.S. Appl. No. 17/023,175, Response filed Sep. 8, 2021 to Non Final Office Action dated Jun. 8, 2021”, 6 pgs.
Provisional Applications (1)
Number Date Country
62105199 Jan 2015 US
Continuations (2)
Number Date Country
Parent 15152975 May 2016 US
Child 16529461 US
Parent 14612692 Feb 2015 US
Child 15152975 US