Secure time-to-live

Information

  • Patent Grant
  • 10567349
  • Patent Number
    10,567,349
  • Date Filed
    Tuesday, June 24, 2014
    10 years ago
  • Date Issued
    Tuesday, February 18, 2020
    4 years ago
Abstract
Determining whether to allow access to a message is disclosed. A message is received from a sender. The message is associated with a first time-to-live (TTL) value. A determination is made that the first time-to-live value has not been exceeded. The determination is made at least in part by obtaining an external master clock time. In response to the determination, access is allowed to the message.
Description
BACKGROUND OF THE INVENTION

Users of electronic devices increasingly desire to communicate privately and securely with one another. Unfortunately, existing approaches to securing communications can be difficult and/or cumbersome to use. As one example, some approaches to data security make use of digital certificates or keys, or pre-shared passwords, which can be tedious to manage. Further, existing approaches are often susceptible to interception (e.g., eavesdropping and man-in-the middle attacks), forensic analysis, and impersonation. Improvements to digital communication techniques are therefore desirable.





BRIEF DESCRIPTION OF THE DRAWINGS

Various embodiments of the invention are disclosed in the following detailed description and the accompanying drawings.



FIG. 1 illustrates an embodiment of an environment in which the exchange of secure communications is facilitated by a security platform.



FIG. 2A illustrates an embodiment of an installation and registration process.



FIG. 2B illustrates an embodiment of a process for generating a pool of keypairs.



FIG. 3 illustrates an example of an interface.



FIG. 4 illustrates an example of a message sending process.



FIG. 5 illustrates an example of a digital security bubble.



FIG. 6 illustrates an example of a digital security bubble.



FIG. 7 illustrates an example of a portion of a digital security bubble.



FIG. 8 illustrates an example of a portion of a digital security bubble.



FIG. 9 illustrates an example of a portion of a digital security bubble.



FIG. 10 illustrates an example of a process for accessing a message included inside a digital security bubble.



FIG. 11 illustrates an example of a registration process.



FIG. 12 illustrates an example of a process for sending a message.



FIG. 13 illustrates an example of a process for performing a synchronous key cache update.



FIG. 14 illustrates an example of a process for performing an asynchronous key cache update.



FIG. 15 illustrates an embodiment of a message composition interface.



FIG. 16 illustrates an embodiment of a message viewing interface.



FIG. 17 illustrates an embodiment of a message viewing interface.



FIG. 18 illustrates an example of a process for determining whether to allow access to a message.



FIG. 19 illustrates an example of a process for determining whether to allow access to a message.



FIG. 20 illustrates an example of a process for determining whether to allow access to a message.





DETAILED DESCRIPTION

The invention can be implemented in numerous ways, including as a process; an apparatus; a system; a composition of matter; a computer program product embodied on a computer readable storage medium; and/or a processor, such as a processor configured to execute instructions stored on and/or provided by a memory coupled to the processor. In this specification, these implementations, or any other form that the invention may take, may be referred to as techniques. In general, the order of the steps of disclosed processes may be altered within the scope of the invention. Unless stated otherwise, a component such as a processor or a memory described as being configured to perform a task may be implemented as a general component that is temporarily configured to perform the task at a given time or a specific component that is manufactured to perform the task. As used herein, the term ‘processor’ refers to one or more devices, circuits, and/or processing cores configured to process data, such as computer program instructions.


A detailed description of one or more embodiments of the invention is provided below along with accompanying figures that illustrate the principles of the invention. The invention is described in connection with such embodiments, but the invention is not limited to any embodiment. The scope of the invention is limited only by the claims and the invention encompasses numerous alternatives, modifications and equivalents. Numerous specific details are set forth in the following description in order to provide a thorough understanding of the invention. These details are provided for the purpose of example and the invention may be practiced according to the claims without some or all of these specific details. For the purpose of clarity, technical material that is known in the technical fields related to the invention has not been described in detail so that the invention is not unnecessarily obscured.



FIG. 1 illustrates an embodiment of an environment in which the exchange of secure communications is facilitated by a security platform (e.g., security platform 102). In the environment shown in FIG. 1, a “digital security bubble” (DSB), described in more detail below, encapsulates or is otherwise provided around a message. The DSB allows information such as encryption information, hardware binding information, message security controls, and decryption information—for multiple recipients (as applicable)—to securely travel with the message. Further, the DSB provides cross-platform support. For example, techniques described herein can be deployed on a variety of operating systems (e.g., Linux, iOS, and Windows), on a variety of smart phone platforms (e.g., iPhone, Android, Windows, Blackberry, etc.), and on a variety of device types (e.g., mobile smart phones, tablets, laptops, desktops, etc.). Using techniques described herein, only intended accounts on intended devices are able to decrypt the messages. Thus, for example, the security platform is unable to decrypt messages. As will be described in more detail below, using the techniques described herein, message participants can maintain a forward secret secure messaging channel, whether communicating synchronously (e.g., where all participants are online or otherwise able to communicate with platform 102) and asynchronously (e.g., where at least one participant is offline or otherwise not in communication with platform 102).


Users of client devices, such as client devices 106-114 communicate securely with one another using techniques described herein. As shown in FIG. 1, client devices include personal computers (110), laptop computers (108), tablets (106), and mobile telephony devices (112, 114). Some client devices, e.g., tablet device 106, make use of techniques described herein via a messaging application (also referred to as an “app”) obtained from a software distribution server 150. Examples of software distribution servers (which can comprise a single server or multiple servers working in cooperation) include app stores (e.g., provided by Apple, Google, Blackberry, Microsoft, Amazon, and/or other entities) and other webservers offering app (and/or other software) downloads. Client devices can also make use of a web interface (e.g., provided by platform 102) instead of or in addition to a dedicated messaging application installed on the device. Other types of devices not depicted in FIG. 1 can also be used in conjunction with the techniques described herein, such as game consoles, camera/video recorders, video players (e.g., incorporating DVD, Blu-ray, Red Laser, Optical, and/or streaming technologies) and other network-connected appliances, as applicable.


Communications are exchanged via one or more networks (depicted collectively in FIG. 1 as network cloud 104). Such networks can include wired, wireless, cellular, and satellite networks. And, such networks can be closed/private networks, as well open networks (e.g., the Internet). Further, as used herein, “communications” and “messages” can take a variety of forms, including: text messages, documents, audiovisual files, SMSes, and voice and video calls. Further, in addition to personal, business, or other types of conversations, the content can pertain to electronic transactions such as credit card security, password protection, directories, and storage drive protection, video on demand security, online gaming, gambling, electronic distribution of music, videos, documents, online learning systems, databases, cloud storage and cloud environments, bank transactions, voting processes, military communications, security of medical records, communication between medically implanted devices and doctors, etc. As will be described in more detail below, the exchange of communications is facilitated by security platform 102 (or embodiments thereof, as applicable).


As will be described in more detail below, a variety of entities can operate embodiments of platform 102. Further, multiple embodiments of platform 102 can exist simultaneously in an environment (with those multiple embodiments operated by a single entity, or different entities) with the techniques described herein adapted as applicable. For example, platform 102 can be operated by a non-profit organization (or an individual, a company, or any other appropriate type of entity or set of entities) for use by the general public (e.g., with arbitrary members of the public able to use platform 102 to exchange communications). As another example, an enterprise organization can operate an embodiment of platform 102 exclusively for use by the employees of the enterprise (and, as applicable, other individuals, such as vendors). As yet another example, a company (or other entity or entities) can operate one or multiple instances of platform 102 on behalf of multiple organizations, such as small business or companies, schools, charitable organizations, etc.


A. Installation/Initialization/Registration

Suppose a user of client device 106 (hereinafter referred to as “Alice”) would like to send a secure message to her friend, Bob (a user of client device 114) in accordance with techniques described herein. In some embodiments, in order to send a message Bob, Alice first obtains a copy of a messaging application suitable for her device. For example, if Alice's tablet device runs iOS, she could obtain an “app” for her tablet from the Apple App Store (an example of software distribution server 106). Bob similarly obtains an appropriate application suitable for his client device 114 (e.g., an Android-based smartphone) from an appropriate location (e.g., the Google Play store or Amazon Appstore). In some embodiments, client devices make use of a web-based application (e.g., made available by platform 102 through interface 118), instead of, or in addition to, a dedicated installed application.


In embodiments where platform 102 is operated on behalf of specific groups of individuals (e.g., on behalf of employees of a company, students/teachers at a school, company stockholders, members of a club, premium customers, etc.), the app can be obtained from a publicly accessible software distribution server as Alice and Bob do above (e.g., from the Google Play store), can be obtained from a privately operated software distribution server (e.g., made available only to company-issued devices or devices otherwise authorized to communicate with the private server), can be provisioned by support personnel associated with the group (e.g., by being directly installed by the support personnel or included in a device image), etc., as applicable. For example, suppose an embodiment of platform 102 is operated by ACME University on behalf of its students and faculty/staff. As mentioned above, the university can itself operate an embodiment of platform 102, or can contract with a third party to make available the embodiment of platform 102 for university users. Freshmen (and other new students/employees, as applicable) at ACME University can be provided with instructions for downloading and installing an ACME University-specific embodiment of the secure messaging application from a university server in conjunction with their new student orientation. As another example, new employees of Beta Corporation can be issued company phones (and/or other devices such as laptops) with an embodiment of the secure messaging application pre-installed and pre-configured by support personnel for Beta Corporation (e.g., where Beta Corporation operates an embodiment of platform 102 on behalf of its employees and business partners). As yet another example, business partners of Beta Corporation (e.g., vendors) can be provided with instructions for provisioning a Beta Corporation-specific embodiment of the secure messaging application via email, or via a website. And, the Beta Corporation-specific embodiment of the secure messaging application can be made available via email, a website, or any other appropriate mechanism.


Returning to the example of Alice (a member of the public, using an embodiment of platform 102 made available to the public), once Alice's tablet 106 has obtained a copy of the secure messaging app, the app is installed, and Alice is able to register for an account. An instance of a messaging app usable in conjunction with the techniques described herein is depicted in FIG. 1 as app 116 (installed on device 106). Examples of events that can occur during an installation/initialization/registration process (200) are illustrated in FIGS. 2A and 2B and will now be described. While the events will be described in one order, events can also be performed in other orders and/or in parallel (instead of in sequence) in other embodiments. Further, various events can be omitted, in some embodiments, as applicable. For example, where an embodiment of platform 102 is made available by an enterprise for use by its employees (or a school on behalf of its student/staff/faculty, etc.), account creation and initialization may at least partially be performed by support personnel (and/or may be performed at least partially in an automated manner based on a new employee/member workflow), instead of being performed by an end user.


In some embodiments, process 200 is performed on a client device, such as Alice's client device 106. The process begins at 202 when a pool of public/private keypairs for the application is generated, on client device 106 (e.g., using RSA, ECDH, or any other appropriate asymmetric encryption algorithms). As one example, the keypairs can be generated using Eliptic Curve Algorithm with Diffie Helman Key Exchange (ECDH). Other cryptographic standards can also be used, such as RSA. In some embodiments, the keypairs are randomly seeded. As will be described in more detail below, each message Alice sends (whether to Bob or anyone else) can be encrypted with a unique, random key that is used only once then destroyed forensically by Alice (the sender's) device. The forensic destruction ensures that the deleted keys cannot be recovered from Alice's device, even via digital forensics methods.



FIG. 2B illustrates an embodiment of a process for generating a plurality of public/private keypairs. In some embodiments, process 250 is performed on a client device (such as client device 106) as portion 202 of process 200. Process 250 begins at 252 when the pool size associated with the client device is initialized. As one example, a default pool size of fifty keys is received as a parameter from platform 102 by application 116. The pool size can also be encoded into application 116 or otherwise provided to device 106 (e.g., configured via an enterprise administrator, where platform 102 is operated on behalf of an enterprise) without requiring the server to transmit the initial pool size. As will be explained in more detail below, the pool size associated with a device can be dynamically adjusted, for example, such that a device (e.g., of a heavy user that is frequently offline) that initially has a pool size of 50 keys can have the size adjusted upward to a pool size of 200 keys (or more).


At 254, a pool of keys (i.e., a number of keypairs equal to the size initialized at 252) is generated on client device 106. As mentioned above, the keypairs can be generated using Eliptic Curve Algorithm with Diffie Helman Key Exchange (ECDH). Other cryptographic standards can also be used, such as RSA.


At 256, a reference value is assigned for each of the respective keypairs. As one example, suppose fifty keypairs are generated at portion 254 of process 250. At 256, fifty respective reference values are assigned to each of the respective keypairs. The reference values will be used to distinguish the various keys in the pool of keys from one another and can be assigned to the keypairs in a variety of ways. As one example, a six digit random number can be generated by device 106 as the first reference value for the first keypair, and each subsequent reference value can be selected as an increment of the first reference value. As another example, every reference value can be randomly selected. Other schemes for selecting/assigning reference values can be employed at 256 as applicable.


At 258, the private keys and reference values are stored (e.g., in a secure database residing on device 106). As will be described in more detail below, the corresponding public keys will be transmitted to platform 102 (along with the associated reference values) and platform 102 will designate one of the public keys in the pool as a reserve key.


Returning to FIG. 2A, at 204, a “random server seed” is generated, and at 206, a “random local seed” is generated. The seeds are used in conjunction with cryptographic key generation, and in some embodiments, the seeds are determined based on captured hardware information (described in more detail below).


At 208, a device identifier (“deviceID”) is created from captured hardware information. Examples of captured hardware information include: hard drive identifiers, motherboard identifiers, CPU identifiers, and MAC addresses for wireless, LAN, Bluetooth, and optical cards. Combinations of information pertaining to device characteristics, such as RAM, CACHE, controller cards, etc., can also be used to uniquely identify the device. Some, or all, of the captured hardware information is run through a cryptographic hash algorithm such as SHA-256, to create a unique deviceID for the device. The captured hardware information can also be used for other purposes, such as to seed cryptographic functions.


At 210, Alice is asked, via an interface provided by app 116, to supply a desired username. Alice enters “Alice” into the interface. A determination is made as to whether the username is available. As one example, app 116 can supply a cryptographic hash of “Alice” to platform 102 for checking. If platform 102 does not already have a record for that hash, the username “Alice” is available for Alice to use. If platform 102 already has a record of that hash, Alice is instructed by the interface to pick an alternate username. Once Alice has selected an available username, she is asked to supply a password. As mentioned above, in some embodiments, portions of process 200 may be omitted (or performed by other entities, as applicable). For example, where a university student at ACME University is getting set up to use an ACME University-specific embodiment of platform 102, the user's name may be preselected or otherwise issued by the University, rather than being selected by the user.


At 212, an application identifier (“appID”) is created. The appID is a unique identifier for the particular installation of the messaging app. If Alice installs the messaging app on multiple devices, each of her devices will have its own unique appID. (And, each of her devices will also have its own unique deviceID.) In some embodiments, the appID is created by hashing Alice's selected password and other information such as device information.


Finally, at 214 Alice's public keys (and reference values), deviceID, and appID are sent to platform 102 in a secure manner. As one example, in some embodiments app 116 is configured to communicate with platform 102 via TLS.


At the conclusion of process 200, Alice is ready to send and receive secure communications.


As mentioned above, alternate versions of processes 200 and/or 250 can be used in accordance with the techniques described herein. As one example, username/password selection (210) can be performed prior to other portions of process 200 (and can be performed by an entity other than the end user of the messaging application, e.g., where an employer determines a username for an employee). As another example, the random server seed generation (204) and random local seed generation (206) can be performed prior to the keypair generation (202), e.g., with the local seed being used in conjunction with the generating of the keypairs. As yet another example, portions of processes 200 and/or 250 can be combined and/or omitted as applicable. For example, instead of generating a pool of fifty key pairs (254), assigning reference values to the pool as a batch operation (256) and storing the keys/values as a batch operation (258), fifty iterations of a process that generates a key pair, assigns a reference value, and stores the information can be performed.


B. Security Platform

As mentioned above, security platform 102 is configured to facilitate the exchange of communications (e.g., among any/all of client devices 106-114). Also as mentioned above, platform 102 can be operated by a variety of entities on behalf of a variety of end users. For example, one embodiment of platform 102 can be made available to members of the public, whether as a public service, or for a fee. As another example, another embodiment of platform 102 can be made available by a business, by a school, by a charitable organization, etc., and its use limited to its employees/students/members, etc., as applicable. Additional detail regarding various aspects of embodiments of platform 102 will now be provided.


Security platform 102 includes one or more interface(s) 118 for communicating with client devices, such as client devices 106-114. As one example, platform 102 provides an application programming interface (API) configured to communicate with apps installed on client devices, such as app 116 and app 138. Platform 102 can also provide other types of interfaces, such as a web interface, or stand alone software programs for desktops and laptops, running on various Operating Systems (OSes). The web interface can allow users of client devices such as client devices 108 and 110 to exchange messages securely (whether with one another or other users), without the need for a separately installed messaging application. The stand alone software program allows users to exchange secure messages via software that is downloaded by each user. As will be discussed in more detail below (e.g., in Section G), in various embodiments, platform 102 makes available (e.g., via one or more interface(s) 118) a master clock time. The master clock time can be used, in various embodiments, to enforce secure time-to-live (TTL) values of messages. The TTL values can be used to enforce (e.g., on behalf of a message sender) time constraints on message access (e.g., by a recipient).


Security platform 102 also includes a database 120. Included in database 120 is a record for each user of platform 102. Each record has associated with it information such as the user's public key pool and associated reference values, deviceID(s), appID(s), and messages. As shown in FIG. 1, database 120 is relational and stores information in a variety of tables, including a table of hashed usernames (124), a table of public keys and reference values (126), a table of deviceIDs (128), a table of appIDs (130), and a table of messages (132). Other techniques can also be used to store the information used by platform 102. For example, messages can be stored in a separate storage 136 instead of being stored within database 120.


Finally, security platform 102 includes a processing engine 134 which performs a variety of tasks, including interacting with database 120 on behalf of interface(s) 118. As will be described in more detail below, one task performed by platform 102 (e.g., by processing engine 134) is to designate one of the keys in the pool of public keys (e.g., received from Alice at the conclusion of portion 214 of process 200) as a “reserve” key. Another task performed by platform 102 (e.g., processing engine 134) is to facilitate the addition of new keys to a user's key pool as the keys are used. Yet another task performed by platform 102 (e.g., processing engine 134) is to dynamically adjust the size of a user's key pool as needed.


The embodiment of platform 102 depicted in FIG. 1 comprises standard commercially available server hardware (e.g., having a multi-core processor(s), 8G+ of RAM, gigabit network interface adaptor(s), and hard drive(s)) running a typical server-class operating system (e.g., Linux). In various embodiments, platform 102 is implemented across a scalable infrastructure comprising multiple such servers, solid state drives, and/or other applicable high-performance hardware.


Whenever platform 102 is described as performing a task, either a single component or a subset of components or all components of platform 102 may cooperate to perform the task. Similarly, whenever a component of platform 102 is described as performing a task, a subcomponent may perform the task and/or the component may perform the task in conjunction with other components.


C. Sending DSB Secured Messages

Returning back to Alice's desire to send a message to Bob: at the conclusion of Section A above, Alice has successfully registered her username (“Alice”) with security platform 102. And, Bob is also a user of platform 102. Suppose Alice would like to send a message to Bob. She starts app 116 and is presented with an interface that includes a “compose” option. Alice selects the compose option and is presented with a message composition interface.


An example message composition interface is shown in FIG. 3. In particular, FIG. 3 depicts interface 300 as rendered on an example tablet device 106, connected to the Internet via an appropriate connection, such as: 3G, 4G or higher cellular connection, WiFi, satellite, wireless or wired LANs, Bluetooth, etc. Tablet device 106 includes a touchscreen. An on-screen keyboard is provided for Alice in region 306. Alice can enter the usernames of one or more recipients in region 302. She can enter message text in region 304. Alice can optionally add attachments by interacting with buttons shown in region 308. Examples of attachments include, but are not limited to: documents, pictures, and audiovisual clips. By selecting button 310, Alice can specify various message control options, such as: the lifetime/expiration of the message (the enforcement of which is described in more detail below in Section G); on which device(s) it can be unencrypted/read; and sharing, saving, forwarding, recalling, and deleting options.


If Alice is satisfied with her message, she can send it to Bob by clicking the send button (314). If she wishes to cancel out of composing the message, she can click the cancel button (312). Suppose Alice clicks send button (314) after composing the message shown in interface 300. An example of the events that occur, in some embodiments, in conjunction with Alice sending a message is illustrated as process 400 in FIG. 4 and will now be described.



FIG. 4 illustrates an example of a process for sending a DSB-secured message. In some embodiments, process 400 is performed on a client device, such as Alice's client device 106. The process begins at 402 when a particular public key (from the user's pool of public keys) and associated reference value, deviceID, and appID of a recipient are obtained from platform 102. As will be explained in more detail below, the recipient's particular public key, deviceID and appID are used in the encryption of the symmetric key used to encrypt data, and in the DSB encapsulation of the message for the hardware/appID binding of the message. As one example, app 116 can request the information from platform 102 via an API (e.g., interface 118). In some embodiments, the information is retrieved when Alice enters the recipient's name into region 302. In other embodiments, the information is retrieved when Alice clicks send button 314, or at any other appropriate time (e.g., while she is composing a message). In the example shown in FIG. 3, Alice is only sending a message to Bob. If she also desires to send the message to other recipients, she can enter their names in region 302 as well, and one of their respective public keys (again selected from their respective pools of public keys) and associated reference values, deviceIDs, and appIDs will also be retrieved at 402.


At 404, a random symmetric encryption key is generated (e.g., by app 116 on device 106). As one example, the symmetric key is an AES 256 bit key. At 406, the symmetric encryption key is used to encrypt the message body, any attachments, and any message control options. In some embodiments, Alice's own information (e.g., public key(s) and associated reference value(s), deviceID(s), and appID(s) are included in the DSB as well. Finally, at 408, the symmetric key is encrypted with the particular public key of each recipient (obtained from the pool of public keys). A DSB encapsulation is then generated, and contains the aforementioned components and reference values of the public keys used to encrypt the symmetric key. Examples of the DSB format are provided in Section D below.


In some cases, a user may own multiple devices. For example, Bob may be the owner of device 114 and 112, both of which are configured with secure messaging apps. Each of Bob's installations will have its own deviceID and appID. When the DSB is created, each of Bob's devices will be considered a separate device under the same username account.


The generated DSB is securely transmitted to platform 102 (e.g., by being encrypted with a symmetric key shared by the app and platform 102, and also encapsulated by TLS as an additional security layer). Irrespective of how many recipients Alice designates for her message (and, e.g., how many recipients there are or how many devices Bob has), only one DSB will be created and transmitted to platform 102. Upon receipt of the DSB, processing engine 134 opens the DSB and determines the recipients of the message. Specifically, the processing engine 134 performs a match against the deviceIDs (in a cryptographic hash and camouflaged representation) included in the DSB and the deviceIDs stored in database 120 as well as the username (in a cryptographic hash and camouflaged representation) in the DSB and the ones stored in the database 120. A cryptographic hash and camouflaged representation means that the hash algorithm (i.e. SHA256) that is used for the deviceID, username, and appID values, is further camouflaged, in some embodiments, by taking multiple hashes of the result values (i.e. multiple rounds of SHA256 of the previous SHA256 value—i.e. SHA(SHA(SHA(SHA . . . ))). Processing engine 134 also creates an entry for the received DSB in message table 132 and notifies the recipient(s) that a new message is available. In various embodiments, other actions are also performed by platform 102 with respect to the DSB. As one example, platform 102 can be configured to remove the DSB as soon as the recipient successfully downloads it. As another example, platform 102 can enforce an expiration time (e.g., seven days) by which, if the DSB has not been accessed by the recipient, the DSB is deleted. Where multiple recipients are included in a DSB, platform 102 can be configured to keep track of which recipients have downloaded a copy of the DSB, and remove it once all recipients have successfully downloaded it (or an expiration event has occurred).


D. DSB Examples


FIG. 5 illustrates an example of a digital security bubble (DSB). DSB 500 is an example of output that can be generated by app 116 as a result of executing process 400. In the example shown, DSB 500 includes a message and optional attachments (502), and one or more message controls (504) encrypted with a key Ek1,1 (encrypted portion 506). In some embodiments, key Ek1,1 is generated by app 116 at portion 404 of process 400. Additional detail regarding portion 506 is shown in FIG. 7, where SSK in FIG. 7 is Ek1,1 of FIG. 5 and represents the sender's symmetric shared key used to encrypt the message and attachments.


DSB 500 also includes, for each message recipient 1-n, the key Ek1,1 encrypted by each of the recipient's respective particular public keys (as shown in region 508). Further, DSB 500 includes a combination of each recipient's respective deviceID, hashed username, appID, and the reference value associated with the particular public key (collectively denoted HWk1-n) in region 510. These constituent parts are also referred to herein as “parameters.” Additional detail regarding the parameters is shown in FIG. 9—namely, a plurality of parameters (such as hashed username, deviceID, and appID) are encrypted using SK2, which is a symmetric key generated by the client and shared with platform 102.


In some embodiments (e.g., as is shown in FIG. 5), a spreading function is used to spread the encrypted symmetric keys inside the DSB (as shown in region 512), by spreading the bits of the encrypted key in a spreading function generated pattern, with the default function being a sequential block or data. The spreading function also contains the cryptographic hashed representation of the recipient usernames that are used by the server to identify the recipients of the message and to set the message waiting flag for each of them. Finally, the DSB is itself encrypted using key Ek1,2 (encrypted portion 514), which is a symmetric key shared between app 116 and platform 102. Additional detail regarding portions 514 and 508 are shown in FIG. 8, where SK1 in FIG. 8 is Ek1,2 in FIG. 5 and represents the symmetric encryption key shared by the app and platform 102, and where User1Pubkey in FIG. 8 is Ek2,1 in FIG. 5 and represents the recipient's particular public key (e.g., selected from the pool of public keys generated at 202).



FIGS. 6-9 illustrate additional examples of the construction of an embodiment of a DSB. FIG. 6 illustrates an example of a DSB 600. DSB 600 encapsulates three subcomponents—part 700 (the encrypted message, attachments, and message controls), part 800 (the symmetric key encrypted with each recipient's particular public key selected from the recipients' respective key pools), and part 900 (encrypted message parameters). As with DSB 500, a symmetric key (shared by app 116 and platform 102) is used to secure the DSB. In addition, the transmission of the DSB to the server is encapsulated with TLS for an additional security layer. FIG. 7 illustrates part 700 of DSB 600. In particular, part 700 includes the message controls (702), message (704), and attachments (706). Part 700 is encrypted using a shared symmetric key SSK (e.g., Ek1,1). FIG. 8 illustrates part 800 of DSB 600. In particular, part 800 includes the shared symmetric key, encrypted to each of the recipients' respective particular public keys (selected from the recipients' respective key pools). Further, the collection of encrypted keys (802-806) is encrypted using symmetric key SK1. FIG. 9 illustrates part 900 of DSB 600. In particular, part 900 includes encrypted message parameters. Part 900 is encrypted using symmetric key SK2.


E. Receiving DSB Secured Messages

As mentioned above, Bob is also a user of platform 102. When Bob loads his copy of the messaging app on his smartphone (i.e., app 138 on device 114), the app communicates with platform 102 (e.g., via interface 118) to determine whether Bob has any new messages. As will be described in more detail below, platform 102 will also determine how many additional keypairs Bob's device should generate to replenish his pool, and facilitate the generation of those keypairs. Since Alice has sent a message to Bob since he last used app 138, a flag is set in database 120, indicating to app 138 that one or messages are available for download.



FIG. 10 illustrates an example of a process for accessing a message included inside a digital security bubble. In some embodiments, process 1000 is performed on a client device, such as Bob's client device 114. The process begins at 1002 when a DSB is received. As one example, a DSB is received at 1002 when app 138 contacts platform 102, determines a flag associated with Bob's account has been set (e.g., indicating he has one or more new messages), and downloads the DSB from platform 102. In such circumstances, upon receipt of the DSB, client 114 is configured to decrypt the DSB using the particular private key of Bob that corresponds to the public key that was selected from his pool at message creation time (and is identifiable by the reference value included in the DSB).


At 1004 (i.e., assuming the decryption was successful), hardware binding parameters are checked. As one example, a determination is made as to whether device information (i.e., collected from device 114) can be used to construct an identical hash to the one included in the received DSB. If the hardware binding parameters fail the check (i.e., an attempt is being made to access Alice's message using Bob's keys on a device that is not Bob's), contents of the DSB will be inaccessible, preventing the decryption of Alice's message. If the hardware binding parameter check is successful, the device is authorized to decrypt the symmetric key (i.e., using Bob's private key generated at 202) which can in turn be used to decrypt Alice's message (1006). As will be described in more detail below (e.g., in Section G), additional controls can be applied (e.g., by Bob's app 138) to Bob's ability to access Alice's message.


F. Additional Example Processes

The following are examples of processes that can be performed by various entities present in environment 100, such as platform 102 and devices 106 and 114 in various embodiments (whether as alternate versions of or additional processes to those described above). The processes can also be performed outside of environment 100, e.g., by other types of platforms and/or devices.



FIG. 11 illustrates an example of a registration process. In some embodiments, process 1100 is performed by device 106. Process 1100 can also be performed by other devices, including devices in environments other than those shown in FIG. 1. Process 1100 begins at 1102 when an initialization value is received. As one example, an initialization value of 50 (corresponding to a target minimum server key cache size of fifty public keys to be stored on platform 102) is received at 1102. In some embodiments, in response to receiving a request from a device, such as device 106, platform 102 sets a server count (C)=0. The server count represents the number of public keys currently stored on platform 102 associated with the device. As device 106 is registering, no keys are present yet on platform 102.


At 1104, a number of keypairs is generated. In this example, a number of asymmetric keypairs equal to the initialization value received at 1102 (e.g., fifty) is generated. In some embodiments, the keypairs are randomly seeded.


At 1106, reference values (e.g., usable to uniquely identify each of the key pairs and described in more detail above) are assigned for each of the keypairs generated at 1104.


At 1108, the private key portion of the key pairs (i.e., the fifty private keys) and associated reference values are securely stored locally (e.g., on device 106). As one example, the private keys are inserted into a database resident on device 106 and secured using an AES key derived from the password selected by Alice at portion 210 in process 200.


Finally, at 1110, the public key portion of the key pairs (i.e., the fifty public keys) and associated reference values are securely transmitted to platform 102. As mentioned above, platform 102 will designate one of the fifty keys as a reserve key (e.g., by setting a flag associated with that particular key).



FIG. 12 illustrates an example of a process for sending a message. In some embodiments, process 1200 is performed by device 114 (e.g., when Bob wants to send a message to Alice). Process 1200 begins at 1202 when device 114 requests a public key associated with Alice from platform 102. If multiple public keys for Alice are present in her pool of keys (i.e., the pool of public keys stored on platform 102 for Alice), the platform will preferentially select (whether randomly, sequentially, or by any other appropriate selection technique) one of the non-reserve keys, and delete the selected key in an atomic operation in conjunction with sending the selected key to device 114. As will be described in more detail below, if only one public key is present for Alice (i.e., only the reserve key remains in the pool), platform 102 will send the reserve key to device 114, but will not delete the reserve key from platform 102 (until such time as the reserve key is replaced with a new key designated as the reserve).


At 1204, a public key is received (e.g., by device 114 from platform 102) along with the reference value associated with the key.


At 1206, the received public key is used to encrypt information, such as a message, or other information (e.g., a symmetric key which in turn is used to encrypt the message). The key reference value associated with the received public key is included in the message metadata or otherwise incorporated into the message payload.


Finally, at 1208, device 114 sends the message (e.g., to platform 102 for retrieval by Alice). Note that using techniques described, Alice's device(s) need not be online (e.g., connected to platform 102) at the time Bob composes and/or sends messages to her.



FIG. 13 illustrates an example of a process for performing a synchronous key cache update. In some embodiments, process 1300 is performed by device 106 (e.g., when Alice connects to platform 102 to retrieve messages). The process begins at 1302 when device 106 connects to platform 102 and retrieves one or more messages.


For each retrieved message (at 1304), read the respective key reference value (e.g., included in the respective message as metadata), retrieve the appropriate private key (i.e., having the key reference value) from local storage on device 106, and decrypt the message(s).


At 1306, device 106 generates additional keypairs (i.e., to replenish public keys used from the pool on platform 102 by Bob). The number of keys to be generated can be determined in a variety of ways. As one example, device 106 can generate a number of new keypairs equal to the number of messages she received at 1302. As another example, device 106 can be instructed (whether by platform 102 or local instructions) to generate the lesser of: A: (the number of messages downloaded at 1302*V), where (V) is a variable impacting the desired expansion rate of the server cache size (e.g. 0.9); or B: the initialization value (e.g., 50 keys, as discussed at 1102 in process 1100).


At 1308 (similar to 1106), reference values (e.g., usable to uniquely identify each of the key pairs and described in more detail above) are assigned for each of the keypairs generated at 1308.


At 1310 (similar to 1108), the private key portion of the key pairs (i.e., the new private keys) and associated reference values are securely stored locally (e.g., on device 106). As one example, the private keys are inserted into a database resident on device 106 and secured using the password selected by Alice at 210 in process 200.


Finally, at 1312 (similar to 1110), the public key portion of the key pairs (i.e., the new public keys) and associated reference values are securely transmitted to platform 102. In this example, suppose Alice's reserve key was not depleted. The key originally designated as her reserve key remains present on platform 102 and remains designated as the reserve key. Now suppose Alice's reserve key was depleted (e.g., because Bob and/or other users of platform 102 sent Alice more than fifty messages prior to her connecting to platform 102). The first 49 messages addressed to Alice would make use of those public keys in her pool not designated as the reserve key. Any additional messages sent to Alice before she can replenish her pool will all make use of her reserve public key (i.e., messages 50, 51, and 52—whether from Bob or others, will all make use of the same public key for Alice—her reserve key). As will be explained below, when Alice's pool has been deleted (i.e., her reserve key is being used), a flag will be set on platform 102 indicating that, in conjunction with her next execution of process 1300 (or portions thereof, as applicable), a new key should be designated as the reserve key, and the existing reserve key be destroyed. Additional actions can also be taken (e.g., by platform 102) in response to Alice depleting her key pool, such as by increasing the size of her pool.



FIG. 14 illustrates an example of a process for performing an asynchronous key cache update. In some embodiments process 1400 is performed by device 106. Process 1400 begins when device 106 connects to platform 102. The connection can be periodic (e.g., app 116 can be configured to connect to platform 102 once a day, once an hour, etc.) and can also be in response to triggering events (e.g., Alice's phone was powered off and has just been powered on, has just connected to a cellular or other network, etc.).


At 1404, the device receives the current server key cache count (i.e., the number of keys presently in the platform's pool for the user). At 1406, the device generates an appropriate number of keypairs (and reference values) and stores/transmits them in accordance with the techniques described above. Further, in the event the server key cache count is zero (i.e., the reserve key is being used by platform 102 due to key pool depletion), one of the newly generated keys will be designated by the server as a replacement reserve key and the old reserve key will be destroyed.


G. Secure Time-to-Live (TTL)

As mentioned above, one example of a message control a sender can specify for a message is a limit on the time period (also referred to herein as a “time-to-live” or “TTL”) during which a recipient is able to access the message (e.g., to view, listen to, or otherwise interact with the message and any attachments). In scenarios such as where the sender is using an embodiment of platform 102 operated by an enterprise on behalf of its employees, the TTL may be selected by an entity other than the sender (e.g., based on a default corporate policy, or based on administrator configurable rules implemented by an enterprise-specific version of the secure messaging application). For example, messages sent by employees to one another can have a first default TTL, and messages sent by employees to vendors (also using the enterprise-specific application) can have a second default TTL. As another example, messages sent by certain employees (e.g., within a particular department such as the legal department, or having certain titles or positions) can be given different default TTLs. In various embodiments, the default TTL can be overridden, if permitted by an administrator configuration.


The TTL is encrypted and sent together with the secure message. When the recipient opens the message (e.g., taps or clicks on the message in an app), the message is decrypted and displayed on the recipient's device. The corresponding TTL is decrypted, and in some embodiments converted into a message expiry time by adding the TTL (e.g., expressed in seconds) to the current time. In various embodiments, the TTL is stored in the recipient's device's secure database and encrypted to prevent tampering with the secure TTL by the device's user. As will be described in more detail below, the current time can also be secured (e.g., against attempts by the recipient to thwart the TTL by adjusting a clock on the recipient's device). Once the TTL has expired, the message is no longer accessible to the recipient (e.g., is removed from the recipient's viewing interface and deleted from the recipient's device's secure database, along with any associated decryption keys).


The sender (or sender's application, as applicable, e.g., where configured by an enterprise administrator) can specify time limits in a variety of ways. As one example, the sender can set a maximum duration (e.g., a one day limit), with the time limit countdown commencing when the recipient first opens the message. The time limit countdown can also be commenced when the sender sends the message. As another example, the sender can specify a fixed start time (e.g., for embargo purposes) before which the recipient is unable to access the message, even if the recipient is already in possession of the message. Once the embargo period ends, as with above, a TTL value can control how long the recipient is able to view the message once opened. This allows, for example, a company to release company news to multiple shareholders in a secure, time-controlled manner, with each shareholder having the same opportunity to open the message at the same start time. This also allows an enterprise to implement rules (e.g., via an enterprise-specific version of the secure messaging application/platform 102) that only allow employees to open messages during certain periods of the day. (E.g., hourly workers can only read messages during business hours; salaried workers have no such prohibition.) As yet another example, the sender can specify a fixed end time after which the recipient is unable to access the message (irrespective of whether the message was also given an “upon opening” TTL, e.g., of ten minutes). Further, in various embodiments, a sender of the message can shorten a limit on an already sent message. For example, if Bob sends Alice a message with a one day limit, and Alice opens that message, Bob can subsequently revoke Alice's ability to continue to read the message (even though the day has not passed) by interacting with his app (e.g., by long pressing on the sent message as it appears to Bob and selecting an “expire now” (immediately expiring the message) or “expire faster” (expiring the message at a new time picked by Bob) option, as applicable).



FIG. 15 illustrates an embodiment of a message composition interface. In particular, FIG. 15 depicts interface 1500 as rendered on Bob's phone 112. In the following example, Bob is composing a message to Alice. In region 1502, Bob has indicated that he would like to send a message to Alice. In region 1504, Bob has provided a message for Alice. Specifically, Bob has provided Alice with information about how to enter a locked gate. By interacting with region 1506, Bob can select an amount of time for which, once Alice opens Bob's message, Alice will be able to view the message. As shown in region 1508, Bob has decided to allow Alice to read the message for six seconds once she opens it. When Bob sends the message (by selecting button 1512), a time value of six seconds (an example of a TTL) will be included as a message control (e.g., an example of message control 504).



FIG. 16 illustrates an embodiment of a message viewing interface. In particular, FIG. 16 depicts interface 1600 as rendered on Alice's tablet 106. In the example of FIG. 16, Alice has just opened the message Bob was composing in interface 1500 of FIG. 15. As indicated in region 1602, Bob sent the message to Alice at 11:41 am. As indicated in region 1604, Alice's device has a time of 11:45 am. As indicated in region 1606, Alice has six seconds to read message 1608. After the six seconds have elapsed, message 1608 will be removed from interface 1600 (and deleted from Alice's device). Also shown in interface 1600 is a message from Bob that Alice has not yet opened (1610). Bob sent message 1610 at 11:42 am, and message 1610 includes one attachment, as indicated in region 1612. Since Alice has not yet opened message 1610, the TTL for message 1610 has not yet been applied to the message. Alice can open message 1610 by clicking on it with her finger. In the event Alice has multiple devices, in some embodiments a received but unopened message (e.g., message 1610) will appear on all of Alice's devices rendered in the manner shown in FIG. 16. In some embodiments, once Alice opens the message on one of the devices, she will be unable to open the message on any of her other devices (i.e., any additional copies of message 1610 will be removed, unopened, from Alice's other devices). In other embodiments, Alice's messages are synchronized across all of her devices, and Alice is allowed to read any of her messages which have not yet expired on any of those devices. In this scenario, the remaining TTL for a given message can be calculated using the time the message is initially opened on a first device/file, and the remaining TTLs reported by all devices on which the message has been opened. For example, suppose Bob sends a message to Alice and sets the TTL to ten minutes. If Alice has three different devices associated with her account (e.g., an iPhone, an Android tablet, and a desktop computer), she is allowed to open the message on any (or all) of her devices as long as the TTL that Bob established (ten minutes in this example) is not exceeded, collectively, across Alice's devices. Suppose Alice opens the message first on her iPhone (e.g., at 11 am) and views it for three minutes. The TTL for the message at that moment is 10−3=7 minutes. If, after two more minutes (e.g., at 11:05 am) Alice opens the same message on her desktop computer, the TTL is now 7−2=5 min. After five more minutes have elapsed (e.g., it is now 11:10 am), if she tries to open the message again on her iPhone, or on her Android tablet, the TTL will be zero, and the message will be deleted from all of Alice's devices. One way of synchronizing the TTL in a multi-device scenario (also referred to herein as the “global TTL” for the message) is for each app to report the remaining TTL for a particular message ID each time the message is opened on that device. The global TTL for that message ID can be synchronized between all of Alice's devices via a variety of mechanisms. For example, Alice's devices could be configured to update one another on the remaining TTL. As another example, platform 102 can receive updates (i.e., reported openings/TTLs) from each of Alice's devices and share that information with other of Alice's devices with the remaining TTL.



FIG. 17 illustrates an embodiment of a message viewing interface. In particular, FIG. 17 depicts interface 1700 as rendered on Alice's tablet 106. In the example of FIG. 17, Alice (at 1:26 pm) has just opened the message that Bob sent her at 11:42 am (i.e., Alice has just opened message 1610 of FIG. 16). Message 1608 is no longer available to Alice as its TTL has expired. In region 1702, Alice is viewing text authored by Bob. Alice can review an attachment that Bob sent by clicking on region 1704. Region 1706 includes a countdown timer that indicates to Alice that she has approximately one hour (59 minutes and 29 seconds) remaining to read Bob's message and view the attachment. When the timer reaches zero, Alice will be unable to view the message or attachment further.



FIG. 18 illustrates an example of a process for determining whether to allow access to a message. In various embodiments, process 1800 is performed on a client device, such as Alice's client device 106. The process begins at 1802 when a message is received. The message has an associated TTL value (or, as explained in more detail below, in some embodiments has a set of associated time-related values). As one example, Bob's message 1610 is received by Alice's device 106 at 1802. Message 1610 has an associated TTL of one hour (3600 seconds), as selected by Bob during message composition. In some embodiments, the TTL is stored in Alice's device's secure database (i.e., preventing tampering with the value by Alice). Next, a determination is made as to whether the TTL has been exceeded. If the TTL has not been exceeded (1804), the message is made available to the recipient (1806). As one example, when Alice initially opens message 1610, the associated TTL (3600 seconds) is decrypted and read by app 116. App 116 begins counting down (using the TTL). So long as the TTL has not been exceeded (e.g., 3601 seconds have elapsed since Alice opened the message), Alice can continue to view the message via app 116. Once the TTL has been exceeded, the message will be removed from her device (e.g., no longer shown on her device's screen and deleted from her device's secure database).



FIG. 19 illustrates an example of a process for determining whether to allow access to a message. Process 1900 is an embodiment of process 1800 and is in some embodiments performed on a client device such as Alice's client device 106. The process begins at 1902 when (as with 1802) a message that has an associated TTL value is received. At 1904, a message open request (e.g., Alice clicking on the lock icon shown in FIG. 16) is received and in response a Current Time is determined. One approach for determining a Current Time is to use the device time. However, a nefarious device user could attempt to circumvent TTL enforcement by modifying the device date, time, and/or time zone settings. A second approach for determining a Current Time is for the recipient's secure messaging app (e.g., app 116 in the case of Alice) to contact platform 102 (or another external time source, such as a dedicated time server) and obtain a Current Time from platform 102 (or the other external time source). In some embodiments, if app 116 is unable to obtain a Current Time (e.g., device 106 is in airplane mode or otherwise offline; or if platform 102 or the other external time source(s) are unreachable), Alice will be unable to open the message (until such time as app 116 is able to obtain a Current Time).


At 1906, the message expiration time (“Expire Time”) is set as the Current Time (determined at 1904) with the TTL (e.g., 3600 seconds) added. Thus for example, when Alice opens message 1610 (e.g., at 1:26 pm), a Current Time is obtained from platform 102 (or another appropriate external time source), and a TTL of 3600 is added to the Current Time, resulting in an Expire Time of 2:26 pm.


At 1908, a determination is made as to whether the Current Time is greater than the Expire Time. If not (1910), Alice is able to view the message (1912), and after a period of time (e.g., one second elapsing), another check of the Current Time vs. the Expire Time is performed (1908). In various embodiments, the Current Time continues to be obtained from an external source (e.g., device 106 contacts platform 102 every second). In other embodiments, app 116 is responsible for maintaining the Current Time, at least a portion of the time, after performing an initial check with platform 102 of the Current Time upon message open. In some embodiments, if a Current Time cannot be obtained from an external source (e.g., platform 102 or another server) during the ongoing checking of portion 1908, the message will cease being available to Alice. So, for example, if Alice temporarily loses connectivity during the one hour window of time Bob has allowed her to read message 1610, Alice will be unable to read message 1610 during that portion of the hour. In some embodiments, the TTL countdown continues, irrespective of whether Alice is offline, meaning that Alice will not be given additional time to view the message to compensate for the period her device lacked connectivity. Eventually (e.g., after one hour has elapsed), the Current Time will exceed the Expire Time (1914), at which point the message is deleted (1916).



FIG. 20 illustrates an example of a process for determining whether to allow access to a message. Process 2000 is an embodiment of process 1800 and is in some embodiments performed on a client device such as Alice's client device 106. The process begins at 2002 when a message that has multiple TTL-related values is received. As one example, a start time (i.e., embargo time) is provided, as is a duration (e.g., 3600 seconds, as per above). Process 2000 can also be adapted to accommodate a hard end time (instead of, or in addition to a start time), as applicable. At 2004, a determination is made (e.g., by contacting platform 102) whether the Current Time exceeds the Start Time. If not (2006), any requests by Alice to open the message will be ignored, as the end of the embargo has not yet been reached. And, additional checks of the Current Time vs. the Start Time will continue until the embargo ends (2008). The remainder of process 2000 continues as per process 1900. E.g., a message open request is received, and a Current Time determined (1904, 2010); the Expire Time is set as the Current Time and TTL (1906, 2012); and the Current Time is checked against the Expire Time (1908, 2014) to determine whether to continue to allow access to the message (1912, 2018) or delete the message (1916, 2022).


Although the foregoing embodiments have been described in some detail for purposes of clarity of understanding, the invention is not limited to the details provided. There are many alternative ways of implementing the invention. The disclosed embodiments are illustrative and not restrictive.

Claims
  • 1. A computing device comprising: an interface configured to: receive a communication, wherein the communication comprises a time-to-live value;one or more processors;memory storing instructions that, when executed by the one or more processors, cause the computing device to:determine an expiration time of the communication by adding the time-to-live value to a first current time;transmit the expiration time to one or more devices associated with a user of the computing device;obtain, based on a request to access the communication, a second current time;compare the second current time to the expiration time; andallow, based on a determination that the second current time is less than the expiration time, access to the communication.
  • 2. The computing device of claim 1, wherein the first time-to-live value is configurable by the sender of the communication.
  • 3. The computing device of claim 1, wherein the instructions cause the one or more processors to: decrypt the communication when allowing access.
  • 4. The computing device of claim 1, wherein allowing access to the communication comprises displaying the communication.
  • 5. The computing device of claim 1, wherein the instructions cause the one or more processors to: deny access to the communication based on a determination that the second current time is greater than the expiration time.
  • 6. The computing device of claim 5 wherein ceasing denying access to the communication comprises deleting the communication.
  • 7. The computing device of claim 5 wherein denying access to the communication comprises deleting a key configured to decrypt the communication.
  • 8. The computing device of claim 1, wherein the communication comprises an embargo time.
  • 9. The computing device of claim 8, wherein the instructions cause the one or more processors to: deny, based on a determination that a third current time is less than the embargo time, access to the communication.
  • 10. A method comprising: receiving, at a first device, a communication, wherein the communication comprises a time-to-live value;determining an expiration time of the communication by adding the time-to-live value to a first current time;transmitting the expiration time to one or more devices associated with a user of the first device;obtaining, based on a request to access the communication, a second current time;comparing the second current time to the expiration time; andallowing, based on a determination that the second current time is less than the expiration time, access to the communication.
  • 11. The method of claim 10 further comprising: denying access to the communication based on a determination is that the second current time is greater than the expiration time.
  • 12. The method of claim 11 wherein denying access to the communication comprises deleting the communication.
  • 13. The method of claim 11 wherein denying access to the communication comprises deleting a key configured to decrypt the communication.
  • 14. The method of claim 10, wherein the first current time is obtained from an external source.
  • 15. The method of claim 10, wherein the second current time is obtained from an external source.
  • 16. A computer program product embodied in a non-transitory tangible computer readable storage medium and comprising computer instructions for: receiving a communication, wherein the communication comprises a time-to-live value;determining an expiration time of the communication by adding the time-to-live value to a first current time;transmitting the expiration time to one or more devices associated with a user of a first device;obtaining, based on a request to access the communication, a second current time;comparing the second current time to the expiration time; andallowing, based on a determination that the second current time is less than the expiration time, access to the communication.
CROSS REFERENCE TO OTHER APPLICATIONS

This application claims priority to U.S. Provisional Patent Application No. 61/839,307 entitled SECURE TIME TO LIVE filed Jun. 25, 2013 which is incorporated herein by reference for all purposes. This application also claims priority to U.S. Provisional Patent Application No. 61/846,568 entitled DIGITAL SECURITY BUBBLE filed Jul. 15, 2013 which is also incorporated herein by reference for all purposes. This application also claims priority to U.S. Provisional Patent Application No. 61/943,826 entitled ENHANCED PERFECT FORWARD SECRECY FOR MULTI-SYNCHRONOUS COMMUNICATION filed Feb. 24, 2014 which is also incorporated herein by reference for all purposes.

US Referenced Citations (384)
Number Name Date Kind
5850443 Van Oorschot et al. Dec 1998 A
5953419 Lohstroh Sep 1999 A
6009173 Sumner Dec 1999 A
6091820 Aziz Jul 2000 A
6112227 Heiner Aug 2000 A
6145079 Mitty et al. Nov 2000 A
6263435 Dondeti Jul 2001 B1
6266420 Langford et al. Jul 2001 B1
6314190 Zimmermann Nov 2001 B1
6324569 Ogilvie et al. Nov 2001 B1
6393464 Dieterman May 2002 B1
6405315 Burns et al. Jun 2002 B1
6457134 Lemke Sep 2002 B1
6463155 Akiyama Oct 2002 B1
6463460 Simonoff Oct 2002 B1
6463463 Godfrey Oct 2002 B1
6711608 Ogilvie Mar 2004 B1
6721784 Leonard et al. Apr 2004 B1
6742116 Matsui May 2004 B1
6801998 Hanna Oct 2004 B1
6804257 Benayoun Oct 2004 B1
6851049 Price, III Feb 2005 B1
6904521 Jivsov Jun 2005 B1
6912656 Perlman Jun 2005 B1
6947556 Matyas Sep 2005 B1
6978284 McBrearty Dec 2005 B2
7024456 Simonoff Apr 2006 B1
7043529 Simonoff May 2006 B1
7139399 Zimmermann Nov 2006 B1
7149893 Leonard et al. Dec 2006 B1
7178021 Hanna Feb 2007 B1
7197537 Koch Mar 2007 B2
7243231 Ellison Jul 2007 B2
7263619 Kim Aug 2007 B1
7308477 Gress et al. Dec 2007 B1
7356564 Hartselle et al. Apr 2008 B2
7366894 Kalimuthu et al. Apr 2008 B1
7383439 Price, III et al. Jun 2008 B2
7526620 McGovern Apr 2009 B1
7584505 Mondri et al. Sep 2009 B2
7640427 Callas et al. Dec 2009 B2
7657037 Callas Feb 2010 B2
7711120 Kimmel May 2010 B2
7715565 Kimmel May 2010 B2
7730309 Zimmermann Jun 2010 B2
7739501 Kimmel Jun 2010 B2
7779394 Homing et al. Aug 2010 B2
7802099 Mitchell et al. Sep 2010 B2
8036221 Fluhrer et al. Oct 2011 B2
8156536 Polk Apr 2012 B2
8175277 Bell et al. May 2012 B2
8291031 Kirkham Oct 2012 B1
8347083 Scudder Jan 2013 B1
8356177 McGrew et al. Jan 2013 B2
8364764 Hartselle et al. Jan 2013 B2
8364984 Jeffries et al. Jan 2013 B2
8379857 Zheng Feb 2013 B1
8380868 Hiie Feb 2013 B2
8407471 Sobel Mar 2013 B1
8412934 De Atley et al. Apr 2013 B2
8429420 Melvin Apr 2013 B1
8433901 De Atley et al. Apr 2013 B2
8392699 Jones et al. May 2013 B2
8447970 Klein et al. May 2013 B2
8478893 Hiie Jul 2013 B2
8489889 Moscaritolo et al. Jul 2013 B1
8503621 Patel et al. Aug 2013 B2
8503681 McGrew et al. Aug 2013 B1
8510552 De Atley et al. Aug 2013 B2
8527769 Kotla et al. Sep 2013 B2
8560843 Moscaritolo et al. Oct 2013 B1
8566350 Chow Oct 2013 B2
8572757 Stamos Oct 2013 B1
8588425 Harwood Nov 2013 B1
8589673 Ackerly Nov 2013 B2
8589680 De Atley et al. Nov 2013 B2
8601114 Blair et al. Dec 2013 B1
8601263 Shankar Dec 2013 B1
8625805 Statica Jan 2014 B1
8631227 Olechowski et al. Jan 2014 B2
8667271 Shkolnikov Mar 2014 B2
8677136 Brokenshire et al. Mar 2014 B2
8682893 Summerlin Mar 2014 B2
8689015 Jeffries et al. Apr 2014 B2
8705565 Kutt et al. Apr 2014 B2
8713311 Roskind Apr 2014 B1
8782409 Murphy et al. Apr 2014 B2
8719233 Gandhi May 2014 B2
8726009 Cook et al. May 2014 B1
8726369 Emigh May 2014 B1
8732452 Byrum et al. May 2014 B2
8756419 De Atley et al. Jun 2014 B2
8762712 Kwan Jun 2014 B1
8788842 Brouwer et al. Jul 2014 B2
8788899 Magnus Hfie Jul 2014 B2
8798614 Bellovin Aug 2014 B2
8824664 Ristock Sep 2014 B1
8837739 Sexton et al. Sep 2014 B1
8862129 Moshir et al. Oct 2014 B2
8863226 Bailey, Jr. Oct 2014 B1
8874902 Ackerly Oct 2014 B2
8913994 Edwards Dec 2014 B2
8918896 Murphy et al. Dec 2014 B2
8925109 Agrawal Dec 2014 B2
8938810 Richardson et al. Jan 2015 B2
8948382 Hassan et al. Feb 2015 B2
8948391 Kritt et al. Feb 2015 B2
8954740 Moscaritolo Feb 2015 B1
8958559 Medina et al. Feb 2015 B2
8996861 Cummings et al. Mar 2015 B1
9027114 Akhter et al. May 2015 B2
9037796 Nagpal May 2015 B2
9059840 Connelly et al. Jun 2015 B2
9071597 Lord et al. Jun 2015 B2
9075638 Barnett et al. Jul 2015 B2
9077525 Chandran et al. Jul 2015 B2
9078127 Kritt et al. Jul 2015 B2
9094405 Buck et al. Jul 2015 B2
9100373 Deluca et al. Aug 2015 B2
9124559 Deluca et al. Sep 2015 B2
9130822 Cooper Sep 2015 B2
9137191 Firstenberg et al. Sep 2015 B2
9154612 Moshir et al. Oct 2015 B2
9197616 Sinha Nov 2015 B2
9210143 Ozzie Dec 2015 B2
9224014 Rios Dec 2015 B2
9225709 Ackerly Dec 2015 B2
9237016 De Atley et al. Jan 2016 B2
9264406 Kipnis et al. Feb 2016 B2
9282192 Laasik et al. Mar 2016 B2
9288047 Brouwer et al. Mar 2016 B2
9326135 Rodrigues et al. Apr 2016 B2
9380044 Zhang et al. Jun 2016 B2
9560006 Prado et al. Jan 2017 B2
20010044901 Grawrock Nov 2001 A1
20020002468 Spagna Jan 2002 A1
20020026487 Ogilvie et al. Feb 2002 A1
20020037736 Kawaguchi Mar 2002 A1
20020038296 Margolus Mar 2002 A1
20020038421 Hamada Mar 2002 A1
20020078380 Lin Jun 2002 A1
20020094085 Roberts Jul 2002 A1
20020095465 Banks Jul 2002 A1
20020154776 Sowa Oct 2002 A1
20020166053 Wilson Nov 2002 A1
20020186703 West Dec 2002 A1
20030021416 Brown et al. Jan 2003 A1
20030028597 Salmi Feb 2003 A1
20030037114 Nishio Feb 2003 A1
20030046533 Olkin Mar 2003 A1
20030084050 Hall May 2003 A1
20030084347 Luzzatto May 2003 A1
20030126215 Udell et al. Jul 2003 A1
20030131060 Hartselle et al. Jul 2003 A1
20030191937 Balissat Oct 2003 A1
20030204720 Schoen Oct 2003 A1
20030208686 Thummalapally Nov 2003 A1
20030217266 Epp Nov 2003 A1
20030229668 Malik Dec 2003 A1
20040013113 Singh Jan 2004 A1
20040025057 Cook Feb 2004 A1
20040054891 Hengeveld Mar 2004 A1
20040064512 Arora Apr 2004 A1
20040078568 Pham Apr 2004 A1
20040107349 Sasselli Jun 2004 A1
20040111612 Choi Jun 2004 A1
20040133520 Callas et al. Jul 2004 A1
20040133775 Callas et al. Jul 2004 A1
20040168055 Lord et al. Aug 2004 A1
20040205248 Little Oct 2004 A1
20040210772 Hooker et al. Oct 2004 A1
20050005106 Chen Jan 2005 A1
20050044059 Samar Feb 2005 A1
20050053232 Bace Mar 2005 A1
20050076061 Cox Apr 2005 A1
20050086527 Jackson Apr 2005 A1
20050102528 Tan May 2005 A1
20050125659 Sarfati Jun 2005 A1
20050149732 Freeman Jul 2005 A1
20050160292 Batthish Jul 2005 A1
20050187966 Iino Aug 2005 A1
20050193199 Asokan Sep 2005 A1
20050198170 LeMay Sep 2005 A1
20050229257 Kim Oct 2005 A1
20050262338 Irwin Nov 2005 A1
20050273592 Pryor Dec 2005 A1
20060020796 Aura Jan 2006 A1
20060031670 Price, III et al. Feb 2006 A1
20060036739 Hagale et al. Feb 2006 A1
20060075234 You Apr 2006 A1
20060090073 Steinberg Apr 2006 A1
20060133346 Chheda Jun 2006 A1
20060147000 Novi Jul 2006 A1
20060149822 Henry et al. Jul 2006 A1
20060212928 Maino Sep 2006 A1
20060224902 Bolt Oct 2006 A1
20060288209 Vogler Dec 2006 A1
20070003065 Schwartz et al. Jan 2007 A1
20070003066 Schwartz et al. Jan 2007 A1
20070005713 LeVasseur Jan 2007 A1
20070011469 Allison Jan 2007 A1
20070016771 Allison Jan 2007 A1
20070022469 Cooper Jan 2007 A1
20070050624 Lord et al. Mar 2007 A1
20070073823 Cohen et al. Mar 2007 A1
20070116269 Nochta May 2007 A1
20070199071 Callas Aug 2007 A1
20070233832 Narayanan Oct 2007 A1
20070276836 Chatterjee Nov 2007 A1
20070277240 Durfee Nov 2007 A1
20080019530 Eldridge et al. Jan 2008 A1
20080022061 Ito et al. Jan 2008 A1
20080026776 Neil Jan 2008 A1
20080091780 Balan et al. Jan 2008 A1
20080049941 Kim Feb 2008 A1
20080065878 Hutson Mar 2008 A1
20080077686 Subhraveti Mar 2008 A1
20080095371 Vataja Apr 2008 A1
20080098038 Motoyama et al. Apr 2008 A1
20080162647 Liang Jul 2008 A1
20080165952 Smith Jul 2008 A1
20080235336 Stern Sep 2008 A1
20080281930 Hartselle et al. Nov 2008 A1
20080304364 Holtzman Dec 2008 A1
20090006851 Freeman Jan 2009 A1
20090052660 Chen et al. Feb 2009 A1
20090069033 Karstens Mar 2009 A1
20090070466 Elbring Mar 2009 A1
20090083112 Bhogal Mar 2009 A1
20090097662 Olechowski Apr 2009 A1
20090100268 Garcia et al. Apr 2009 A1
20090132822 Chen May 2009 A1
20090154705 Price et al. Jun 2009 A1
20090198997 Yeap Aug 2009 A1
20090222668 Zaccone Sep 2009 A1
20090265552 Moshir et al. Oct 2009 A1
20090281829 Hansen et al. Nov 2009 A1
20090299200 Eggenberger Dec 2009 A1
20100002882 Rieger et al. Jan 2010 A1
20100017602 Bussard Jan 2010 A1
20100031038 Kruegel Feb 2010 A1
20100100967 Douglas et al. Apr 2010 A1
20100138619 Benavides Jun 2010 A1
20100138903 Medvinsky Jun 2010 A1
20100138934 Minoshima Jun 2010 A1
20100161817 Xiao Jun 2010 A1
20100174911 Isshiki Jul 2010 A1
20100199340 Jonas Aug 2010 A1
20100205444 Mehta Aug 2010 A1
20100211541 Deetz et al. Aug 2010 A1
20100217984 Hill Aug 2010 A1
20100223470 Lord et al. Sep 2010 A1
20100239087 Chaisson Sep 2010 A1
20100250939 Adams Sep 2010 A1
20100250946 Korte Sep 2010 A1
20100262714 Hiie Oct 2010 A1
20100275007 Kutt et al. Oct 2010 A1
20100296655 Solow Nov 2010 A1
20100304766 Goyal Dec 2010 A1
20110035591 Dudziak et al. Feb 2011 A1
20110131406 Jones et al. Jun 2011 A1
20110145564 Moshir et al. Jun 2011 A1
20110145571 Schmidt-Karaca et al. Jun 2011 A1
20110151903 Moshir et al. Jun 2011 A1
20110173272 Carvalho Neto Jul 2011 A1
20110202598 Evans et al. Aug 2011 A1
20110246942 Misawa Oct 2011 A1
20110307707 Fielder Dec 2011 A1
20120030183 Deetz et al. Feb 2012 A1
20120030743 Semba Feb 2012 A1
20120045064 Rembarz Feb 2012 A1
20120096035 Spector Apr 2012 A1
20120098639 Ijas Apr 2012 A1
20120102317 Mathur et al. Apr 2012 A1
20120102326 Palekar Apr 2012 A1
20120117568 Plotkin May 2012 A1
20120150921 Yamakawa Jun 2012 A1
20120159164 Brown et al. Jun 2012 A1
20120173353 Rausch Jul 2012 A1
20120173880 Swaminathan Jul 2012 A1
20120179905 Ackerly Jul 2012 A1
20120184309 Cohen Jul 2012 A1
20120209924 Evans et al. Aug 2012 A1
20120231770 Clarke Sep 2012 A1
20120240037 Migos et al. Sep 2012 A1
20120284805 Katsikas Nov 2012 A1
20120297451 Ozzie Nov 2012 A1
20120300331 Deetz Nov 2012 A1
20120311329 Medina et al. Dec 2012 A1
20120331073 Williams Dec 2012 A1
20130034229 Sauerwald et al. Feb 2013 A1
20130036302 Lord et al. Feb 2013 A1
20130046828 Grewal et al. Feb 2013 A1
20130070045 Meek Mar 2013 A1
20130073850 Zaverucha Mar 2013 A1
20130077774 Lynch, III Mar 2013 A1
20130091298 Ozzie et al. Apr 2013 A1
20130097688 Charles et al. Apr 2013 A1
20130136255 Brown May 2013 A1
20130159436 Hartselle et al. Jun 2013 A1
20130194301 Robbins Aug 2013 A1
20130198522 Kohno Aug 2013 A1
20130208725 Wakumoto Aug 2013 A1
20130219035 Detienne Aug 2013 A1
20130232209 Fernandez et al. Sep 2013 A1
20130252585 Moshir et al. Sep 2013 A1
20130254537 Bogorad Sep 2013 A1
20130275842 Peach Oct 2013 A1
20130290416 Nelson Oct 2013 A1
20130304833 St. Clair Nov 2013 A1
20130305049 Krakowski Nov 2013 A1
20130305057 Greco et al. Nov 2013 A1
20130308628 Marueli et al. Nov 2013 A1
20130326220 Connelly Dec 2013 A1
20130332724 Walters Dec 2013 A1
20130339465 Zhang Dec 2013 A1
20140033286 Zhang Jan 2014 A1
20140052982 Ackerly Feb 2014 A1
20140053150 Barnett et al. Feb 2014 A1
20140068254 Scharf Mar 2014 A1
20140090042 Short Mar 2014 A1
20140129785 Deetz et al. May 2014 A1
20140129838 Kim May 2014 A1
20140136842 Kitze et al. May 2014 A1
20140157435 Stamos Jun 2014 A1
20140162601 Kim Jun 2014 A1
20140164776 Hook Jun 2014 A1
20140169554 Scarisbrick Jun 2014 A1
20140181221 Kushtagi Jun 2014 A1
20140196140 Gong Jul 2014 A1
20140206285 Jance et al. Jul 2014 A1
20140207887 Hartselle et al. Jul 2014 A1
20140207900 Liu Jul 2014 A1
20140233736 Zhang Aug 2014 A1
20140258719 Cidon Sep 2014 A1
20140304508 Murphy et al. Oct 2014 A1
20140304515 Feuerman Oct 2014 A1
20140307543 Marueli et al. Oct 2014 A1
20140325622 Luk Oct 2014 A1
20140325667 Sun Oct 2014 A1
20140351586 Hook et al. Nov 2014 A1
20150006884 Ackerly Jan 2015 A1
20150039882 Watanabe Feb 2015 A1
20150082391 Lerman et al. Mar 2015 A1
20150121549 Baessler et al. Apr 2015 A1
20150143546 Charles et al. May 2015 A1
20150154206 Kumar et al. Jun 2015 A1
20150156177 Murphy et al. Jun 2015 A1
20150169615 Batchu Jun 2015 A1
20150169893 Desai Jun 2015 A1
20150195239 Firstenberg et al. Jul 2015 A1
20150215291 Abdunabi et al. Jul 2015 A1
20150227761 Cohen Aug 2015 A1
20150229612 Hassan et al. Aug 2015 A1
20150244658 Speyer et al. Aug 2015 A1
20150244684 Ng Aug 2015 A1
20150264005 Brooks et al. Sep 2015 A1
20150264042 Lord et al. Sep 2015 A1
20150264540 Brooks et al. Sep 2015 A1
20150271653 Brooks et al. Sep 2015 A1
20150281185 Cooley Oct 2015 A1
20150282117 Serna Pozuelo Oct 2015 A1
20150304849 Moom Oct 2015 A1
20150312260 Kim Oct 2015 A1
20150326731 Laasik et al. Nov 2015 A1
20150332332 Darmaki et al. Nov 2015 A1
20150347770 Whalley Dec 2015 A1
20150350036 Cooper et al. Dec 2015 A1
20150350119 Thirumalai et al. Dec 2015 A1
20150350247 Adler et al. Dec 2015 A1
20150350895 Brander et al. Dec 2015 A1
20150365395 Enriquez et al. Dec 2015 A1
20150370918 Patterson et al. Dec 2015 A1
20150373021 Tussy et al. Dec 2015 A1
20160013939 Jun Jan 2016 A1
20160034692 Singler et al. Feb 2016 A1
20160036791 Sauerwald et al. Feb 2016 A1
20160055325 Ozzie Feb 2016 A1
20160057156 Lin et al. Feb 2016 A1
20160063258 Ackerly Mar 2016 A1
20160065530 Prado et al. Mar 2016 A1
20160072781 Zhang et al. Mar 2016 A1
20160191513 Tomlinson Jun 2016 A1
20160265005 Zelle et al. Sep 2016 A1
Foreign Referenced Citations (3)
Number Date Country
2018015 Jan 2009 EP
2011101784 Aug 2011 WO
2014140736 Sep 2014 WO
Non-Patent Literature Citations (22)
Entry
Author Unknown, “MyENIGMA: Whitepaper”, Qnective AG, 2013.
Moscaritolo et al., “Silent Circle Instant Messaging Protocol Protocol Specification”, Silent Circle Engineering, Dec. 5, 2012, Version 1.0.
Author Unknown, “Perfect Forward Secrecy”, IBM Security SiteProtector System, http://pic.dhe.ibm.com/infocenter/sprotect/v2r8m0/index.jsp?topic=%2Fcom.ibm.siteprotector.doe%2Freferences%2Fsp_agenthelp_perfect_forward_secrecy.htm.
Sun et al., Password-based authentication and key distribution protocols with perfect forward secrecy, Journal of Computer and System Sciences 72, pp. 1002-1011, 2006.
Sui et al., “An Improved Authenticated Key Agreement Protocol with Perfect Forward Secrecy for Wireless Mobile Communication”, WCNC, pp. 2088-2093, 2005.
Cremers et al., “One-round strongly secure key exchange with perfect forward secrecy and deniability”, IACR Cryptology ePrint Archive, version 3, 2011.
Diffie et al., “Authentication and authenticated key exchanges”, Designs, Codes and Cryptography, vol. 2, Issue 2, pp. 107-125, Jun. 1992.
Lin et al., “An Improvement on Secure E-mail Protocols Providing Perfect Forward Secrecy”, JCIS, 2006.
B. Kaliski, “PKCS #7: Cryptographic Message Syntax”, Version 1.5, Mar. 1998, http://tools.ietf.org/html/rfc2315.
S. Kent, “Privacy Enhancement for Internet Electronic Mail: Part 2: Certificate-Based Key Managment”, Feb. 1993, http://tools.ietf.org/html/rfc1422.
Pomian & Corella, LLC, “Submission of Prior Art under 37 CFR 1.501 for U.S. Pat. No. 8,625,805”, Apr. 18, 2014.
Wong et al., “Email Protocols with Perfect Forward Secrecy”, Int. J. Security and Networks, vol. 7, No. 1, 2012.
Green et al., “Dancing on the Lip of the Volcano—Chosen Ciphertext Attacks on Apple iMessage” Mar. 21, 2016.
Leach et al. “A Universally Unique Identifier (UUID) URN Namespace”, Standards Track, Jul. 2005.
Oikonomidis et al. “Identity Based Protocols for Secure Electronic Content Distribution and Licensing Proceedings of the Fourth International Conference on Web Delivering of Music”, Wedelmusic 2004. http://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber= 1358105.
Pei et al. “An Intelligent Digital Content Protection Framework between Home Network Receiver Devices”, 2006 Interational Conference on Computational Intelligence and Security. http://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=4076116.
Yampolskiy, Roman V. “Mimicry Attack on Strategy-Based Behavioral Biometric”, Fifth International Conference on Information Technology: New Generations, 2008. ITNG 2008. http://ieeexplore.ieee.org/stamp/stamp.jsp?arnumber=4492601.
Ms. Smith, “How to easily encrypt email with Virtru for free: Gmail, Hotmail, Outlook, Yahoo”, Network World, Jan. 28, 2014.
Author Unknown, “Threema Cyrptography Whitepaper” Threema., Mar. 9, 2015.
Toorani et al.,“SSMS—A Secure SMS Messaging Protocol for the M-Payment Systems”, Proceedings of the 13th IEEE Symposium on Computers and Communications, IEEE, Jul. 2008.
Callas et al., “OpenPGP Message Format”, RFC 4880, Nov. 2007.
Stedman et al., “A User Study of the Off-the-Record Messaging”, Symposium on Unstable Privacy and Security (SOUPS), Jul. 2008.
Related Publications (1)
Number Date Country
20180367507 A1 Dec 2018 US
Provisional Applications (3)
Number Date Country
61839307 Jun 2013 US
61846568 Jul 2013 US
61943826 Feb 2014 US