1. Field of the Disclosure
The technology of the disclosure relates generally to Web Real-Time Communications (WebRTC) interactive sessions.
2. Technical Background
Web Real-Time Communications (WebRTC) represents an ongoing effort to develop industry standards for integrating real-time communications functionality into web clients, such as web browsers, to enable direct interaction with other web clients. This real-time communications functionality is accessible by web developers via standard markup tags, such as those provided by version 5, of the Hypertext Markup Language (HTML5), and client-side scripting Application Programming Interfaces (APIs), such as JavaScript APIs. More information regarding WebRTC may be found in “WebRTC: APIs and RTCWEB Protocols of the HTML5, Real-Time Web,” by Alan B. Johnston and Daniel C. Burnett (2012, Digital Codex LLC), which is incorporated herein in its entirety by reference.
WebRTC provides built-in capabilities for establishing real-time video, audio, and/or data streams in both point-to-point interactive sessions and multi-party interactive sessions. The WebRTC standards are currently under joint development by the World Wide Web Consortium (W3C) and the Internet Engineering Task Force (IETF). Information on the current state of WebRTC standards can be found at, e.g., http://www.w3c.org and http://www/ietf.org.
To establish a WebRTC interactive session (e.g., a real-time video, audio, and/or data exchange), two web clients may retrieve WebRTC-enabled web applications, such as HTML5/JavaScript web applications, from a WebRTC application server. Through the web applications, the two web clients engage in a media negotiation to communicate and reach an agreement on parameters that define characteristics of the WebRTC interactive session. This media negotiation is known as a WebRTC “offer/answer” exchange. Once the WebRTC offer/answer exchange is complete, the web clients may then establish a direct peer connection with one another, and may begin a real-time exchange of media or data. The peer connection between the web clients typically employs the Secure Real-time Transport Protocol (SRTP) to transport real-time media flows, and may utilize various other protocols for real-time data interchange.
In multi-party WebRTC interactive sessions, each participating web client may be directly connected to every other participating web client in what is referred to as a “full mesh” or “fully distributed” architecture. However, the “full mesh” architecture may impose significant burdens on computing resources and bandwidth utilization. Multi-party WebRTC interactive sessions may also be handled by a central media server to which every web client is connected. Implementing such a central media server may pose challenges due to new media extensions employed by WebRTC, as well as variations between implementations of a WebRTC media stack in different web client types and client versions.
Embodiments disclosed in the detailed description provide scalable Web Real-Time Communications (WebRTC) media engines. Related methods, systems, and computer-readable media are also disclosed. In some embodiments, a WebRTC server, through which two or more WebRTC clients seek to establish a WebRTC interactive session, provides a scalable WebRTC media engine. The WebRTC media engine instantiates one or more virtual WebRTC agents, and causes each of the two or more WebRTC clients to establish a WebRTC interactive flow with the one or more virtual WebRTC agents. The WebRTC media engine then synthesizes audio and/or video content from the established WebRTC interactive flows, and provides the resulting synthesized content to the WebRTC interactive flows via the one or more virtual WebRTC agents. In this manner, the WebRTC media engine may provide synthesizing of WebRTC interactive flow content in a scalable fashion without requiring a full mesh architecture, and without placing an excess burden on local computing resources.
In this regard, in one embodiment, a method for providing a scalable WebRTC media engine is provided. The method comprises instantiating, by a WebRTC media engine executing on a computing device, one or more virtual WebRTC agents, each corresponding to one or more of a plurality of WebRTC clients. The method further comprises establishing a plurality of WebRTC interactive flows, each connecting one of the one or more virtual WebRTC agents with the corresponding one or more of the plurality of WebRTC clients. The method also comprises receiving contents of the plurality of WebRTC interactive flows as input from the one or more virtual WebRTC agents. The method additionally comprises synthesizing the contents of the plurality of WebRTC interactive flows. The method further comprises directing the synthesized contents as output to one or more of the plurality of WebRTC interactive flows via the one or more virtual WebRTC agents.
In another embodiment, a system for providing a scalable WebRTC media engine is provided. The system comprises at least one communications interface, and a WebRTC server associated with the at least one communications interface. The WebRTC server comprises a WebRTC media engine configured to instantiate one or more virtual WebRTC agents, each corresponding to one or more of a plurality of WebRTC clients. The WebRTC media engine is further configured to establish a plurality of WebRTC interactive flows, each connecting one of the one or more virtual WebRTC agents with the corresponding one or more of the plurality of WebRTC clients. The WebRTC media engine is also configured to receive contents of the plurality of WebRTC interactive flows as input from the one or more virtual WebRTC agents. The WebRTC media engine is additionally configured to synthesize the contents of the plurality of WebRTC interactive flows. The WebRTC media engine is further configured to direct the synthesized contents as output to one or more of the plurality of WebRTC interactive flows via the one or more virtual WebRTC agents.
In another embodiment, a non-transitory computer-readable medium is provided. The non-transitory computer-readable medium has stored thereon computer-executable instructions to cause a processor to implement a method comprising instantiating one or more virtual WebRTC agents, each corresponding to one or more of a plurality of WebRTC clients. The method implemented by the computer-executable instructions further comprises establishing a plurality of WebRTC interactive flows, each connecting one of the one or more virtual WebRTC agents with the corresponding one or more of the plurality of WebRTC clients. The method implemented by the computer-executable instructions also comprises receiving contents of the plurality of WebRTC interactive flows as input from the one or more virtual WebRTC agents. The method implemented by the computer-executable instructions additionally comprises synthesizing the contents of the plurality of WebRTC interactive flows. The method implemented by the computer-executable instructions further comprises directing the synthesized contents as output to one or more of the plurality of WebRTC interactive flows via the one or more virtual WebRTC agents.
The accompanying drawing figures incorporated in and forming a part of this specification illustrate several aspects of the disclosure, and together with the description serve to explain the principles of the disclosure.
With reference now to the drawing figures, several exemplary embodiments of the present disclosure are described. The word “exemplary” is used herein to mean “serving as an example, instance, or illustration.” Any embodiment described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other embodiments.
Embodiments disclosed in the detailed description provide scalable Web Real-Time Communications (WebRTC) media engines. Related methods, systems, and computer-readable media are also disclosed. In some embodiments, a WebRTC server, through which two or more WebRTC clients seek to establish a WebRTC interactive session, provides a scalable WebRTC media engine. The WebRTC media engine instantiates one or more virtual WebRTC agents, and causes each of the two or more WebRTC clients to establish a WebRTC interactive flow with the one or more virtual WebRTC agents. The WebRTC media engine then synthesizes audio, video, and/or data content from the established WebRTC interactive flows, and provides the resulting synthesized content to the WebRTC interactive flows via the one or more virtual WebRTC agents. In this manner, the WebRTC media engine may provide synthesizing of WebRTC interactive flow content in a scalable fashion without requiring a full mesh architecture, and without placing an excess burden on local computing resources.
In this regard, in one embodiment, a method for providing a scalable WebRTC media engine is provided. The method comprises instantiating, by a WebRTC media engine executing on a computing device, one or more virtual WebRTC agents, each corresponding to one or more of a plurality of WebRTC clients. The method further comprises establishing a plurality of WebRTC interactive flows, each connecting one of the one or more virtual WebRTC agents with the corresponding one or more of the plurality of WebRTC clients. The method also comprises receiving contents of the plurality of WebRTC interactive flows as input from the one or more virtual WebRTC agents. The method additionally comprises synthesizing the contents of the plurality of WebRTC interactive flows. The method further comprises directing the synthesized contents as output to one or more of the plurality of WebRTC interactive flows via the one or more virtual WebRTC agents.
In
The computing devices 20(1)-20(N) of
As seen in
To establish a WebRTC interactive session, each of the WebRTC clients 26(1)-26(N) downloads a WebRTC web application (not shown) from a WebRTC application provider 36 of the WebRTC server 12 via corresponding Hyper Text Transfer Protocol (HTTP)/Hyper Text Transfer Protocol Secure (HTTPS) connections 38(1)-38(N). In some embodiments, the WebRTC web application may comprise an HTML5/JavaScript web application that provides a rich user interface using HTML5, and uses JavaScript to handle user input and to communicate with the WebRTC application provider 36. The WebRTC clients 26(1)-26(N) then engage in a WebRTC offer/answer exchange by exchanging WebRTC session description objects (not shown) via the WebRTC application provider 36. The exchanged WebRTC session description objects are used to determine the media types and capabilities for the desired WebRTC interactive session.
Typically, after the WebRTC offer/answer exchange is complete, WebRTC interactive flows are established directly between the WebRTC clients 26(1)-26(N) via one or more peer connections. For example, in a full mesh architecture, the WebRTC clients 26(1)-26(N) are all connected to each other, with each of the WebRTC clients 26(1)-26(N) receiving and synthesizing the WebRTC interactive flows received from the other WebRTC clients 26. However, this may result in the consumption of significant computing resources and/or network bandwidth, and may provide limited scalability and compatibility between and among the WebRTC clients 26(1)-26(N).
In this regard, virtual WebRTC agents 40(1)-40(X) are provided by the scalable WebRTC media engine 14. In some embodiments, the number X of virtual WebRTC agents 40 may be less than or equal to the number N of WebRTC clients 26 (i.e., 1≦X≦N). In the example of
After instantiation, the virtual WebRTC agents 40(1)-40(X) are each directed by the scalable WebRTC media engine 14 to download a virtual WebRTC application (not shown) from a virtual WebRTC application provider 42. Some embodiments may provide that the virtual WebRTC application provider 42 is communicatively coupled to the scalable WebRTC media engine 14. In some embodiments, the virtual WebRTC application provider 42 may be integrated into or otherwise constitute an element of the scalable WebRTC media engine 14 and/or the WebRTC application provider 36. The virtual WebRTC application includes specialized instructions for interfacing with the WebRTC APIs of the virtual WebRTC agents 40(1)-40(X). The virtual WebRTC agents 40(1)-40(X) may communicate via the virtual WebRTC application with the WebRTC clients 26(1)-26(N), respectively, and with the scalable WebRTC media engine 14.
The scalable WebRTC media engine 14 causes the virtual WebRTC agents 40(1)-40(X) to establish the WebRTC interactive flows 18(1)-18(N) with the corresponding WebRTC clients 26(1)-26(N). In doing so, the scalable WebRTC media engine 14 and/or the virtual WebRTC agents 40(1)-40(X) may modify the data being passed between the WebRTC clients 26(1)-26(N) to facilitate the establishment of the WebRTC interactive flows 18(1)-18(N). For example, data related to the location and/or identity of the endpoints of each of the WebRTC interactive flows 18(1)-18(N) may be modified to permit the virtual WebRTC agents 40(1)-40(X) to act as proxies for their corresponding WebRTC clients 26(1)-26(N).
Once the WebRTC interactive flows 18(1)-18(N) are established, the scalable WebRTC media engine 14 receives content of the WebRTC interactive flows 18(1)-18(N) as input from the virtual WebRTC agents 40(1)-40(X). In some embodiments, the scalable WebRTC media engine 14 provides virtual audio receivers (Rx) 44(1)-44(X), virtual video receivers 46(1)-46(X), and/or virtual data receivers 48(1)-48(X), to which the corresponding virtual WebRTC agents 40(1)-40(X) are communicatively coupled. As the WebRTC interactive flows 18(1)-18(N) commence, the virtual WebRTC agents 40(1)-40(X) forward audio signals received from the corresponding WebRTC interactive flows 18(1)-18(N) to the virtual audio receivers 44(1)-44(X). The virtual WebRTC agents 40(1)-40(X) also forward video signals received from the corresponding WebRTC interactive flows 18(1)-18(N) to the virtual video receivers 46(1)-46(X), and forward data received from the corresponding WebRTC interactive flows 18(1)-18(N) to the virtual data receivers 48(1)-48(X).
The scalable WebRTC media engine 14 then synthesizes the content of the WebRTC interactive flows 18(1)-18(N) received from the virtual WebRTC agents 40(1)-40(X). Some embodiments may provide that the synthesizing of the content of the WebRTC interactive flows 18(1)-18(N) is carried out by a media synthesizing functionality provider 49 of the scalable WebRTC media engine 14. Synthesizing the content of the WebRTC interactive flows 18(1)-18(N) may include selectively combining audio, video, and/or data content of the WebRTC interactive flows 18(1)-18(N) to generate synthesized content. Synthesizing the content of the WebRTC interactive flows 18(1)-18(N) may also include generating and adding audio, video, and/or data content to the content of the WebRTC interactive flows 18(1)-18(N). For example, the scalable WebRTC media engine 14 may inject content into the WebRTC interactive flows 18(1)-18(N) by adding content to an input from the virtual audio receivers 44(1)-44(X), the virtual video receivers 46(1)-46(X), and/or the virtual data receivers 48(1)-48(X). In some embodiments, synthesizing the content of the WebRTC interactive flows 18(1)-18(N) may include applying an audio processing algorithm. As non-limiting examples, an N loudest algorithm (i.e., including the N loudest participants in the audio content), a loudest and longest algorithm (prioritizing audio sources that are loudest or that have been generating output for the longest amount of time), or an N−1, algorithm (including all audio sources except those associated with a receiving endpoint), or combinations thereof, may be applied to an audio content of the WebRTC interactive flows 18(1)-18(N). Some embodiments may provide that synthesizing the contents of the plurality of WebRTC interactive flows may include selecting a video content of one or more of the plurality of WebRTC interactive flows, combining video contents of two or more of the plurality of WebRTC interactive flows, modifying a video content of one or more of the plurality of WebRTC interactive flows, or combinations thereof. According to some embodiments disclosed herein, custom synthesized content may be generated for specific ones of the virtual WebRTC agents 40(1)-40(X) and/or the WebRTC clients 26(1)-26(N).
The synthesized content is then directed by the scalable WebRTC media engine 14 as output to the WebRTC interactive flows 18(1)-18(N) via the virtual WebRTC agents 40(1)-40(X). In some embodiments, the media synthesizing functionality provider 49 is configured to direct the synthesized content to virtual audio transmitters (Tx) 50(1)-50(X), virtual video transmitters 52(1)-52(X), and/or virtual data transmitters 54(1)-54(X) to which the corresponding virtual WebRTC agents 40(1)-40(X) are communicatively coupled.
In the example of
Conversely, in situations in which the WebRTC clients 26(1)-26(N) are known to be compatible, the scalable WebRTC media engine 14 may provide a WebRTC interactive session between the WebRTC clients 26(1)-26(N) using a single virtual WebRTC agent. Providing a single virtual WebRTC agent may conserve computing resources of the computing device 16 and further increase the scalability of the WebRTC server 12. This embodiment is discussed in greater detail below with respect to
To generally describe exemplary operations of the scalable WebRTC media engine 14 of
The scalable WebRTC media engine 14 then establishes a plurality of WebRTC interactive flows 18(1)-18(N), each connecting one of the one or more virtual WebRTC agents 40(1)-40(X) with a corresponding one or more of the plurality of WebRTC clients 26(1)-26(N) (block 58). According to some embodiments herein, the scalable WebRTC media engine 14 and/or the virtual WebRTC agents 40(1)-40(X) may modify the data being passed between the WebRTC clients 26(1)-26(N) to facilitate the establishment of the WebRTC interactive flows 18(1)-18(N). For example, data relating to the location and/or identity of the endpoints of each of the WebRTC interactive flows 18(1)-18(N) may be modified to permit the virtual WebRTC agents 40(1)-40(X) to act as proxies for their corresponding WebRTC clients 26(1)-26(N).
The scalable WebRTC media engine 14 next receives contents of the plurality of WebRTC interactive flows 18(1)-18(N) as input from the one or more virtual WebRTC agents 40(1)-40(X) (block 60). In some embodiments, this may be accomplished through the use of virtual audio receivers 44, virtual video receivers 46, and virtual data receivers 48 provided by the scalable WebRTC media engine 14, as illustrated in
The scalable WebRTC media engine 14 then directs the synthesized contents as output to the plurality of WebRTC interactive flows 18(1)-18(N) via the one or more virtual WebRTC agents 40(1)-40(X) (block 64). Some embodiments may provide that the scalable WebRTC media engine 14 directs the synthesized content using virtual audio transmitters 50(1)-50(X), virtual video transmitters 52(1)-52(X), and/or virtual data transmitters 54(1)-54(X) to which the corresponding virtual WebRTC agents 40(1)-40(X) are communicatively coupled. Thus, the scalable WebRTC media engine 14 may provide synthesizing of content from the plurality of WebRTC interactive flows 18(1)-18(N) in an efficient and scalable manner, while automatically resolving any potential incompatibility issues between the plurality of WebRTC clients 26(1)-26(N).
To illustrate exemplary communications flows during the establishment of a WebRTC interactive session using the virtual WebRTC agents 40(1)-40(X) of
As seen in
The WebRTC server 12 (i.e., the scalable WebRTC media engine 14) instantiates the virtual WebRTC agent 40(X) corresponding to the WebRTC client 26(N), as indicated by arrow 68. In some embodiments, a client type and/or a client version of the WebRTC client 26(N) may be determined based on the SDP Object A, a query/response exchange between the scalable WebRTC media engine 14 and the WebRTC client 26(N), an HTTP header, or other data provided by the WebRTC client 26(N). The virtual WebRTC agent 40(X) may be instantiated having a client type and/or a client version that is known to be compatible with the WebRTC client 26(N). The WebRTC server 12 then forwards the SDP Object A to the WebRTC client 26(1), as indicated by arrow 70.
After the WebRTC client 26(1) receives the SDP Object A from the WebRTC server 12, the WebRTC client 26(1) in response sends a WebRTC session description object, referred to as SDP Object B, via HTTPS to the WebRTC server 12, as indicated by arrow 72. The SDP Object B in this example represents the “answer” in the WebRTC offer/answer exchange. The WebRTC server 12 then instantiates the virtual WebRTC agent 40(1) corresponding to the WebRTC client 26(1), as indicated by arrow 74. Some embodiments may provide that a client type and/or a client version of the WebRTC client 26(1) may be determined based on the SDP Object B, a query/response exchange between the scalable WebRTC media engine 14 and the WebRTC client 26(1), an HTTP header, or other data provided by the WebRTC client 26(1). The virtual WebRTC agent 40(1) may be instantiated having a client type and/or a client version known to be compatible with the WebRTC client 26(1). The WebRTC server 12, in turn, forwards the SDP Object B to the WebRTC client 26(N), as shown by arrow 76.
With continuing reference to
Once the ICE hole punching indicated by arrows 78 and 80 is successful, the WebRTC clients 26(1) and 26(N) begin key negotiations to establish a secure peer connection. In a typical peer-to-peer architecture, the key negotiations take place directly between the WebRTC clients 26(1) and 26(N). However, in this example, each of the WebRTC clients 26(1) and 26(N) negotiates with its corresponding virtual WebRTC agent 40(1) and 40(X), as indicated by bidirectional arrows 82 and 84. If key negotiations are successfully concluded, peer connections are established between the WebRTC client 26(1) and the virtual WebRTC agent 40(1), and between the WebRTC client 26(N) and the virtual WebRTC agent 40(X), as indicated by bidirectional arrows 86 and 88.
Upon establishing peer connections with their corresponding virtual WebRTC agents 40(1) and 40(X), the WebRTC clients 26(1) and 26(N) begin exchanging WebRTC media and/or data flows. As seen in
As noted above with respect to
In the example of
In
If the first WebRTC client 26(1) and the second WebRTC client 26(N) are determined to be compatible, the scalable WebRTC media engine 14 may enable a WebRTC interactive session between the first WebRTC client 26(1) and the second WebRTC client 26(N) using a single virtual WebRTC agent 106, as discussed above with respect to
Returning to decision block 110 of
The scalable WebRTC media engine 14 then establishes a first WebRTC interactive flow 18(1) between the first WebRTC client 26(1) and the first virtual WebRTC agent 40(1) (block 124). The scalable WebRTC media engine 14 also establishes a second WebRTC interactive flow 18(N) between the second WebRTC client 26(N) and the second virtual WebRTC agent 40(X) (block 126). Because each of the first WebRTC client 26(1) and the second WebRTC client 26(N) are interacting directly with their respective compatible virtual WebRTC agents 40(1) and 40(X), no compatibility issues should arise with respect to the WebRTC interactive flows 18(1) and 18(N). Processing then resumes at block 118 of
Referring now to
The scalable WebRTC media engine 14 then directs the synthesized contents as output to the WebRTC interactive flows 18(1) and 18(N) via the virtual WebRTC agents 40(1) and 40(X), or the virtual WebRTC agent 106, as appropriate (block 130). Some embodiments may provide that the scalable WebRTC media engine 14 directs the synthesized content using virtual audio transmitters 50, virtual video transmitters 52, and/or virtual data transmitters 54 to which the corresponding virtual WebRTC agents 40(1)-40(X) are communicatively coupled. The scalable WebRTC media engine 14 next determines whether there are at least two active WebRTC interactive flows 18 among the WebRTC interactive flows 18(1) and 18(N) (block 132). If the WebRTC interactive flows 18(1) and 18(N) are still active, processing returns to block 132 of
The exemplary computer system 138 includes a processing device or processor 140, a main memory 142 (as non-limiting examples, read-only memory (ROM), flash memory, dynamic random access memory (DRAM) such as synchronous DRAM (SDRAM), etc.), and a static memory 144 (as non-limiting examples, flash memory, static random access memory (SRAM), etc.), which may communicate with each other via a bus 146. Alternatively, the processing device 140 may be connected to the main memory 142 and/or the static memory 144 directly or via some other connectivity means.
The processing device 140 represents one or more processing devices, such as a microprocessor, central processing unit (CPU), or the like. More particularly, the processing device 140 may be a complex instruction set computing (CISC) microprocessor, a reduced instruction set computing (RISC) microprocessor, a very long instruction word (VLIW) microprocessor, a processor implementing other instruction sets, or a processor implementing a combination of instruction sets. The processing device 140 is configured to execute processing logic in instructions 148 and/or cached instructions 150 for performing the operations and steps discussed herein.
The computer system 138 may further include a communications interface in the form of a network interface device 152. It also may or may not include an input 154 to receive input and selections to be communicated to the computer system 138 when executing the instructions 148, 150. It also may or may not include an output 156, including but not limited to display(s) 158. The display(s) 158 may be a video display unit (as non-limiting examples, a liquid crystal display (LCD) or a cathode ray tube (CRT)), an alphanumeric input device (as a non-limiting example, a keyboard), a cursor control device (as a non-limiting example, a mouse), and/or a touch screen device (as a non-limiting example, a tablet input device or screen).
The computer system 138 may or may not include a data storage device 160 that includes using drive(s) 162 to store the functions described herein in a computer-readable medium 164, on which is stored one or more sets of instructions 166 (e.g., software) embodying any one or more of the methodologies or functions described herein. The functions can include the methods and/or other functions of the processing system 136, a participant user device, and/or a licensing server, as non-limiting examples. The one or more sets of instructions 166 may also reside, completely or at least partially, within the main memory 142 and/or within the processing device 140 during execution thereof by the computer system 138. The main memory 142 and the processing device 140 also constitute machine-accessible storage media. The instructions 148, 150, and/or 166 may further be transmitted or received over a network 168 via the network interface device 152. The network 168 may be an intra-network or an inter-network.
While the computer-readable medium 164 is shown in an exemplary embodiment to be a single medium, the term “machine-accessible storage medium” should be taken to include a single medium or multiple media (as non-limiting examples, a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions 166. The term “machine-accessible storage medium” shall also be taken to include any medium that is capable of storing, encoding, or carrying a set of instructions 148, 150, and/or 166 for execution by the machine, and that cause the machine to perform any one or more of the methodologies disclosed herein. The term “machine-accessible storage medium” shall accordingly be taken to include, but not be limited to, solid-state memories, optical and magnetic media, and carrier wave signals.
The embodiments disclosed herein may be embodied in hardware and in instructions that are stored in hardware, and may reside, as non-limiting examples, in Random Access Memory (RAM), flash memory, Read Only Memory (ROM), Electrically Programmable ROM (EPROM), Electrically Erasable Programmable ROM (EEPROM), registers, a hard disk, a removable disk, a CD-ROM, or any other form of computer readable medium known in the art. An exemplary storage medium is coupled to the processor such that the processor can read information from, and write information to, the storage medium. In the alternative, the storage medium may be integral to the processor. The processor and the storage medium may reside in an Application Specific Integrated Circuit (ASIC). The ASIC may reside in a remote station. In the alternative, the processor and the storage medium may reside as discrete components in a remote station, base station, or server.
It is also noted that the operational steps described in any of the exemplary embodiments herein are described to provide examples and discussion. The operations described may be performed in numerous different sequences other than the illustrated sequences. Furthermore, operations described in a single operational step may actually be performed in a number of different steps. Additionally, one or more operational steps discussed in the exemplary embodiments may be combined. It is to be understood that the operational steps illustrated in the flow chart diagrams may be subject to numerous different modifications as will be readily apparent to one of skill in the art. Those of skill in the art would also understand that information and signals may be represented using any of a variety of different technologies and techniques. As non-limiting examples, data, instructions, commands, information, signals, bits, symbols, and chips that may be referenced throughout the above description may be represented by voltages, currents, electromagnetic waves, magnetic fields or particles, optical fields or particles, or any combination thereof.
The previous description of the disclosure is provided to enable any person skilled in the art to make or use the disclosure. Various modifications to the disclosure will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other variations without departing from the spirit or scope of the disclosure. Thus, the disclosure is not intended to be limited to the examples and designs described herein, but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.
Number | Name | Date | Kind |
---|---|---|---|
6714967 | Horvitz | Mar 2004 | B1 |
7107316 | Brown et al. | Sep 2006 | B2 |
7145898 | Elliott | Dec 2006 | B1 |
7266591 | Johnston | Sep 2007 | B1 |
7379993 | Valdes et al. | May 2008 | B2 |
7636348 | Bettis et al. | Dec 2009 | B2 |
7730309 | Zimmermann | Jun 2010 | B2 |
8015484 | Backer | Sep 2011 | B2 |
8300632 | Davis et al. | Oct 2012 | B2 |
8467308 | Johnston | Jun 2013 | B2 |
8605711 | Sinnreich et al. | Dec 2013 | B1 |
8606950 | Glatron et al. | Dec 2013 | B2 |
8693392 | Cooper et al. | Apr 2014 | B2 |
8695077 | Gerhard et al. | Apr 2014 | B1 |
8737596 | Kelley et al. | May 2014 | B2 |
8832271 | McCarty | Sep 2014 | B2 |
8856236 | Moore et al. | Oct 2014 | B2 |
8861692 | Phelps et al. | Oct 2014 | B1 |
8867731 | Lum et al. | Oct 2014 | B2 |
20030112766 | Riedel et al. | Jun 2003 | A1 |
20030131245 | Linderman | Jul 2003 | A1 |
20050084082 | Horvitz et al. | Apr 2005 | A1 |
20050177380 | Pritchard et al. | Aug 2005 | A1 |
20060155814 | Bennett et al. | Jul 2006 | A1 |
20060159063 | Kumar | Jul 2006 | A1 |
20070083929 | Sprosts et al. | Apr 2007 | A1 |
20070143408 | Daigle | Jun 2007 | A1 |
20070255662 | Tumminaro | Nov 2007 | A1 |
20070283423 | Bradley et al. | Dec 2007 | A1 |
20080162642 | Bachiri et al. | Jul 2008 | A1 |
20080192646 | Song et al. | Aug 2008 | A1 |
20110102930 | Johnston et al. | May 2011 | A1 |
20110238862 | Chaturvedi et al. | Sep 2011 | A1 |
20120001932 | Burnett et al. | Jan 2012 | A1 |
20120079031 | Matthews et al. | Mar 2012 | A1 |
20120158862 | Mosko et al. | Jun 2012 | A1 |
20130091286 | Spencer | Apr 2013 | A1 |
20140013202 | Schlumberger | Jan 2014 | A1 |
20140043994 | Bansal et al. | Feb 2014 | A1 |
20140095633 | Yoakum | Apr 2014 | A1 |
20140095724 | Yoakum et al. | Apr 2014 | A1 |
20140108594 | Siegel et al. | Apr 2014 | A1 |
20140126708 | Sayko | May 2014 | A1 |
20140126714 | Sayko | May 2014 | A1 |
20140161237 | Tolksdorf | Jun 2014 | A1 |
20140201820 | Li et al. | Jul 2014 | A1 |
20140223452 | Santhanam et al. | Aug 2014 | A1 |
20140237057 | Khodorenko | Aug 2014 | A1 |
20140245143 | Saint-Marc | Aug 2014 | A1 |
20140258822 | Li et al. | Sep 2014 | A1 |
20140270104 | O'Connor | Sep 2014 | A1 |
20140280734 | Chaturvedi et al. | Sep 2014 | A1 |
20140282054 | Yoakum | Sep 2014 | A1 |
20140282135 | Segre | Sep 2014 | A1 |
20140282399 | Gorelik et al. | Sep 2014 | A1 |
20140282903 | Singh et al. | Sep 2014 | A1 |
20140344169 | Phelps et al. | Nov 2014 | A1 |
20140348044 | Narayanan et al. | Nov 2014 | A1 |
20140365676 | Yoakum | Dec 2014 | A1 |
20140379931 | Gaviria | Dec 2014 | A1 |
20150002619 | Johnston et al. | Jan 2015 | A1 |
20150006610 | Johnston et al. | Jan 2015 | A1 |
20150006611 | Johnston et al. | Jan 2015 | A1 |
20150026473 | Johnston et al. | Jan 2015 | A1 |
20150036690 | Pastro | Feb 2015 | A1 |
20150039687 | Waxman et al. | Feb 2015 | A1 |
20150039760 | Yoakum | Feb 2015 | A1 |
Number | Date | Country |
---|---|---|
1615386 | Jan 2006 | EP |
2468758 | Sep 2010 | GB |
2468759 | Sep 2010 | GB |
2014060008 | Apr 2014 | WO |
2014123738 | Aug 2014 | WO |
2014190094 | Nov 2014 | WO |
Entry |
---|
Johnston, Alan et al., “Taking on WebRTC in an Enterprise,” IEEE Communications Magazine, Apr. 2013, pp. 48-54, vol. 51, Issue 4. |
Search Report for British patent application GB1317121.0 mailed Mar. 14, 2014, 3 pages. |
Search Report for British patent application GB1317122.8 mailed Mar. 11, 2014, 3 pages. |
Andreasen et al., “Session Description Protocol (SDP): Security Descriptions for Media Streams,” Network Working Group, Request for Comments: 4568, Standards Track, The Internet Society, Jul. 2006, 40 pages. |
Baugher et al., “The Secure Real-time Transport Protocol (SRTP),” Network Working Group, Request for Comments: 3711, Standards Track, The Internet Society, Mar. 2004, 51 pages. |
Johnston et al., “WebRTC: APIs and RTCWEB Protocols of the HTML5 Real-Time Web,” (Book), Second Edition, Smashwords Edition, Digital Codex LLC, Jun. 2013, 254 pages. |
Mahy et al., “Traversal Using Relays around NAT (TURN) : Relay Extensions to Session Traversal Utilities for NAT (STUN),” Internet Engineering Task Force, Request for Comments: 5766, Standards Track, IETF Trust, Apr. 2010, 61 pages. |
McGrew et al., “Datagram Transport Layer Security (DTLS) Extension to Establish Keys for the Secure Real-Time Transport Protocol (SRTP),” Internet Engineering Task Force, Request for Comments: 5764, Standards Track, IETF Trust, May 2010, 24 pages. |
Zimmermann et al., “ZRTP: Media Path Key Agreement for Unicast Secure RTP,” Internet Engineering Task Force, Request for Comments: 6189, Informational, IETF Trust, Apr. 2011, 102 pages. |
Singh, Kundan et al., “Building Communicating Web Applications Leveraging Endpoints and Cloud Resource Service,” Presented at the Sixth International Conference on Cloud Computing, Jun. 28, 2013, Santa Clara, California, IEEE Computer Society, pp. 486-493. |
Singh, Kundan et al., “Private Overlay of Enterprise Social Data and Interactions in the Public Web Context,” presented at the 9th IEEE International Conference on Collaborative Computing: Networking, Applications and Worksharing (Collaboratecom), Oct. 20-23, 2013, Austin, Texas, IEEE, 10 pages. |
Berners-Lee, Tim, “Socially Aware Cloud Storage,” Notes on web design, Aug. 17, 2009, http://www.w3.org/DesignIssues/CloudStorage.html, 9 pages. |
Chandra, Ramesh et al., “Separating Web Applications from User Data Storage with BStore,” presented at the USENIX Conference on Web Application Development, Jun. 2010, Boston, Massachusettes, 13 pages. |
Davids, Carol et al., “SIP APIs for Voice and Video Communications on the Web,” presented at the International Conference on Principles, Systems and Applications of IP Telecommunications (IPTcomm), Aug. 2011, Chicago, Illinois, 7 pages. |
Geambasu, Roxana et al., “Organizing and Sharing Distributed Personal Web-Service Data,” presented at the International World Wide Web Conference, Apr. 21-25, 2008, Beijing, China, International World Wide Web Conference Committee, pp. 755-764. |
Hsu, F. et al., “Secure File System Services for Web 2.0 Applications,” presented at the ACM Cloud Computing Security Workshop, Nov. 13, 2009, Chicago, Illinois, Association for Computing Machinery, 7 pages. |
Josh!, R., “Data-Oriented Architecture: A Loosley Coupled Real-Time SOA,” Whitepaper, Aug. 2007, Real-Time Innovations, Inc., http://rtcgroup.com/whitepapers/files/RTI—DataOrientedArchitecture—WhitePaper.pdf, 54 pages. |
Vahdat, Amin et al., “WebFS: A Global Cache Coherent File System,” UC Berkeley, Dec. 1996, retrieved Sep. 16, 2014 from https://www.cs.duke.edu/˜vahdat/webfs/webfs.html, 12 pages. |
Non-Final Office Action for U.S. Appl. No. 14/037,440, mailed Sep. 12, 2014, 15 pages. |
Loreto, Salvatore et al., “Real-Time Communications in the Web: Issues, Achievements, and Ongoing Standardization Efforts,” IEEE Internet Computing, vol. 16, Issue 5, IEEE Computer Society, Oct. 2, 2012, pp. 68-73. |
Search Report for British patent application GB1411584.4 mailed Dec. 30, 2014, 4 pages. |
Search Report for British patent application GB1411580.2 mailed Dec. 30, 2014, 4 pages. |
Non-Final Office Action for U.S. Appl. No. 13/835,913, mailed Nov. 20, 2014, 15 pages. |
Non-Final Office Action for U.S. Appl. No. 13/803,292, mailed Jan. 27, 2015, 13 pages. |
Non-Final Office Action for U.S. Appl. No. 14/050,891, mailed Jan. 29, 2015, 9 pages. |
Non-Final Office Action for U.S. Appl. No. 13/955,023, mailed Feb. 2, 2015, 12 pages. |
Final Office Action for U.S. Appl. No. 13/835,913, mailed Mar. 26, 2015, 17 pages. |
Final Office Action for U.S. Appl. No. 14/037,440, mailed Feb. 11, 2015, 19 pages. |
Non-Final Office Action for U.S. Appl. No. 14/068,839, mailed Feb. 20, 2015, 15 pages. |
Fischl, J. et al., “Framework for Establishing a Secure Real-time Transport Protocol (SRTP) Security Context Using Datagram Transport Layer Security (DTLS),” Internet Engineering Task Force (IETF), Request for Comments (RFC) 5763, May 2010, pp. 1-26. |
Jesup, R. et al., “DTLS Encapsulationof SCTP Packets for RTCWEB,” IETF: Network Working Group, Internet Draft, Feb. 16, 2013, pp. 1-6. |
Rescorla, E., “Security Considerations for RTC-Web,” IETF RTCWEB, Internet Draft, Jan. 22, 2013, pp. 1-16. |
Rescorla, E., “WebRTC Security Architecture,” IETF RTCWEB, Internet Draft, Jul. 14, 2013, pp. 1-30. |
Notice of Allowance for U.S. Appl. No. 13/944,368, mailed Apr. 1, 2015, 8 pages. |
Non-Final Office Action for U.S. Appl. No. 13/931,967, mailed May 5, 2015, 10 pages. |
Non-Final Office Action fo U.S. Appl. No. 13/931,970, mailed May 7, 2015, 9 pages. |
Number | Date | Country | |
---|---|---|---|
20150002619 A1 | Jan 2015 | US |