The technology of the disclosure relates generally to Web Real-Time Communications (WebRTC) interactive sessions.
Web Real-Time Communications (WebRTC) is an ongoing effort to develop industry standards for integrating real-time communications functionality into web clients, such as web browsers, to enable direct interaction with other web clients. This real-time communications functionality is accessible by web developers via standard markup tags, such as those provided by version 5 of the Hypertext Markup Language (HTML5), and client-side scripting Application Programming Interfaces (APIs) such as JavaScript APIs. More information regarding WebRTC may be found in “WebRTC: APIs and RTCWEB Protocols of the HTML5 Real-Time Web,” by Alan B. Johnston and Daniel C. Burnett (2012 Digital Codex LLC), which is incorporated herein by reference in its entirety.
WebRTC provides built-in capabilities for establishing real-time video, audio, and/or data streams in both point-to-point and multi-party interactive sessions. The WebRTC standards are currently under joint development by the World Wide Web Consortium (W3C) and the Internet Engineering Task Force (IETF). Information on the current state of WebRTC standards can be found at, e.g., http://www.w3c.org and http://www/ietf.org.
To establish a WebRTC interactive session (e.g., a real-time video, audio, and/or data exchange), two web clients may retrieve WebRTC-enabled web applications, such as HTML5/JavaScript web applications, from a web application server. Through the web applications, the two web clients then engage in an initiation dialogue to communicate and reach an agreement on parameters that define characteristics of the WebRTC interactive session. This initiation dialogue may take the form of a WebRTC “offer/answer” exchange. In an offer/answer exchange, a first web client on a sender device sends an “offer” to a second web client on a recipient device. The offer includes a WebRTC session description object (also referred to as a “token”) that specifies media types and capabilities that the first web client supports and prefers for use in the WebRTC interactive session. The second web client then responds with a WebRTC session description object “answer” that indicates which of the offered media types and capabilities are supported and acceptable by the second web client for use in the WebRTC interactive session. Once the WebRTC offer/answer exchange is complete, the web clients may then establish a direct peer connection with one another, and may begin an exchange of media or data packets transporting real-time communications. The peer connection between the web clients typically employs the Secure Real-time Transport Protocol (SRTP) to transport real-time media flows, and may utilize various other protocols for real-time data interchange.
The quality of the audio and/or video streams of a WebRTC interactive session may depend in large part on the capabilities of media components such as microphones, speakers, and/or webcams provided by computing devices on which WebRTC clients execute. Often, a computing device executing a WebRTC client is poorly equipped for providing high-quality audio and/or video in WebRTC interactive sessions. For instance, built-in microphones and speakers typically provided by conventional computers and laptops may create echoes and/or background noises. Likewise, webcams attached to or integrated into conventional computers may not provide sufficient resolution, frame rate, and/or color depth, as non-limiting examples. As a result, the quality of the user experience in the WebRTC interactive session may be significantly diminished.
Embodiments disclosed in the detailed description provide enhancing media characteristics during Web Real-Time Communications (WebRTC) interactive sessions by using Session Initiation Protocol (SIP) endpoints. Related methods, systems, and computer-readable media are also disclosed. In this regard, in one embodiment, a media redirection agent is provided (e.g., as part of a WebRTC client) to direct an audio and/or a video portion of a WebRTC interactive session to a SIP endpoint, thus providing enhanced media characteristics for the audio and/or video. The SIP endpoint may be, for instance, telephony or teleconferencing equipment or other SIP-enabled communications equipment that is more optimally equipped for presenting audio and/or video than a computing device on which the WebRTC client executes. The media redirection agent may be configured to intercept a WebRTC initiation token generated by the WebRTC client or received from a remote endpoint. Based on the WebRTC initiation token, the media redirection agent may generate a SIP endpoint WebRTC token, and may send the SIP endpoint WebRTC token to the remote endpoint. A WebRTC interactive session may then be established between the remote endpoint and the SIP endpoint based on the SIP endpoint WebRTC token. In some embodiments, the established WebRTC interactive session may include only an audio stream or a video stream, while some embodiments may provide that the established WebRTC interactive session includes both an audio stream and a video stream. According to some embodiments described herein, the SIP endpoint may provide one of an audio stream and a video stream of the established WebRTC interactive session, while the WebRTC client may provide the other. Some embodiments may further provide using an intermediately located media element and/or a SIP feature server to establish the WebRTC interactive session between the SIP endpoint and the remote endpoint. In this manner, such embodiments may provide enhanced audio and/or video for the WebRTC interactive session, resulting in an enhanced user experience.
In another embodiment, a method for enhancing media characteristics during WebRTC interactive sessions by using SIP endpoints is provided. The method comprises intercepting, by a media redirection agent of a WebRTC client executing on a computing device, a WebRTC initiation token. The method further comprises generating a SIP endpoint WebRTC token based on the WebRTC initiation token. The method also comprises sending the SIP endpoint WebRTC token to a remote endpoint. The method additionally comprises establishing a WebRTC interactive session between the remote endpoint and a SIP endpoint based on the SIP endpoint WebRTC token.
In another embodiment, a system for enhancing media characteristics during WebRTC interactive sessions by using SIP endpoints is provided. The system comprises at least one communications interface and a SIP endpoint. The system further comprises a first computing device communicatively coupled to the SIP endpoint via the at least one communications interface and executing a WebRTC client. The WebRTC client comprises a media redirection agent configured to intercept a WebRTC initiation token. The media redirection agent is further configured to generate a SIP endpoint WebRTC token based on the WebRTC initiation token. The media redirection agent is also configured to send the SIP endpoint WebRTC token to a remote endpoint. The media redirection agent is additionally configured to establish a WebRTC interactive session between the remote endpoint and the SIP endpoint based on the SIP endpoint WebRTC token.
In another embodiment, a non-transitory computer-readable medium is provided, having stored thereon computer-executable instructions to cause a processor to intercept a WebRTC initiation token. The computer-executable instructions further cause the processor to generate a SIP endpoint WebRTC token based on the WebRTC initiation token. The computer-executable instructions also cause the processor to send the SIP endpoint WebRTC token to a remote endpoint. The computer-executable instructions additionally cause the processor to establish a WebRTC interactive session between the remote endpoint and a SIP endpoint based on the SIP endpoint WebRTC token.
The accompanying drawing figures incorporated in and forming a part of this specification illustrate several aspects of the disclosure, and together with the description serve to explain the principles of the disclosure.
With reference now to the drawing figures, several exemplary embodiments of the present disclosure are described. The word “exemplary” is used herein to mean “serving as an example, instance, or illustration.” Any embodiment described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other embodiments.
Embodiments disclosed in the detailed description provide enhancing media characteristics during Web Real-Time Communications (WebRTC) interactive sessions by using Session Initiation Protocol (SIP) endpoints. Related methods, systems, and computer-readable media are also disclosed. In this regard, in one embodiment, a media redirection agent is provided (e.g., as part of a WebRTC client) to direct an audio and/or a video portion of a WebRTC interactive session to a SIP endpoint, thus providing enhanced media characteristics for the audio and/or video. The SIP endpoint may be, for instance, telephony or teleconferencing equipment or other SIP-enabled communications equipment that is more optimally equipped for presenting audio and/or video than a computing device on which the WebRTC client executes. The media redirection agent may be configured to intercept a WebRTC initiation token generated by the WebRTC client or received from a remote endpoint. Based on the WebRTC initiation token, the media redirection agent may generate a SIP endpoint WebRTC token, and may send the SIP endpoint WebRTC token to the remote endpoint. A WebRTC interactive session may then be established between the remote endpoint and the SIP endpoint based on the SIP endpoint WebRTC token. In some embodiments, the established WebRTC interactive session may include only an audio stream or a video stream, while some embodiments may provide that the established WebRTC interactive session includes both an audio stream and a video stream. According to some embodiments described herein, the SIP endpoint may provide one of an audio stream and a video stream of the established WebRTC interactive session, while the WebRTC client may provide the other. Some embodiments may further provide using an intermediately located media element and/or a SIP feature server to establish the WebRTC interactive session between the SIP endpoint and the remote endpoint. In this manner, such embodiments may provide enhanced audio and/or video for the WebRTC interactive session, resulting in an enhanced user experience.
In this regard,
Some embodiments may provide that the computing device 16 on which the WebRTC client 12 executes may be any computing device having network communications capabilities, such as a smartphone, a tablet computer, a dedicated web appliance, a media server, a desktop or server computer, or a purpose-built communications device, as non-limiting examples. The computing device 16 includes a communications interface 20 for connecting the computing device 16 to one or more public and/or private networks (not shown). In some embodiments, the elements of the computing device 16 may be distributed across more than one computing device 16.
In the embodiment illustrated in
In the example of
The network 30 in
External to the network 30 is a WebRTC application server 44, which serves WebRTC applications 46 and 48 to the WebRTC client 12 and to a remote endpoint 50 to requesting web clients. In some embodiments, the WebRTC application server 44 may be a single server, while in some applications the WebRTC application server 44 may comprise multiple servers that are communicatively coupled to each other. The remote endpoint 50 may be any computing or communications device having network communications capabilities, such as a smartphone, a tablet computer, a dedicated web appliance, or a desktop computer, as non-limiting examples, and may include a WebRTC client having functionality corresponding to that of the WebRTC client 12. It is to be understood that either or both of the WebRTC application server 44 and the remote endpoint 50 may reside within the network 30, or may reside within a public or private network communicatively coupled to the network 30.
To establish a WebRTC interactive session, the WebRTC client 12 and the remote endpoint 50 each downloads the WebRTC applications 46 and 48, respectively, from the WebRTC application server 44 (e.g., via Hyper Text Transfer Protocol (HTTP)/Hyper Text Transfer Protocol Secure (HTTPS) connections). In some embodiments, the WebRTC applications 46 and 48 may comprise HTML5/JavaScript web applications that provide rich user interfaces using HTML5, and that use JavaScript to handle user input and to communicate with the WebRTC application server 44. It is to be understood that one or both of the WebRTC applications 46 and 48 may comprise multiple, interoperable WebRTC applications tailored for specific characteristics (such as operating systems and/or platforms) of the WebRTC client 12 and/or the remote endpoint 50.
In a typical WebRTC interactive session, the WebRTC client 12 and the remote endpoint 50 engage in an initiation dialogue 52 with one another to negotiate media types and capabilities of the desired WebRTC interactive session. In some embodiments, the initiation dialogue 52 may include a WebRTC offer/answer exchange in which WebRTC session description objects (not shown) are exchanged between the WebRTC client 12 and the remote endpoint 50 via the WebRTC application server 44. After the initiation dialogue 52 is complete, a WebRTC interactive flow (not shown), comprising audio, video, and/or data streams, is established directly between the WebRTC client 12 and the remote endpoint 50 via a peer connection (not shown).
However, media characteristics (e.g., an audio quality and/or video quality) of the WebRTC interactive flow for a user (not shown) of the WebRTC client 12 may depend on the capabilities of the computing device 16 on which the WebRTC client 12 executes. For example, if the computing device 16 is a laptop or tablet computer, built-in microphone and speakers (not shown) provided by the computing device 16 may be poorly suited for providing high-quality audio in WebRTC interactive sessions. Similarly, a webcam (not shown) attached to or integrated into the computing device 16 may provide relatively inferior video quality. Consequently, a quality of user experience for the user of the WebRTC client 12 in the WebRTC interactive session may not be enhanced.
In this regard, the media redirection agent 14 is provided to enhance media characteristics during WebRTC interactive sessions by using the SIP endpoint 18. The SIP endpoint 18 may comprise, for instance, telephony or teleconferencing equipment (or other SIP-enabled equipment) that is more optimally equipped for presenting audio and/or video than the computing device 16. As described in greater detail below, the media redirection agent 14 may be configured to intercept a WebRTC session initiation token (not shown) generated by the WebRTC client 12 or received from the remote endpoint 50. The media redirection agent 14 may then generate a SIP endpoint WebRTC token (not shown) based on the WebRTC session initiation token. Exemplary operations for generating the SIP endpoint WebRTC token in various scenarios (using the media element 36 and/or the SIP feature server 32, as non-limiting examples) are discussed below. The media redirection agent 14 causes the SIP endpoint WebRTC token to be sent to the remote endpoint 50, and a WebRTC interactive session may then be established between the remote endpoint 50 and the SIP endpoint 18 based on the SIP endpoint WebRTC token. The audio and/or video capabilities of the SIP endpoint 18 may thus be employed to enhance the media characteristics of the audio and/or video stream of the WebRTC interactive session.
In some embodiments, if the SIP endpoint 18 is configured to provide WebRTC functionality natively, the SIP endpoint 18 and the remote endpoint 50 may be directly linked via a communications connection 54 over which the WebRTC interactive session is conducted. However, if the SIP endpoint 18 does not natively support WebRTC functionality, the SIP endpoint 18 may be linked to the remote endpoint 50 by a communications connection 56 passing through the media element 36. The media element 36 may be configured to route an audio and/or video portion of the WebRTC interactive session to the SIP endpoint 18 (as indicated by bidirectional arrow 58). In some embodiments, one of an audio stream and a video stream of the WebRTC interactive session may be routed to the SIP endpoint 18, while the other of the audio stream and the video stream of the WebRTC interactive session may be routed to the WebRTC functionality provider 24 of the WebRTC client 12.
To illustrate communications flows and operations of various embodiments of the present disclosure, the following figures are provided.
To illustrate exemplary high-level communications flows among elements of the WebRTC interactive system 10 of
As seen in
Based on the WebRTC initiation token 60, the media redirection agent 14 generates a SIP endpoint WebRTC token 66. The SIP endpoint WebRTC token 66 includes connection information for establishing a WebRTC connection between the SIP endpoint 18 and the remote endpoint 50. In embodiments wherein the WebRTC initiation token 60 represents a WebRTC offer generated by the scripting engine 22 of the WebRTC client 12, the SIP endpoint WebRTC token 66 may represent a WebRTC offer to be sent to the remote endpoint 50. Conversely, in embodiments wherein the WebRTC initiation token 60 represents a WebRTC offer generated by the remote endpoint 50, the SIP endpoint WebRTC token 66 may represent a WebRTC answer to be sent to the remote endpoint 50.
Embodiments of the media redirection agent 14 disclosed herein may generate the SIP endpoint WebRTC token 66 with the assistance of one or more of the other elements shown in
Generation of the SIP endpoint WebRTC token 66 may also be facilitated by the SIP feature server 32. In such embodiments, the media redirection agent 14 may exchange SIP messages with the SIP feature server 32 (as indicated by bidirectional arrow 74). The SIP feature server 32 then communicates with the media element 36 (bidirectional arrow 76), and the media element 36 engages in SIP communications with the SIP endpoint 18 (bidirectional arrow 78).
After generating the SIP endpoint WebRTC token 66, the media redirection agent 14 sends the SIP endpoint WebRTC token 66 to the remote endpoint 50, as indicated by arrow 80. It is to be understood that sending the SIP endpoint WebRTC token 66 to the remote endpoint 50 may include sending the SIP endpoint WebRTC token 66 to the WebRTC application server 44, which may then relay the SIP endpoint WebRTC token 66 to the remote endpoint 50. If the SIP endpoint WebRTC token 66 represents a WebRTC offer generated by the WebRTC client 12, the remote endpoint 50 may respond with a WebRTC answer (not shown).
Based on the SIP endpoint WebRTC token 66, a WebRTC interactive session 82 may be established between the SIP endpoint 18 and the remote endpoint 50. The WebRTC interactive session 82 may include only an audio stream or a video stream, or may include both audio and video. In embodiments where the WebRTC interactive session 82 includes only one of an audio stream and a video stream, a separate split stream 84 including the other of the audio stream and the video stream may be established between the remote endpoint 50 and the WebRTC functionality provider 24 of the WebRTC client 12. In this manner, as a non-limiting example, the capabilities of the SIP endpoint 18 may be used to enhance audio characteristics of the WebRTC interactive session 82, while the WebRTC client 12 provides video via the split stream 84.
To illustrate exemplary operations of the media redirection agent 14 of
The media redirection agent 14 then generates a SIP endpoint WebRTC token 66 based on the WebRTC initiation token 60 (block 88). In some embodiments, the SIP endpoint WebRTC token 66 may represent a WebRTC offer, or may constitute a WebRTC answer to be sent in response to a WebRTC offer received from the remote endpoint 50. The media redirection agent 14 sends the SIP endpoint WebRTC token 66 to the remote endpoint 50 (block 90). Based on the SIP endpoint WebRTC token 66, a WebRTC interactive session 82 is established between the remote endpoint 50 and a SIP endpoint 18 (block 92).
In some embodiments, the SIP endpoint 18 may provide native WebRTC support, enabling the SIP endpoint 18 to generate a WebRTC offer or answer itself, and to engage in a WebRTC interactive session 82. Accordingly, the media redirection agent 14 in such embodiments may communicate directly with the SIP endpoint 18 (i.e., without the need for the media element 36) to generate the SIP endpoint WebRTC token 66. To illustrate exemplary communications flows and operations in such embodiments,
In
The SIP endpoint WebRTC token 66 is then sent to the remote endpoint 50 via the WebRTC application server 44 as a WebRTC offer (arrows 102 and 104). If a user of the remote endpoint 50 wishes to participate in a WebRTC interactive session, the remote endpoint 50 returns a WebRTC response token 106 to the media redirection agent 14 via the WebRTC application server 44 (arrows 108 and 110). The media redirection agent 14 then forwards the WebRTC response token 106 to the SIP endpoint 18 (arrow 112). Subsequently, a WebRTC interactive session (not shown) is established between the SIP endpoint 18 and the remote endpoint 50 based on the SIP endpoint WebRTC token 66 and the WebRTC response token 106.
According to some embodiments disclosed herein, the SIP endpoint 18 may not provide native WebRTC functionality, necessitating the use of the media element 36 by the media redirection agent 14 in generating the SIP endpoint WebRTC token 66. In particular, it may be necessary for the media element 36 to act as a “go-between” between the SIP endpoint 18 and the remote endpoint 50 in a WebRTC interactive session. This may be accomplished by establishing a SIP-based communications session between the SIP endpoint 18 and the media element 36, and establishing a WebRTC interactive session between the media element 36 and the remote endpoint 50. In this regard,
As seen in
Using the media element SIP offer 138, the media redirection agent 14 generates and sends a SIP initiation request 142 to the SIP endpoint 18 (arrow 144). The SIP initiation request 142 may comprise, for instance, a SIP INVITE message. The media redirection agent 14 then receives a SIP answer message 146 from the SIP endpoint 18 (arrow 148). As a non-limiting example, the SIP answer message 146 may comprise a SIP 200 OK message. The SIP answer message 146 is then relayed to the media element 36 (arrow 150). In this manner, a SIP-based communications session (not shown) may be established between the media element 36 and the SIP endpoint 18.
It is to be understood that the sequence of communications flows for establishing the SIP-based communications session between the media element 36 and the SIP endpoint 18 described above is one non-limiting example. Other embodiments within the scope of the present disclosure may employ a different sequence of communications flows to establish the SIP-based communications session. For example, the media redirection agent 14 may request and receive the media element SIP offer 138 from the SIP endpoint 18, send the SIP initiation request 142 to and receive the SIP answer message 146 from the media element 36, and send the SIP answer message 146 to the SIP endpoint 18.
With continuing reference to
To illustrate further exemplary operations of the media redirection agent 14 of
In
The media redirection agent 14 next receives a SIP answer message 146 (block 176). The SIP answer message 146 is sent to the first of the media element 36 and the SIP endpoint 18 (block 178). Subsequently, the media redirection agent 14 requests the SIP endpoint WebRTC token 66 based on the WebRTC initiation token 60 from the media element 36 (block 180). The media redirection agent 14 receives the SIP endpoint WebRTC token 66 from the media element 36 (block 182). The SIP endpoint WebRTC token 66 may then be used to establish a WebRTC interactive session (not shown) between the SIP endpoint 18 and the remote endpoint 50 via the media element 36, as described above.
In the example of
The media redirection agent 14 in
As with
Exemplary operations for providing the communications flows described in
The media redirection agent 14 receives the media element SIP offer 196 (block 220). The media redirection agent 14 then generates a SIP initiation request 200 based on the media element SIP offer 196 (block 222). The SIP initiation request 200 is sent by the media redirection agent 14 to a second one selected from the group consisting of the media element 36 and the SIP endpoint 18 (block 224). The media redirection agent 14 next receives a SIP answer message 204 (block 226). The SIP answer message 204 is sent to the first of the media element 36 and the SIP endpoint 18 (block 228).
In some embodiments, it may be desirable for the audio and video streams received from the remote endpoint 50 during a WebRTC interactive session to be split between the SIP endpoint 18 and the WebRTC functionality provider 24 of the WebRTC client 12. For instance, some embodiments disclosed herein may use the SIP endpoint 18 to enhance an audio stream of a WebRTC interactive session, while the WebRTC functionality provider 24 of the WebRTC client 12 provides the video stream of the WebRTC interactive session. To accomplish this, the media redirection agent 14 may be configured to use the media element 36 to establish a SIP-based audio communications connection with the SIP endpoint 18, establish a WebRTC video flow with the WebRTC client 12, and “split” the incoming WebRTC audio and video flows from the remote endpoint 50 to redirect them to the SIP endpoint 18 and the WebRTC client 12, respectively.
In this regard,
In
With continued reference to
It is to be understood that the sequence of communications flows for establishing a WebRTC interactive video session between the media element 36 and the WebRTC client 12 is one non-limiting example, and that other embodiments within the scope of the present disclosure may employ a different sequence of communications flows. For instance, the media redirection agent 14 may send the WebRTC video initiation token 232 to the WebRTC client 12, receive the WebRTC video response token 236 from the WebRTC client 12, and send the WebRTC video response token 236 to the media element 36. It is to be further understood that the sequence of communications flows for establishing the SIP-based communications session between the media element 36 and the SIP endpoint 18 described in
Referring now to
The media redirection agent 14 in
Referring now to
The sequence of communications flows for establishing a WebRTC interactive video session between the media element 36 and the WebRTC client 12 shown in
In
The media redirection agent 14 then carries out operations to establish a WebRTC interactive session between the remote endpoint 50 and the SIP endpoint 18 based on the SIP endpoint WebRTC token 66 (block 328). As seen in
In some embodiments, communications between the media redirection agent 14 and the media element 36 may be facilitated by a SIP feature server 32. In this regard,
In the example of
With continued reference to
Referring now to
The exemplary computer system 398 includes a processing device or processor 400, a main memory 402 (as non-limiting examples, read-only memory (ROM), flash memory, dynamic random access memory (DRAM) such as synchronous DRAM (SDRAM), etc.), and a static memory 404 (as non-limiting examples, flash memory, static random access memory (SRAM), etc.), which may communicate with each other via a bus 405. Alternatively, the processing device 400 may be connected to the main memory 402 and/or the static memory 404 directly or via some other connectivity means.
The processing device 400 represents one or more processing devices such as a microprocessor, central processing unit (CPU), or the like. More particularly, the processing device 400 may be a complex instruction set computing (CISC) microprocessor, a reduced instruction set computing (RISC) microprocessor, a very long instruction word (VLIW) microprocessor, a processor implementing other instruction sets, or processors implementing a combination of instruction sets. The processing device 400 is configured to execute processing logic in instructions 406 and/or cached instructions 408 for performing the operations and steps discussed herein.
The computer system 398 may further include a communications interface in the form of a network interface device 410. It also may or may not include an input 412 to receive input and selections to be communicated to the computer system 398 when executing the instructions 406, 408. It also may or may not include an output 414, including but not limited to display(s) 416. The display(s) 416 may be a video display unit (as non-limiting examples, a liquid crystal display (LCD) or a cathode ray tube (CRT)), an alphanumeric input device (as a non-limiting example, a keyboard), a cursor control device (as a non-limiting example, a mouse), and/or a touch screen device (as a non-limiting example, a tablet input device or screen).
The computer system 398 may or may not include a data storage device 418 that includes using drive(s) 420 to store the functions described herein in a computer-readable medium 422, on which is stored one or more sets of instructions 424 (e.g., software) embodying any one or more of the methodologies or functions described herein. The functions can include the methods and/or other functions of the processing system 396, a participant user device, and/or a licensing server, as non-limiting examples. The one or more sets of instructions 424 may also reside, completely or at least partially, within the main memory 402 and/or within the processing device 400 during execution thereof by the computer system 398. The main memory 402 and the processing device 400 also constitute machine-accessible storage media. The instructions 406, 408, and/or 424 may further be transmitted or received over a network 426 via the network interface device 410. The network 426 may be an intra-network or an inter-network.
While the computer-readable medium 422 is shown in an exemplary embodiment to be a single medium, the term “machine-accessible storage medium” should be taken to include a single medium or multiple media (as non-limiting examples, a centralized or distributed database, and/or associated caches and servers) that store the one or more sets of instructions 424. The term “machine-accessible storage medium” shall also be taken to include any medium that is capable of storing, encoding, or carrying a set of instructions for execution by the machine, and that cause the machine to perform any one or more of the methodologies disclosed herein. The term “machine-accessible storage medium” shall accordingly be taken to include, but not be limited to, solid-state memories, optical and magnetic media, and carrier wave signals.
The embodiments disclosed herein may be embodied in hardware and in instructions that are stored in hardware, and may reside, as non-limiting examples, in Random Access Memory (RAM), flash memory, Read Only Memory (ROM), Electrically Programmable ROM (EPROM), Electrically Erasable Programmable ROM (EEPROM), registers, a hard disk, a removable disk, a CD-ROM, or any other form of computer readable medium known in the art. An exemplary storage medium is coupled to the processor such that the processor can read information from, and write information to, the storage medium. In the alternative, the storage medium may be integral to the processor. The processor and the storage medium may reside in an Application Specific Integrated Circuit (ASIC). The ASIC may reside in a remote station. In the alternative, the processor and the storage medium may reside as discrete components in a remote station, base station, or server.
It is also noted that the operational steps described in any of the exemplary embodiments herein are described to provide examples and discussion. The operations described may be performed in numerous different sequences other than the illustrated sequences. Furthermore, operations described in a single operational step may actually be performed in a number of different steps. Additionally, one or more operational steps discussed in the exemplary embodiments may be combined. It is to be understood that the operational steps illustrated in the flow chart diagrams may be subject to numerous different modifications as will be readily apparent to one of skill in the art. Those of skill in the art would also understand that information and signals may be represented using any of a variety of different technologies and techniques. As non-limiting examples, data, instructions, commands, information, signals, bits, symbols, and chips that may be referenced throughout the above description may be represented by voltages, currents, electromagnetic waves, magnetic fields or particles, optical fields or particles, or any combination thereof.
The previous description of the disclosure is provided to enable any person skilled in the art to make or use the disclosure. Various modifications to the disclosure will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other variations without departing from the spirit or scope of the disclosure. Thus, the disclosure is not intended to be limited to the examples and designs described herein, but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.
Number | Name | Date | Kind |
---|---|---|---|
6714967 | Horvitz | Mar 2004 | B1 |
7107316 | Brown et al. | Sep 2006 | B2 |
7145898 | Elliott | Dec 2006 | B1 |
7266591 | Johnston | Sep 2007 | B1 |
7379993 | Valdes et al. | May 2008 | B2 |
7636348 | Bettis et al. | Dec 2009 | B2 |
7730309 | Zimmermann | Jun 2010 | B2 |
8015484 | Backer | Sep 2011 | B2 |
8250635 | Chari et al. | Aug 2012 | B2 |
8300632 | Davis et al. | Oct 2012 | B2 |
8467308 | Johnston | Jun 2013 | B2 |
8494507 | Tedesco et al. | Jul 2013 | B1 |
8601144 | Ryner | Dec 2013 | B1 |
8605711 | Sinnreich et al. | Dec 2013 | B1 |
8606950 | Glatron et al. | Dec 2013 | B2 |
8693392 | Cooper et al. | Apr 2014 | B2 |
8695077 | Gerhard et al. | Apr 2014 | B1 |
8737596 | Kelley et al. | May 2014 | B2 |
8744147 | Torti | Jun 2014 | B2 |
8832271 | McCarty | Sep 2014 | B2 |
8856236 | Moore et al. | Oct 2014 | B2 |
8861692 | Phelps et al. | Oct 2014 | B1 |
8867731 | Lum et al. | Oct 2014 | B2 |
20020161685 | Dwinnell | Oct 2002 | A1 |
20030112766 | Riedel et al. | Jun 2003 | A1 |
20030120599 | Agboatwalla et al. | Jun 2003 | A1 |
20030131245 | Linderman | Jul 2003 | A1 |
20030188193 | Venkataramappa | Oct 2003 | A1 |
20040019494 | Ridgeway et al. | Jan 2004 | A1 |
20040081173 | Feather | Apr 2004 | A1 |
20040093515 | Reeves, Jr. | May 2004 | A1 |
20040167984 | Herrmann | Aug 2004 | A1 |
20040216173 | Horoszowski et al. | Oct 2004 | A1 |
20050084082 | Horvitz et al. | Apr 2005 | A1 |
20050177380 | Pritchard et al. | Aug 2005 | A1 |
20060104526 | Gringeler et al. | May 2006 | A1 |
20060155814 | Bennett et al. | Jul 2006 | A1 |
20060159063 | Kumar | Jul 2006 | A1 |
20060200855 | Willis | Sep 2006 | A1 |
20060230438 | Shappir et al. | Oct 2006 | A1 |
20070083929 | Sprosts et al. | Apr 2007 | A1 |
20070143408 | Daigle | Jun 2007 | A1 |
20070255662 | Tumminaro | Nov 2007 | A1 |
20070283423 | Bradley et al. | Dec 2007 | A1 |
20080046414 | Haub et al. | Feb 2008 | A1 |
20080046457 | Haub et al. | Feb 2008 | A1 |
20080046838 | Haub et al. | Feb 2008 | A1 |
20080127137 | Becker et al. | May 2008 | A1 |
20080162642 | Bachiri et al. | Jul 2008 | A1 |
20080192646 | Song et al. | Aug 2008 | A1 |
20080270541 | Keener et al. | Oct 2008 | A1 |
20090070477 | Baum et al. | Mar 2009 | A1 |
20090094684 | Chinnusamy et al. | Apr 2009 | A1 |
20090219920 | Brandstatter | Sep 2009 | A1 |
20090300060 | Beringer et al. | Dec 2009 | A1 |
20100011282 | Dollard et al. | Jan 2010 | A1 |
20100023519 | Kailash et al. | Jan 2010 | A1 |
20100024019 | Backlund | Jan 2010 | A1 |
20100037324 | Grant et al. | Feb 2010 | A1 |
20100118700 | Blum et al. | May 2010 | A1 |
20100246571 | Geppert et al. | Sep 2010 | A1 |
20110102930 | Johnston et al. | May 2011 | A1 |
20110206013 | Aramoto et al. | Aug 2011 | A1 |
20110238862 | Chaturvedi et al. | Sep 2011 | A1 |
20120001932 | Burnett et al. | Jan 2012 | A1 |
20120079031 | Matthews et al. | Mar 2012 | A1 |
20120137231 | Maxfield et al. | May 2012 | A1 |
20120158862 | Mosko et al. | Jun 2012 | A1 |
20120192086 | Ghods et al. | Jul 2012 | A1 |
20120221725 | Schroeder, Jr. | Aug 2012 | A1 |
20130002799 | Mock | Jan 2013 | A1 |
20130078972 | Levien et al. | Mar 2013 | A1 |
20130091286 | Spencer | Apr 2013 | A1 |
20130138829 | Bulava | May 2013 | A1 |
20130321340 | Seo et al. | Dec 2013 | A1 |
20130325934 | Fausak et al. | Dec 2013 | A1 |
20130346329 | Alib-Bulatao et al. | Dec 2013 | A1 |
20140013202 | Schlumberger | Jan 2014 | A1 |
20140043994 | Bansal et al. | Feb 2014 | A1 |
20140095633 | Yoakum | Apr 2014 | A1 |
20140095724 | Yoakum et al. | Apr 2014 | A1 |
20140095731 | Carey et al. | Apr 2014 | A1 |
20140108594 | Siegel et al. | Apr 2014 | A1 |
20140126708 | Sayko | May 2014 | A1 |
20140126714 | Sayko | May 2014 | A1 |
20140126715 | Lum | May 2014 | A1 |
20140143823 | Manchester et al. | May 2014 | A1 |
20140149512 | Leitch | May 2014 | A1 |
20140161237 | Tolksdorf | Jun 2014 | A1 |
20140201820 | Li et al. | Jul 2014 | A1 |
20140219167 | Santhanam et al. | Aug 2014 | A1 |
20140222893 | Gangadharan | Aug 2014 | A1 |
20140222894 | Gangadharan | Aug 2014 | A1 |
20140222930 | Gangadharan | Aug 2014 | A1 |
20140223452 | Santhanam | Aug 2014 | A1 |
20140237057 | Khodorenko | Aug 2014 | A1 |
20140241215 | Massover | Aug 2014 | A1 |
20140244235 | Michaelis | Aug 2014 | A1 |
20140245143 | Saint-Marc | Aug 2014 | A1 |
20140258822 | Li et al. | Sep 2014 | A1 |
20140259127 | Shaw | Sep 2014 | A1 |
20140269326 | Westin et al. | Sep 2014 | A1 |
20140270104 | O'Connor | Sep 2014 | A1 |
20140280734 | Chaturvedi et al. | Sep 2014 | A1 |
20140282054 | Yoakum | Sep 2014 | A1 |
20140282135 | Segre | Sep 2014 | A1 |
20140282399 | Gorelik et al. | Sep 2014 | A1 |
20140282765 | Casey et al. | Sep 2014 | A1 |
20140282903 | Singh et al. | Sep 2014 | A1 |
20140324979 | Gao | Oct 2014 | A1 |
20140325078 | Shan | Oct 2014 | A1 |
20140344169 | Phelps | Nov 2014 | A1 |
20140348044 | Narayanan | Nov 2014 | A1 |
20140365676 | Yoakum | Dec 2014 | A1 |
20140379931 | Gaviria | Dec 2014 | A1 |
20150002614 | Zino | Jan 2015 | A1 |
20150002619 | Johnston et al. | Jan 2015 | A1 |
20150006610 | Johnston et al. | Jan 2015 | A1 |
20150006611 | Johnston et al. | Jan 2015 | A1 |
20150026473 | Johnston et al. | Jan 2015 | A1 |
20150033300 | Timariu | Jan 2015 | A1 |
20150036690 | Pastro | Feb 2015 | A1 |
20150039687 | Waxman et al. | Feb 2015 | A1 |
20150039760 | Yoakum | Feb 2015 | A1 |
20150052067 | Thiyagarajan et al. | Feb 2015 | A1 |
20150082021 | Mandyam | Mar 2015 | A1 |
20150106438 | Fan | Apr 2015 | A1 |
20150106528 | Somes | Apr 2015 | A1 |
20150180748 | Ding | Jun 2015 | A1 |
20150180825 | Ren | Jun 2015 | A1 |
20150195309 | Opsenica | Jul 2015 | A1 |
20150229635 | Saridaki | Aug 2015 | A1 |
20150264106 | Baek | Sep 2015 | A1 |
20150280963 | Bollapalli | Oct 2015 | A1 |
20150281234 | Sylvain | Oct 2015 | A1 |
20150334136 | Gao | Nov 2015 | A1 |
20150365244 | Schmitz | Dec 2015 | A1 |
20150373056 | Bouthemy | Dec 2015 | A1 |
20160065623 | Whynot | Mar 2016 | A1 |
20160191523 | Jian | Jun 2016 | A1 |
20160315938 | Kunz | Oct 2016 | A1 |
20160323325 | Nie | Nov 2016 | A1 |
20160359927 | Li | Dec 2016 | A1 |
20170104704 | Choi | Apr 2017 | A1 |
20170141934 | Ding | May 2017 | A1 |
Number | Date | Country |
---|---|---|
1732665 | Feb 2006 | CN |
103580986 | Feb 2014 | CN |
1615386 | Jan 2006 | EP |
2529316 | Dec 2012 | EP |
2295747 | Jun 1996 | GB |
2468758 | Sep 2010 | GB |
2468759 | Sep 2010 | GB |
2517833 | Mar 2015 | GB |
2002207683 | Jul 2002 | JP |
2002374318 | Dec 2002 | JP |
2005346556 | Dec 2005 | JP |
2006050407 | Feb 2006 | JP |
2011504665 | Feb 2011 | JP |
WO 2004062229 | Jul 2004 | WO |
2014060008 | Apr 2014 | WO |
2014123738 | Aug 2014 | WO |
2014190094 | Nov 2014 | WO |
2015032277 | Mar 2015 | WO |
Entry |
---|
Rodriguez, Pedro et al., “Advanced Videoconferencing Services Based on WebRTC,” IADIS International Conferences Web Based Communities and Social Media 2012 and Collaborative Technologies 2012, Jul. 17-23, 2012, pp. 180-184, http://www.iadisportal.org/wbc-2012-proceedings. |
Examination Report for British Patent Application GB1411584.4, dated Aug. 21, 2015, 1 page. |
Examination Report for British Patent Application GB1411580.2, dated Aug. 21, 2015, 1 page. |
Notification of Reasons for Refusal for Japanese Patent Application 2013-201221, dated Aug. 25, 2015, 8 pages. |
Non-Final Office Action for U.S. Appl. No. 13/835,913, dated Sep. 3, 2015, 19 pages. |
Non-Final Office Action for U.S. Appl. No. 13/912,520, dated Sep. 9, 2015, 21 pages. |
Non-Final Office Action for U.S. Appl. No. 13/863,662, dated Sep. 25, 2015, 23 pages. |
Advisory Action for U.S. Appl. No. 13/803,292, dated Aug. 21, 2015, 3 pages. |
Non-Final Office Action for U.S. Appl. No. 13/803,292, dated Oct. 9, 2015, 13 pages. |
Advisory Action and Examiner-Initiated Interview Summary for U.S. Appl. No. 14/050,891, dated Sep. 29, 2015, 4 pages. |
Final Office Action for U.S. Appl. No. 14/068,839, dated Sep. 9, 2015, 17 pages. |
Final Office Action for U.S. Appl. No. 13/931,967, dated Aug. 20, 2015, 12 pages. |
Final Office Action for U.S. Appl. No. 13/931,970, dated Aug. 27, 2015, 10 pages. |
Bergkvist, Adam et al., “WebRTC 1.0: Real-time Communication Between Browsers,” W3C Working Draft, Feb. 9, 2012, http://www.w3.org/TR/2012/WD-webrtc-20120209/, 42 pages. |
Notice of Reason for Refusal for Japanese Patent Application 2013-201228, dated Jun. 11, 2015, 8 pages. |
Advisory Action for U.S. Appl. No. 13/835,913, dated Jun. 10, 2015, 3 pages. |
Final Office Action for U.S. Appl. No. 13/803,292, dated Jun. 12, 2015, 17 pages. |
Final Office Action and Examiner Initiated Interview Summary for U.S. Appl. No. 14/050,891, dated Jun. 29, 2015, 11 pages. |
Final Office Action for U.S. Appl. No. 13/955,023, dated Jul. 20, 2015, 17 pages. |
Corrected Notice of Allowability for U.S. Appl. No. 13/944,368, dated Jul. 23, 2015, 4 pages. |
Non-Final Office Action for U.S. Appl. No. 14/141,798, dated Jul. 17, 2015, 13 pages. |
Extended European Search Report for European Patent Application 15161452.6, dated Jun. 23, 2015, 5 pages. |
Non-Final Office Action for U.S. Appl. No. 13/955,711, dated Nov. 9, 2015, 10 pages. |
Notice of Allowance for U.S. Appl. No. 14/050,891, dated Nov. 10, 2015, 10 pages. |
Non-Final Office Action for U.S. Appl. No. 13/955,023, dated Dec. 9, 2015, 13 pages. |
Non-Final Office Action for U.S. Appl. No. 14/037,440, dated Oct. 22, 2015, 15 pages. |
Advisory Action for U.S. Appl. No. 13/931,967, dated Nov. 3, 2015, 3 pages. |
Advisory Action for U.S. Appl. No. 13/931,970, dated Nov. 5, 2015, 3 pages. |
Non-Final Office Action for U.S. Appl. No. 14/068,943, dated Dec. 2, 2015, 16 pages. |
Non-Final Office Action for U.S. Appl. No. 14/255,429, dated Nov. 9, 2015, 26 pages. |
Search Report for British Patent Application No. GB1423089.0, dated Jul. 6, 2015, 4 pages. |
Author Unknown, “WebRTC,” WebRTC.org, Date Accessed: Jan. 26, 2016, 4 pages, http://webrtc.org/. |
Notice of Allowance for U.S. Appl. No. 13/863,662, dated Feb. 1, 2016, 17 pages. |
Advisory Action and Examiner-Initiated Interview Summary for U.S. Appl. No. 14/068,839, dated Feb. 12, 2016, 4 pages. |
Final Office Action for U.S. Appl. No. 14/141,798, dated Dec. 24, 2015, 10 pages. |
Non-Final Office Action for U.S. Appl. No. 14/174,371, dated Feb. 18, 2016, 18 pages. |
Non-Final Office Action for U.S. Appl. No. 13/931,970, dated Feb. 23, 2016, 11 pages. |
Loreto, Salvatore et al., “Real-Time Communications in the Web: Issues, Achievements, and Ongoing Standardization Efforts,” IEEE Internet Computing, vol. 16, Issue 5, IEEE Computer Society, Oct. 2, 2012, pp. 68-73. |
Search Report for British patent application GB1411584.4 dated Dec. 30, 2014, 4 pages. |
Search Report for British patent application GB1411580.2 dated Dec. 30, 2014, 4 pages. |
Non-Final Office Action for U.S. Appl. No. 13/931,968, dated Dec. 8, 2014, 6 pages. |
Non-Final Office Action for U.S. Appl. No. 13/835,913, dated Nov. 20, 2014, 15 pages. |
Non-Final Office Action for U.S. Appl. No. 13/803,292, dated Jan. 27, 2015, 13 pages. |
Non-Final Office Action for U.S. Appl. No. 14/050,891, dated Jan. 29, 2015, 9 pages. |
Non-Final Office Action for U.S. Appl. No. 13/955,023, dated Feb. 2, 2015, 12 pages. |
Barth, A. “The Web Origin Concept,” Internet Engineering Task Force (IETF), Request for Comments 6454 (RFC 6454), Dec. 2011, 19 pages, http://www.ietf.org/rfc/rfc6454.txt. |
Fischl, J. et al., “Framework for Establishing a Secure Real-time Transport Protocol (SRTP) Security Context Using Datagram Transport Layer Security (DTLS),” Internet Engineering Task Force (IETF), Request for Comments (RFC) 5763, May 2010, 26 pages. |
Jesup, R. et al., “DTLS Encapsulation of SCTP Packets for RTCWEB,” IETF: Network Working Group, Internet Draft, Feb. 16, 2013, 6 pages. |
Johnston, A. et al., “An Origin Attribute for the STUN Protocol,” Internet Engineering Task Force (IETF), Internet-Draft, Jul. 20, 2014, 14 pages, https://tools.ietf.org/html/draft-ietf-tram-stun-origin-00. |
Rescorla, E., “Security Considerations for RTC-Web,” IETF RTCWEB, Internet Draft, Jan. 22, 2013, 16 pages. |
Rescorla, E., “WebRTC Security Architecture,” IETF RTCWEB, Internet Draft, Jul. 14, 2013, 30 pages. |
Corrected Notice of Allowability for U.S. Appl. No. 13/931,968, dated Apr. 24, 2015, 4 pages. |
Advisory Action for U.S. Appl. No. 14/037,440, dated May 20, 2015, 3 pages. |
Search Report for British Patent Application GB1419338.7, dated Apr. 27, 2015, 4 pages. |
Search Report for British Patent Application GB1419334.6, dated Apr. 28, 2015, 6 pages. |
Notice of Allowance for U.S. Appl. No. 13/944,368, dated Apr. 1, 2015, 8 pages. |
Non-Final Office Action for U.S. Appl. No. 13/931,967, dated May 5, 2015, 10 pages. |
Non-Final Office Action for U.S. Appl. No. 13/931,970, dated May 7, 2015, 9 pages. |
Notice of Allowance for U.S. Appl. No. 13/931,968, dated Mar. 23, 2015, 7 pages. |
Final Office Action for U.S. Appl. No. 13/835,913, dated Mar. 26, 2015, 17 pages. |
Final Office Action for U.S. Appl. No. 14/037,440, dated Feb. 11, 2015, 19 pages. |
Non-Final Office Action for U.S. Appl. No. 14/068,839, dated Feb. 20, 2015, 15 pages. |
Andreasen et al., “Session Description Protocol (SDP): Security Descriptions for Media Streams,” Network Working Group, Request for Comments: 4568, Standards Track, The Internet Society, Jul. 2006, 40 pages. |
Baugher et al., “The Secure Real-time Transport Protocol (SRTP),” Network Working Group, Request for Comments: 3711, Standards Track, The Internet Society, Mar. 2004, 51 pages. |
Johnston, Alan et al., “Taking on WebRTC in an Enterprise,” IEEE Communications Magazine, Apr. 2013, pp. 48-54, vol. 51, Issue 4. |
Johnston et al., “WebRTC: APIs and RTCWEB Protocols of the HTML5 Real-Time Web,” (Book), Second Edition, Smashwords Edition, Digital Codex LLC, Jun. 2013, 254 pages. |
Mahy et al., “Traversal Using Relays around NAT (TURN) : Relay Extensions to Session Traversal Utilities for NAT (STUN),” Internet Engineering Task Force, Request for Comments: 5766, Standards Track, IETF Trust, Apr. 2010, 61 pages. |
McGrew et al., “Datagram Transport Layer Security (DTLS) Extension to Establish Keys for the Secure Real-Time Transport Protocol (SRTP),” Internet Engineering Task Force, Request for Comments: 5764, Standards Track, IETF Trust, May 2010, 24 pages. |
Singh, Kundan et al., “Building Communicating Web Applications Leveraging Endpoints and Cloud Resource Service,” Presented at the Sixth International Conference on Cloud Computing, Jun. 28, 2013, Santa Clara, California, IEEE Computer Society, pp. 486-493. |
Singh, Kundan et al., “Private Overlay of Enterprise Social Data and Interactions in the Public Web Context,” presented at the 9th IEEE International Conference on Collaborative Computing: Networking, Applications and Worksharing (Collaboratecom), Oct. 20-23, 2013, Austin, Texas, IEEE, 10 pages. |
Zimmermann et al., “ZRTP: Media Path Key Agreement for Unicast Secure RTP,” Internet Engineering Task Force, Request for Comments: 6189, Informational, IETF Trust, Apr. 2011, 102 pages. |
Search Report for British patent application GB1317121.0 dated Mar. 14, 2014, 3 pages. |
Search Report for British patent application GB1317122.8 dated Mar. 11, 2014, 3 pages. |
Berners-Lee, Tim, “Socially Aware Cloud Storage,” Notes on web design, Aug. 17, 2009, http://www.w3.org/DesignIssues/CloudStorage.html, 9 pages. |
Chandra, Ramesh et al., “Separating Web Applications from User Data Storage with BStore,” presented at the USENIX Conference on Web Application Development, Jun. 2010, Boston, Massachusettes, 13 pages. |
Davids, Carol et al., “SIP APIs for Voice and Video Communications on the Web,” presented at the International Conference on Principles, Systems and Applications of IP Telecommunications (IPTcomm), Aug. 2011, Chicago, Illinois, 7 pages. |
Geambasu, Roxana et al., “Organizing and Sharing Distributed Personal Web-Service Data,” presented at the International World Wide Web Conference, Apr. 21-25, 2008, Beijing, China, International World Wide Web Conference Committee, pp. 755-764. |
Hsu, F. et al., “Secure File System Services for Web 2.0 Applications,” presented at the ACM Cloud Computing Security Workshop, Nov. 13, 2009, Chicago, Illinois, Association for Computing Machinery, 7 pages. |
Joshi, R., “Data-Oriented Architecture: A Loosley Coupled Real-Time SOA,” Whitepaper, Aug. 2007, Real-Time Innovations, Inc., http://rtcgroup.com/whitepapers/files/RTI—DataOrientedArchitecture—WhitePaper.pdf, 54 pages. |
Vahdat, Amin et al., “WebFS: A Global Cache Coherent File System,” UC Berkeley, Dec. 1996, retrieved Sep. 16, 2014 from https://www.cs.duke.edu/˜vahdat/webfs/webfs.html, 12 pages. |
Non-Final Office Action for U.S. Appl. No. 14/037,440, dated Sep. 12, 2014, 15 pages. |
Number | Date | Country | |
---|---|---|---|
20150373057 A1 | Dec 2015 | US |