The present disclosure relates to recording media and more particularly to enabling quality-aware media recording in a large-scale, multi-tenant, elastic collaboration cloud service operation using a service function chain architecture.
Call recording is a key capability in contact center environments. Call recording helps businesses to identify service delivery quality improvement opportunities, comply with legal regulations and promote knowledge reuse for learning/training purposes. Currently, a significant portion of the install base uses on-premise contact center deployments with call recording provided by a dedicated set of Session Initiation Protocol (SIP) based recording servers. As enterprises and small to midsized businesses move towards subscription-based consumption models, service providers are now offering contact center as a service (CCaaS) by hosting multi-tenant, elastic, large-scale Unified Communications (UC) and contact center infrastructure in the cloud. These cloud deployments typically replicate the on-premise architecture and provision several virtual instances of session border controllers (SBCs), and recording servers to support the high call volume in the operator's network. This approach comes with processing inefficiencies such as individual SIP signaling session between Session Recording Client (SRC) and Session Recording Server (SRS) for each recorded session.
In order to describe the manner in which the above-recited and other advantages and features of the disclosure can be obtained, a more particular description of the principles briefly described above will be rendered by reference to specific embodiments thereof which are illustrated in the appended drawings. Understanding that these drawings depict only exemplary embodiments of the disclosure and are not therefore to be considered to be limiting of its scope, the principles herein are described and explained with additional specificity and detail through the use of the accompanying drawings in which:
Various embodiments of the disclosure are discussed in detail below. While specific implementations are discussed, it should be understood that this is done for illustration purposes only. A person skilled in the relevant art will recognize that other components and configurations may be used without parting from the spirit and scope of the disclosure.
There are several issues with the above described contact center as a service (CCaaS) which is hosted by a multi-tenant, elastic, large-scale Unified Communications (UC) and contact center infrastructure in the cloud. The first issue relates to scaling. The current architecture does not enable a plausible approach to scaling the call recording feature in a highly multitenant environment. The second issue relates to quality. Again, the current architecture does not provide a sufficient approach for maintaining good quality. The question is can the system dynamically switch to another service if there is a media quality issue. The architecture needs to be media quality aware. Further, there is an inability to take real-time action to prevent recording media quality degrade.
For example, in today's architecture a network can include a recording server and a separate control server. The SIP protocol can be used to set up the communication paths. In one example, assume a call comes into a company for customer support. The company received a call at a call controlling system. If the company desires to record the call, then a separate call would be established between the call controlling system of the company and the recording server. The SIP protocol would also likely be used to establish the communication (a separate call) between the call controlling system and the recording server. Thus, in the current scenario, the recording function is enabled through this set up of two separate calls between the various entities, the caller, the call controlling system of the company, and a separate recording server.
Additional features and advantages of the disclosure will be set forth in the description which follows, and in part will be obvious from the description, or can be learned by practice of the herein disclosed principles. The features and advantages of the disclosure can be realized and obtained by means of the instruments and combinations particularly pointed out in the appended claims. These and other features of the disclosure will become more fully apparent from the following description and appended claims, or can be learned by the practice of the principles set forth herein. The scheme proposed addresses the issues raised above by leveraging the Service Function Chaining (SFC) Architecture.
Application Ser. No. 14/989,132 (CPOL-998347), filed Jan. 6, 2016, and incorporated herein by reference, is entitled Network Service Header (NSH) Metadata-Based End-to-End Multimedia Session Identification and Multimedia Service Optimization. The '132 application describes the propagation of session-ID value as NSH metadata in the Service Function Chain (SFC) architecture. That disclosure suggests that session-ID metadata can be used to determine inclusion of Session Recording Server (SRS) Service Function (SF) in the Service Function Chain (SFC). The '132 application also proposes reporting of media flow statistics aggregated at the session-ID level. The concepts proposed here build on the foundation of the '132 application by also including a stream ID in the NSH metadata and also goes on to describe the end-to-end procedure used to realize real-time session recording service as a Network Service Function (NSF). The contents of application Ser. No. 14/989,132 are incorporated herein by reference.
In one aspect, an application can provide call recording capabilities. In one approach, the application does not leverage SIP signaling nor SFC to establish recording session. Instead it uses proprietary application APIs to invoke recording to a call, which is not at all an optimal approach to solve this problem for the industry (large clouds, multi-vendor deployments, etc.). The concept disclosed herein leverages and builds on the standards-based SIP Recording (SIPREC) architecture that was defined in the IETF and whose various architectural components are described here: https://datatracker.ietforg/wg/siprec/documents/, incorporated herein by reference. The solution described in this submission makes use of the Service function Chaining (SFC) architecture (defined in RFC 7665) and the metadata field provided by the Network Service Header (NSH) (defined in draft-ietf-sfc-nsh). This disclosure make use of these technologies to deliver a unique and novel solution to deliver real-time and quality-aware session recording for large scale cloud deployments.
This disclosure provides a novel method to enable quality-aware media recording in large scale, multi-tenant, elastic collaboration cloud service operation networks using service function chain architecture. The recording functionality is provided as a service by recording network service function (NSF) using the stream Ii) of the media stream passed as metadata in the network service header (NSH). Recording NSF also reports media statistics to a control plane and a real-time transport protocol (RTP) stream classifier, which is part of a media plane, enabling them to take real-time actions to mitigate or minimize quality impairments in the recorded media. The control plane takes care of connecting callers (from wherever they are calling from) and the call set up. The media plane includes the RTP stream classifier and the network service functions that perform various processing on the media streams in real time, such as recording. The media plane can manage audio, video, media sharing, and so forth. The architecture disclosed herein enables the control plane and the media plane to communicate in such ways as using an API as well as enabling the media plane to be broken up into individual components, logical functions or service functions.
An example method includes establishing a communication session between a first participant and a second participant, programming, via a control plane, an RTP stream classifier which is to process RTP packets associated with the communication session with classification logic. The method includes receiving a first packet at the stream classifier and, when the communication session requires recording, applying the classification logic at the stream classifier to route the first packet into a chosen service function path of a plurality of service function paths, wherein the chosen service function path includes a recording service function. The recording service function can report media quality data to the control plane. Based on the media quality data, the control plane can update the classification logic programmed in the RTP stream classifier to migrate the communication session to a new chosen service function path to yield updated classification logic. The RTP stream classifier then receives subsequent RTP packets of the same session and routes them, according to the updated classification logic, to the new chosen service function path.
Several advantages of the concepts disclosed herein include the ability to eliminate the need to have SIP signaling between the Session Recording Client (SRC) and the Session Recording Server (SRS) and to detect low quality recording conditions and take actions to prevent continued quality degradation. This is useful for environments that require zero or very minimal loss recording for regulatory compliance. While one aspect of the idea applies to applications using that SIP protocol, the concepts disclosed therein are broader and any particular discussion of the concepts disclosed herein should not be presumed to be in the context of SIP unless explicitly stated. Thus, any signaling protocol and/or any recording architecture could apply to the principles disclosed herein.
Further advantages can include enabling storage location flexibility in a multi-tenant environment. The storage location can be different for different tenants and can be dynamically changed by the control plane and increasing scalability by moving recording as a data plane functionality.
The present disclosure addresses the issues raised above. The disclosure provides a system, method and computer-readable storage device embodiments. First a general example system shall be disclosed in
To enable user interaction with the computing device 100, an input device 145 can represent any number of input mechanisms, such as a microphone for speech, a touch-sensitive screen for gesture or graphical input, keyboard, mouse, motion input, speech and so forth. An output device 135 can also be one or more of a number of output mechanisms known to those of skill in the art. In some instances, multimodal systems can enable a user to provide multiple types of input to communicate with the computing device 100. The communications interface 140 can generally govern and manage the user input and system output. There is no restriction on operating on any particular hardware arrangement and therefore the basic features here may easily be substituted for improved hardware or firmware arrangements as they are developed.
Storage device 130 is a non-volatile memory and can be a hard disk or other types of computer readable media which can store data that are accessible by a computer, such as magnetic cassettes, flash memory cards, solid state memory devices, digital versatile disks, cartridges, random access memories (RAMs) 125, read only memory (ROM) 120, and hybrids thereof.
The storage device 130 can include software modules 132, 134, 136 for controlling the processor 110. Other hardware or software modules are contemplated. The storage device 130 can be connected to the system bus 105. In one aspect, a hardware module that performs a particular function can include the software component stored in a computer-readable medium in connection with the necessary hardware components, such as the processor 110, bus 105, display 135, and so forth, to carry out the function.
Example functionality for the individual service functions 223, 224, 226, 228 could include converting from RTP to SRTP, a transcoding component such as changing the payload from one codec to another, social media functions, and so forth. In one aspect, individual service functions 223, 224, 226, 228 can also provide unidirectional or bidirectional communication with the control plane 204. For example, information about the respective processing performed by these individual service functions can affect the quality of recording the session. Accordingly, the control plane may not only receive quality related data from the recording function 222, but can receive other data which can directly indicate packet loss or other degradation of quality directly or could include tangential information from which it could be inferred that a loss of quality or a reduction in quality may or is being experienced with respect to the recording. Such information can be utilized independently or in combination with quality information for the recording function 222 to arrive at decisions on whether to update the classification logic for the RTP stream classifier 208 and if so, how to revise the logic so as to cause a migration of a stream from one service function path to another.
In another example, the SF1223 in SFP1 (214) can add metadata to a packet or packet header. The added metadata can be passed on by SF2 (224) to the recording module 222 which can be processed or passed on to the control plane 204 to increase the understanding of potential packet loss or other issues which can affect the quality of a recording. In one aspect, SF2 (224) can add to the metadata, and perform a function based on the metadata, remove the metadata, or modify the metadata. The broader point is that with respect to providing quality aware recording, other service functions in the service function path 214 can perform specific analyses or processing with respect to their functionality which can impact the quality of recording. This information can be generated, processed and passed on such that it can ultimately be utilized in potentially updating classification logic transferred to the RTP stream classifier 208 for making classification decisions. Ultimately, any one or more of the following pieces of information can be utilized to provide quality-aware recording: the source port and/or address, the destination port and/or address, what information is in the RTP packet, how the classification rule is applied, external information about performance or triggering events, user profile information, service level agreement data, social media data associated with participants or the communication, and so forth.
The RTP stream classifier 208 processes incoming RTP packets 206 and determines, based on analysis of the packets, a chosen service function path for processing or managing the packets. The optional paths are logical chains of service functions as a shown in
The system could also use alternate nodes or entities beyond the structure shown in
Examples of sessions established by the control plane 204 can include internal calls between endpoints registered to the same cloud provider, inbound calls from external SIP service provider to an endpoint registered to the cloud and outbound calls from an endpoint registered within the cloud to an external public switched telephone network (PSTN) caller. A user or participant can also be a device such as a speech processing system.
For sessions that need to be recorded, the control plane 204 sets up the signaling session such that RTP streams flow through its media plane. For sessions that do not require additional media processing, a direct media connection is established between the source and destination endpoints/devices or a Service function path that doesn't have the Recording service function is utilized.
The following describe the method to deliver recording as a service. First, the control plane 204 maintains information 202 for all active media streams and exposes this information to the Network Service Function via an API interface. This information can include: 1. Flow details, such as one or more of the following parameters: {Src IP, Src port, Dest IP, Dest port, Protocol}; 2. Stream ID, which is a unique identifier for the media flow; 3. Session ID, which is the session ID of the communication session to which the media flow belongs to; 4. A boolean parameter indicating whether the media flow needs to be recorded or not; 5. Parent Tenant IDs, which can be a list of tenants to which the media stream is associated with (origination and destination tenants of a media stream may be different); 6. A recording tenant ID, which is a subset of parent tenant IDs that require this stream to be recorded; 7. A recording profile parameters (recording format); and/or 8. A secure real-time transport protocol (SRTP) key parameters required to decrypt the media flow.
The session ID is something specific to signaling that allows the system to track a call end-to-end even if there are different call legs that are set up. Assume that a user calls into a contact center, the call may be processed by several different products. From an end-user perspective, all of the call legs which are established to connect to the various products appear to be one single call. Previously, one would have to manually collate all of these call legs. Now, there is a session ID, there are some unique IDs that are used, that allow the call to be tracked end-to-end, even if there are multiple call legs. The system can track each product that the call is connected to. This is a unique identifier for a call. The session ID is a particular identifier configured for a particular protocol, such as a SIP signaling protocol, and identifies the session from an end-user perspective. However, other similar functionality could be built into a different version of a session ID or whatever protocol may be used. Thus, an identifier could be used to identify a session, independent of a specific protocol that is used to establish, maintain and manage that session.
The stream ID relates to an individual media stream. The stream ID can be structured in a tuple with a source address of the destination address as well as other information. It can maintain who was sending what, whether the communication is bidirectional, and so forth. It maintains a high level metadata table.
The control plane 204 uses one or more pieces of the above information to determine the chain of Network Service Functions, or the Service Function Path (SFP), for each media flow. For example, the control plane 204 can include the Recording 222 Network Service Function in one service function path SFP1 (214) if the media flow needs to be recorded. Once the path 214, 218 is determined, the system configures the RTP stream classifier 208 at the media plane with {Src IP, Src port, Dest IP, Dest port, Protocol} Service Function Path (SFP) mapping logic. There may be more than one SFP for a given media flow in which case each SFP is associated with a priority or an identification of which path should be used for which portion of a flow. Note in
Next is described the life of a RTP packet 206. After the communication session is established by the control plane 204, the control plane 204 programs the RTP Stream classifier 208 with {Src IP, Src port, Dest IP, Dest port, Protocol} or a Service Function Path (SFP) mapping. In the example of
The RTP stream classifier 208 is the entry point for all RTP packets 206 handled by the media plane 200. The classifier 208 looks up {Src IP, Src port, Dest IP, Dest port, Protocol}—SFP mapping table 210 and selects the mapped SFP with the highest priority or other parameter for selecting a mapped path. The RTP stream classifier 208 forms a Network Service Header 212 with the selected SFP ID along with stream ID as metadata. The NSH header 212 and RTP packet 206 is encapsulated in an outer header (GRE for example) and sent to the Service Function chain SFP1 (214) as a NSH packet. Other formats could be used as well in terms of the protocol or structure of the packet besides the NSH structure. If the media needs to be recorded, then the classifier 208 will choose a service function path that includes the service function recording feature or module. The logic to perform this functionality is provided by the control plane 204 to the RTP stream classifier 208. Thus, to enable the functionality of the classifier 208 being able to route a particular stream to the appropriate service function path 214, 218, the system needs two pieces of information. The first piece of information is the stream ID and the next piece of information is the recorder profile. As the packet is received, the classifier 208 adds an NSH header which includes the stream ID and the recorder profile.
Assume that the NSH packet is provided to the service function path 214, which includes the recording feature 222. The NSH packet traverses the SFs 223, 224 in the Service Function chain 214. The packet reaches the recording service function 222 which uses stream ID value from the NSH metadata as a key to look up a list of active recording buffers. If a buffer is found, the RTP packet is copied to the buffer. If lookup is not successful, it creates a new recording buffer in the list with stream ID as the key and adds the RTP packet to the buffer. Note the buffers can be stored in a cloud storage 220 where the recorded data is stored encrypted. After a period of time of buffering packets in the buffer, the recording service function 222 begins to transmit the files to cloud storage 220. For example, after one minute of buffering a media stream, the recording service function 222 could begin to transmit the recording content to the cloud storage 220. The approach disclosed in application Ser. No. 14/643,802, incorporated herein by reference, can be used to store the data in cloud in an encrypted manner.
The recording service function 222 uses an application programming interface (API) with the control plane 204 to retrieve a tenant ID and/or a recording profile information for the given media stream. Thus, there can be bidirectional communication between the recording service function 222 and the control plane 204. The recording profile can also be included as part of the NSH 212. The recording profile can include such information as storage logic, a chosen codec, a preferred codec, and/or a list of prioritized codecs. The recording service function 222 converts the RTP packets in the media stream's local recording buffer into a recording file as per the format specified in the recording profile. The recording service function 222 then uploads the contents of the recording file to the storage location (Box, Google etc.) 220 of each Tenant that has requested recording. The conversion and upload task can happen as the service function receives RTP packets or the service function 222 can store it in a temporary buffer and do it periodically (e.g. 1 min). If the recording file for this stream already exists in the cloud, the new media section may be appended to the existing file. Alternatively the recording SF 222 can create multiple files but index them using the stream ID as a key. All the related files of a given stream stored in cloud would have same stream ID key and be therefore connected or chained together. A search based on stream ID would return the list of files.
In another aspect, the NSH can include a tenant ID. The tenant ID could identify which cloud storage provider the system should be utilized for recording that stream. The tenant ID can be identified in the NSH, or communicated from the control plane 204 to the recording service function 222. For example, if the tenant ID identifies Google, Box, or Amazon, and so forth, as a particular cloud storage provider, the recording service function 222 can establish the communication with the chosen provider for storing the data associated with the stream ID. This approach allows the system to scale. In the existing architecture, the call controller and the recording element must have a SIP call communication established which can be limiting in terms of scale. The improvement disclosed herein is that there is a logical component (which may need some API interfaces) that is processing on the media and thus, from a scaling perspective, the new approach can make scaling much easier.
Further, how does one make the system intelligent enough such that it knows when something is happening with respect to the quality of the media. There are some scenarios where there is a requirement to have zero loss in the recording. If the network is experiencing some loss, then the system should be dynamic enough to pull in other resources to provide for zero loss. In the existing architecture, the control plane is not aware of such loss. Each of these service functions can provide feedback back to the control plane 204. The feedback represented from the recording function 222 and the control plane 204 can also represent feedback from the other functions 223, 224 or the service function path 214. It is noted as well that the communication can be bidirectional between the control plane and/or the service function path 214 or other individual service functions. In other words, data, feedback, or any other communication can be exchanged between the control plane 204 and the other components. The recording service function 222 can also maintain statistics of the success of receiving packets, or details about jitter, and so forth. The recording service function 222 can report that back to the control plane 204 and dynamically report that a particular SFP 214 is experiencing problems. In this case, the system can dynamically change the classification logic in the RTP stream classifier 208 as directed from the control plane 204 so that the stream is no longer going to go to SFP 1 (214), but is rerouted to go to a new SFP 11 (not shown). The SFP 11 (now shown) could be in a particular cluster of processing elements that is geographically separate or in improved position. Perhaps the processing is being moved closer to the user's location. The new cluster of service functions could be performing at a higher performance level than the initial path. The position of the new service functions could also provide improved bandwidth, improved hardware processing, and so forth. Any improvement parameter could be the basis upon which the control plane 204 provides updated classification logic to the RTP stream classifier 208 in order to migrate a particular stream from an underperforming service function path 214 to a new service function path, which can include the same or equivalent service functions as required by the stream, but which can provide the necessary quality for recording the media associated with the stream.
Included herein would be all of the necessary control elements to pause a recording from a first recording service function 222 and reengage or reestablish the recording from a new SFP 11. Utilizing the various data about the stream, such as the stream ID, the recorder profile, the tenant ID, and so forth, and enable the system to migrate from one service function path to another. The storage provider 220, for example, may be in the middle of recording a certain media stream when the control plane 204 migrates the recording from one service path to another. The storage provider 220 can identify a location at which the stream is paused or stopped and subsequently received from the new recording service function in SFP 11, in order to either start a new file for the recording or to expand and continue recording using the existing file from the first SFP.
There is no perceived change from the user standpoint, but there is a change in service function path to which a stream is assigned. In one example, assume a call is being set up at 10 MB per second, and the recording of the call is very important. If the system experiences packet loss which can affect the quality of the recording, then the system can take some action to lower the bandwidth, or increase the bandwidth, or some other action.
In addition to forking RTP and storing in the cloud, the recording service function 222 can use the control plane's API interface to retrieve metadata required and store them along with the recorded data in the cloud. The metadata may be indexed using SIP session ID as key information. Alternatively, some other identifier (like conference ID, participant ID etc. which can be shared with playback application) can be used as an index key to store the metadata. The metadata can follow the format defined in draft-ietf-siprec-metadata. If the SIPREC metadata is used, the associated stream ID is stored as metadata's<stream>XML element's UUID.
After processing the NSH packet, the recording SF 222 passes it to next SF in the chain 214. The last SF in the chain removes the actual RTP packet from the encapsulated packet (GRE) and sends it to next Layer 3 network hop.
Next is discussed the concept of quality aware recording. The recording SF 222 also computes and maintains media quality statistics for each active recording session. It periodically reports media quality statistics information to the control plane 204. The recording service function 222 could also report media quality data directly to the stream classifier 208 and/or any other node. The reporting of the data enables the control plane 204 to be recording-quality aware and invoke specific actions such as triggering a signaling update to negotiate a low bandwidth codec for the recording stream.
The recording SF 222 could also send the recording quality information to the RTP stream classifier 208 which can invoke specific actions such as choosing the next preferred SFP if the current recording SF 222 is experiencing poor media quality
There are several ways to stop a recording session: (i) The recording SF 222 can detect end of stream by means of having a configurable media inactivity timer. If no RTP packets are received within the expiration time, the stream is considered inactive and the recording session is stopped; and (ii) Alternatively, the control plane 204, on receiving a disconnect for a communication session, can request the recording service function 222 to stop the recording session by passing the corresponding stream IDs.
When the recording is stopped, the remaining RTP packets of the recording buffer are appended to the existing recording file in the storage location.
Another aspect of this disclosure is how to retrieve recordings from the storage provider 220. The stream ID can be used as a key for identifying the stream for playback. A playback application 230 can use the session ID and stream ID to fetch the list of recording files from the cloud 220 for a given communication session. If the streams are stored encrypted in the cloud, the mechanism mentioned in application Ser. No. 14/643,802 (CPOL-995420), filed Mar. 10, 2015, entitled Recording Encrypted Media Session, can be used by playback application to decrypt the files and play the recordings. The contents of application Ser. No. 14/643,802 are incorporated herein by reference
A participant in the communication session can utilize a site or an application on a device such as a desktop computer or a mobile device, and later retrieve the recorded file associated with the communication session from storage 220. The device 230 can store the session ID and stream ID or can access over a network, a database (not shown) of such information. The database could communicate with the control plane 204 and/or the cloud storage entity to 220 to obtain such information. Furthermore, the database can organize and index the session ID and stream ID such that a user via device 230 could provide information such as the other participant in the communication, the date of the communication, or other identifying information which can be used to retrieve a session ID and/or stream ID which were provided to the storage entity for retrieving the recording.
As recordings between individuals may be desired to be kept private, more secure access can be provided. For example, a password can be provided, or information about the other participants in the conversation can be requested. The system can request identification of the date or time of the recording of the communication or when the communication session occurred, and so forth.
In one aspect, assume that the buffer associated with recording module 222 is one minute long. Accordingly, one minute groupings of packets are transmitted to the storage facility 220. The groupings can be organized into a single file in the storage facility 220 which has an associated stream ID that is a key for later access. The system could also record individual one minute long files in the storage facility 220 for later retrieval. These files could be identified by the stream ID plus a chronological ID that the user could identify the 14th minute of recording within a particular stream ID. Automatic speech recognition or other speech processing could also be applied to the audio such that content-based searching could also be performed. The stream ID is used to retrieve all associated files. The session ID could also be used as well. For example, if a user called a call center and talked with several different individuals, there may be multiple stream IDs associated with the overall experience. The session ID can encompass multiple stream IDs and be associated with the overall session.
It is also noted that, as a shown in
One aspect of this disclosure relates to social media.
At this stage, the communication would be handled and managed as set forth above with respect to providing quality aware recording in a storage entity 220.
Following the end of the audible communication, either of user 302 or user 306, or only one of the users 302 and 306 under certain circumstances, may retrieve the recording in the storage entity 220. For example, the retrieval method can be implemented via the social media communication interface. An object could be presented via the social network 304 which a user can interact with to retrieve the recording.
In other scenarios, say when multiple users are commenting on a news article, individual users can engage in a conversation amidst other users. In such a scenario, the interface can provide the ability of one user to request a call or a new mode of communication with another user. For example, if Mary, John and Jane each provide a comment about a news article or posting, and Joe responds to Jane's comment, and Jane in turn response to Joe, a mechanism is presented which would enable Jane and/or Joe to request a telephone call with the other person. Once the participants are identified, the available means of vindication established and confirmed, and a communication session then initiated, the flow can continue as described above with respect to which service function path(s) is/are utilized to achieve the particular functionality desired by the users, including recording the call.
Other service functions in a path 214, 218 could also be implemented based on the particular functionality of a social media network. For example, if two users engage in a conversation initiated from Facebook or Instagram, a path could be chosen which includes additional functionality which is presented to the users. For example, relevant pictures to the conversation could be coordinated and presented to the user for posting on the social media network.
Also shown in
In another aspect, the button 402 could be used to initiate the recording of a call that is going to take place. For example, the button 402 could appear based on an analysis of the language of the text session (shown in
Of course while the example set forth above primarily discuss recording and audible medication such as a call, other forms of communication can also be recorded in a similar manner. Facetime video calls, Skype video calls, screen interactions, virtual reality experiences, display images, could all be recorded depending on the appropriate context of a communication between two people. Furthermore, interactions between humans and machines such as through speech recognition systems or interactions with avatars could also be recorded in a similar manner.
Of the various embodiments disclosed herein, the functionality that can be claimed by way of example can be addressed from different viewpoints. For example, the functionality performed by the control plane tool for can provide one claim set. Another claim could be focused on the functionality from the media plane, which includes the RTP stream classifier 208 and the various network service functions 214, 218 that perform the various processing on the media streams in real time. One embodiment could be from the standpoint of the storage provider 220. The various signals that are transmitted and received in the functionality performed by each of the separate entities can be separately claimed. All of the functionality that would be necessary to perform any such steps from the respective standpoint can be included within this disclosure even if not expressly described. For example,
When the communication session requires recording, the method includes applying the classification logic at the stream classifier to route the first packet into a chosen service function path of a plurality of service function paths, wherein the chosen service function path comprises a recording service function (508), and reporting media quality data from the recording service function in the chosen service function path to the control plane (510). Based on the media quality data, the method can include updating the classification logic programmed in the stream classifier by the control plane to migrate the communication session to a new chosen service function path to yield updated classification logic (512) receiving subsequent packets at the stream classifier and routing them, according to the updated classification logic, to the new chosen service function path (514).
The method can further include forming a network service header identifying the chosen service function path and a stream ID associated with the communication session. This step can be performed at the RTP stream classifier 208 or in another module. In one aspect, the first packet and the second packet are real-time transport protocol packets. The structure of the packets can also be other than RTP packets. The recording service function can retrieve a tenant ID that identifies a recording service provider 220 to record the communication session. The tenant ID can be retrieved from the control plane via an application programming interface or included in a network service header. It is noted that any two components communicating herein can communicate via an API. For example, the control plane, the media plane, a social media entity, an application on a mobile device that is utilized to access the recording, a call center accessing a recording, or any other entity that receives or transmits data in order to implement qualities aware recording using service function changes disclosed herein can communicate via a customized API with any other entity.
Routing the first packet to the chosen service function path can include routing the first packet with a recording profile that identifies at least one parameter associated with recording the communication session. The recording service function can retrieve data required to store a recording of the communication session. Such data can include at least one of a conference ID, a participant ID, a session ID, a stream ID, a date, a time, social media data, external data, content associated with the communication session, security information, policy information associated with retrieving the recording of the communication session, and a document ID for a document associated with the communication session.
In another aspect, the media quality data further can include data received from a service function in the chosen service function path, wherein the service function differs from the recording service function.
The method further includes transmitting a request to the requesting party to provide verification information (602). The verification information may include, but is not limited to, information corresponding to the date and time of the audio session, a user's password for logging into the social network, one or more personalized security questions, a fingerprint, etc. In one example, the request for verification information may be provided as a pop up screen to the requesting party on a screen of the requesting party's device. In another aspect, the system may be able to identify a sufficient level of certainty, a particular recording session and not need specific verification information. For example, the two parties communicating in a social media network may only have a single recording session and simply making the request by one party is sufficient to identify the session.
Upon receiving the requesting party's response to the verification information, the social network server processes the received verification information to determine if the requesting party is authorized to retrieve the audio session (604). The social network server determines whether the processed information at 604 indicates that the requesting party is authorized to access the recorded audio session or not (606). If the social network server determines that the requesting party is authorized, the social network server retrieves the recorded audio session from the storage 220, as described above with reference to
In one example, there may be more than one recorded audio session between users 302 and 306. Accordingly, after receiving the request at 600 and/or concurrently with requesting the verification information at 602, the social network server provides the requesting party with the option of choosing one of the available audio sessions (e.g., by presenting a drop down menu on the requesting party's device).
Referring back to 606, if the social network server determines that the requesting party is not authorized, the social network server determines whether a number of times the requesting party has provided invalid/insufficient verification information is equal to or greater than a threshold or not (612). The threshold may be an adjustable parameter (e.g., set to 3) that limits a number of attempts by the requesting party (possibly an unauthorized requesting party) to access the recorded audio session, for security purposes.
If the social network server determines that the number of attempts equals or exceeds the threshold, the social network server denies the requesting party access to the recorded audio session by displaying a pertinent message on a screen/display of the requesting party's device (614). In one example, the requesting party may then have to wait for a period of time (e.g., a few hours, 24 hours, etc.) before attempting to retrieve the recorded audio session. Alternatively, the requesting party may be asked to log out of the social network service and log back in, in order to attempt retrieving the recorded audio session.
If the social network server determines, at 612, that the number of attempts is less than the threshold, the social network server reverts back to 602 and re-transmits the request for verification information to the requesting party. In one example, the social network server may alter the type of verification information requested from the requesting party. For example, if on the first attempt, the requesting party is asked to provide his or her password for the social network and the provided password is incorrect, then on the second attempt, the social network server may request a different kind of verification information to be furnished by the requesting party (e.g., one or more security questions such as date of birth, name of your first pet, etc.). Thereafter, the social network server repeats the processes 604 to 614, as appropriate.
In one aspect, any of the functionality described above with respect to approaches to initiating a recording between two users of the social network 304 or retrieving a stored recording between two individuals can be performed by one or more of the social network 304, the control plane tool for, and stores 220. The overall concept relates to how to connect users of the social network into the system of
In some embodiments the computer-readable storage devices, mediums, and memories can include a cable or wireless signal containing a bit stream and the like. However, when mentioned, non-transitory computer-readable storage media expressly exclude media such as energy, carrier signals, electromagnetic waves, and signals per se.
Methods according to the above-described examples can be implemented using computer-executable instructions that are stored or otherwise available from computer readable media. Such instructions can comprise, for example, instructions and data which cause or otherwise configure a general purpose computer, special purpose computer, or special purpose processing device to perform a certain function or group of functions. Portions of computer resources used can be accessible over a network. The computer executable instructions may be, for example, binaries, intermediate format instructions such as assembly language, firmware, or source code. Examples of computer-readable media that may be used to store instructions, information used, and/or information created during methods according to described examples include magnetic or optical disks, flash memory, USB devices provided with non-volatile memory, networked storage devices, and so on.
Devices implementing methods according to these disclosures can comprise hardware, firmware and/or software, and can take any of a variety of form factors. Typical examples of such form factors include laptops, smart phones, small form factor personal computers, personal digital assistants, rack mount devices, standalone devices, and so on. Functionality described herein also can be embodied in peripherals or add-in cards. Such functionality can also be implemented on a circuit board among different chips or different processes executing in a single device, by way of further example.
The instructions, media for conveying such instructions, computing resources for executing them, and other structures for supporting such computing resources are means for providing the functions described in these disclosures.
Although a variety of examples and other information was used to explain aspects within the scope of the appended claims, no limitation of the claims should be implied based on particular features or arrangements in such examples, as one of ordinary skill would be able to use these examples to derive a wide variety of implementations. Further and although some subject matter may have been described in language specific to examples of structural features and/or method steps, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to these described features or acts. For example, such functionality can be distributed differently or performed in components other than those identified herein. Rather, the described features and steps are disclosed as examples of components of systems and methods within the scope of the appended claims. Moreover, claim language reciting “at least one of” a set indicates that one member of the set or multiple members of the set satisfy the claim.
It should be understood that features or configurations herein with reference to one embodiment or example can be implemented in, or combined with, other embodiments or examples herein. That is, terms such as “embodiment”, “variation”, “aspect”, “example”, “configuration”, “implementation”, “case”, and any other terms which may connote an embodiment, as used herein to describe specific features or configurations, are not intended to limit any of the associated features or configurations to a specific or separate embodiment or embodiments, and should not be interpreted to suggest that such features or configurations cannot be combined with features or configurations described with reference to other embodiments, variations, aspects, examples, configurations, implementations, cases, and so forth. In other words, features described herein with reference to a specific example (e.g., embodiment, variation, aspect, configuration, implementation, case, etc.) can be combined with features described with reference to another example. Precisely, one of ordinary skill in the art will readily recognize that the various embodiments or examples described herein, and their associated features, can be combined with each other.
A phrase such as an “aspect” does not imply that such aspect is essential to the subject technology or that such aspect applies to all configurations of the subject technology. A disclosure relating to an aspect may apply to all configurations, or one or more configurations. A phrase such as an aspect may refer to one or more aspects and vice versa. A phrase such as a “configuration” does not imply that such configuration is essential to the subject technology or that such configuration applies to all configurations of the subject technology. A disclosure relating to a configuration may apply to all configurations, or one or more configurations. A phrase such as a configuration may refer to one or more configurations and vice versa. The word “exemplary” is used herein to mean “serving as an example or illustration.” Any aspect or design described herein as “exemplary” is not necessarily to be construed as preferred or advantageous over other aspects or designs.
Moreover, claim language reciting “at least one of” a set indicates that one member of the set or multiple members of the set satisfy the claim. For example, claim language reciting “at least one of A, B, and C” or “at least one of A, B, or C” means A alone, B alone, C alone, A and B together, A and C together, B and C together, or A, B and C together.
Number | Name | Date | Kind |
---|---|---|---|
5812773 | Norin | Sep 1998 | A |
5889896 | Meshinsky et al. | Mar 1999 | A |
6108782 | Fletcher et al. | Aug 2000 | A |
6178453 | Mattaway et al. | Jan 2001 | B1 |
6298153 | Oishi | Oct 2001 | B1 |
6343290 | Cossins et al. | Jan 2002 | B1 |
6587463 | Hebb | Jul 2003 | B1 |
6643260 | Kloth et al. | Nov 2003 | B1 |
6683873 | Kwok et al. | Jan 2004 | B1 |
6721804 | Rubin et al. | Apr 2004 | B1 |
6733449 | Krishnamurthy et al. | May 2004 | B1 |
6735631 | Oehrke et al. | May 2004 | B1 |
6885670 | Regula | Apr 2005 | B1 |
6996615 | McGuire | Feb 2006 | B1 |
7054930 | Cheriton | May 2006 | B1 |
7058706 | Lyer et al. | Jun 2006 | B1 |
7062571 | Dale et al. | Jun 2006 | B1 |
7076397 | Ding et al. | Jul 2006 | B2 |
7111177 | Chauvel et al. | Sep 2006 | B1 |
7212490 | Kao et al. | May 2007 | B1 |
7277948 | Igarashi et al. | Oct 2007 | B2 |
7313667 | Pullela et al. | Dec 2007 | B1 |
7379846 | Williams et al. | May 2008 | B1 |
7480672 | Hahn et al. | Jan 2009 | B2 |
7496043 | Leong et al. | Feb 2009 | B1 |
7536476 | Alleyne | May 2009 | B1 |
7567504 | Darling et al. | Jul 2009 | B2 |
7606147 | Luft et al. | Oct 2009 | B2 |
7647594 | Togawa | Jan 2010 | B2 |
7684322 | Sand et al. | Mar 2010 | B2 |
7773510 | Back et al. | Aug 2010 | B2 |
7808897 | Mehta et al. | Oct 2010 | B1 |
7881957 | Cohen et al. | Feb 2011 | B1 |
7917647 | Cooper et al. | Mar 2011 | B2 |
8010598 | Tanimoto | Aug 2011 | B2 |
8028071 | Mahalingam et al. | Sep 2011 | B1 |
8040801 | Lin | Oct 2011 | B2 |
8041714 | Aymeloglu et al. | Oct 2011 | B2 |
8121117 | Amdahl et al. | Feb 2012 | B1 |
8171415 | Appleyard et al. | May 2012 | B2 |
8234377 | Cohn | Jul 2012 | B2 |
8244559 | Horvitz et al. | Aug 2012 | B2 |
8250215 | Stienhans et al. | Aug 2012 | B2 |
8280880 | Aymeloglu et al. | Oct 2012 | B1 |
8284664 | Aybay et al. | Oct 2012 | B1 |
8284776 | Petersen | Oct 2012 | B2 |
8301746 | Head et al. | Oct 2012 | B2 |
8345692 | Smith | Jan 2013 | B2 |
8406141 | Couturier et al. | Mar 2013 | B1 |
8407413 | Yucel et al. | Mar 2013 | B1 |
8448171 | Donnellan et al. | May 2013 | B2 |
8477610 | Zuo et al. | Jul 2013 | B2 |
8495252 | Lais et al. | Jul 2013 | B2 |
8495356 | Ashok et al. | Jul 2013 | B2 |
8510469 | Portolani | Aug 2013 | B2 |
8514868 | Hill | Aug 2013 | B2 |
8532108 | Li et al. | Sep 2013 | B2 |
8533687 | Greifeneder et al. | Sep 2013 | B1 |
8547974 | Guruswamy et al. | Oct 2013 | B1 |
8560639 | Murphy et al. | Oct 2013 | B2 |
8560663 | Baucke et al. | Oct 2013 | B2 |
8589543 | Dutta et al. | Nov 2013 | B2 |
8590050 | Nagpal et al. | Nov 2013 | B2 |
8611356 | Yu et al. | Dec 2013 | B2 |
8612625 | Andreis et al. | Dec 2013 | B2 |
8630291 | Shaffer et al. | Jan 2014 | B2 |
8639787 | Lagergren et al. | Jan 2014 | B2 |
8656024 | Krishnan et al. | Feb 2014 | B2 |
8660129 | Brendel et al. | Feb 2014 | B1 |
8719804 | Jain | May 2014 | B2 |
8775576 | Hebert et al. | Jul 2014 | B2 |
8797867 | Chen et al. | Aug 2014 | B1 |
8805951 | Faibish et al. | Aug 2014 | B1 |
8850182 | Fritz et al. | Sep 2014 | B1 |
8856339 | Mestery et al. | Oct 2014 | B2 |
8909780 | Dickinson et al. | Dec 2014 | B1 |
8909928 | Ahmad et al. | Dec 2014 | B2 |
8918510 | Gmach et al. | Dec 2014 | B2 |
8924720 | Raghuram et al. | Dec 2014 | B2 |
8930747 | Levijarvi et al. | Jan 2015 | B2 |
8938775 | Roth et al. | Jan 2015 | B1 |
8959526 | Kansal et al. | Feb 2015 | B2 |
8977754 | Curry, Jr. et al. | Mar 2015 | B2 |
9009697 | Breiter et al. | Apr 2015 | B2 |
9015324 | Jackson | Apr 2015 | B2 |
9043439 | Bicket et al. | May 2015 | B2 |
9049115 | Rajendran et al. | Jun 2015 | B2 |
9063789 | Beaty et al. | Jun 2015 | B2 |
9065727 | Liu et al. | Jun 2015 | B1 |
9075649 | Bushman et al. | Jul 2015 | B1 |
9104334 | Madhusudana et al. | Aug 2015 | B2 |
9164795 | Vincent | Oct 2015 | B1 |
9167050 | Durazzo et al. | Oct 2015 | B2 |
9201701 | Boldyrev et al. | Dec 2015 | B2 |
9201704 | Chang et al. | Dec 2015 | B2 |
9203784 | Chang et al. | Dec 2015 | B2 |
9223634 | Chang et al. | Dec 2015 | B2 |
9244776 | Koza et al. | Jan 2016 | B2 |
9251114 | Ancin et al. | Feb 2016 | B1 |
9264478 | Hon et al. | Feb 2016 | B2 |
9313048 | Chang et al. | Apr 2016 | B2 |
9361192 | Smith et al. | Jun 2016 | B2 |
9380075 | He et al. | Jun 2016 | B2 |
9432294 | Sharma et al. | Aug 2016 | B1 |
9444744 | Sharma et al. | Sep 2016 | B1 |
9473365 | Melander et al. | Oct 2016 | B2 |
9503530 | Niedzielski | Nov 2016 | B1 |
9558078 | Farlee et al. | Jan 2017 | B2 |
9613078 | Vermeulen et al. | Apr 2017 | B2 |
9628471 | Sundaram et al. | Apr 2017 | B1 |
9632858 | Sasturkar et al. | Apr 2017 | B2 |
9658876 | Chang et al. | May 2017 | B2 |
9692802 | Bicket et al. | Jun 2017 | B2 |
9727359 | Tsirkin | Aug 2017 | B2 |
9736063 | Wan et al. | Aug 2017 | B2 |
9755858 | Bagepalli et al. | Sep 2017 | B2 |
9792245 | Raghavan et al. | Oct 2017 | B2 |
9804988 | Ayoub et al. | Oct 2017 | B1 |
9954783 | Thirumurthi et al. | Apr 2018 | B1 |
20020004900 | Patel | Jan 2002 | A1 |
20020073337 | Ioele et al. | Jun 2002 | A1 |
20020143928 | Maltz et al. | Oct 2002 | A1 |
20020166117 | Abrams et al. | Nov 2002 | A1 |
20020174216 | Shorey et al. | Nov 2002 | A1 |
20030012147 | Buckman | Jan 2003 | A1 |
20030018591 | Komisky | Jan 2003 | A1 |
20030056001 | Mate et al. | Mar 2003 | A1 |
20030228585 | Inoko et al. | Dec 2003 | A1 |
20040004941 | Malan et al. | Jan 2004 | A1 |
20040095237 | Chen et al. | May 2004 | A1 |
20040131059 | Ayyakad et al. | Jul 2004 | A1 |
20040264481 | Darling et al. | Dec 2004 | A1 |
20050060418 | Sorokopud | Mar 2005 | A1 |
20050125424 | Herriott et al. | Jun 2005 | A1 |
20060059558 | Selep et al. | Mar 2006 | A1 |
20060104286 | Cheriton | May 2006 | A1 |
20060120575 | Ahn et al. | Jun 2006 | A1 |
20060126665 | Ward et al. | Jun 2006 | A1 |
20060146825 | Hofstaedter et al. | Jul 2006 | A1 |
20060155875 | Cheriton | Jul 2006 | A1 |
20060168338 | Bruegl et al. | Jul 2006 | A1 |
20060294207 | Barsness et al. | Dec 2006 | A1 |
20070011330 | Dinker et al. | Jan 2007 | A1 |
20070174663 | Crawford et al. | Jul 2007 | A1 |
20070223487 | Kajekar et al. | Sep 2007 | A1 |
20070242830 | Conrado et al. | Oct 2007 | A1 |
20080005293 | Bhargava et al. | Jan 2008 | A1 |
20080084880 | Dharwadkar | Apr 2008 | A1 |
20080165778 | Ertemalp | Jul 2008 | A1 |
20080198752 | Fan et al. | Aug 2008 | A1 |
20080201711 | Amir Husain | Aug 2008 | A1 |
20080235755 | Blaisdell et al. | Sep 2008 | A1 |
20090006527 | Gingell, Jr. et al. | Jan 2009 | A1 |
20090010277 | Halbraich | Jan 2009 | A1 |
20090019367 | Cavagnari et al. | Jan 2009 | A1 |
20090031312 | Mausolf et al. | Jan 2009 | A1 |
20090083183 | Rao et al. | Mar 2009 | A1 |
20090138763 | Arnold | May 2009 | A1 |
20090177775 | Radia et al. | Jul 2009 | A1 |
20090178058 | Stillwell, III et al. | Jul 2009 | A1 |
20090182874 | Morford et al. | Jul 2009 | A1 |
20090265468 | Annambhotla et al. | Oct 2009 | A1 |
20090265753 | Anderson et al. | Oct 2009 | A1 |
20090293056 | Ferris | Nov 2009 | A1 |
20090300608 | Ferris et al. | Dec 2009 | A1 |
20090313562 | Appleyard et al. | Dec 2009 | A1 |
20090323706 | Germain et al. | Dec 2009 | A1 |
20090328031 | Pouyadou et al. | Dec 2009 | A1 |
20100042720 | Stienhans et al. | Feb 2010 | A1 |
20100061250 | Nugent | Mar 2010 | A1 |
20100115341 | Baker et al. | May 2010 | A1 |
20100131765 | Bromley et al. | May 2010 | A1 |
20100191783 | Mason et al. | Jul 2010 | A1 |
20100192157 | Jackson et al. | Jul 2010 | A1 |
20100205601 | Abbas et al. | Aug 2010 | A1 |
20100211782 | Auradkar et al. | Aug 2010 | A1 |
20100217886 | Seren | Aug 2010 | A1 |
20100293270 | Augenstein et al. | Nov 2010 | A1 |
20100318609 | Lahiri et al. | Dec 2010 | A1 |
20100325199 | Park et al. | Dec 2010 | A1 |
20100325257 | Goel et al. | Dec 2010 | A1 |
20100325441 | Laurie et al. | Dec 2010 | A1 |
20100333116 | Prahlad et al. | Dec 2010 | A1 |
20110016214 | Jackson | Jan 2011 | A1 |
20110035754 | Srinivasan | Feb 2011 | A1 |
20110055396 | Dehaan | Mar 2011 | A1 |
20110055398 | Dehaan et al. | Mar 2011 | A1 |
20110055470 | Portolani | Mar 2011 | A1 |
20110072489 | Parann-Nissany | Mar 2011 | A1 |
20110075667 | Li et al. | Mar 2011 | A1 |
20110110382 | Jabr et al. | May 2011 | A1 |
20110116443 | Yu et al. | May 2011 | A1 |
20110126099 | Anderson et al. | May 2011 | A1 |
20110138055 | Daly et al. | Jun 2011 | A1 |
20110145413 | Dawson et al. | Jun 2011 | A1 |
20110145657 | Bishop et al. | Jun 2011 | A1 |
20110173303 | Rider | Jul 2011 | A1 |
20110185063 | Head et al. | Jul 2011 | A1 |
20110199902 | Leavy et al. | Aug 2011 | A1 |
20110213687 | Ferris et al. | Sep 2011 | A1 |
20110213966 | Fu et al. | Sep 2011 | A1 |
20110219434 | Betz et al. | Sep 2011 | A1 |
20110231715 | Kunii et al. | Sep 2011 | A1 |
20110231899 | Pulier et al. | Sep 2011 | A1 |
20110239039 | Dieffenbach et al. | Sep 2011 | A1 |
20110252327 | Awasthi et al. | Oct 2011 | A1 |
20110261811 | Battestilli et al. | Oct 2011 | A1 |
20110261828 | Smith | Oct 2011 | A1 |
20110276675 | Singh et al. | Nov 2011 | A1 |
20110276951 | Jain | Nov 2011 | A1 |
20110295998 | Ferris et al. | Dec 2011 | A1 |
20110305149 | Scott et al. | Dec 2011 | A1 |
20110307531 | Gaponenko et al. | Dec 2011 | A1 |
20110320870 | Kenigsberg et al. | Dec 2011 | A1 |
20120005724 | Lee | Jan 2012 | A1 |
20120023418 | Frields et al. | Jan 2012 | A1 |
20120054367 | Ramakrishnan et al. | Mar 2012 | A1 |
20120072318 | Akiyama et al. | Mar 2012 | A1 |
20120072578 | Alam | Mar 2012 | A1 |
20120072581 | Tung et al. | Mar 2012 | A1 |
20120072985 | Davne et al. | Mar 2012 | A1 |
20120072992 | Arasaratnam et al. | Mar 2012 | A1 |
20120084445 | Brock et al. | Apr 2012 | A1 |
20120084782 | Chou et al. | Apr 2012 | A1 |
20120096134 | Suit | Apr 2012 | A1 |
20120102193 | Rathore et al. | Apr 2012 | A1 |
20120102199 | Hopmann et al. | Apr 2012 | A1 |
20120131174 | Ferris et al. | May 2012 | A1 |
20120137215 | Kawara | May 2012 | A1 |
20120158967 | Sedayao et al. | Jun 2012 | A1 |
20120159097 | Jennas, II et al. | Jun 2012 | A1 |
20120166649 | Watanabe et al. | Jun 2012 | A1 |
20120167094 | Suit | Jun 2012 | A1 |
20120173541 | Venkataramani | Jul 2012 | A1 |
20120173710 | Rodriguez | Jul 2012 | A1 |
20120179909 | Sagi et al. | Jul 2012 | A1 |
20120180044 | Donnellan et al. | Jul 2012 | A1 |
20120182891 | Lee et al. | Jul 2012 | A1 |
20120185632 | Lais et al. | Jul 2012 | A1 |
20120185913 | Martinez et al. | Jul 2012 | A1 |
20120192016 | Gotesdyner et al. | Jul 2012 | A1 |
20120192075 | Ebtekar et al. | Jul 2012 | A1 |
20120201135 | Ding et al. | Aug 2012 | A1 |
20120203908 | Beaty et al. | Aug 2012 | A1 |
20120204169 | Breiter et al. | Aug 2012 | A1 |
20120204187 | Breiter et al. | Aug 2012 | A1 |
20120214506 | Skaaksrud et al. | Aug 2012 | A1 |
20120222106 | Kuehl | Aug 2012 | A1 |
20120236716 | Anbazhagan et al. | Sep 2012 | A1 |
20120240113 | Hur | Sep 2012 | A1 |
20120265976 | Spiers et al. | Oct 2012 | A1 |
20120272025 | Park et al. | Oct 2012 | A1 |
20120281706 | Agarwal et al. | Nov 2012 | A1 |
20120281708 | Chauhan et al. | Nov 2012 | A1 |
20120290647 | Ellison et al. | Nov 2012 | A1 |
20120297238 | Watson et al. | Nov 2012 | A1 |
20120311106 | Morgan | Dec 2012 | A1 |
20120311568 | Jansen | Dec 2012 | A1 |
20120324092 | Brown et al. | Dec 2012 | A1 |
20120324114 | Dutta et al. | Dec 2012 | A1 |
20130003567 | Gallant et al. | Jan 2013 | A1 |
20130013248 | Brugler et al. | Jan 2013 | A1 |
20130036213 | Hasan et al. | Feb 2013 | A1 |
20130044636 | Koponen et al. | Feb 2013 | A1 |
20130066940 | Shao | Mar 2013 | A1 |
20130069950 | Adam et al. | Mar 2013 | A1 |
20130080509 | Wang | Mar 2013 | A1 |
20130080624 | Nagai et al. | Mar 2013 | A1 |
20130091557 | Gurrapu | Apr 2013 | A1 |
20130097601 | Podvratnik et al. | Apr 2013 | A1 |
20130104140 | Meng et al. | Apr 2013 | A1 |
20130111540 | Sabin | May 2013 | A1 |
20130117337 | Dunham | May 2013 | A1 |
20130124712 | Parker | May 2013 | A1 |
20130125124 | Kempf et al. | May 2013 | A1 |
20130138816 | Kuo et al. | May 2013 | A1 |
20130144978 | Jain et al. | Jun 2013 | A1 |
20130152076 | Patel | Jun 2013 | A1 |
20130152175 | Hromoko et al. | Jun 2013 | A1 |
20130159097 | Schory et al. | Jun 2013 | A1 |
20130159496 | Hamilton et al. | Jun 2013 | A1 |
20130160008 | Cawlfield et al. | Jun 2013 | A1 |
20130162753 | Hendrickson et al. | Jun 2013 | A1 |
20130169666 | Pacheco et al. | Jul 2013 | A1 |
20130179941 | McGloin et al. | Jul 2013 | A1 |
20130182712 | Aguayo et al. | Jul 2013 | A1 |
20130185413 | Beaty et al. | Jul 2013 | A1 |
20130185433 | Zhu et al. | Jul 2013 | A1 |
20130191106 | Kephart et al. | Jul 2013 | A1 |
20130198050 | Shroff et al. | Aug 2013 | A1 |
20130198374 | Zalmanovitch et al. | Aug 2013 | A1 |
20130204849 | Chacko | Aug 2013 | A1 |
20130232491 | Radhakrishnan et al. | Sep 2013 | A1 |
20130232492 | Wang | Sep 2013 | A1 |
20130246588 | Borowicz et al. | Sep 2013 | A1 |
20130250770 | Zou et al. | Sep 2013 | A1 |
20130254415 | Fullen et al. | Sep 2013 | A1 |
20130262347 | Dodson | Oct 2013 | A1 |
20130283364 | Chang et al. | Oct 2013 | A1 |
20130297769 | Chang et al. | Nov 2013 | A1 |
20130318240 | Hebert et al. | Nov 2013 | A1 |
20130318546 | Kothuri et al. | Nov 2013 | A1 |
20130339693 | Bonanno | Dec 2013 | A1 |
20130339949 | Spiers et al. | Dec 2013 | A1 |
20140006481 | Frey et al. | Jan 2014 | A1 |
20140006535 | Reddy | Jan 2014 | A1 |
20140006585 | Dunbar et al. | Jan 2014 | A1 |
20140019639 | Ueno | Jan 2014 | A1 |
20140040473 | Ho et al. | Feb 2014 | A1 |
20140040883 | Tompkins | Feb 2014 | A1 |
20140052877 | Mao | Feb 2014 | A1 |
20140059310 | Du et al. | Feb 2014 | A1 |
20140074850 | Noel et al. | Mar 2014 | A1 |
20140075048 | Yuksel et al. | Mar 2014 | A1 |
20140075108 | Dong et al. | Mar 2014 | A1 |
20140075357 | Flores et al. | Mar 2014 | A1 |
20140075501 | Srinivasan et al. | Mar 2014 | A1 |
20140089727 | Cherkasova et al. | Mar 2014 | A1 |
20140098762 | Ghai et al. | Apr 2014 | A1 |
20140108985 | Scott et al. | Apr 2014 | A1 |
20140122560 | Ramey et al. | May 2014 | A1 |
20140136779 | Guha et al. | May 2014 | A1 |
20140140211 | Chandrasekaran et al. | May 2014 | A1 |
20140141720 | Princen et al. | May 2014 | A1 |
20140156557 | Zeng et al. | Jun 2014 | A1 |
20140160924 | Pfautz | Jun 2014 | A1 |
20140164486 | Ravichandran et al. | Jun 2014 | A1 |
20140188825 | Muthukkaruppan et al. | Jul 2014 | A1 |
20140189095 | Lindberg et al. | Jul 2014 | A1 |
20140189125 | Amies et al. | Jul 2014 | A1 |
20140215471 | Cherkasova | Jul 2014 | A1 |
20140222953 | Karve et al. | Aug 2014 | A1 |
20140244851 | Lee | Aug 2014 | A1 |
20140245298 | Zhou et al. | Aug 2014 | A1 |
20140269266 | Filsfils et al. | Sep 2014 | A1 |
20140269446 | Lum et al. | Sep 2014 | A1 |
20140280805 | Sawalha | Sep 2014 | A1 |
20140282536 | Dave et al. | Sep 2014 | A1 |
20140282611 | Campbell et al. | Sep 2014 | A1 |
20140282669 | McMillan | Sep 2014 | A1 |
20140282889 | Ishaya et al. | Sep 2014 | A1 |
20140289200 | Kato | Sep 2014 | A1 |
20140297569 | Clark et al. | Oct 2014 | A1 |
20140297835 | Buys | Oct 2014 | A1 |
20140314078 | Jilani | Oct 2014 | A1 |
20140317261 | Shatzkamer et al. | Oct 2014 | A1 |
20140366155 | Chang et al. | Dec 2014 | A1 |
20140372567 | Ganesh et al. | Dec 2014 | A1 |
20150006470 | Mohan | Jan 2015 | A1 |
20150033086 | Sasturkar et al. | Jan 2015 | A1 |
20150043335 | Testicioglu | Feb 2015 | A1 |
20150043576 | Dixon et al. | Feb 2015 | A1 |
20150052247 | Threefoot et al. | Feb 2015 | A1 |
20150052517 | Raghu et al. | Feb 2015 | A1 |
20150058382 | St. Laurent et al. | Feb 2015 | A1 |
20150058459 | Amendjian et al. | Feb 2015 | A1 |
20150058557 | Madhusudana et al. | Feb 2015 | A1 |
20150070516 | Shoemake et al. | Mar 2015 | A1 |
20150071285 | Kumar et al. | Mar 2015 | A1 |
20150089478 | Cheluvaraju et al. | Mar 2015 | A1 |
20150100471 | Curry, Jr. et al. | Apr 2015 | A1 |
20150106802 | Ivanov et al. | Apr 2015 | A1 |
20150106805 | Melander et al. | Apr 2015 | A1 |
20150109923 | Hwang | Apr 2015 | A1 |
20150117199 | Chinnaiah Sankaran et al. | Apr 2015 | A1 |
20150117458 | Gurkan et al. | Apr 2015 | A1 |
20150120914 | Wada et al. | Apr 2015 | A1 |
20150149828 | Mukerji et al. | May 2015 | A1 |
20150178133 | Phelan et al. | Jun 2015 | A1 |
20150215819 | Bosch et al. | Jul 2015 | A1 |
20150227405 | Jan et al. | Aug 2015 | A1 |
20150242204 | Hassine et al. | Aug 2015 | A1 |
20150249709 | Teng et al. | Sep 2015 | A1 |
20150271199 | Bradley et al. | Sep 2015 | A1 |
20150280980 | Bitar | Oct 2015 | A1 |
20150281067 | Wu | Oct 2015 | A1 |
20150281113 | Siciliano et al. | Oct 2015 | A1 |
20150309908 | Pearson et al. | Oct 2015 | A1 |
20150319063 | Zourzouvillys et al. | Nov 2015 | A1 |
20150326524 | Tankala et al. | Nov 2015 | A1 |
20150339210 | Kopp et al. | Nov 2015 | A1 |
20150373108 | Fleming et al. | Dec 2015 | A1 |
20150379062 | Vermeulen et al. | Dec 2015 | A1 |
20160011925 | Kulkarni et al. | Jan 2016 | A1 |
20160013990 | Kulkarni et al. | Jan 2016 | A1 |
20160062786 | Meng et al. | Mar 2016 | A1 |
20160065417 | Sapuram et al. | Mar 2016 | A1 |
20160094398 | Choudhury et al. | Mar 2016 | A1 |
20160094480 | Kulkarni et al. | Mar 2016 | A1 |
20160094643 | Jain et al. | Mar 2016 | A1 |
20160094894 | Inayatullah et al. | Mar 2016 | A1 |
20160099847 | Melander et al. | Apr 2016 | A1 |
20160099873 | Gerö et al. | Apr 2016 | A1 |
20160103838 | Sainani et al. | Apr 2016 | A1 |
20160105393 | Thakkar et al. | Apr 2016 | A1 |
20160127184 | Bursell | May 2016 | A1 |
20160134557 | Steinder et al. | May 2016 | A1 |
20160147676 | Cha et al. | May 2016 | A1 |
20160162436 | Raghavan et al. | Jun 2016 | A1 |
20160164914 | Madhav et al. | Jun 2016 | A1 |
20160188527 | Cherian et al. | Jun 2016 | A1 |
20160234071 | Nambiar et al. | Aug 2016 | A1 |
20160239399 | Babu et al. | Aug 2016 | A1 |
20160253078 | Ebtekar et al. | Sep 2016 | A1 |
20160254968 | Ebtekar et al. | Sep 2016 | A1 |
20160261564 | Foxhoven et al. | Sep 2016 | A1 |
20160277368 | Narayanaswamy et al. | Sep 2016 | A1 |
20160292027 | Moyer | Oct 2016 | A1 |
20160292611 | Boe et al. | Oct 2016 | A1 |
20160352682 | Chang | Dec 2016 | A1 |
20160378389 | Hrischuk et al. | Dec 2016 | A1 |
20170005948 | Melander et al. | Jan 2017 | A1 |
20170024260 | Chandrasekaran et al. | Jan 2017 | A1 |
20170026470 | Bhargava et al. | Jan 2017 | A1 |
20170034199 | Zaw | Feb 2017 | A1 |
20170041342 | Efremov et al. | Feb 2017 | A1 |
20170054659 | Ergin et al. | Feb 2017 | A1 |
20170063674 | Maskalik et al. | Mar 2017 | A1 |
20170097841 | Chang et al. | Apr 2017 | A1 |
20170099188 | Chang et al. | Apr 2017 | A1 |
20170104755 | Arregoces et al. | Apr 2017 | A1 |
20170126583 | Xia | May 2017 | A1 |
20170147297 | Krishnamurthy et al. | May 2017 | A1 |
20170163569 | Koganti | Jun 2017 | A1 |
20170171158 | Hoy et al. | Jun 2017 | A1 |
20170192823 | Karaje et al. | Jul 2017 | A1 |
20170264663 | Bicket et al. | Sep 2017 | A1 |
20170302521 | Lui et al. | Oct 2017 | A1 |
20170310556 | Knowles et al. | Oct 2017 | A1 |
20170317932 | Paramasivam | Nov 2017 | A1 |
20170339070 | Chang et al. | Nov 2017 | A1 |
20180069885 | Patterson et al. | Mar 2018 | A1 |
20180173372 | Greenspan et al. | Jun 2018 | A1 |
20180174060 | Velez-Rojas et al. | Jun 2018 | A1 |
Number | Date | Country |
---|---|---|
101719930 | Jun 2010 | CN |
101394360 | Jul 2011 | CN |
102164091 | Aug 2011 | CN |
104320342 | Jan 2015 | CN |
105740084 | Jul 2016 | CN |
2228719 | Sep 2010 | EP |
2439637 | Apr 2012 | EP |
2645253 | Nov 2014 | EP |
10-2015-0070676 | May 2015 | KR |
M394537 | Dec 2010 | TW |
WO 2009155574 | Dec 2009 | WO |
WO 2010030915 | Mar 2010 | WO |
WO 2013158707 | Oct 2013 | WO |
WO 2016118052 | Jul 2016 | WO |
Entry |
---|
Amedro, Brian, et al., “An Efficient Framework for Running Applications on Clusters, Grids and Cloud,” 2010, 17 pages. |
Author Unknown, “A Look at DeltaCloud: The Multi-Cloud API,” Feb. 17, 2012, 4 pages. |
Author Unknown, “About Deltacloud,” Apache Software Foundation, Aug. 18, 2013, 1 page. |
Author Unknown, “Architecture for Managing Clouds, A White Paper from the Open Cloud Standards Incubator,” Version 1.0.0, Document No. DSP-IS0102, Jun. 18, 2010, 57 pages. |
Author Unknown, “Cloud Infrastructure Management Interface—Common Information Model (CIMI-CIM),” Document No. DSP0264, Version 1.0.0, Dec. 14, 2012, 21 pages. |
Author Unknown, “Cloud Infrastructure Management Interface (CIMI) Primer,” Document No. DSP2027, Version 1.0.1, Sep. 12, 2012, 30 pages. |
Author Unknown, “cloudControl Documentation,” Aug. 25, 2013, 14 pages. |
Author Unknown, “Interoperable Clouds, A White Paper from the Open Cloud Standards Incubator,” Version 1.0.0, Document No. DSP-IS0101, Nov. 11, 2009, 21 pages. |
Author Unknown, “Microsoft Cloud Edge Gateway (MCE) Series Appliance,” Iron Networks, Inc., 2014, 4 pages. |
Author Unknown, “Use Cases and Interactions for Managing Clouds, A White Paper from the Open Cloud Standards Incubator,” Version 1.0.0, Document No. DSP-ISO0103, Jun. 16, 2010, 75 pages. |
Author Unknown, “Apache Ambari Meetup What's New,” Hortonworks Inc., Sep. 2013, 28 pages. |
Author Unknown, “Introduction,” Apache Ambari project, Apache Software Foundation, 2014, 1 page. |
Citrix, “Citrix StoreFront 2.0” White Paper, Proof of Concept Implementation Guide, Citrix Systems, Inc., 2013, 48 pages. |
Citrix, “CloudBridge for Microsoft Azure Deployment Guide,” 30 pages. |
Citrix, “Deployment Practices and Guidelines for NetScaler 10.5 on Amazon Web Services,” White Paper, citrix.com, 2014, 14 pages. |
Gedymin, Adam, “Cloud Computing with an emphasis on Google App Engine,” Sep. 2011, 146 pages. |
Good, Nathan A., “Use Apache Deltacloud to administer multiple instances with a single API,” Dec. 17, 2012, 7 pages. |
Kunz, Thomas, et al., “OmniCloud—The Secure and Flexible Use of Cloud Storage Services,” 2014, 30 pages. |
Logan, Marcus, “Hybrid Cloud Application Architecture for Elastic Java-Based Web Applications,” F5 Deployment Guide Version 1.1, 2016, 65 pages. |
Lynch, Sean, “Monitoring cache with Claspin” Facebook Engineering, Sep. 19, 2012, 5 pages. |
Meireles, Fernando Miguel Dias, “Integrated Management of Cloud Computing Resources,” 2013-2014, 286 pages. |
Mu, Shuai, et al., “uLibCloud: Providing High Available and Uniform Accessing to Multiple Cloud Storages,” 2012 IEEE, 8 pages. |
Sun, Aobing, et al., “IaaS Public Cloud Computing Platform Scheduling Model and Optimization Analysis,” Int. J. Communications, Network and System Sciences, 2011, 4, 803-811, 9 pages. |
Szymaniak, Michal, et al., “Latency-Driven Replica Placement”, vol. 47 No. 8, IPSJ Journal, Aug. 2006, 12 pages. |
Toews, Everett, “Introduction to Apache jclouds,” Apr. 7, 2014, 23 pages. |
Von Laszewski, Gregor, et al., “Design of a Dynamic Provisioning System for a Federated Cloud and Bare-metal Environment,” 2012, 8 pages. |
Ye, Xianglong, et al., “A Novel Blocks Placement Strategy for Hadoop,” 2012 IEEE/ACTS 11th International Conference on Computer and Information Science, 2012 IEEE, 5 pages. |
Author Unknown, “5 Benefits of a Storage Gateway in the Cloud,” Blog, TwinStrata, Inc., Jul. 25, 2012, XP055141645, 4 pages, https://web.archive.org/web/20120725092619/http://blog.twinstrata.com/2012/07/10//5-benefits-of-a-storage-gateway-in-the-cloud. |
Author Unknown, “Joint Cisco and VMWare Solution for Optimizing Virtual Desktop Delivery: Data Center 3.0: Solutions to Accelerate Data Center Virtualization,” Cisco Systems, Inc. and VMware, Inc., Sep. 2008, 10 pages. |
Author Unknown, “Open Data Center Alliance Usage: Virtual Machine (VM) Interoperability in a Hybrid Cloud Environment Rev. 1.2,” Open Data Center Alliance, Inc., 2013, 18 pages. |
Author Unknown, “Real-Time Performance Monitoring on Juniper Networks Devices, Tips and Tools for Assessing and Analyzing Network Efficiency,” Juniper Networks, Inc., May 2010, 35 pages. |
Beyer, Steffen, “Module “Data::Locations?!”,” YAPC::Europe, London, UK,ICA, Sep. 22-24, 2000, XP002742700, 15 pages. |
Borovick, Lucinda, et al., “Architecting the Network for the Cloud,” IDC White Paper, Jan. 2011, 8 pages. |
Bosch, Greg, “Virtualization,” last modified Apr. 2012 by B. Davison, 33 pages. |
Broadcasters Audience Research Board, “What's Next,” http://lwww.barb.co.uk/whats-next, accessed Jul. 22, 2015, 2 pages. |
Cisco Systems, Inc. “Best Practices in Deploying Cisco Nexus 1000V Series Switches on Cisco UCS B and C Series Cisco UCS Manager Servers,” Cisco White Paper, Apr. 2011, 36 pages, http://www.cisco.com/en/US/prod/collateral/switches/ps9441/ps9902/white_paper_c11-558242.pdf. |
Cisco Systems, Inc., “Cisco Unified Network Services: Overcome Obstacles to Cloud-Ready Deployments,” Cisco White Paper, Jan. 2011, 6 pages. |
Cisco Systems, Inc., “Cisco Intercloud Fabric: Hybrid Cloud with Choice, Consistency, Control and Compliance,” Dec. 10, 2014, 22 pages. |
Cisco Technology, Inc., “Cisco Expands Videoscape TV Platform Into the Cloud,” Jan. 6, 2014, Las Vegas, Nevada, Press Release, 3 pages. |
CSS Corp, “Enterprise Cloud Gateway (ECG)—Policy driven framework for managing multi-cloud environments,” original published on or about Feb. 11, 2012; 1 page; http://www.css-cloud.com/platform/enterprise-cloud-gateway.php. |
Fang K., “LISP MAC-EID-TO-RLOC Mapping (LISP based L2VPN),” Network Working Group, Internet Draft, CISCO Systems, Jan. 2012, 12 pages. |
Herry, William, “Keep It Simple, Stupid: OpenStack nova-scheduler and its algorithm”, May 12, 2012, IBM, 12 pages. |
Hewlett-Packard Company, “Virtual context management on network devices”, Research Disclosure, vol. 564, No. 60, Apr. 1, 2011, Mason Publications, Hampshire, GB, Apr. 1, 2011, 524. |
Juniper Networks, Inc., “Recreating Real Application Traffic in Junosphere Lab,” Solution Brief, Dec. 2011, 3 pages. |
Kenhui, “Musings on Cloud Computing and IT-as-a-Service: [Updated for Havana] Openstack Computer for VSphere Admins, Part 2: Nova-Scheduler and DRS”, Jun. 26, 2013, Cloud Architect Musings, 12 pages. |
Kolyshkin, Kirill, “Virtualization in Linux,” Sep. 1, 2006, XP055141648, 5 pages, https://web.archive.org/web/20070120205111/http://download.openvz.org/doc/openvz-intro.pdf. |
Lerach, S.R.O., “Golem,” http://www.lerach.cz/en/products/golem, accessed Jul. 22, 2015, 2 pages. |
Linthicum, David, “VM Import could be a game changer for hybrid clouds”, InfoWorld, Dec. 23, 2010, 4 pages. |
Naik, Vijay K., et al., “Harmony: A Desktop Grid for Delivering Enterprise Computations,” Grid Computing, 2003, Fourth International Workshop on Proceedings, Nov. 17, 2003, pp. 1-11. |
Nair, Srijith K. et al., “Towards Secure Cloud Bursting, Brokerage and Aggregation,” 2012, 8 pages, www.flexiant.com. |
Nielsen, “SimMetry Audience Measurement—Technology,” http://www.nielsen-admosphere.eu/products-and-services/simmetry-audience-measurement-technology/, accessed Jul. 22, 2015, 6 pages. |
Nielsen, “Television,” http://www.nielsen.com/us/en/solutions/measurement/television.html, accessed Jul. 22, 2015, 4 pages. |
Open Stack, “Filter Scheduler,” updated Dec. 17, 2017, 5 pages, accessed on Dec. 18, 2017, https://docs.openstack.org/nova/latest/user/filter-scheduler.html. |
Rabadan, J., et al., “Operational Aspects of Proxy-ARP/ND in EVPN Networks,” BESS Worksgroup Internet Draft, draft-snr-bess-evpn-proxy-arp-nd-02, Oct. 6, 2015, 22 pages. |
Saidi, Ali, et al., “Performance Validation of Network-Intensive Workloads on a Full-System Simulator,” Interaction between Operating System and Computer Architecture Workshop, (IOSCA 2005), Austin, Texas, Oct. 2005, 10 pages. |
Shunra, “Shunra for HP Software; Enabling Confidence in Application Performance Before Deployment,” 2010, 2 pages. |
Son, Jungmin, “Automatic decision system for efficient resource selection and allocation in inter-clouds,” Jun. 2013, 35 pages. |
Wikipedia, “Filter (software)”, Wikipedia, Feb. 8, 2014, 2 pages, https://en.wikipedia.org/w/index.php?title=Filter_%28software%29&oldid=594544359. |
Wikipedia; “Pipeline (Unix)”, Wikipedia, May 4, 2014, 4 pages, https://en.wikipedia.org/w/index.php?title=Pipeline2/028Unix%29&oldid=606980114. |
Extended European Search Report from the European Patent Office, dated Mar. 29, 2018, 9 pages, for the corresponding European Patent Application No. EP 17206917.1. |
Al-Harbi, S.H., et al., “Adapting k-means for supervised clustering,” Jun. 2006, Applied Intelligence, vol. 24, Issue 3, pp. 219-226. |
Bohner, Shawn A., “Extending Software Change Impact Analysis into COTS Components,” 2003, IEEE, 8 pages. |
Hood, C. S., et al., “Automated Proactive Anomaly Detection,” 1997, Springer Science and Business Media Dordrecht , pp. 688-699. |
Vilalta R., et al., “An efficient approach to external cluster assessment with an application to martian topography,” Feb. 2007, 23 pages, Data Mining and Knowledge Discovery 14.1: 1-23. New York: Springer Science & Business Media. |
Number | Date | Country | |
---|---|---|---|
20180176281 A1 | Jun 2018 | US |