The advancement of global communications has allowed groups of individuals (e.g., co-workers, classmates, social club members, etc.) to interact with one another via electronic methods. In particular, voice conferencing applications executed on electronic devices, such as computers, smartphones, etc., allow real-time communications between groups of remotely located individuals. In many professional settings, a large number of group meetings may be conducted using such voice conferencing applications.
Various examples will be described below referring to the following figures:
Voice conferencing applications (or more simply “conferencing applications”) allow a plurality of individuals to conduct real-time voice conferencing sessions via electronic devices. Voice conferencing sessions may comprise audio or audio-visual data streams that are directed to and from the plurality of individuals to facilitate real-time conversation. More particularly, a voice conferencing session may comprise a voice call (e.g., such as those conducted over a voice over internet protocol (VOIP) system), a video conference, etc. An electronic device for executing a voice conferencing session may include any suitable device that may execute machine-readable instructions. In some examples, an electronic device for executing a voice conferencing session may comprise, for instance, a computer (e.g., desktop computer, laptop computer, all-in-one computer), a smartphone, etc.
In some circumstances, a user of an electronic device may wish to multi-task while participating in a voice conferencing session. For instance, a user may wish to perform another task during the voice conferencing session, or may even wish to participate in multiple voice conferencing sessions simultaneously. In either case, the user may not actively listen for the entire duration of the voice conference session(s), and may therefore fail to answer a question, provide requested input in a timely manner, or otherwise miss cues for response or feedback.
Accordingly, examples disclosed herein provide systems and methods for monitoring the content of one (or a plurality of) voice conferencing sessions being executed on an electronic device, and for cuing the user when a response is requested or appropriate. In some examples, the systems and methods may comprise machine-readable instructions that are stored and executed on the user's electronic device (which may be referred to as the “client device”). This arrangement may be distinct from situations where machine-readable instructions are stored and/or executed on a server or other electronic device operated by the voice conferencing service that is hosting the voice conferencing session. As a result, the systems and methods may operate independently of the conferencing application itself and may be utilized to monitor voice conferences provided by different voice conference applications being executed on the electronic device. Thus, through use of the example systems and methods described herein, a user may more effectively multi-task during a voice conferencing session.
Referring now to
The processor 12 may comprise any suitable processing device, such as a microprocessor or central processing unit (CPU). The processor 12 executes machine-readable instructions (e.g. machine-readable instructions 30) stored on memory 14, thereby causing the processor 12 (and, more generally, electronic device 10) to perform some or all of the actions attributed herein to the processor 12 (and, more generally, to electronic device 10). The memory 14 (e.g., a non-transitory machine-readable medium) may comprise volatile storage (e.g., random access memory (RAM)), non-volatile storage (e.g., flash storage, etc.), or combinations of both volatile and non-volatile storage. Data read or written by the processor 12 when executing machine-readable instructions can also be stored on memory 14.
The processor 12 may comprise one processing device or a plurality of processing devices that are distributed within electronic device 10. Likewise, the memory 14 may comprise one memory device or a plurality of memory devices that are distributed within the electronic device 10.
The electronic device 10 may communicate with other devices via a network 22, such as, for instance, the Internet, a telecommunications network, etc. For instance, the network interface 16 may be coupled to an antenna 20 that communicates wirelessly with network 22 (or, more specifically, a node or gateway of network 22). In some examples, network interface 16 may communicate with network 22 via a wired connection (e.g., via Ethernet cable).
The electronic device 10 also includes (or is coupled to) a display 18. Accordingly, processor 12 may cause images to be presented on display 18 during operation. Display 18 may comprise any suitable display device, such as, for instance, a liquid crystal display (LCD), an organic light emitting diode (OLED) display, a micro-LED display, a plasma display, etc.
In addition, electronic device 10 also includes (or is coupled to) a speaker 28. In some examples, the speaker 28 may comprise a speaker (or collection of speakers) for emitting audible sound waves into an environment (e.g., such as into an environment immediately surrounding the electronic device 10). In some examples, the speaker 28 may be incorporated within a headset that is wearable by the user (not shown).
During operation, electronic device 10 may execute one or a plurality of voice conferencing sessions. More specifically, in some examples, the electronic device 10 may receive data for a first voice conferencing session from a first conferencing service 24 and may receive data for a second voice conferencing session from a second conferencing service 26 via the network 22. The first conferencing service 24 may or may not be different from the second conferencing service 26.
The first voice conferencing session is executed on electronic device 10 via a first conferencing application 38, and the second voice conferencing session is executed on electronic device 10 via a second conferencing application 40. The first conferencing application 38 may comprise machine-readable instructions for processing data associated with the first voice conferencing session, and the second conferencing application 40 may comprise machine-readable instructions for processing data associated with the second voice conferencing session. As shown in
Referring specifically to the first voice conferencing session, data is provided to the electronic device 10 from the first conferencing service 24 via network 22, antenna 20 and network interface 16. The processor 12 may process the incoming data according to the machine-readable instructions of the first conferencing application 38 and output corresponding audio and/or visual signals associated with the first voice conferencing session to the speaker 28 and display 18, respectively. Conversely, audio and/or visual inputs from the user to the first voice conferencing session (e.g., such as visual inputs captured by a camera and audio inputs captured by a microphone) may be processed by the processor 12 according to the machine-readable instructions of the first conferencing application 38 and may be communicated to the first conferencing service 24 via network interface 16, antenna 20, and network 22.
Referring specifically to the second voice conferencing session, data is provided to the electronic device 10 from the second conferencing service 26 via network 22, antenna 20, and network interface 16. The processor 12 may process the incoming data according to the machine-readable instructions of the second conferencing application 40 and output corresponding audio and/or visual signals associated with the second voice conferencing session to the speaker 28 and display 18, respectively. Conversely, audio and/or visual inputs from the user to the second voice conferencing session may be processed by the processor 12 according to the machine-readable instructions of the second conferencing application 40 and may be communicated to the second conferencing service 26 via network interface 16, antenna 20, and network 22.
Referring still to
More particularly, machine-readable instructions 30 may cause processor 12 to determine the process identifiers (IDs) assigned by electronic device 10 (particularly processor 12) to the processes associated with the first voice conferencing session and the second voice conferencing session (e.g., such as the processes generated by execution of the machine-readable instructions of the first conferencing application 38 and the second conferencing application 40 as described above). In some examples, the process IDs may comprise numeric codes or other unique identifiers that are assigned by the electronic device 10 to distinguish multiple simultaneous processes being executed on processor 12. In some examples, the process IDs may be determined by analyzing some or all of the processes being executed by the electronic device 10 to determine which of these processes may be associated with the first conferencing application 38 and/or the second conferencing application 40.
Once the process IDs of the first voice conferencing session and second voice conferencing session are identified, machine-readable instructions 30 further cause the processor 12 to determine the thread IDs of the thread(s) associated with the identified process IDs that comprise the audio data for the first voice conferencing session and the second voice conferencing session. For instance, reference is now made to
During operation, once the process ID of the first process 50 and the process ID of the second process 52 are identified as previously described, the processor 12 (e.g.,
Referring again to
The keyword(s) 36 may be selected (e.g., by the user and/or the electronic device 10) to indicate when the user is being cued for providing input to the first voice conferencing session or the second voice conferencing session. In some examples, the keyword(s) 36 may comprise the user's name (or a part thereof), a name of the user's supervisor or co-worker, a word associated with a particular subject-matter, a department name, etc.
If a match is detected between the audio data of the first voice conferencing session or the second voice conferencing session and the keyword(s) 36, the processor 12 may then generate a notification that may be output to the display 18 and/or the speaker 28 to cue the user to provide input to the first voice conferencing session or the second voice conferencing session. For instance, the notification may comprise a pop-up window on the display 18 that may obstruct or minimize other content being presented thereon. In some examples, the pop-up window may comprise a user-interface for the first voice conferencing session or the second voice conferencing session. In some examples, the notification may comprise an audio feed from the corresponding voice conferencing session (e.g., the first voice conferencing session or the second voice conferencing session) to the speaker 28. However, any suitable method of notifying the user that a response is needed or appropriate in one of the voice conferencing sessions may be used in other examples.
Referring now to
Machine-readable instructions 100 comprise identifying a process ID assigned by the electronic device 10 to a voice conferencing session executed on the electronic device 10 at block 102. In some examples, the voice conferencing session may comprise the first voice conferencing session or the second voice conferencing session described above. In addition, identifying the process ID assigned to the voice conferencing session may comprise analyzing the processes executed on the electronic device 10 to determine which process(es) are associated with a voice conferencing application as described above.
In addition, machine-readable instructions 100 include identifying a thread associated with the process ID that includes audio data for the voice conferencing session at block 104. For instance, as described above for the processes 50, 52 shown in
Further, machine-readable instructions 100 may comprise transcribing the audio data into text at block 108 and comparing the text to a keyword stored in the memory 14 at block 110. For instance, as described above, the processor 12 may transcribe the audio data into text via a suitable method, system, or application, and then compare the transcribed text to stored keyword(s) to determine if there is a match.
Still further, machine-readable instructions 100 may comprise, at block 112, determining that a user of the electronic device is being cued in the voice conferencing session based on the comparison at block 110. Accordingly, at block 114, machine-readable instructions 100 may comprise generating a notification for the user that input is requested in the voice conferencing session. The notification may comprise a visual and/or audio notification that may alert the user to the need for input to the voice conferencing session in the manner described above.
Referring now to
Machine-readable instructions 200 may include identifying a process ID assigned to a voice conferencing session being executed on the electronic device at block 202. In some examples, the voice conferencing session may comprise the first voice conferencing session or the second voice conferencing session described above. In addition, identifying the process ID at block 202 may be carried out in the manner described above for block 102 of machine-readable instructions 100 (e.g.,
In addition, machine-readable instructions 200 include identifying an audio thread using the process ID at block 204, and identifying an audio thread ID of the audio thread at block 206. For instance, in some examples, blocks 204 and 206 may be carried out in the manner described above for blocks 104 and 106 of machine-readable instructions 100 (e.g.,
Further, machine-readable instructions 200 include using the audio thread ID to compare the audio data to a keyword to identify a match at block 208. For instance, in some examples, block 208 may be carried out in the manner described above for blocks 108 and 110 of machine-readable instructions 100 (e.g.,
Still further, machine-readable instructions 200 include, at block 210, generating a notification for a user of the electronic device 10 in response to the match from block 208. The notification may comprise a visual and/or audio notification that may alert the user to the need for input to the voice conferencing session in the manner described above.
Referring now to
Method 300 includes identifying process IDs assigned by an electronic device for multiple voice conferencing sessions executed simultaneously on the electronic device at block 302. For instance, as described above for the electronic device 10 of
In addition, method 300 includes sampling data from threads associated with the process IDs at block 304, determining that threads include audio data for the voice conferencing sessions based on the sampling at block 306, and identifying audio thread IDs for the threads at block 308. For instance, as was described above for electronic device 10 and shown in
Further, method 300 includes using the audio thread IDs to analyze the audio data for each of the threads at block 310. For instance, as described above for electronic device 10, the processor 12 may analyze audio data of the identified audio thread IDs by transcribing the audio data into text via a suitable method, system, or application, and then comparing the transcribed text to stored keyword(s) to determine if there is a match.
Still further, method 300 includes cuing a user of the electronic device to provide input to one of the voice conferencing sessions at block 312 based on the analysis at block 310. In some examples, the cuing at block 312 may comprise generating a notification such as is described above for electronic device 10. However, any suitable cuing method may be utilized to alert the user that input is requested and/or appropriate in one of the voice conferencing sessions at block 312.
Example systems and methods have been described herein for monitoring the content of one (or a plurality of) voice conferencing sessions being executed on an electronic device, and for cuing the user when a response is requested or appropriate. Thus, through use of the examples systems and methods described herein, a user may more effectively multi-task during a voice conference.
In the figures, certain features and components disclosed herein may be shown exaggerated in scale or in somewhat schematic form, and some details of certain elements may not be shown in the interest of clarity and conciseness. In some of the figures, in order to improve clarity and conciseness, a component or an aspect of a component may be omitted.
In the following discussion and in the claims, the terms “including” and “comprising” are used in an open-ended fashion, and thus should be interpreted to mean “including, but not limited to . . . .” Also, the term “couple” or “couples” is intended to be broad enough to encompass both indirect and direct connections. Thus, if a first device couples to a second device, that connection may be through a direct connection or through an indirect connection via other devices, components, and connections.
As used herein, including in the claims, the word “or” is used in an inclusive manner. For example, “A or B” means any of the following: “A” alone, “B” alone, or both “A” and “B.”
The above discussion is meant to be illustrative of the principles and various examples of the present disclosure. Numerous variations and modifications will become apparent to those skilled in the art once the above disclosure is fully appreciated. It is intended that the following claims be interpreted to embrace all such variations and modifications.
Filing Document | Filing Date | Country | Kind |
---|---|---|---|
PCT/US2020/064339 | 12/10/2020 | WO |