The present invention relates generally to the fields of telephony and conferencing systems.
A participant to a conference held over a network may be interrupted, i.e., pulled out of real-time listening or participation, for a variety of reasons. For example, the participant may need to step away from the conference to answer another call, or briefly discuss a matter of immediate urgency with a co-worker. Alternatively, the participant may invoke some feature of the conference service that results in the conference audio being interrupted, such as playing out the roster, receiving whisper information, moving to a sidebar conversation, etc. Upon returning to the live conference the participant usually wants to listen to what transpired in the conference during the time he was gone. Simply listening to a recording of the missed portions of the conference, however, is inadequate insomuch as the participant is permanently shifted away from real-time participation.
A number of prior art conferencing systems include a feature that allows portions of a conference session to be recorded and then played back at an accelerated rate, in what is commonly referred to as “catch-up” mode. The problem with these systems is that the participant frequently forgets that he is listening to the conference in catch-up mode, and attempts to speak as a participant in the conference. Comments made during catch-up mode, however, cannot be mixed into the real-time (“live”) conference because the live conference has already progressed beyond the point in time of the catch-up mode playback.
The present invention will be understood more fully from the detailed description that follows and from the accompanying drawings, which however, should not be taken to limit the invention to the specific embodiments shown, but are for explanation and understanding only.
In the following description specific details are set forth, such as device types, system configurations, protocols, methods, etc., in order to provide a thorough understanding of the present invention. However, persons having ordinary skill in the relevant arts will appreciate that these specific details may not be needed to practice the present invention.
In the context of the present application, endpoint devices represent an end user, client, or person who wishes to initiate or participate in a conference session and via conferencing system 19 via IP network 20. Other endpoint devices not specifically shown in
In the example of
Conferencing system 19 comprises a server with a processor subsystem for running a conferencing program including software (or firmware) plug-ins, modules, or enhancements that implement the various features and functions described herein. In a specific implementation, for example, conferencing system 19 may run a modified or enhanced IP communication system software product such as Cisco's MeetingPlace™ conferencing application that allows users to schedule and attend meeting conferences. In the embodiment shown, conferencing system 19 may also include a digital signal processor (DSP) or firmware/software-based system that mixes and/or switches audio/video signals received at its input ports under the control of the server. The audio signals received at the conference server ports originate from each of the conference or meeting participant endpoints (e.g., endpoint devices 16, 22 and 14), and possibly from an interactive voice response (IVR) system (not shown). In certain embodiments, conference server may also incorporate or be associated with a natural language automatic speech recognition (ASR) module for interpreting and parsing speech of the participants, and standard speech-to-text (STT) and text-to-speech (TTS) converter modules.
Practitioners in the art will appreciate that the actual media paths are established by conferencing system 19. In other words, conferencing system 19 handles all of the control plane functions of the conference session, and is responsible for engaging the necessary media components/resources to satisfy the media requirements of each of the endpoints (i.e., endpoint 16) participating in a conference session. In operation, endpoint devices 14, 16 and 22 shown in
Conference mixer 25 mixes or switches the data (audio, video, and/or web collaboration) packet streams and also sends real-time media for recording (via path 45) to server 27. As will be described in more detail below, server 27 may also transmit media playback of triggered comments to mixer 25 on request via connection path 46. As participants talk during a conference, real-time media data transmissions occur via paths 39, 41 and 43 between conference mixer 25 and respective endpoints 14, 22 and 16. Conference mixer 25 mixes the real-time media according to control signals/messages provided by conference control unit 23. For instance, conference mixer 25 may be directed to apply any one of a wide variety of audio/video mixing or switching algorithms to produce an output media stream delivered to the endpoints devices.
Note that in certain embodiments, conference mixer 25 may filter out annoyances by dynamically suppressing audio streams that meet predefined annoyance criteria (e.g., white noise, pink noise, snoring, etc.). It is appreciated that for non-verbal audience reaction, the audio mix can be of relatively low-fidelity. However, for applications where verbal reaction is not to be filtered out as an annoyance, the verbal audience reaction should be of relatively high-fidelity.
Participant 18 may drop out of the real-time conference session by entering a command (e.g., pause button) on endpoint 16. In response, recording and playback server 27 begins recording real-time media (via path 45) from the conference session. Assume, for example, that during the session participant 18 may have a question or a comment for moderator/lecturer 11, with the moderator requesting that all questions be held until the end of the lecture. Participant 18 may click on an icon button of a user interface running on endpoint 16 to signal that he is dropping out of real-time listening mode to record his question/comment. Concurrently, server 27 begins recording the real-time media occurring in lecture hall 12 while endpoint 16 records his question/comment. That is, recording and playback server 27 may also record the question/comment received from endpoint 16 simultaneous with the recording of the real-time media.
Alternatively, detection of speech by the conferencing system may cause the participant to automatically drop out of real-time listening mode and into recording mode. That is, audible speech of a participant who otherwise would not be able to participate directly in the conference may trigger recording of that speech. When the speech stops for a predetermined period (e.g., 5 seconds of silence), the system may automatically stop the recording and enter catch-up mode for that participant.
In a specific implementation, recording and playback server 27 may also automatically record a predetermined time period (e.g., 5-10 seconds) of the conference session just prior to the point in time when endpoint 16 dropped out of the conference and attach it to his recorded question/comment. This additional recording provides a context for the recorded question/comment. The recorded speech and the recorded real-time media may be stored in a multimedia storage unit 50. The combination of the context media and the question/comment may be placed in a moderated question and answer queue, allowing the lecturer to subsequently manage or trigger individual questions or comments.
When participant 18 has finished recording his question/comment (indicated by an appropriate command entered on endpoint 16), server 27 starts transmitting playback of the catch-up media recording (shown by path 52) to endpoint 16. The catch-up media is played back at a faster rate (e.g., 1.5×-2× with pauses and silences removed or reduced) than real-time until endpoint 16 catches up to the real-time conference session. Later, after the lecture portion of the conference has ended and the moderator is ready to respond to questions/comments, server 27 retrieves the triggered questions/comments from multimedia storage 50. The questions/comments are played back to conference mixer 25 and transmitted to each of endpoints 16, 14 and 22. Note that the additional 5-10 seconds of recorded media preceding the question/comment may help the listeners to understand and identify the context of the question/comment.
After the question/comment has been recorded, the system may optionally prompt the participant as to whether he wishes to edit the recording (block 67). If so, then the participant may edit his comments by recording over his previous comments (block 68). Other optional prompts may include the system asking whether the participant wants to attach any documents such as charts, diagrams, pictures, etc., (blocks 69 & 70); or whether the participant wants to adjust the marked context point (blocks 71 & 72), e.g., by changing the time mark when the comment is triggered in the conference session (for instance, the participant may indicate that he wants twenty seconds of the recorded real-time media played back before his comment).
The system may also give the participant the option of submitting/storing his question/comment (blocks 73 & 75). If the participant declines (e.g., he reconsiders and determines that his comment is unnecessary or does not significantly add to the discussion), the system discards the recorded comment and resumes catch-up playback (block 74). In the case of a submitted question/comment, the system may further query whether the conference is a moderated conference (block 76). For moderated conferences, the comment may be entered into the floor-control queue (block 77); that is, when the floor is opened for questions and/or comments, the comment will be heard by the participants in on ordered manner. Once the comment has been submitted, the system may then resume catch-up playback (78). If the conference is not moderated, the comment may be stored for subsequent triggering (block 79) after which time the system resumes catch-up playback (80).
In a specific implementation, when the context is triggered to pop-up on a participant's web browser, the agenda context may appear (e.g., “the context for this comment is Agendum 53,” etc.). If the context involves web sharing, a slide that was used when the comment was made may automatically pop-up on the participant's computer screen. Additionally, an excerpted portion of the audio playback preceding the comment may accompany the information presented graphically on the user's PC monitor screen. After the full context has been played out, the actual comment is then played (block 117).
In certain embodiments, the system may be configured to allow user-controlled marking of context preceding a comment or question. This feature may apply to playback of recorded conferences, playback of catch-up media while the live, moderated conference session continues (i.e., in real-time), or during a real-time moderated conference. A variety of different options may be provided. For instance, in one embodiment a participant might hear something in a real-time conference session, or playback of a recorded session, and want to set a mark at that point using a GUI or VUI. In a specific example, a participant who has already recorded a comment/question may desire to mark a point in the session preceding where his comment/question occurred. In other words, instead of using a default context setting (e.g., 10 seconds of playback preceding the comment), the system may allow the commentator to determine the appropriate context and set the context accordingly. For instance, the commentator may determine that 25 seconds of preceding recording is needed to fully understand his comment. In other cases, the commentator may decide to attach a report, graph, or web link instead of, or in addition to, a selected number of seconds of contextual playback.
User-controlled marking of context may also be provided in the conferencing system of
In another example of participant-controlled marking, the user may insert provisional marks while remaining a real time participant in the conference. For example, the participant may be listening to a lecture and begin to hear something that he doesn't understand. If misunderstood point is subsequently clarified, the participant can cancel the provisional mark. However, if the participant then leaves a comment or question, the provisional mark can act as the actual beginning of the context for the question.
It should be understood that elements of the present invention may also be provided as a computer program product which may include a machine-readable medium having stored thereon instructions which may be used to program a computer (e.g., a processor or other electronic device) to perform a sequence of operations. Alternatively, the operations may be performed by a combination of hardware and software. The machine-readable medium may include, but is not limited to, floppy diskettes, optical disks, CD-ROMs, and magneto-optical disks, ROMs, RAMs, EPROMs, EEPROMs, magnet or optical cards, or other type of machine-readable medium suitable for storing electronic instructions.
Additionally, although the present invention has been described in conjunction with specific embodiments, numerous modifications and alterations are well within the scope of the present invention. Accordingly, the specification and drawings are to be regarded in an illustrative rather than a restrictive sense.
Number | Name | Date | Kind |
---|---|---|---|
5483587 | Hogan et al. | Jan 1996 | A |
5600366 | Schulman | Feb 1997 | A |
5673253 | Shaffer | Sep 1997 | A |
5729687 | Rothrock et al. | Mar 1998 | A |
5917830 | Chen et al. | Jun 1999 | A |
5963217 | Grayson et al. | Oct 1999 | A |
6044081 | Bell et al. | Mar 2000 | A |
6137834 | Wine et al. | Oct 2000 | A |
6141324 | Abbott et al. | Oct 2000 | A |
6236854 | Bradshaw | May 2001 | B1 |
6269107 | Jong | Jul 2001 | B1 |
6332153 | Cohen | Dec 2001 | B1 |
6501739 | Cohen | Dec 2002 | B1 |
6505169 | Bhagavath et al. | Jan 2003 | B1 |
6570926 | Agrawal et al. | May 2003 | B1 |
6608820 | Bradshaw | Aug 2003 | B1 |
6624841 | Buchner | Sep 2003 | B1 |
6643298 | Brunheroto et al. | Nov 2003 | B1 |
6643496 | Shimoyama et al. | Nov 2003 | B1 |
6650652 | Valencia | Nov 2003 | B1 |
6671262 | Kung et al. | Dec 2003 | B1 |
6675216 | Quatrano et al. | Jan 2004 | B1 |
6718553 | Kenworthy | Apr 2004 | B2 |
6735572 | Landesmann | May 2004 | B2 |
6744785 | Robinett et al. | Jun 2004 | B2 |
6771644 | Brassil et al. | Aug 2004 | B1 |
6771657 | Elstermann | Aug 2004 | B1 |
6775247 | Shaffer et al. | Aug 2004 | B1 |
6816469 | Kung et al. | Nov 2004 | B1 |
6865540 | Faber et al. | Mar 2005 | B1 |
6876734 | Summers et al. | Apr 2005 | B1 |
6925068 | Stanwood et al. | Aug 2005 | B1 |
6931001 | Deng | Aug 2005 | B2 |
6931113 | Ortel | Aug 2005 | B2 |
6937569 | Sarkar et al. | Aug 2005 | B1 |
6947417 | Laursen et al. | Sep 2005 | B2 |
6956600 | Gaylord | Oct 2005 | B1 |
6956828 | Simard et al. | Oct 2005 | B2 |
6959075 | Cutaia et al. | Oct 2005 | B2 |
6976055 | Shaffer et al. | Dec 2005 | B1 |
6989856 | Firestone et al. | Jan 2006 | B2 |
7003086 | Shaffer et al. | Feb 2006 | B1 |
7007098 | Smyth et al. | Feb 2006 | B1 |
7031348 | Gazit | Apr 2006 | B1 |
7084898 | Firestone et al. | Aug 2006 | B1 |
7127487 | Wang et al. | Oct 2006 | B1 |
7209763 | Martin et al. | Apr 2007 | B2 |
7379653 | Yap et al. | May 2008 | B2 |
7660849 | Shaffer et al. | Feb 2010 | B1 |
20010000540 | Cooper et al. | Apr 2001 | A1 |
20020004841 | Sawatari | Jan 2002 | A1 |
20020006165 | Kato | Jan 2002 | A1 |
20020052214 | Maggenti et al. | May 2002 | A1 |
20020087976 | Kaplan et al. | Jul 2002 | A1 |
20020163918 | Cline | Nov 2002 | A1 |
20030016632 | Refai et al. | Jan 2003 | A1 |
20030025786 | Norsworthy | Feb 2003 | A1 |
20030076850 | Jason, Jr. | Apr 2003 | A1 |
20030198195 | Li | Oct 2003 | A1 |
20040057449 | Black | Mar 2004 | A1 |
20040165527 | Gu et al. | Aug 2004 | A1 |
20040165710 | DelHoyo et al. | Aug 2004 | A1 |
20040184586 | Coles et al. | Sep 2004 | A1 |
20040199659 | Ishikawa et al. | Oct 2004 | A1 |
20040213152 | Matuoka et al. | Oct 2004 | A1 |
20040250201 | Caspi | Dec 2004 | A1 |
20040255328 | Baldwin et al. | Dec 2004 | A1 |
20050069102 | Chang | Mar 2005 | A1 |
20050078171 | Firestone et al. | Apr 2005 | A1 |
20050081244 | Barrett et al. | Apr 2005 | A1 |
20050138372 | Kajihara et al. | Jun 2005 | A1 |
20050259803 | Khartabil | Nov 2005 | A1 |
20060020995 | Opie et al. | Jan 2006 | A1 |
20060067250 | Boyer et al. | Mar 2006 | A1 |
20060072672 | Holcomb et al. | Apr 2006 | A1 |
20060090166 | Dhara et al. | Apr 2006 | A1 |
20060104458 | Kenoyer | May 2006 | A1 |
20060146734 | Wenger et al. | Jul 2006 | A1 |
20060189337 | Farrill et al. | Aug 2006 | A1 |
20060259755 | Kenoyer | Nov 2006 | A1 |
20070110029 | Gilmore et al. | May 2007 | A1 |
20070123284 | Schliwa-Bertling et al. | May 2007 | A1 |
20070133435 | Eneroth et al. | Jun 2007 | A1 |
20080043140 | Herpel et al. | Feb 2008 | A1 |
Number | Date | Country |
---|---|---|
1 553 735 | Jul 2005 | EP |
Number | Date | Country | |
---|---|---|---|
20080137558 A1 | Jun 2008 | US |