The disclosure relates generally to videoconferencing, and more particularly to a videoconferencing device and method for providing more engaging, dynamic and immersive video conferencing experiences.
In a typical videoconference, a local video image and audio are captured by a camera and microphone, respectively. In an attempt to capture a better local video image, the camera is often manually manipulated to zoom in or zoom out to capture a wide view showing all participants and as much of their environment as possible, assuming that such the wide view would be the best for the remote participants. However, the optimal local view may drastically differ from one remote device to another. For example, when displayed on a small display device, the wide local view may not display enough detail in a small display. On the other hand, when the camera is zoomed in to capture a narrow local view, certain areas of the local site that could give relevant context may not be captured at all. Accordingly, there still remain significant areas for new and improved ideas for more effective videoconferencing schemes.
A first device is disclosed, which is located at a first site and in communication with a second device located at a second site geographically separated from the first site. The first device includes a controller configured to automatically control selection of a view defining an area of the second site selectable to be displayed by the first device, the controller automatically selecting the view based on a display configuration of the first device. The controller is configured to automatically control selection of a sound of the second site to be rendered by the first device, the controller automatically selecting the sound based on the selected view. The first device also includes a display configured to display the selected view of the second site based on a video image of the second site captured by the second device, the display having the display configuration. The first device further includes a speaker configured to render the selected sound of the second site based on an audio of the second site captured by the second device.
In another implementation, a method for operating a first device is disclosed, which is located at a first site and in communication with a second device located at a second site geographically separated from the first site. The method includes automatically selecting, via a controller of the first device, a view defining an area of the second site selectable to be displayed by a display of the first device, based on a display configuration of the remote device; based on a selected view, automatically selecting, via the controller of the first device, a sound of the second site to be rendered by a speaker of the first device; displaying, via the display of the first device, the selected view of the second site based on a video image of the second site captured by the second device; and rendering, via the speaker of the first device, the selected sound of the second site based on an audio of the second site captured by the second device.
In another implementation, a first device is disclosed, which is located at a first site and in communication with a second device located at a second site geographically separated from the first site. The first device includes means for selecting, based on a display configuration of the first device, a view defining an area of the second site selectable to be displayed by the first device; means for selecting, based on the selected view, a sound of the second site; means for displaying the selected view of the second site based on a video image of the second site captured by the second device; and means for rendering the selected sound of the second site based on an audio of the second site captured by the second device.
This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter. Furthermore, the claimed subject matter is not limited to implementations that solve any or all disadvantages noted in any part of this disclosure.
The drawing figures depict one or more implementations in accord with the present teachings, by way of example only, not by way of limitation. In the figures, like reference numerals refer to the same or similar elements. Furthermore, it should be understood that the drawings are not necessarily to scale.
In the following detailed description, numerous specific details are set forth by way of examples in order to provide a thorough understanding of the relevant teachings. However, it should be apparent that the present teachings may be practiced without such details. In other instances, well known methods, procedures, components, and/or circuitry have been described at a relatively high-level, without detail, in order to avoid unnecessarily obscuring aspects of the present teachings. In the following material, indications of direction, such as “top” or “left,” are merely to provide a frame of reference during the following discussion, and are not intended to indicate a required, desired, or intended orientation of the described articles unless expressly indicated.
This description is directed to a videoconferencing device and a method for operating the same to provide more engaging, dynamic and immersive videoconferencing experiences. The videoconferencing device may be a remote device that receives a video image and audio from a near end device. The remote device may have a complete control over selecting how the near end site should be displayed at the remote device, which is referred to as a view selection. The “view” may define an area of the near end site that is selectable to be displayed by the remote device. The remote device may also have a complete control over selecting how a sound of the near end site should be rendered by the remote device, which is referred to as a sound selection. For example, the remote device may automatically control the view selection based on its own display configuration. Hence, the view of the near end site displayed by the remote device may be optimized for the display of the remote device, providing visually improved videoconferencing experiences. Also, the remote device may automatically control the sound selection based on the selected view to provide more immersive videoconferencing experiences. Further, the remote device may be configured to periodically check if a new view and sound selection is necessary and, if necessary, select and switch to the new view and sound.
The near end device 200 may be a single-unit device having one or more cameras and one or more microphones to capture and process the near end video image and audio. For example, the near end device 200 may be a videoconference device equipped with one or more cameras and one or more microphones to capture the near end video image and audio. Alternatively, the near end device 200 may be a combination of separate units configured to capture and process the near end video image and audio. For example, the near end device 200 may include a plurality of cameras and microphones and a console in a multicasting studio. Implementations of the near end device 200 are not limited to the particular examples described in this disclosure.
The remote device 100 may be configured to receive the near end video and audio data 206, 208 and display and render the selected view and sound of the near end site 202 based on the received near end video and audio data 206, 208. As described above, the remote device 100 may have a complete control over the view selection and sound selection. For example, the remote device 100 may be configured to automatically select a view of the near end site 202 to be displayed by the remote device 100. The remote device 100 may also automatically select a sound of the near end site 202 to be rendered by the remote device 100. In an implementation, the view selection by the remote device 100 may be automatically controlled based on a display configuration of the remote device 100, such as a display size or resolution of the remote device 100. The sound selection may also be automatically controlled by the remote device 100 based on the selected view. The remote device 100 may also be configured to automatically select a new view and sound, when necessary.
The remote and near end devices 100, 200 may be connected directly to each other via the network 10. Alternatively, an additional device, such as a cloud server 20 connected to the network 10, may be used to administrate and/or mediate the connections and communications between the remote and near end devices 100, 200. The cloud server 20 may be operated by, for example, a videoconferencing service provider, multicasting studio or the like, to enable secure videoconferencing or multicasting to various remote devices 100 of different configurations and capabilities. For example, when the near end device 200 is used to multicast an event occurring at the near end site 202, the cloud server 20 may exclusively receive the near end video and audio data 206, 208 from the near end device 200 and provide the received near end video and audio data 206, 208 to a number of remote devices 100 via the network 10.
As illustrated in
In some instances, the actual size and/or resolution of the near end video image displayed on the remote device 100 may be different from the display size and resolution of the remote device 100. For example, the remote user 104 may run a videoconferencing application that opens a window (shown in
Due to the actual display size and/or resolution differences among the different remote devices 100, different distances of the remote users 104 from the remote devices 100, different bandwidth/connection conditions and/or the like, the same near end video image may be displayed drastically different on different remote devices 100. For example, the near end device 200 may capture and process the near end video image to provide a full view 310 (shown in
The near end device 200 may capture and process the near end video image to provide a narrow view 330 illustrated in
Furthermore, when the near end audio captured at the near end site 202 is not matched with a view of the near end video image displayed on the remote device 100, the remote user 104 may become confused. For example, a narrow view prominently showing a head of an active speaker is displayed on the remote device 100 and the sound of the active speaker rendered by the remote device 100 is mixed with voices from other near end subjects 204 not shown in the narrow view, the remote user 104 may become confused as to which voice originates form the active speaker. Also, when two or more near end subjects 204 speak simultaneously and the sound of the each near end subject 204 speaking simultaneously is not aligned with a position of each near end subject 204 displayed on the remote device 100, the remote user 104 may feel disoriented and even nauseated.
According to the disclosure, each remote device 100 may be configured to have a complete control over the view and sound selection. As described above, the view selection refers to how the near end site 202 should be displayed by the remote device 100, and the sound selection refers to how the sound of the near end site 202 should be rendered by the remote device 100. For example, the remote device 100 may be configured to automatically select a view of the near end site 202 to be displayed by the remote device 100. The view may be defined as an area of the near end site 202 that is selectable to be displayed by the remote device 100. Also, the remote device 100 may be configured to automatically select a sound of the near end site 202 to be rendered by the remote device 100. In an implementation, the remote device 100 may automatically control the view selection based on its display configuration, such as display size and resolution, to select the view optimized for the display configuration. The remote device 100 may automatically control the sound selection based on the selected view such that selected sound matches the selected view.
The remote device 100 may also be configured to automatically switch the view and sound when an event occurs at the near end site 202 that requires a new view and sound for the remote device 100. For example, when one of the near end subjects 204 becomes an active speaker, the remote device 100 may switch the view from the full view 310 to the narrow view 330 to prominently display the active speaker. The near end audio may also be rendered to switch from a full auditory sound to a narrow beam sound focusing on the voice of the active speaker. When the active speaker stops speaking and none of the remaining near end subjects 204 speak, the remote device 100 may switch back to the full view 310. The sound of the near end audio may also be switched back to the full auditory sound.
The communications unit 110 may be configured to establish communications between the remote device 100 and the near end device 100 via the network 10. The communications unit 100 may also establish communication between the remote device 100 and the cloud server 20 (illustrated in
The controller 120 may receive display configuration information, such as the display size and resolution. For example, the display configuration information may be stored in the data storage unit 170, and the controller 120 may be connected to the data storage unit 170 to retrieve the display configuration information 172. The controller 120 may be connected to the user interface 180, such as a keyboard, mouse, touch screen, remote control, switches, buttons and the like, to receive a user input 182 from the remote user 104. The remote user 104 may use the user interface 180 to control the operations of the remote device 100. For example, the remote user 104 may use the user interface 180 to manually select the view and sound to be displayed and rendered by the remote device 100. When the user input 182 for manual view and sound selection is received, the controller 120 may cease to automatically control the view and sound selection.
The controller 120 may have a complete control over selection of a view of the video image captured by the near end device 200 to be displayed at the display 140 of the remote device 100 and a sound of the audio captured by the near end device 200 to be rendered by the speakers 160A, 160B of the remote device 100. The controller 120 may also select the view and sound by itself. Alternatively, the controller 120 and the near end device 200 may share the view and sound selection responsibilities, or the view and sound selection may be carried out by the near end device 200.
In an implementation, the controller 120 may receive a plurality of near end video images captured by the near end device 200. Each near end video image may be processed by the near end device 200 to have a different view. The controller 120 may then select one of the plurality of near end video images to display the view selected based on the display configuration information. The sound of the near end audio may be selected in the same manner. For example, the controller 120 may receive a plurality of near end audio streams captured by the near end device 200. Each near end audio stream may be processed by the near end device 200 to have a different sound. The controller 120 may then select one of the plurality of near end video images to render the sound to match the selected view. In this implementation, the view and sound selection is carried out by the controller 120 while the processing of the captured near end video image and audio is mostly done locally at the near end device 100. An example of this implementation is shown in
In another implementation, the controller 120 may receive a near end video image from the near end device 200. The near end video image may contain coordinate data for a plurality of views selectable by the controller 120. The controller 120 may then process the near end video image to generate the view selected based on the display configuration information of the remote device 100. The controller 120 may also receive and process a near end audio captured by the near end device 200 to generate the sound that matches the selected view. In this implementation, the controller 120 may perform both the view and sound selection and the near end video image and audio processing to generate the selected view and sound to be displayed and rendered by the remote device 100. An example of this implementation is shown in
In yet another implementation, the controller 120 may select a view of the near end video image based on the display configuration information and select a sound of near end audio based on the selected view. The controller 120 may then send a control signal 122 to the near end device 200 to indicate the selected view and sound. In response to the control signal 122, the near end device 200 may send the selected view of the near end video image and the selected sound of the near end audio. An example of this implementation is shown in
In yet another implementation, the cloud server 20 may be used to administrate and/or mediate the selection of the near end view and sound for the remote device 100. For example, the cloud server 20 may be connected to one or more remote devices 100 to receive the control data 106 indicating the view and sound selected by each remote device 100. As described above, the different remote devices 100 may require and select different views and sounds. The cloud server 20 may then receive a plurality of near end video data streams and audio data streams. Based on the view and sound selected by each remote device 100, the cloud server 20 may select, from the plurality of near end video data streams and audio data streams, a video steam and an audio stream that match the view and sound selected by each remote device 100 and provide the selected video and audio streams to each remote device 100. Alternatively, the cloud server 20 may receive video and audio streams having coordinate information from the near end device 200 and process the video and audio streams to generate the view and sound selected by each remote device 100. The cloud server 20 may then send video and audio streams of the selected view and sound to each remote device 100.
Referring to
In another implementation, the controller 120 may output the control signal 122 to the communications unit 110. The control signal 122 may include selection information indicating the view selected to be displayed on the display 150 and the sound selected to be rendered by the speakers 160A, 160B. The control signal 122 may be included in control data 106 output from the communications unit 100 to the near end device 200 or cloud server 20. Based on the selection information in the received control data 106, the near end device 200 or cloud server 20 may select the view and sound of the near end site 202 to the remote device 100.
In another implementation, the near end device 100 may include a plurality of cameras 210 having different capturing angles, such as wide, medium and narrow capturing angles X, Y and Z, respectively, in order to capture different views, such as full, fit and narrow views, respectively. The cameras 210 having different view angles may simultaneously capture the near end site 202 to generate a plurality of different views of the near end site 202, which may be sent to the remote device 100 for automatic selection by the controller 120. Alternatively, the near end device 200 may receive control data 106 containing the control signal 122 from the controller 120. Based on the control signal 122, the near end device 20 may capture the near end video image in the view indicated by the control signal 122.
The controller 120 may select the fit views 320, 350 to more prominently display two or more the near end subjects 204. For example, when the fit view 320 is selected, all of the near end subjects 204 may be displayed to fit the display 140. When two neighboring near end subjects 204A, 204B are speaking to each other, the controller 120 may select the fit view 350 to show the near end subjects 204A, 204B to fit the display 140. In an embodiment, the fit views 320, 350 may be generated by cropping and enlarging a portion of the full view 310, which may be performed by the remote device 100 or the near end device 200. Alternatively, the camera 210 may be rotated, tilted, zoomed in or zoomed out to capture the fit views 320, 350.
The narrow views 330, 340 may be selected to more prominently display a single near end subject 204, which may be an active speaker. For example, the landscape narrow view 330 may be selected to prominently display the near end subject 204C on the computer monitor 100B or the projection device having a landscape display. The portrait narrow view 340 may be selected to prominently display the near end subject 204E on the mobile phone or tablet 100A having a portrait display. When the mobile phone or tablet 100A is operated in a landscape display mode, the controller 120 may select the landscape narrow view 330. Similar to the fit views 320, 350, the narrow views 330, 340 may be generated by cropping and enlarging a portion of the full view 310. Alternatively, the narrow views 330, 340 may be captured by rotating, titling and/or zooming in the camera 210.
When the narrow view 340 (illustrated in
In the implementation shown in
In the implementation shown in
During a course of the videoconference or multicasting, an evident may occur and the remote device 100 may need to display a different view. When the controller 120 determines that a different view needs to be displayed (YES at Step 550), the controller 120 may select a new view (Step 560). For example, when the near end subjects 204E is actively speaking, the controller 120 may select the narrow view 340 to prominently display the near end subject 204E. The controller 120 then select a new sound (Step 570). As described above, the new sound may be selected to match the newly selected view (at Step 560). For example, when the narrow view 340 showing the near end subject 204E is selected as the new view, the controller 120 may select the narrow beam sound as the new sound to center the beam sound 410E, as shown in
In another implementation, once the near end view and sound are selected (Steps 520, 530) based on the display configuration, the controller 120 may not change the view and sound. For example, when the display size is too small to display the fit or full view on the remote device 100, such as the mobile device 100A shown in
To determine the angular head width Θ of the remote user 204C, the controller 120 may determine a proportion P1 of the head width WH to a width WW of the window 142 and a proportion P2 of the window width WW to a width WD of the display 140. The rendered head width W may be calculated by multiplying the display width WD (m), the proportion P1 (%) and the proportion P2 (%). The angular head width Θ may be determined from Θ=arctan (W/D). The distance D may be detected by using a depth sensor or camera. Alternatively, the distance D may be estimated based on the type of the remote device 100. For example, the estimated distance D for the mobile phone or tablet 100A (illustrated in
When the angular head width Θ is not less than 40 arcminutes (NO at Step 620), the controller 120 may determine whether the angular head width Θ is greater than 60 arcminutes (Step 625). When the angular head width Θ is not greater than 60 arcminutes (NO at Step 625), the angular head width Θ is determined to be within in the predetermined range and the process 600 may proceed to Step 670. When the angular head width Θ is greater than 60 arcminutes (YES at Step 625), the controller 120 may determine whether the narrow view is currently being displayed (Step 650). When the narrow view is displayed (YES at Step 650), the controller 120 may switch the view to the fit view to decrease the angular head width Θ (Step 652). When the current view is not the narrow view (NO at Step 650), the process 600 may move to Step 670. After switching to the fit view, the controller 120 may check if the angular head width Θ is still greater than 60 arcminutes (Step 660). If the angular head width Θ has become less than 60 arcminutes (NO at Step 660), the process 600 may move to Step 670. If the angular head width Θ is still greater than 60 arcminutes (YES at Step 660), the controller 120 may switch to the full view to further decrease the angular head width Θ (Step 662) and the process 600 may move to Step 670. Hence, the controller 120 may be configured to periodically determine whether the remote device 100 is displaying a view optimized for the specific display configuration of the remote device 100 and automatically select and switch to a new view optimized for the remote device 100.
As such, according to the description, the remote device 100 may have a complete control over how the near end site 202 should be displayed and how the sound of the near end site 202 should be rendered at the remote device 100. Since the view and sound of the near end site 202 may be selected based on its own display configuration, the remote device 100 may provide an improved videoconferencing or multicasting experience to the remote user 104. Also, the sound of the near end site may be selected to match the selected view of the near end site, which may provide more immersive videoconferencing or multicasting experiences. Further, the remote device 100 may periodically check if new view and sound need to be selected and switch to the new view and sound if necessary. Hence, the remote device 100 may provide more engaging and dynamic videoconferencing or multicasting experiences to the remote user 104.
The computer system 1500 may further include a read only memory (ROM) 1508 or other static storage device coupled to the bus 1502 for storing static information and instructions for the processor 1504. A storage device 1510, such as a flash or other non-volatile memory can be coupled to the bus 1502 for storing information and instructions.
The computer system 1500 may be coupled via bus 1502 to a display 1512, such as a liquid crystal display (LCD), for displaying information. One or more user input devices, such as the example user input device 1514 can be coupled to the bus 1502, and can be configured for receiving various user inputs, such as user command selections and communicating these to the processor 1504, or to the main memory 1506. The user input device 1514 can include physical structure, or virtual implementation, or both, providing user input modes or options, for controlling, for example, a cursor, visible to a user through display 1512 or through other techniques, and such modes or operations can include, for example virtual mouse, trackball, or cursor direction keys.
The computer system 1500 can include respective resources of the processor 1504 executing, in an overlapping or interleaved manner, respective program instructions. Instructions may be read into the main memory 1506 from another machine-readable medium, such as the storage device 1510. In some examples, hard-wired circuitry may be used in place of or in combination with software instructions. The term “machine-readable medium” as used herein refers to any medium that participates in providing data that causes a machine to operate in a specific fashion. Such a medium may take forms, including but not limited to, non-volatile media, volatile media, and transmission media. Non-volatile media can include, for example, optical or magnetic disks, such as the storage device 1510. Transmission media can include optical paths, or electrical or acoustic signal propagation paths, and can include acoustic or light waves, such as those generated during radio-wave and infra-red data communications, that are capable of carrying instructions detectable by a physical mechanism for input to a machine.
The computer system 1500 can also include a communication interface 1518 coupled to the bus 1502, for two-way data communication coupling to a network link 1520 connected to a local network 1522. The network link 1520 can provide data communication through one or more networks to other data devices. For example, the network link 1520 may provide a connection through the local network 1522 to a host computer 1524 or to data equipment operated by an Internet Service Provider (ISP) 1526 to access through the Internet 1528 a server 1530, for example, to obtain code for an application program.
While various embodiments have been described, the description is intended to be exemplary, rather than limiting, and it is understood that many more embodiments and implementations are possible that are within the scope of the embodiments. Although many possible combinations of features are shown in the accompanying figures and discussed in this detailed description, many other combinations of the disclosed features are possible. Any feature of any embodiment may be used in combination with or substituted for any other feature or element in any other embodiment unless specifically restricted. Therefore, it will be understood that any of the features shown and/or discussed in the present disclosure may be implemented together in any suitable combination. Accordingly, the embodiments are not to be restricted except in light of the attached claims and their equivalents. Also, various modifications and changes may be made within the scope of the attached claims.
While the foregoing has described what are considered to be the best mode and/or other examples, it is understood that various modifications may be made therein and that the subject matter disclosed herein may be implemented in various forms and examples, and that the teachings may be applied in numerous applications, only some of which have been described herein. It is intended by the following claims to claim any and all applications, modifications and variations that fall within the true scope of the present teachings.
Unless otherwise stated, all measurements, values, ratings, positions, magnitudes, sizes, and other specifications that are set forth in this specification, including in the claims that follow, are approximate, not exact. They are intended to have a reasonable range that is consistent with the functions to which they relate and with what is customary in the art to which they pertain.
The scope of protection is limited solely by the claims that now follow. That scope is intended and should be interpreted to be as broad as is consistent with the ordinary meaning of the language that is used in the claims when interpreted in light of this specification and the prosecution history that follows and to encompass all structural and functional equivalents. Notwithstanding, none of the claims are intended to embrace subject matter that fails to satisfy the requirement of Sections 101, 102, or 103 of the Patent Act, nor should they be interpreted in such a way. Any unintended embracement of such subject matter is hereby disclaimed.
Except as stated immediately above, nothing that has been stated or illustrated is intended or should be interpreted to cause a dedication of any component, step, feature, object, benefit, advantage, or equivalent to the public, regardless of whether it is or is not recited in the claims.
It will be understood that the terms and expressions used herein have the ordinary meaning as is accorded to such terms and expressions with respect to their corresponding respective areas of inquiry and study except where specific meanings have otherwise been set forth herein. Relational terms such as first and second and the like may be used solely to distinguish one entity or action from another without necessarily requiring or implying any actual such relationship or order between such entities or actions. The terms “comprises,” “comprising,” or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. An element proceeded by “a” or “an” does not, without further constraints, preclude the existence of additional identical elements in the process, method, article, or apparatus that comprises the element.
The Abstract of the Disclosure is provided to allow the reader to quickly ascertain the nature of the technical disclosure. It is submitted with the understanding that it will not be used to interpret or limit the scope or meaning of the claims. In addition, in the foregoing Detailed Description, it can be seen that various features are grouped together in various examples for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the claims require more features than are expressly recited in each claim. Rather, as the following claims reflect, inventive subject matter lies in less than all features of a single disclosed example. Thus, the following claims are hereby incorporated into the Detailed Description, with each claim standing on its own as a separately claimed subject matter.
Number | Name | Date | Kind |
---|---|---|---|
3118340 | Iwerks et al. | Jan 1964 | A |
4611342 | Miller et al. | Sep 1986 | A |
4658425 | Julstrom | Apr 1987 | A |
4752961 | Kahn | Jun 1988 | A |
4929824 | Miyazaki | May 1990 | A |
4980761 | Natori | Dec 1990 | A |
5021818 | Satoh et al. | Jun 1991 | A |
5033087 | Bahl et al. | Jul 1991 | A |
5121426 | Baumhauer et al. | Jun 1992 | A |
5206721 | Ashida et al. | Apr 1993 | A |
5264940 | Komiya et al. | Nov 1993 | A |
5271057 | Addeo et al. | Dec 1993 | A |
5281995 | Terashita et al. | Jan 1994 | A |
5307405 | Sih | Apr 1994 | A |
5335011 | Addeo et al. | Aug 1994 | A |
5343246 | Arai et al. | Aug 1994 | A |
5353058 | Takei | Oct 1994 | A |
5382972 | Kannes | Jan 1995 | A |
5404397 | Janse et al. | Apr 1995 | A |
5412738 | Brunelli et al. | May 1995 | A |
5436896 | Anderson et al. | Jul 1995 | A |
5450481 | Penzias | Sep 1995 | A |
5539483 | Nalwa | Jul 1996 | A |
5561668 | Genter | Oct 1996 | A |
5572278 | Hamada et al. | Nov 1996 | A |
5572728 | Tada et al. | Nov 1996 | A |
5587998 | Velardo et al. | Dec 1996 | A |
5625410 | Washino et al. | Apr 1997 | A |
5631697 | Nishimura et al. | May 1997 | A |
5686957 | Baker | Nov 1997 | A |
5710591 | Bruno et al. | Jan 1998 | A |
5737011 | Lukacs | Apr 1998 | A |
5745305 | Nalwa | Apr 1998 | A |
5751285 | Kashiwagi et al. | May 1998 | A |
5767897 | Howell | Jun 1998 | A |
5778082 | Chu et al. | Jul 1998 | A |
5787183 | Chu et al. | Jul 1998 | A |
5790182 | St. Hilaire | Aug 1998 | A |
5793527 | Nalwa | Aug 1998 | A |
5796819 | Romesburg | Aug 1998 | A |
5808663 | Okaya | Sep 1998 | A |
5818514 | Duttweiler et al. | Oct 1998 | A |
5844599 | Hildin | Dec 1998 | A |
5850250 | Konopka et al. | Dec 1998 | A |
5877801 | Martin et al. | Mar 1999 | A |
5959662 | Shaffer et al. | Sep 1999 | A |
5959667 | Maeng | Sep 1999 | A |
5963250 | Parker et al. | Oct 1999 | A |
5990934 | Nalwa | Nov 1999 | A |
5999208 | McNerney et al. | Dec 1999 | A |
5999630 | Iwamatsu | Dec 1999 | A |
6002430 | McCall et al. | Dec 1999 | A |
6005610 | Pingali | Dec 1999 | A |
6005611 | Gullichsen et al. | Dec 1999 | A |
6040860 | Tamura et al. | Mar 2000 | A |
6043837 | Driscoll et al. | Mar 2000 | A |
6072522 | Ippolito et al. | Jun 2000 | A |
6111702 | Nalwa | Aug 2000 | A |
6115176 | Nalwa | Sep 2000 | A |
6118474 | Nayar | Sep 2000 | A |
6128143 | Nalwa | Oct 2000 | A |
6133941 | Ono | Oct 2000 | A |
6141145 | Nalwa | Oct 2000 | A |
6144501 | Nalwa | Nov 2000 | A |
6157403 | Nagata | Dec 2000 | A |
6175454 | Hoogland et al. | Jan 2001 | B1 |
6179426 | Rodriguez et al. | Jan 2001 | B1 |
6192196 | Keller | Feb 2001 | B1 |
6192342 | Akst | Feb 2001 | B1 |
6195204 | Nalwa | Feb 2001 | B1 |
6208963 | Martinez et al. | Mar 2001 | B1 |
6211912 | Shahraray | Apr 2001 | B1 |
6219086 | Murata | Apr 2001 | B1 |
6219089 | Driscoll et al. | Apr 2001 | B1 |
6219090 | Nalwa | Apr 2001 | B1 |
6219639 | Bakis et al. | Apr 2001 | B1 |
6219640 | Basu et al. | Apr 2001 | B1 |
6222683 | Hoogland et al. | Apr 2001 | B1 |
6275258 | Chim | Aug 2001 | B1 |
6285365 | Nalwa | Sep 2001 | B1 |
6285392 | Satoda et al. | Sep 2001 | B1 |
6304285 | Geng | Oct 2001 | B1 |
6313865 | Driscoll et al. | Nov 2001 | B1 |
6331869 | Furlan et al. | Dec 2001 | B1 |
6337708 | Furlan et al. | Jan 2002 | B1 |
6341044 | Driscoll et al. | Jan 2002 | B1 |
6346950 | Jouppi | Feb 2002 | B1 |
6346967 | Gullichsen et al. | Feb 2002 | B1 |
6353662 | Foladare et al. | Mar 2002 | B1 |
6356296 | Driscoll et al. | Mar 2002 | B1 |
6356397 | Nalwa | Mar 2002 | B1 |
6369818 | Hoffman et al. | Apr 2002 | B1 |
6373642 | Wallerstein et al. | Apr 2002 | B1 |
6377918 | Series | Apr 2002 | B1 |
6377919 | Burnett et al. | Apr 2002 | B1 |
6388820 | Wallerstein et al. | May 2002 | B1 |
6392687 | Driscoll et al. | May 2002 | B1 |
6405166 | Huang et al. | Jun 2002 | B1 |
6424377 | Driscoll et al. | Jul 2002 | B1 |
6426774 | Driscoll et al. | Jul 2002 | B1 |
6442272 | Osovets | Aug 2002 | B1 |
6459451 | Driscoll et al. | Oct 2002 | B2 |
6466254 | Furlan et al. | Oct 2002 | B1 |
6469217 | Saito et al. | Oct 2002 | B2 |
6469732 | Chang et al. | Oct 2002 | B1 |
6480229 | Driscoll et al. | Nov 2002 | B1 |
6480300 | Aoyama | Nov 2002 | B1 |
6483532 | Girod | Nov 2002 | B1 |
6493032 | Wallerstein et al. | Dec 2002 | B1 |
6515696 | Driscoll et al. | Feb 2003 | B1 |
6539547 | Driscoll et al. | Mar 2003 | B2 |
6545699 | Satoda | Apr 2003 | B2 |
6549230 | Tosaya | Apr 2003 | B2 |
6577333 | Tai et al. | Jun 2003 | B2 |
6583815 | Driscoll et al. | Jun 2003 | B1 |
6593956 | Potts et al. | Jul 2003 | B1 |
6593969 | Driscoll et al. | Jul 2003 | B1 |
6597520 | Wallerstein et al. | Jul 2003 | B2 |
6600511 | Kaneko et al. | Jul 2003 | B1 |
6611281 | Strubbe | Aug 2003 | B2 |
6611601 | Terada et al. | Aug 2003 | B2 |
6618073 | Lambert et al. | Sep 2003 | B1 |
6628767 | Wellner et al. | Sep 2003 | B1 |
6700711 | Nalwa | Mar 2004 | B2 |
6731334 | Maeng et al. | May 2004 | B1 |
6741250 | Furlan et al. | May 2004 | B1 |
6756990 | Koller | Jun 2004 | B2 |
6757442 | Avinash | Jun 2004 | B1 |
6766035 | Gutta | Jul 2004 | B1 |
6771819 | DeYong et al. | Aug 2004 | B2 |
6774934 | Belt et al. | Aug 2004 | B1 |
6795106 | Cooper | Sep 2004 | B1 |
6826284 | Benesty et al. | Nov 2004 | B1 |
6839067 | Liu et al. | Jan 2005 | B2 |
6842516 | Armbruster | Jan 2005 | B1 |
6844990 | Artonne et al. | Jan 2005 | B2 |
6868157 | Okuda | Mar 2005 | B1 |
6882971 | Craner | Apr 2005 | B2 |
6885509 | Wallerstein et al. | Apr 2005 | B2 |
6904182 | Simard et al. | Jun 2005 | B1 |
6914622 | Smith et al. | Jul 2005 | B1 |
6914989 | Janse et al. | Jul 2005 | B2 |
6922403 | Yoo | Jul 2005 | B1 |
6924832 | Shifter et al. | Aug 2005 | B1 |
6934370 | Leban et al. | Aug 2005 | B1 |
6937266 | Rui et al. | Aug 2005 | B2 |
6937744 | Toyama | Aug 2005 | B1 |
6940540 | Beal et al. | Sep 2005 | B2 |
6950141 | Mori et al. | Sep 2005 | B2 |
6968064 | Ning | Nov 2005 | B1 |
6970796 | Tashev | Nov 2005 | B2 |
6987856 | Feng et al. | Jan 2006 | B1 |
7010119 | Marton et al. | Mar 2006 | B2 |
7020337 | Viola et al. | Mar 2006 | B2 |
7024033 | Li et al. | Apr 2006 | B2 |
7031499 | Viola et al. | Apr 2006 | B2 |
7039199 | Rui | May 2006 | B2 |
7081915 | Hamilton | Jul 2006 | B1 |
7084801 | Balan et al. | Aug 2006 | B2 |
7092001 | Schulz | Aug 2006 | B2 |
7092002 | Ferren et al. | Aug 2006 | B2 |
7099510 | Jones et al. | Aug 2006 | B2 |
7099821 | Visser et al. | Aug 2006 | B2 |
7102669 | Skow | Sep 2006 | B2 |
7113201 | Taylor et al. | Sep 2006 | B1 |
7116791 | Matsuo | Oct 2006 | B2 |
7117157 | Taylor et al. | Oct 2006 | B1 |
7119829 | Leonard et al. | Oct 2006 | B2 |
7123727 | Elko et al. | Oct 2006 | B2 |
7127071 | Rui et al. | Oct 2006 | B2 |
7133530 | Poletti | Nov 2006 | B2 |
7149315 | Johnston et al. | Dec 2006 | B2 |
7158645 | June et al. | Jan 2007 | B2 |
7174029 | Agostinelli et al. | Feb 2007 | B2 |
7197186 | Jones et al. | Mar 2007 | B2 |
7203323 | Tashev | Apr 2007 | B2 |
7206809 | Ludwig et al. | Apr 2007 | B2 |
7212651 | Viola et al. | May 2007 | B2 |
7242763 | Etter | Jul 2007 | B2 |
7242810 | Chang | Jul 2007 | B2 |
7260257 | Zhang et al. | Aug 2007 | B2 |
7305078 | Kardos | Dec 2007 | B2 |
7343289 | Cutler et al. | Mar 2008 | B2 |
7349005 | Rui et al. | Mar 2008 | B2 |
7362368 | Steinberg et al. | Apr 2008 | B2 |
7403231 | Kawanishi | Jul 2008 | B2 |
7428000 | Cutler et al. | Sep 2008 | B2 |
7430333 | Yu et al. | Sep 2008 | B2 |
7454010 | Ebenezer | Nov 2008 | B1 |
7508413 | Nose | Mar 2009 | B2 |
7515172 | Rui et al. | Apr 2009 | B2 |
7538802 | Lee | May 2009 | B2 |
7542600 | Yu et al. | Jun 2009 | B2 |
7559026 | Girish et al. | Jul 2009 | B2 |
7570785 | Breed | Aug 2009 | B2 |
7580054 | Rui et al. | Aug 2009 | B2 |
7586513 | Muren et al. | Sep 2009 | B2 |
7613313 | Juppi et al. | Nov 2009 | B2 |
7636662 | Dimtrova et al. | Dec 2009 | B2 |
7742055 | Phelps | Jun 2010 | B2 |
7778483 | Messina et al. | Aug 2010 | B2 |
7782366 | Imai | Aug 2010 | B2 |
7796831 | Tanaka | Sep 2010 | B2 |
7876367 | Muramatsu | Jan 2011 | B2 |
7948524 | Endo et al. | May 2011 | B2 |
8024189 | Zhang et al. | Sep 2011 | B2 |
8111282 | Cutler et al. | Feb 2012 | B2 |
8189807 | Cutler | May 2012 | B2 |
8219387 | Cutler et al. | Jul 2012 | B2 |
8234113 | Zhang et al. | Jul 2012 | B2 |
8248448 | Feng et al. | Aug 2012 | B2 |
8300080 | Cutler | Oct 2012 | B2 |
8314829 | Cutler | Nov 2012 | B2 |
8315366 | Basart et al. | Nov 2012 | B2 |
8465416 | Kitano et al. | Jun 2013 | B2 |
8510110 | Zhang et al. | Aug 2013 | B2 |
8526632 | Cutler | Sep 2013 | B2 |
8576272 | Wu et al. | Nov 2013 | B2 |
8614734 | Cutler | Dec 2013 | B2 |
8717402 | Cutler | May 2014 | B2 |
8749650 | Cutler | Jun 2014 | B2 |
8773499 | Watson et al. | Jul 2014 | B2 |
9071895 | Cutler | Jun 2015 | B2 |
9307191 | Berrett et al. | Apr 2016 | B2 |
9621795 | Whyte et al. | Apr 2017 | B1 |
9648346 | Zhang et al. | May 2017 | B2 |
9883140 | Johnson et al. | Jan 2018 | B2 |
9980040 | Whyte et al. | May 2018 | B2 |
20010028719 | Hayashi | Oct 2001 | A1 |
20010046309 | Kamei | Nov 2001 | A1 |
20010052422 | Kobayashi et al. | Dec 2001 | A1 |
20020033791 | Arakawa | Mar 2002 | A1 |
20020034020 | Wallerstein et al. | Mar 2002 | A1 |
20020052735 | Nagano et al. | May 2002 | A1 |
20020063802 | Gullichsen et al. | May 2002 | A1 |
20020094132 | Hoffman et al. | Jul 2002 | A1 |
20020097885 | Birchfield et al. | Jul 2002 | A1 |
20020101505 | Gutta et al. | Aug 2002 | A1 |
20020105598 | Tai et al. | Aug 2002 | A1 |
20020140804 | Colmenarez et al. | Oct 2002 | A1 |
20020154417 | Wallerstein et al. | Oct 2002 | A1 |
20020172350 | Edwards et al. | Nov 2002 | A1 |
20020184012 | Burnett et al. | Dec 2002 | A1 |
20020191071 | Rui et al. | Dec 2002 | A1 |
20020196327 | Rui et al. | Dec 2002 | A1 |
20030018475 | Basu et al. | Jan 2003 | A1 |
20030037109 | Newman et al. | Feb 2003 | A1 |
20030043260 | Yap et al. | Mar 2003 | A1 |
20030059061 | Tsuji et al. | Mar 2003 | A1 |
20030088832 | Agostinelli et al. | May 2003 | A1 |
20030110147 | Li et al. | Jun 2003 | A1 |
20030118200 | Beaucoup et al. | Jun 2003 | A1 |
20030123659 | Forstrom et al. | Jul 2003 | A1 |
20030142402 | Carbo et al. | Jul 2003 | A1 |
20030160862 | Charlier et al. | Aug 2003 | A1 |
20030174146 | Kenoyer | Sep 2003 | A1 |
20030184645 | Biegelsen et al. | Oct 2003 | A1 |
20030185424 | Sato et al. | Oct 2003 | A1 |
20030187632 | Menich | Oct 2003 | A1 |
20030193606 | Driscoll et al. | Oct 2003 | A1 |
20030193607 | Driscoll et al. | Oct 2003 | A1 |
20030198328 | Li | Oct 2003 | A1 |
20030220971 | Kressin | Nov 2003 | A1 |
20030234772 | Zhang et al. | Dec 2003 | A1 |
20040001137 | Cutler et al. | Jan 2004 | A1 |
20040008407 | Wallerstein et al. | Jan 2004 | A1 |
20040008423 | Driscoll et al. | Jan 2004 | A1 |
20040021764 | Driscoll et al. | Feb 2004 | A1 |
20040032796 | Chu et al. | Feb 2004 | A1 |
20040101038 | Etter | May 2004 | A1 |
20040120510 | Leblanc | Jun 2004 | A1 |
20040170164 | Leblanc et al. | Sep 2004 | A1 |
20040172255 | Aoki et al. | Sep 2004 | A1 |
20040201698 | Keenan et al. | Oct 2004 | A1 |
20040228215 | Ichikawa et al. | Nov 2004 | A1 |
20040228474 | Taniguchi et al. | Nov 2004 | A1 |
20040243416 | Gardos | Dec 2004 | A1 |
20040252384 | Wallerstein et al. | Dec 2004 | A1 |
20040252845 | Tashev | Dec 2004 | A1 |
20040254982 | Hoffman et al. | Dec 2004 | A1 |
20040263636 | Cutler et al. | Dec 2004 | A1 |
20040267521 | Cutler et al. | Dec 2004 | A1 |
20050008169 | Muren et al. | Jan 2005 | A1 |
20050018828 | Nierhaus et al. | Jan 2005 | A1 |
20050046703 | Cutler | Mar 2005 | A1 |
20050057666 | Hu et al. | Mar 2005 | A1 |
20050117015 | Cutler | Jun 2005 | A1 |
20050117034 | Cutler | Jun 2005 | A1 |
20050140779 | Schulz et al. | Jun 2005 | A1 |
20050151837 | Cutler | Jul 2005 | A1 |
20050157866 | Marton et al. | Jul 2005 | A1 |
20050175216 | Bloom et al. | Aug 2005 | A1 |
20050178953 | Worthington et al. | Aug 2005 | A1 |
20050228673 | Nefian et al. | Oct 2005 | A1 |
20050246167 | Nakajima et al. | Nov 2005 | A1 |
20050262201 | Rudolph et al. | Nov 2005 | A1 |
20050265562 | Rui | Dec 2005 | A1 |
20050265607 | Chang | Dec 2005 | A1 |
20050276270 | Buinevicius et al. | Dec 2005 | A1 |
20050278171 | Suppappola et al. | Dec 2005 | A1 |
20050280700 | Rui et al. | Dec 2005 | A1 |
20050285933 | Rui et al. | Dec 2005 | A1 |
20050285943 | Cutler | Dec 2005 | A1 |
20050286724 | Yamada et al. | Dec 2005 | A1 |
20060017805 | Rodman | Jan 2006 | A1 |
20060023075 | Cutler | Feb 2006 | A1 |
20060075422 | Choi et al. | Apr 2006 | A1 |
20060133619 | Curry | Jun 2006 | A1 |
20060146177 | Cutler | Jul 2006 | A1 |
20060147063 | Chen | Jul 2006 | A1 |
20060204023 | Stinson et al. | Sep 2006 | A1 |
20060221915 | Gatta et al. | Oct 2006 | A1 |
20060233353 | Beaucoup et al. | Oct 2006 | A1 |
20060239443 | Oxford et al. | Oct 2006 | A1 |
20070035632 | Silvernail et al. | Feb 2007 | A1 |
20070036343 | Sudo et al. | Feb 2007 | A1 |
20070047442 | Snyder | Mar 2007 | A1 |
20070058879 | Cutler et al. | Mar 2007 | A1 |
20070127819 | Lee et al. | Jun 2007 | A1 |
20070136053 | Ebenezer | Jun 2007 | A1 |
20070160357 | Lin et al. | Jul 2007 | A1 |
20070206875 | Ida et al. | Sep 2007 | A1 |
20070263849 | Stokes et al. | Nov 2007 | A1 |
20070266092 | Schweitzer, III | Nov 2007 | A1 |
20070297682 | Zhang et al. | Dec 2007 | A1 |
20080037802 | Posa et al. | Feb 2008 | A1 |
20080069403 | Breed | Mar 2008 | A1 |
20080089552 | Nakamura et al. | Apr 2008 | A1 |
20080107281 | Togami et al. | May 2008 | A1 |
20080112598 | Gabara | May 2008 | A1 |
20080127260 | Hong et al. | May 2008 | A1 |
20080143819 | Rodman et al. | Jun 2008 | A1 |
20080170717 | Liu et al. | Jul 2008 | A1 |
20080184124 | Agarwal et al. | Jul 2008 | A1 |
20080232608 | Ullmann | Sep 2008 | A1 |
20080247565 | Elko et al. | Oct 2008 | A1 |
20080255840 | Cutler | Oct 2008 | A1 |
20080273683 | Cohen et al. | Nov 2008 | A1 |
20080292140 | Morris et al. | Nov 2008 | A1 |
20090002476 | Cutler | Jan 2009 | A1 |
20090002477 | Cutler | Jan 2009 | A1 |
20090002480 | Cutler | Jan 2009 | A1 |
20090003678 | Cutler | Jan 2009 | A1 |
20090046139 | Cutler et al. | Feb 2009 | A1 |
20090073275 | Awazu | Mar 2009 | A1 |
20090150149 | Culter et al. | Jun 2009 | A1 |
20090322915 | Cutler | Dec 2009 | A1 |
20090323981 | Cutler | Dec 2009 | A1 |
20100039497 | Cutler | Feb 2010 | A1 |
20100123770 | Friel et al. | May 2010 | A1 |
20100142815 | Sim | Jun 2010 | A1 |
20100150360 | Beaucoup | Jun 2010 | A1 |
20100245624 | Beaucoup | Sep 2010 | A1 |
20110085017 | Robinson et al. | Apr 2011 | A1 |
20110090311 | Fang et al. | Apr 2011 | A1 |
20110096140 | Samadani et al. | Apr 2011 | A1 |
20110164141 | Tico et al. | Jul 2011 | A1 |
20110242277 | Do et al. | Oct 2011 | A1 |
20110249190 | Nguyen et al. | Oct 2011 | A1 |
20110313766 | Zhang et al. | Dec 2011 | A1 |
20120026277 | Malzbender et al. | Feb 2012 | A1 |
20120038627 | Sung et al. | Feb 2012 | A1 |
20120050458 | Mauchly et al. | Mar 2012 | A1 |
20120216129 | Ng et al. | Aug 2012 | A1 |
20120218371 | Cutler | Aug 2012 | A1 |
20120262536 | Chen et al. | Oct 2012 | A1 |
20120274736 | Robinson et al. | Nov 2012 | A1 |
20120278077 | Zhang et al. | Nov 2012 | A1 |
20120327179 | Watson et al. | Dec 2012 | A1 |
20130027506 | Cutler | Jan 2013 | A1 |
20130063539 | Sakuraba | Mar 2013 | A1 |
20130271559 | Feng | Oct 2013 | A1 |
20140133665 | Xiang et al. | May 2014 | A1 |
20140184727 | Xiao et al. | Jul 2014 | A1 |
20140205270 | Kelly et al. | Jul 2014 | A1 |
20150116451 | Xu et al. | Apr 2015 | A1 |
20150189455 | Donaldson | Jul 2015 | A1 |
20160219248 | Reznik | Jul 2016 | A1 |
20170195815 | Christoph | Jul 2017 | A1 |
20170201825 | Whyte et al. | Jul 2017 | A1 |
20180165520 | Meisser | Jun 2018 | A1 |
Number | Date | Country |
---|---|---|
101072332 | Nov 2007 | CN |
101282452 | Oct 2008 | CN |
201426153 | Mar 2010 | CN |
0821522 | Jan 1998 | EP |
1377041 | Jan 2004 | EP |
1443498 | Aug 2004 | EP |
1587294 | Oct 2005 | EP |
1677535 | Jul 2006 | EP |
S5972295 | Apr 1984 | JP |
H06276514 | Sep 1994 | JP |
H08125835 | May 1996 | JP |
H09307870 | Nov 1997 | JP |
H10304329 | Nov 1998 | JP |
2001296343 | Oct 2001 | JP |
2003304589 | Oct 2003 | JP |
2004032782 | Jan 2004 | JP |
2005523667 | Aug 2005 | JP |
2007005969 | Jan 2007 | JP |
20000044065 | Jul 2000 | KR |
20000044079 | Jul 2000 | KR |
200354365 | Jun 2004 | KR |
9607979 | Mar 1996 | WO |
9847291 | Jan 1999 | WO |
0013419 | Mar 2000 | WO |
0137552 | May 2001 | WO |
2005122632 | Dec 2005 | WO |
2006030995 | Mar 2006 | WO |
2007003061 | Jan 2007 | WO |
2009006004 | Jan 2009 | WO |
2017195514 | Nov 2017 | WO |
2018025458 | Feb 2018 | WO |
2018075090 | Apr 2018 | WO |
Entry |
---|
“Acoustic Echo Canceller Data Sheet”, In Miket Dsp Solutions Data Sheet, Oct. 3, 2004, 49 Pages. |
“Acoustic Echo Canceller Demo Using the Block Frequency Domain Adaptive Filter Algorithm”, Retrieved From: https://web.archive.org/web/20070407005509/https://www.dspalgorithms.com/bfdafaec/bfdafaec11.html, Retrieved on: Apr. 7, 2007, 8 Pages. |
“EF400”, Retrieved From: https://web.archive.org/web/20070205093444/http://www.aspi.com/products/echofree/ef400.html, Retrieved on: Feb. 5, 2007, 2 Pages. |
“Microphone Array Project in Microsoft Research”, Retrieved From: https://web.archive.org/web/20071215030921/http://research.microsoft.com/users/ivantash/MicrophoneArrayProject.aspx, Retrieved on: Nov. 5, 2007, 3 Pages. |
“Microsoft Eyes Future of Teleconferencing With RoundTable”, Retrieved From: http://ducknetweb.blogspot.com/2007/09/microsoft-eyes-future-of.html, Sep. 13, 2007, 6 Pages. |
“Microsoft RoundTable”, Retrieved From: http://www.fullview.com/Microsoft_RoundTable_Datasheet.pdf, Aug. 16, 2007, 2 Pages. |
“Revolabs Solo Desktop Wireless Conference Microphone System”, Retrieved From: https://web.archive.org/web/20080519230321/http://www.goelectronic.com/Merchant2/merchant.mvc?Screen=PROD&Product_Code=REVOLABS+SOLO+DESKTOP, Retrieved on: May 19, 2008, 2 Pages. |
“Office Action Issued in European Patent Application No. 078356981”, dated Mar. 22, 2011, 4 Pages. |
“Search Report Issued in European Patent Application No. 07835698.7”, dated Jul. 22, 2010, 8 Pages. |
“Office Action Issued in European Patent Application No. 08769995.5”, dated Jan. 23, 2012, 4 Pages. |
“Search Report Issued in European Patent Application No. 08769995.5”, dated Feb. 10, 2011, 5 Pages. |
“Search Report Issued in European Patent Application No. 08771122.2”, dated Jun. 25, 2014, 7 Pages. |
“Final Office Action Issued in U.S. Appl. No. 09/681,843”, dated Jun. 18, 2003, 15 Pages. |
“Final Office Action Issued in U.S. Appl. No. 09/681,843”, dated Dec. 28, 2004, 14 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 09/681,843”, dated Nov. 10, 2003, 15 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 09/681,843”, dated May 6, 2004, 16 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 09/681,843”, dated Jul. 5, 2002, 12 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 09/681,843”, dated Dec. 31, 2002, 12 Pages. |
“Final Office Action Issued in U.S. Appl. No. 10/608,313”, dated Jul. 5, 2007, 20 Pages. |
“Non-Final Office Action Issued in U.S. Appl. No. 10/608,313”, dated Feb. 8, 2007, 16 Pages. |
“Non-Final Office Action Issued in U.S. Appl. No. 10/608,313”, dated Sep. 28, 2007, 21 Pages. |
“Notice of Allowance Issued in Korean Patent Application No. 10-2008-7030540”, dated Sep. 27, 2013, 3 Pages. |
“Office Action Issued in Korean Patent Application No. 10-2008-7030540”, dated Apr. 19, 2013, 4 Pages. |
“Notice of Allowance Issued in Korean Patent Application No. 10-2009-7026719”, dated Nov. 20, 2014, 3 Pages. |
“Office Action Issued in Korean Patent Application No. 10-2009-7026719”, dated May 30, 2014, 4 Pages. |
“Notice of Allowance Issued in Korean Patent Application No. 10-2009-7027254”, dated Nov. 26, 2014, 3 Pages. |
“Office Action Issued in Korean Patent Application No. 10-2009-7027254”, dated Jun. 18, 2014, 4 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 11/193,702”, dated May 23, 2008, 18 Pages. |
“Ex-parte Quayle Action Issued in U.S. Appl. No. 11/193,976”, dated Jan. 12, 2009, 6 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 11/425,967”, dated Apr. 21, 2010, 8 Pages. |
“Non-Final Office Action Issued in U.S. Appl. No. 11/425,967”, dated Dec. 6, 2010, 8 pages. |
“Final Office Action Issued in U.S. Appl. No. 11/735,674”, dated Apr. 5, 2011, 18 pages. |
“Non Final Office Action Issued in U.S. Appl. No. 11/735,674”, dated Oct. 27, 2010, 20 Pages. |
“Non-Final Office Action Issued in U.S. Appl. No. 11/770,713”, dated Nov. 4, 2011, 13 Pages. |
“Non-Final Office Action Issued in U.S. Appl. No. 11/770,713”, dated Oct. 3, 2012, 11 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 11/771,786”, dated Jun. 6, 2012, 8 Pages. |
“Final Office Action Issued in U.S. Appl. No. 11/771,802”, dated May 20, 2011, 26 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 11/771,802”, dated Nov. 9, 2010, 25 Pages. |
“Non-Final Office Action Issued in U.S. Appl. No. 11/824,412”, dated Feb. 9, 2012, 7 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 11/953,764”, dated Jan. 19, 2012, 6 Pages. |
“Notice of Allowance Issued in U.S. Appl. No. 12/163,451”, dated Jan. 27, 2012, 9 Pages. |
“Non-Final Office Action Issued in U.S. Appl. No. 12/190,585”, dated Feb. 24, 2012, 11 Pages. |
“Notice of Allowance Issued in U.S. Appl. No. 12/190,585”, dateed Jul. 16, 2012, 9 Pages. |
“Non-Final Office Action Issued in U.S. Appl. No. 12/191,270”, dated Apr. 28, 2011, 17 pages. |
“Notice of Allowance Issued in U.S. Appl. No. 12/191,270”, dated Sep. 21, 2011, 10 pages. |
“Notice of Allowance Issued in U.S. Appl. No. 13/168,656”, dated Mar. 6, 2014, 6 Pages. |
“Non Final Office Action Issued in U.S. Appl. No. 13/221,640”, dated Nov. 3, 2011, 12 Pages. |
“Non-Final Office Action Issued in U.S. Appl. No. 13/546,153”, dated Oct. 18, 2012, 12 Pages. |
“Non-Final Office Action Issued in U.S. Appl. No. 13/644,866”, dated Apr. 16, 2013, 8 Pages. |
“Final Office Action Issued in U.S. Appl. No. 13/681,075”, dated Feb. 24, 2015, 5 Pages. |
“Second Written Opinion Issued in PCT Application No. PCT/US2016/068612”, dated Nov. 15, 2017, 5 Pages. |
Rick, et al., “Integration of Automatic Microphone Selection With Acoustic Echo Cancellation”, Retrieved From: http://www.aes.org/e-lib/browse.cfm?elib=10220, Oct. 1996, 1 Page. |
Rui, et al., “Sound Source Localization for Circular Arrays of Directional Microphones”, In Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing, Mar. 23, 2005, 4 Pages. |
Rui, et al., “Viewing Meetings Captured by an Omni-Directional Camera”, In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Mar. 31, 2001, 8 Pages. |
Schapire, Robert E., “The Boosting Approach to Machine Learning: An overview”, In Proceedings of MSRI Workshop on Nonlinear Estimation and Classification, Dec. 19, 2001, pp. 1-23. |
Szeliski, et al., “Creating Full View Panoramic Image Mosaics and Environment Maps”, In Proceedings of the 24th Annual Conference on Computer graphics and Interactive Techniques, Aug. 3, 1997, 8 Pages. |
Tapia, et al., “Concept and Partial Prototype Video: Ubiquitous Video Communication With the Perception of Eye contact”, Retrieved From: http://www.ccs.neu.edu/home/intille/papers-files/TapiaIntilleRebulaStoddard03.pdf, Jan. 2003, 2 Pages. |
Valin, et al., “Robust Sound Source Localization Using a Microphone Array on a Mobile Robot”, In Proceedings of International Conference on Intelligent Robots and Systems, vol. 2, Oct. 27, 2003, 6 Pages. |
Vaseghi, Saeed V., “Echo Cancellation”, In Book Advanced Signal Processing and Digital Noise Reduction, Jan. 2000, pp. 1-20. |
Viola, et al., “Fast and Robust Classification Using Asymmetric AdaBoost and a Detector Cascade”, In Proceedings of the 14th International Conference on Neural Information Processing Systems: Natural and Synthetic, Dec. 3, 2001, 8 Pages. |
Viola, et al., “Multiple Instance Boosting for Object Detection”, In Proceedings of Advances in Neural Information Processing Systems, Jan. 2006, 8 Pages. |
Viola, et al., “Robust Real-Time Face Detection”, In Proceedings of International Journal of computer vision, vol. 57, Issue 2, May 1, 2004, pp. 137-154. |
Wang, et al., “Voice Source Localization for Automatic Camera Pointing System in Videoconferencing”, In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing, vol. 1, Apr. 22, 1997, pp. 187-190. |
Wark, et al., “The Use of Temporal Speech and Lip Information for Multi-Modal Speaker Identification Via Multi-Stream HMM's”, In Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing, Jun. 5, 2000, pp. 2389-2392. |
Wilson, et al., “Combining Multiple Depth Cameras and Projectors for Interactions on, Above, and Between Surfaces”, In Proceedings of the 23rd Annual ACM Symposium on User Interface Software and Technology, Oct. 3, 2010, pp. 273-282. |
Yokoe, et al., “Audio-Based Estimation of Speakers Directions for Multimedia Meeting Logs”, In Proceedings of 2007 IEEE International Conference on Multimedia and Expo, Jul. 2, 2007, pp. 212-215. |
Yoshimi, Billibon H., “A Multimodal Speaker Detection and Tracking System for Teleconferencing”, In Proceedings of the Tenth ACM International Conference on Multimedia, Dec. 1, 2002, pp. 427-428. |
Zhang, et al., “Boosting-Based Multimodal Speaker Detection for Distributed Meetings”, In Proceedings of IEEE Eighth Workshop on Multimedia Signal Processing, Oct. 3, 2006, 6 Pages. |
Zhang, et al., “Maximum Likelihood Sound Source Localization for Multiple Directional Microphones”, In Proceedings of 2007 IEEE International Conference on Acoustics, Speech and Signal Processing, Apr. 15, 2007, 4 Pages. |
“Non-Final Office Action Issued in U.S. Appl. No. 13/681,075”, dated Jun. 20, 2014, 8 Pages. |
“Non-Final Office Action Issued in U.S. Appl. No. 13/681,075”, dated Oct. 24, 2014, 7 Pages. |
“Notice of Allowance Issued in U.S. Appl. No. 14/991,847”, dated Nov. 29, 2016, 11 Pages. |
“Ex-parte Quayle Action Issued in U.S. Appl. No. 15/441,793”, dated Oct. 20, 2017, 8 Pages. |
“First Office Action Issued in Chinese Patent Application No. 200780022873.0”, dated Oct. 12, 2010, 7 Pages. |
“Third Office Action Issued in Chinese Patent Application No. 200780022873.0”, dated May 3, 2012, 6 Pages. |
“Notice of Allowance Issued in Japanese Patent Application No. 2010-514923”, dated Oct. 19, 2012, 6 Pages. |
“Notice of Allowance Issued in Japanese Patent Application No. 2010-514961”, dated Jun. 14, 2013, 6 Pages. |
“Office Action Issued in Japanese Patent Application No. 2010-514961”, dated Jan. 25, 2013, 6 Pages. |
“Office Action Issued in Japanese Patent Application No. 2010-514961”, dated Aug. 3, 2012, 6 Pages. |
“Office Action Issued in Canadian Patent Application No. 2653278”, dated Dec. 3, 2013, 3 Pages. |
“Office Action Issued in Indian Patent Application No. 6300/CHENP/2008”, dated Oct. 21, 2016, 8 Pages. |
Asano, et al., “Detection and Separation of Speech Event Using Audio and Video Information Fusion and Its Application to Robust Speech Interface”, In EURASIP Journal on Advances in Signal Processing, Dec. 1, 2004, pp. 1727-1738. |
Besson, et al., “Information Theoretic Optimization of Audio Features for Multimodal Speaker Detection”, In Technical Report No. 08/2005, Feb. 2005, 19 Pages. |
Busso, et al., “Smart Room: Participant and Speaker Localization and Identification”, In Proceedings of IEEE International Conference on Acoustics, Speech, and Signal Processing, Mar. 18, 2005, 4 Pages. |
Capin, et al., “Camera-Based Virtual Environment Interaction on Mobile Devices”, In Proceedings of International Symposium on Computer and Information Sciences, Nov. 1, 2006, 9 Pages. |
Chang, et al., “Automatic Head-Size Equalization in Panorama Images for Video Conferencing”, In Technical Report, MSR-TR-2005-48, May 2005, 42 Pages. |
Chaudhari, et al., “Information Fusion and Decision Cascading for Audio-Visual Speaker Recognition Based on Time-Varying Stream Reliability Prediction”, In Proceedings of International Conference on Multimedia and Expo, Jul. 6, 2003, pp. 9-12. |
Cutler, et al., “Distributed Meetings: A Meeting Capture and Broadcasting System”, In Proceedings of the Tenth ACM International Conference on Multimedia, Dec. 1, 2002, pp. 503-512. |
Eneman, et al., “Real-Time Implementation of an Acoustic Echo Canceller”, In Proceedings of the COST#254, Workshop on Intelligent Communications, Jun. 4, 1998, 7 Pages. |
Gaspar, et al., “New Depth from Focus Method for 3D PTZ Camera Target Tracking”, Retrieved From: http://www.dem.ist.utl.pt/poliveira/Courses/dsfps0910/docs/prts/DSFPS2010%20-%20DFF%20-%20TG.pdf, Retrieved on: Apr. 4, 2011, 4 Pages. |
Green, et al., “Panocam: Combining Panoramic Video with Acoustic Beamforming for Videoconferencing”, In Proceedings of the Canadian Acoustic Association Conference, Oct. 6, 2004, 3 pages. |
Hampapur, et al., “Face Cataloger: Multi-Scale Imaging for Relating Identity to Location”, In Proceedings of the IEEE Conference on Advanced Video and Signal Based Surveillance, Jul. 21, 2003, 8 Pages. |
He, et al., “Exploring Benefits of Non-Linear Time Compression”, In Proceedings of the Ninth ACM International Conference on Multimedia, Sep. 30, 2001, pp. 382-391. |
He, Li-Wei, “Why Take Notes, Use the Whiteboard Capture System”, In Microsoft Technical Report, MSR-TR-2002-89, Sep. 1, 2002, 11 Pages. |
Hu, et al., “Principal Axis-Based Correspondence Between Multiple Cameras for People Tracking”, In Proceedings of IEEE Transactions on Pattern Analysis and Machine Intelligence vol. 28, Issue 4, Apr. 2006, pp. 663-671. |
Hughes, et al., “The DICE Multi-Site Satellite Videoconference System”, Retrieved From: http://www.esaint/esapub/bulletin/bullet80/hughe80.htm, Retrieved on: Nov. 5, 2007, 7 Pages. |
Jiang, et al., “Adaptive Noise Reduction of Speech Signals”, In Microsoft Technical Report, MSR-TR-2000-86, Jul. 1, 2000, 9 Pages. |
Kapralos, et al., “Audio-Visual Localization of Multiple Speakers in a Video Teleconferencing Setting”, In International Journal of Imaging Systems and Technology, vol. 13, Issue 1, Jun. 2, 2003, pp. 1-32. |
Kozielski, et al., “Online Speaker Recognition for Teleconferencing Systems”, In Technical Report, Technical University of Munich, Apr. 14, 2014, 67 Pages. |
Lathoud, et al., “AV16.3: An Audio-Visual Corpus for Speaker Localization and Tracking”, In International Workshop on Machine Learning for Multimodal Interaction, Jun. 21, 2004, pp. 182-195. |
Li, et al “Learning to Detect Multi-View Faces in Real-Time”, In Proceedings of 2nd International Conference on Development and Learning, Jun. 12, 2002, pp. 1-6. |
Lindstrom, et al., “A Hybrid Acoustic Echo Canceller and Suppressor”, In Journal of Signal Processing, vol. 87, Issue 4, Apr. 1, 2007, pp. 739-749. |
Lindstrom, “Improving the Performance of a Low-Complexity Doubletalk Detector by a Subband Approach”, In Proceedings of International Conference on Systems, Signals and Devices, Mar. 21, 2005, 5 Pages. |
Liu, et al. “Automating Camera Management for Lecture Room Environments”, In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Mar. 31, 2001, 8 Pages. |
Liu, et al., “FLYSPEC: A Multi-User Video Camera System With Hybrid Human and Automatic Control”, In Proceedings of the Tenth ACM International Conference on Multimedia., Dec. 1, 2002, pp. 484-492. |
Mansoori, et al., “Solving Infinite-Horizon Optimal Control Problems Using Haar Wavelet Collocation Method”, In Journal of the Australian & New Zealand Industrial & Applied Mathematics, Mar. 18, 2014, pp. 1-5. |
Marti, et al., “Real Time Speaker Localization and Detection System for Camera Steering in Multiparticipant Videoconferencing Environments”, In IEEE International Conference on Acoustics, Speech and Signal Processing, May 27, 2011, pp. 2592-2595. |
MEIR, et al., “An Introduction to Boosting and Leveraging”, In Proceedings of Advanced Lectures on Machine Learning, Jan. 1, 2003, pp. 118-183. |
Minotto, et al., “Simultaneous-Speaker Voice Activity Detection and Localization Using Mid-Fusion of SVM and HMMs”, In Proceedings of IEEE Transactions on Multimedia, vol. 16, Issue 4,, Jun. 2014, pp. 1032-1044. |
Mondal, et al., “SureCall: Towards Glitch-Free Real-Time AudioNideo Conferencing”, In Proceedings of 18th International Workshop on Quality of Service, Jun. 16, 2010, 9 Pages. |
Nanda, et al., “Practical Calibrations for a Real-Time Digital Omnidirectional Camera”, In Proceedings of Computer Vision and Pattern Recognition Technical Sketches, vol. 20, Dec. 2001, pp. 1-4. |
Nguyen, et al., “More Than Face-To-Face Empathy Effects of Video Framing”, In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, Apr. 6, 2009, pp. 423-432. |
Oh, et al., “The ubiTV Application for a Family in ubiHome”, In Proceedings of 2nd Ubiquitous Home Workshop, Dec. 6, 2005, 10 Pages. |
Pavlovic, et al., “Multimodal Speaker Detection Using Error Feedback Dynamic Bayesian Networks”, In Proceedings IEEE Conference on Computer Vision and Pattern Recognition, Jun. 15, 2000, 8 Pages. |
“International Search Report and Written Opinion Issued in PCT Application No. PCT/US2007/003715”, dated Jan. 31, 2008, 12 Pages. |
“International Search Report and Written Opinion Issued in PCT Application No. PCT/US2008/065579”, dated Jan. 6, 2009, 10 Pages. |
“International Search Report and Written Opinion Issued in PCT Application No. PCT/US2008/067035”, dated Oct. 29, 2008, 10 Pages. |
“International Preliminary Report on Patentability Issued in PCT Application No. PCT/US2016/068612”, dated Mar. 21, 2018, 7 Pages. |
“International Search Report and Written Opinion Issued in PCT Application No. PCT/US2016/068612”, dated Mar. 16, 2017, 13 Pages. |
“Office Action and Search Report Issued in Chinese Patent Application No. 201680070404.5”, dated Oct. 29, 2019, 17 Pages. |
Cooperstock, Jeremy R.., “Multimodal Telepresence Systems”, In Proceedings of IEEE Signal Processing Magazine vol. 28 , Issue: 1, Jan. 1, 2011, pp. 77-86. |
“International Search Report and Written Opinion Issued in PCT Application No. PCT/US2019/032328”, dated Aug. 5, 2019, 15 Pages. |
“Office Action Issued in European Patent Application No. 16826846.4”, dated Aug. 20, 2019, 6 Pages. |
Jiang, et al., “A Simple Microphone Array for Source Direction and Distance Estimation”, In Proceedings of 6th IEEE Conference on Industrial Electronics and Applications, Jun. 21, 2011, 4 Pages. |
“Summon to Attend Oral Proceedings Issued in European Patent Application No. 16826846.4”, dated Apr. 3, 2020, 10 Pages. |
“Office Action and Search Report Issued in Chinese Patent Application No. 201680070404.5”, dated May 26, 2020, 18 Pages. |
“Industrial Robot”, Published by Southwest Jiaotong University, Jan. 31, 2011, 16 Pages. |
“Third Office Action Issued in Chinese Patent Application No. 201680070404.5”, dated Nov. 4, 2020, 10 Pages. |
Number | Date | Country | |
---|---|---|---|
20190373216 A1 | Dec 2019 | US |