The disclosure is related to consumer goods and, more particularly, to methods, systems, products, features, services, and other elements directed to media playback or some aspect thereof.
Options for accessing and listening to digital audio in an out-loud setting were limited until in 2003, when SONOS, Inc. filed for one of its first patent applications, entitled “Method for Synchronizing Audio Playback between Multiple Networked Devices,” and began offering a media playback system for sale in 2005. The Sonos Wireless HiFi System enables people to experience music from many sources via one or more networked playback devices. Through a software control application installed on a smartphone, tablet, or computer, one can play what he or she wants in any room that has a networked playback device. Additionally, using the controller, for example, different songs can be streamed to each room with a playback device, rooms can be grouped together for synchronous playback, or the same song can be heard in all rooms synchronously.
Given the ever growing interest in digital media, there continues to be a need to develop consumer-accessible technologies to further enhance the listening experience.
Features, aspects, and advantages of the presently disclosed technology may be better understood with regard to the following description, appended claims, and accompanying drawings where:
The drawings are for the purpose of illustrating example embodiments, but it is understood that the inventions are not limited to the arrangements and instrumentality shown in the drawings.
I. Overview
Listening to media content out loud can be a social activity that involves family, friends, and guests. Media content may include, for instance, talk radio, books, audio from television, music stored on a local drive, music from media sources (e.g. Pandora® Radio, Spotify®, Slacker®, Radio, Google Play™, iTunes Radio), and other audible material. In a household, for example, people may play music out loud at parties and other social gatherings. In such an environment, people may wish to play the music in one listening zone or multiple listening zones simultaneously, such that the music in each listening zone may be synchronized, without audible echoes or glitches. Such an experience may be further enriched when people can use voice commands to control an audio playback device or system. For example, a person may wish to change the audio content, playlist, or listening zone, add a music track to a playlist or playback queue, or change a playback setting (e.g. play, pause, next track, previous track, playback volume, and EQ settings, among others).
Listening to media content out loud can also be an individual experience. For example, an individual may play music out loud for themselves in the morning before work, during a workout, in the evening during dinner, or at other times throughout the day at home or at work. For these individual experiences, the individual may choose to limit the playback of audio content to a single listening zone or area. Such an experience may be further enriched when an individual can use a voice command to choose a listening zone, audio content, and playback settings, among other settings.
Identifying the person trying to execute the voice command can also be an important element of the experience. It may be desirable to execute a voice command based on who the person is and what the person wants the media playback device or system to do. By way of illustration, at a party or a social gathering in a household, the host or household owner may want to prevent certain guests from using a voice command to change the audio content, listening zone, or playback settings. In some cases, the host or household owner may want to allow certain guests to use voice commands to change the audio content, listening zone, or playback settings, while preventing other guests from making such changes. User identification based on user profiles or voice configuration settings can help distinguish a household owner's voice from a guest's voice.
In another example, user identification can be used to distinguish an adult's voice from a child's voice. In some cases, the household owner may want to prevent a child from using a voice command to listen to audio content inappropriate for the child. In other cases, a household owner may want to prevent a child from changing the listening zone, or playback settings. For example, the household owner may want to listen to audio content at a certain volume and prevent a child from changing the volume of the audio content. User identification may help set parental control settings or restriction settings that would prevent a child from accessing certain content or changing the listening zone, or playback settings. For example, user identification based on user profiles or voice configuration settings may help determine who the child is, what the child is allowed to listen to, or what settings the child is allowed to change.
In yet another example, user identification may be used to prevent unintentional voice commands. For example, the household owner may want to prevent audio from the television or any other audio content from unintentionally triggering a voice command. Many other examples, similar and different from the above, are described herein and illustrate different types of actions based on voice recognition.
Some embodiments described herein include a media playback system (or perhaps one or more components thereof) receiving a voice command and determining an appropriate action for the media playback system to execute based on user identification.
One aspect includes receiving a voice command for a media playback system. In some embodiments, the media playback system includes one or more media playback devices alone or in combination with a computing device, such as a media playback system server. In some embodiments, the media playback system may include or communicate with a networked microphone system server and one or more network microphone devices (NMDs). In some embodiments, the media playback system server and/or the networked microphone system server may be cloud-based server systems. Any one or a combination of these devices and/or servers may receive a voice command for the media playback system.
In some embodiments, one or more functions may be performed by the networked microphone system individually or in combination with the media playback system. In some embodiments, receiving a voice command includes the networked microphone system receiving a voice command via one or more of NMDs, and transmitting the voice command to the media playback system for further processing. In some embodiments, the media playback system may then convert the voice command to an equivalent text command, and parse the text command to identify a command. In some embodiments, the networked microphone system may convert the voice command to an equivalent text command and transmit the text command to the media playback system to parse the text command and identify a command.
A voice command may be a command to control any of the media playback system controls discussed herein. For example, in some embodiments, the voice command may be a command for the media playback system to play media content via one or more playback devices of the media playback system. In some embodiments, the voice command may be a command to modify a playback setting for one or more media playback devices of the media playback system. Playback settings may include, for example, playback volume, playback transport controls, music source selection, and grouping, among other possibilities.
After receiving a voice command, the computing device of the media playback system determines whether the voice command was received from a registered user of the media playback system. In some embodiments, the media playback system may be registered to a particular user or one or more users in a household. In some embodiments, the computing device of the media playback system may be configured to link or associate a voice command to a registered user based on user profiles stored in the computing device. A registered user or users may have created a user profile stored in the computing device. The user profile may contain information specific to the user. For example, the user profile may contain information about the user's age, location, preferred playback settings, preferred playlists, preferred audio content, access restrictions set on the user, and information identifying the user's voice, among other possibilities.
In some embodiments, the computing device of the media playback system may be configured to link or associate a voice command to a user based on voice configuration settings set by a user. In some embodiments, the media playback system may ask a user to provide voice inputs or a series of voice inputs. The computing device of the media playback system may then process the voice inputs, associate the voice inputs to the user, and store the information so that the media playback system can recognize voice commands from the user.
In some embodiments, the computing device of the media playback system may be configured to determine a confidence level associated with a voice command, which may further help determine that the voice command was received from a registered user. A confidence level may be determined based on user history, location, individually or in combination with any other information generally found in a user profile.
In response to determining that the voice command was received from a registered user, the computing device of the media playback system may configure an instruction or a set of instructions for the media playback system. The instructions may be based on content from the voice command and information in a user profile for the registered user. Additionally or alternatively, the instructions may be based on content from the voice command and voice configuration settings stored on the computing device.
In some embodiments, the content from the voice command may include a command for one or more playback devices to play media content. In some embodiments, based on the command for one or more playback devices to play media content and information in a user profile for the registered user, the computing device of the media playback system may configure an instruction or a set of instructions to cause one or more playback devices to obtain media content from a preferred media source of a registered user. In some embodiments, based on the command for one or more playback devices to play media content and information in a user profile for the registered user, the computing device may configure an instruction or a set of instructions to cause the media playback system to play the media content via one or more playback devices of the media playback system. In some embodiments, based on the command for the one or more playback devices to play media content and information in a user profile for the registered user, the computing device may include instructions to (i) configure the one or more playback devices with one or more of the registered user's preferred playback settings and (ii) cause the one or more playback devices to play the media content via the media playback system with the registered user's preferred playback settings.
In some embodiments, the content from the voice command may include a command for one or more playback devices to play media content but may not identify a particular listening zone or playback zone of the media playback system. Based on the content from the voice command and information in a user profile for the registered user, the computing device may configure an instruction or a set of instructions to cause one or more playback devices to play the media content via one or more media playback devices within the particular playback zone of the media playback system.
In some embodiments, the content from the voice command may include a command for the media playback system to modify a playback setting. Based on the content from the voice command and information in a user profile for the registered user, the computing device may configure an instruction or a set of instructions to cause the media playback system to modify the playback setting for one or more playback devices of the media playback system.
Some embodiments include the media playback system determining an order of preference to resolve conflicting voice commands received from different users. A conflicting voice commands may be, for example, a voice command received from a user to play a song and a subsequent voice command received from another user to stop playing the song. Many other examples, similar and different from the above, are described herein. In some embodiments, the media playback system may assign an order of preference in which voice commands received from registered guests have a higher priority than nonregistered guests.
Additionally, the media playback system may take actions based on receiving a wakeup word or wakeup phrase, associated with a registered user or a registered guest user. A wakup word or wakeup phrase (e.g., “Hey Sonos”) may be used to trigger a time period during which the system will accept additional commands from a user based on the specific command or wakeup word received. For example, a host or authorized guest may send a voice command to add songs to a play queue (e.g., “Hey Sonos, let's queue up songs”), which may open a time period or window (e.g., 5 minutes) for the host or authorized guest to send additional voice commands to add specific songs to a play queue. Many other examples, similar and different from the above, are described herein.
After configuring an instruction or set of instructions for the media playback system, some embodiments of the computing device may send the instruction or set of instructions to one or more playback devices of the media playback system.
Some embodiments include the computing device of the media playback system determining whether the voice command was received from a child. In some embodiments, the computing device may distinguish between an adult and a child based on information in a user profile or a guest profile. In some embodiments, the computing device may distinguish between an adult and a child based on the tone or frequency of the user's voice.
In response to determining that the voice command was received from a child, some embodiments may prevent one or more playback devices from playing given media that may be inappropriate for the child. Some embodiments may prevent the computing device and/or one or more playback devices from modifying a playback setting based on the content of a child's voice command.
Some embodiments include actions based on determining whether a voice command was received from a guest user instead of a registered user of the media playback system. In some embodiments, a registered user may have created a guest profile for the guest user. The guest profile may include any information included in a user profile. In some embodiments, the computing device of the media playback system may determine that a voice command was not received from a registered user, and may then ask the registered user if the voice command came from a guest of the registered user.
In response to determining that the voice command was received from a guest user, the computing device of the media playback system may (1) assign a restriction setting for the guest user, (2) configure an instruction for one or more playback devices based on content from the voice command and the assigned restriction setting for the guest user, and (3) send the instruction to one or more playback devices. A restriction setting may be any setting that limits the control of the media playback system.
While some examples described herein may refer to functions performed by given actors such as “users” and/or other entities, it should be understood that this is for purposes of explanation only. The claims should not be interpreted to require action by any such example actor unless explicitly required by the language of the claims themselves. It will be understood by one of ordinary skill in the art that this disclosure includes numerous other embodiments.
II. Example Operating Environment
Further discussions relating to the different components of the example media playback system 100 and how the different components may interact to provide a user with a media experience may be found in the following sections. While discussions herein may generally refer to the example media playback system 100, technologies described herein are not limited to applications within, among other things, the home environment as shown in
a. Example Playback Devices
In one example, the processor 202 may be a clock-driven computing component configured to process input data according to instructions stored in the memory 206. The memory 206 may be a tangible computer-readable medium configured to store instructions executable by the processor 202. For instance, the memory 206 may be data storage that can be loaded with one or more of the software components 204 executable by the processor 202 to achieve certain functions. In one example, the functions may involve the playback device 200 retrieving audio data from an audio source or another playback device. In another example, the functions may involve the playback device 200 sending audio data to another device or playback device on a network. In yet another example, the functions may involve pairing of the playback device 200 with one or more playback devices to create a multi-channel audio environment.
Certain functions may involve the playback device 200 synchronizing playback of audio content with one or more other playback devices. During synchronous playback, a listener will preferably not be able to perceive time-delay differences between playback of the audio content by the playback device 200 and the one or more other playback devices. U.S. Pat. No. 8,234,395 entitled, “System and method for synchronizing operations among a plurality of independently clocked digital data processing devices,” which is hereby incorporated by reference, provides in more detail some examples for audio playback synchronization among playback devices.
The memory 206 may further be configured to store data associated with the playback device 200, such as one or more zones and/or zone groups the playback device 200 is a part of, audio sources accessible by the playback device 200, or a playback queue that the playback device 200 (or some other playback device) may be associated with. The data may be stored as one or more state variables that are periodically updated and used to describe the state of the playback device 200. The memory 206 may also include the data associated with the state of the other devices of the media system, and shared from time to time among the devices so that one or more of the devices have the most recent data associated with the system. Other embodiments are also possible.
The audio processing components 208 may include one or more digital-to-analog converters (DAC), an audio preprocessing component, an audio enhancement component or a digital signal processor (DSP), and so on. In one embodiment, one or more of the audio processing components 208 may be a subcomponent of the processor 202. In one example, audio content may be processed and/or intentionally altered by the audio processing components 208 to produce audio signals. The produced audio signals may then be provided to the audio amplifier(s) 210 for amplification and playback through speaker(s) 212. Particularly, the audio amplifier(s) 210 may include devices configured to amplify audio signals to a level for driving one or more of the speakers 212. The speaker(s) 212 may include an individual transducer (e.g., a “driver”) or a complete speaker system involving an enclosure with one or more drivers. A particular driver of the speaker(s) 212 may include, for example, a subwoofer (e.g., for low frequencies), a mid-range driver (e.g., for middle frequencies), and/or a tweeter (e.g., for high frequencies). In some cases, each transducer in the one or more speakers 212 may be driven by an individual corresponding audio amplifier of the audio amplifier(s) 210. In addition to producing analog signals for playback by the playback device 200, the audio processing components 208 may be configured to process audio content to be sent to one or more other playback devices for playback.
Audio content to be processed and/or played back by the playback device 200 may be received from an external source, such as via an audio line-in input connection (e.g., an auto-detecting 3.5 mm audio line-in connection) or the network interface 214.
The network interface 214 may be configured to facilitate a data flow between the playback device 200 and one or more other devices on a data network. As such, the playback device 200 may be configured to receive audio content over the data network from one or more other playback devices in communication with the playback device 200, network devices within a local area network, or audio content sources over a wide area network such as the Internet. In one example, the audio content and other signals transmitted and received by the playback device 200 may be transmitted in the form of digital packet data containing an Internet Protocol (IP)-based source address and IP-based destination addresses. In such a case, the network interface 214 may be configured to parse the digital packet data such that the data destined for the playback device 200 is properly received and processed by the playback device 200.
As shown, the network interface 214 may include wireless interface(s) 216 and wired interface(s) 218. The wireless interface(s) 216 may provide network interface functions for the playback device 200 to wirelessly communicate with other devices (e.g., other playback device(s), speaker(s), receiver(s), network device(s), control device(s) within a data network the playback device 200 is associated with) in accordance with a communication protocol (e.g., any wireless standard including IEEE 802.11a, 802.11b, 802.11g, 802.11n, 802.11ac, 802.15, 4G mobile communication standard, and so on). The wired interface(s) 218 may provide network interface functions for the playback device 200 to communicate over a wired connection with other devices in accordance with a communication protocol (e.g., IEEE 802.3). While the network interface 214 shown in
The microphone(s) 220 may be arranged to detect sound in the environment of the playback device 200. For instance, the microphone(s) may be mounted on an exterior wall of a housing of the playback device. The microphone(s) may be any type of microphone now known or later developed such as a condenser microphone, electret condenser microphone, or a dynamic microphone. The microphone(s) may be sensitive to a portion of the frequency range of the speaker(s) 220. One or more of the speaker(s) 220 may operate in reverse as the microphone(s) 220. In some aspects, the playback device 200 might not have microphone(s) 220.
In one example, the playback device 200 and one other playback device may be paired to play two separate audio components of audio content. For instance, playback device 200 may be configured to play a left channel audio component, while the other playback device may be configured to play a right channel audio component, thereby producing or enhancing a stereo effect of the audio content. The paired playback devices (also referred to as “bonded playback devices”) may further play audio content in synchrony with other playback devices.
In another example, the playback device 200 may be sonically consolidated with one or more other playback devices to form a single, consolidated playback device. A consolidated playback device may be configured to process and reproduce sound differently than an unconsolidated playback device or playback devices that are paired, because a consolidated playback device may have additional speaker drivers through which audio content may be rendered. For instance, if the playback device 200 is a playback device designed to render low frequency range audio content (i.e. a subwoofer), the playback device 200 may be consolidated with a playback device designed to render full frequency range audio content. In such a case, the full frequency range playback device, when consolidated with the low frequency playback device 200, may be configured to render only the mid and high frequency components of audio content, while the low frequency range playback device 200 renders the low frequency component of the audio content. The consolidated playback device may further be paired with a single playback device or yet another consolidated playback device.
By way of illustration, SONOS, Inc. presently offers (or has offered) for sale certain playback devices including a “PLAY:1,” “PLAY:3,” “PLAY:5,” “PLAYBAR,” “CONNECT:AMP,” “CONNECT,” and “SUB.” Any other past, present, and/or future playback devices may additionally or alternatively be used to implement the playback devices of example embodiments disclosed herein. Additionally, it is understood that a playback device is not limited to the example illustrated in
b. Example Playback Zone Configurations
Referring back to the media playback system 100 of
As shown in
In one example, one or more playback zones in the environment of
As suggested above, the zone configurations of the media playback system 100 may be dynamically modified, and in some embodiments, the media playback system 100 supports numerous configurations. For instance, if a user physically moves one or more playback devices to or from a zone, the media playback system 100 may be reconfigured to accommodate the change(s). For instance, if the user physically moves the playback device 102 from the balcony zone to the office zone, the office zone may now include both the playback device 118 and the playback device 102. The playback device 102 may be paired or grouped with the office zone and/or renamed if so desired via a control device such as the control devices 126 and 128. On the other hand, if the one or more playback devices are moved to a particular area in the home environment that is not already a playback zone, a new playback zone may be created for the particular area.
Further, different playback zones of the media playback system 100 may be dynamically combined into zone groups or split up into individual playback zones. For instance, the dining room zone and the kitchen zone 114 may be combined into a zone group for a dinner party such that playback devices 112 and 114 may render audio content in synchrony. On the other hand, the living room zone may be split into a television zone including playback device 104, and a listening zone including playback devices 106, 108, and 110, if the user wishes to listen to music in the living room space while another user wishes to watch television.
c. Example Control Devices
The processor 302 may be configured to perform functions relevant to facilitating user access, control, and configuration of the media playback system 100. The memory 304 may be data storage that can be loaded with one or more of the software components executable by the processor 302 to perform those functions. The memory 304 may also be configured to store the media playback system controller application software and other data associated with the media playback system 100 and the user.
In one example, the network interface 306 may be based on an industry standard (e.g., infrared, radio, wired standards including IEEE 802.3, wireless standards including IEEE 802.11a, 802.11b, 802.11g, 802.11n, 802.11ac, 802.15, 4G mobile communication standard, and so on). The network interface 306 may provide a means for the control device 300 to communicate with other devices in the media playback system 100. In one example, data and information (e.g., such as a state variable) may be communicated between control device 300 and other devices via the network interface 306. For instance, playback zone and zone group configurations in the media playback system 100 may be received by the control device 300 from a playback device or another network device, or transmitted by the control device 300 to another playback device or network device via the network interface 306. In some cases, the other network device may be another control device.
Playback device control commands such as volume control and audio playback control may also be communicated from the control device 300 to a playback device via the network interface 306. As suggested above, changes to configurations of the media playback system 100 may also be performed by a user using the control device 300. The configuration changes may include adding/removing one or more playback devices to/from a zone, adding/removing one or more zones to/from a zone group, forming a bonded or consolidated player, separating one or more playback devices from a bonded or consolidated player, among others. Accordingly, the control device 300 may sometimes be referred to as a controller, whether the control device 300 is a dedicated controller or a network device on which media playback system controller application software is installed.
Control device 300 may include microphone(s) 310. Microphone(s) 310 may be arranged to detect sound in the environment of the control device 300. Microphone(s) 310 may be any type of microphone now known or later developed such as a condenser microphone, electret condenser microphone, or a dynamic microphone. The microphone(s) may be sensitive to a portion of a frequency range. Two or more microphones 310 may be arranged to capture location information of an audio source (e.g., voice, audible sound) and/or to assist in filtering background noise.
The user interface 308 of the control device 300 may be configured to facilitate user access and control of the media playback system 100, by providing a controller interface such as the controller interface 400 shown in
The playback control region 410 may include selectable (e.g., by way of touch or by using a cursor) icons to cause playback devices in a selected playback zone or zone group to play or pause, fast forward, rewind, skip to next, skip to previous, enter/exit shuffle mode, enter/exit repeat mode, enter/exit cross fade mode. The playback control region 410 may also include selectable icons to modify equalization settings, and playback volume, among other possibilities.
The playback zone region 420 may include representations of playback zones within the media playback system 100. In some embodiments, the graphical representations of playback zones may be selectable to bring up additional selectable icons to manage or configure the playback zones in the media playback system, such as a creation of bonded zones, creation of zone groups, separation of zone groups, and renaming of zone groups, among other possibilities.
For example, as shown, a “group” icon may be provided within each of the graphical representations of playback zones. The “group” icon provided within a graphical representation of a particular zone may be selectable to bring up options to select one or more other zones in the media playback system to be grouped with the particular zone. Once grouped, playback devices in the zones that have been grouped with the particular zone will be configured to play audio content in synchrony with the playback device(s) in the particular zone. Analogously, a “group” icon may be provided within a graphical representation of a zone group. In this case, the “group” icon may be selectable to bring up options to deselect one or more zones in the zone group to be removed from the zone group. Other interactions and implementations for grouping and ungrouping zones via a user interface such as the user interface 400 are also possible. The representations of playback zones in the playback zone region 420 may be dynamically updated as playback zone or zone group configurations are modified.
The playback status region 430 may include graphical representations of audio content that is presently being played, previously played, or scheduled to play next in the selected playback zone or zone group. The selected playback zone or zone group may be visually distinguished on the user interface, such as within the playback zone region 420 and/or the playback status region 430. The graphical representations may include track title, artist name, album name, album year, track length, and other relevant information that may be useful for the user to know when controlling the media playback system via the user interface 400.
The playback queue region 440 may include graphical representations of audio content in a playback queue associated with the selected playback zone or zone group. In some embodiments, each playback zone or zone group may be associated with a playback queue containing information corresponding to zero or more audio items for playback by the playback zone or zone group. For instance, each audio item in the playback queue may comprise a uniform resource identifier (URI), a uniform resource locator (URL) or some other identifier that may be used by a playback device in the playback zone or zone group to find and/or retrieve the audio item from a local audio content source or a networked audio content source, possibly for playback by the playback device.
In one example, a playlist may be added to a playback queue, in which case information corresponding to each audio item in the playlist may be added to the playback queue. In another example, audio items in a playback queue may be saved as a playlist. In a further example, a playback queue may be empty, or populated but “not in use” when the playback zone or zone group is playing continuously streaming audio content, such as Internet radio that may continue to play until otherwise stopped, rather than discrete audio items that have playback durations. In an alternative embodiment, a playback queue can include Internet radio and/or other streaming audio content items and be “in use” when the playback zone or zone group is playing those items. Other examples are also possible.
When playback zones or zone groups are “grouped” or “ungrouped,” playback queues associated with the affected playback zones or zone groups may be cleared or re-associated. For example, if a first playback zone including a first playback queue is grouped with a second playback zone including a second playback queue, the established zone group may have an associated playback queue that is initially empty, that contains audio items from the first playback queue (such as if the second playback zone was added to the first playback zone), that contains audio items from the second playback queue (such as if the first playback zone was added to the second playback zone), or a combination of audio items from both the first and second playback queues. Subsequently, if the established zone group is ungrouped, the resulting first playback zone may be re-associated with the previous first playback queue, or be associated with a new playback queue that is empty or contains audio items from the playback queue associated with the established zone group before the established zone group was ungrouped. Similarly, the resulting second playback zone may be re-associated with the previous second playback queue, or be associated with a new playback queue that is empty, or contains audio items from the playback queue associated with the established zone group before the established zone group was ungrouped. Other examples are also possible.
Referring back to the user interface 400 of
The audio content sources region 450 may include graphical representations of selectable audio content sources from which audio content may be retrieved and played by the selected playback zone or zone group. Discussions pertaining to audio content sources may be found in the following section.
d. Example Audio Content Sources
As indicated previously, one or more playback devices in a zone or zone group may be configured to retrieve for playback audio content (e.g. according to a corresponding URI or URL for the audio content) from a variety of available audio content sources. In one example, audio content may be retrieved by a playback device directly from a corresponding audio content source (e.g., a line-in connection). In another example, audio content may be provided to a playback device over a network via one or more other playback devices or network devices.
Example audio content sources may include a memory of one or more playback devices in a media playback system such as the media playback system 100 of
In some embodiments, audio content sources may be regularly added or removed from a media playback system such as the media playback system 100 of
The above discussions relating to playback devices, controller devices, playback zone configurations, and media content sources provide only some examples of operating environments within which functions and methods described below may be implemented. Other operating environments and configurations of media playback systems, playback devices, and network devices not explicitly described herein may also be applicable and suitable for implementation of the functions and methods.
e. Example Plurality of Networked Devices
Each of the plurality of devices 500 may be network-capable devices that can establish communication with one or more other devices in the plurality of devices according to one or more network protocols, such as NFC, Bluetooth, Ethernet, and IEEE 802.11, among other examples, over one or more types of networks, such as wide area networks (WAN), local area networks (LAN), and personal area networks (PAN), among other possibilities.
As shown, the computing devices 504, 506, and 508 may be part of a cloud network 502. The cloud network 502 may include additional computing devices. In one example, the computing devices 504, 506, and 508 may be different servers. In another example, two or more of the computing devices 504, 506, and 508 may be modules of a single server. Analogously, each of the computing device 504, 506, and 508 may include one or more modules or servers. For ease of illustration purposes herein, each of the computing devices 504, 506, and 508 may be configured to perform particular functions within the cloud network 502. For instance, computing device 508 may be a source of audio content for a streaming music service.
As shown, the computing device 504 may be configured to interface with NMDs 512, 514, and 516 via communication path 542. NMDs 512, 514, and 516 may be components of one or more “Smart Home” systems. In one case, NMDs 512, 514, and 516 may be physically distributed throughout a household, similar to the distribution of devices shown in
In one example, one or more of the NMDs 512, 514, and 516 may be devices configured primarily for audio detection. In another example, one or more of the NMDs 512, 514, and 516 may be components of devices having various primary utilities. For instance, as discussed above in connection to
As shown, the computing device 506 may be configured to interface with CR 522 and PBDs 532, 534, 536, and 538 via communication path 544. In one example, CR 522 may be a network device such as the network device 200 of
In one example, as with NMDs 512, 514, and 516, CR522 and PBDs 532, 534, 536, and 538 may also be components of one or more “Smart Home” systems. In one case, PBDs 532, 534, 536, and 538 may be distributed throughout the same household as the NMDs 512, 514, and 516. Further, as suggested above, one or more of PBDs 532, 534, 536, and 538 may be one or more of NMDs 512, 514, and 516.
The NMDs 512, 514, and 516 may be part of a local area network, and the communication path 542 may include an access point that links the local area network of the NMDs 512, 514, and 516 to the computing device 504 over a WAN (communication path not shown). Likewise, each of the NMDs 512, 514, and 516 may communicate with each other via such an access point.
Similarly, CR 522 and PBDs 532, 534, 536, and 538 may be part of a local area network and/or a local playback network as discussed in previous sections, and the communication path 544 may include an access point that links the local area network and/or local playback network of CR 522 and PBDs 532, 534, 536, and 538 to the computing device 506 over a WAN. As such, each of the CR 522 and PBDs 532, 534, 536, and 538 may also communicate with each over such an access point.
In one example, communication paths 542 and 544 may comprise the same access point. In an example, each of the NMDs 512, 514, and 516, CR 522, and PBDs 532, 534, 536, and 538 may access the cloud network 502 via the same access point for a household.
As shown in
In one example, CR 522 may communicate with NMD 512 over Bluetooth™, and communicate with PBD 534 over another local area network. In another example, NMD 514 may communicate with CR 522 over another local area network, and communicate with PBD 536 over Bluetooth. In a further example, each of the PBDs 532, 534, 536, and 538 may communicate with each other according to a spanning tree protocol over a local playback network, while each communicating with CR 522 over a local area network, different from the local playback network. Other examples are also possible.
In some cases, communication means between the NMDs 512, 514, and 516, CR 522, and PBDs 532, 534, 536, and 538 may change depending on types of communication between the devices, network conditions, and/or latency demands. For instance, communication means 546 may be used when NMD 516 is first introduced to the household with the PBDs 532, 534, 536, and 538. In one case, the NMD 516 may transmit identification information corresponding to the NMD 516 to PBD 538 via NFC, and PBD 538 may in response, transmit local area network information to NMD 516 via NFC (or some other form of communication). However, once NMD 516 has been configured within the household, communication means between NMD 516 and PBD 538 may change. For instance, NMD 516 may subsequently communicate with PBD 538 via communication path 542, the cloud network 502, and communication path 544. In another example, the NMDs and PBDs may never communicate via local communications means 546. In a further example, the NMDs and PBDs may communicate primarily via local communications means 546. Other examples are also possible.
In an illustrative example, NMDs 512, 514, and 516 may be configured to receive voice inputs to control PBDs 532, 534, 536, and 538. The available control commands may include any media playback system controls previously discussed, such as playback volume control, playback transport controls, music source selection, and grouping, among other possibilities. In one instance, NMD 512 may receive a voice input to control one or more of the PBDs 532, 534, 536, and 538. In response to receiving the voice input, NMD 512 may transmit via communication path 542, the voice input to computing device 504 for processing. In one example, the computing device 504 may convert the voice input to an equivalent text command, and parse the text command to identify a command. Computing device 504 may then subsequently transmit the text command to the computing device 506. In another example, the computing device 504 may convert the voice input to an equivalent text command, and then subsequently transmit the text command to the computing device 506. The computing device 506 may then parse the text command to identify one or more playback commands.
For instance, if the text command is “Play ‘Track 1’ by ‘Artist 1’ from ‘Streaming Service 1’ in ‘Zone 1’,” The computing device 506 may identify (i) a URL for “Track 1” by “Artist 1” available from “Streaming Service 1,” and (ii) at least one playback device in “Zone 1.” In this example, the URL for “Track 1” by “Artist 1” from “Streaming Service 1” may be a URL pointing to computing device 508, and “Zone 1” may be the bonded zone 530. As such, upon identifying the URL and one or both of PBDs 536 and 538, the computing device 506 may transmit via communication path 544 to one or both of PBDs 536 and 538, the identified URL for playback. One or both of PBDs 536 and 538 may responsively retrieve audio content from the computing device 508 according to the received URL, and begin playing “Track 1” by “Artist 1” from “Streaming Service 1.”
One having ordinary skill in the art will appreciate that the above is just one illustrative example, and that other implementations are also possible. In one case, operations performed by one or more of the plurality of devices 500, as described above, may be performed by one or more other devices in the plurality of device 500. For instance, the conversion from voice input to the text command may be alternatively, partially, or wholly performed by another device or devices, such as NMD 512, computing device 506, PBD 536, and/or PBD 538. Analogously, the identification of the URL may be alternatively, partially, or wholly performed by another device or devices, such as NMD 512, computing device 504, PBD 536, and/or PBD 538.
f. Example Network Microphone Device
The processor 602 may include one or more processors and/or controllers, which may take the form of a general or special-purpose processor or controller. For instance, the processing unit 602 may include microprocessors, microcontrollers, application-specific integrated circuits, digital signal processors, and the like. The memory 604 may be data storage that can be loaded with one or more of the software components executable by the processor 602 to perform those functions. Accordingly, memory 604 may comprise one or more non-transitory computer-readable storage mediums, examples of which may include volatile storage mediums such as random access memory, registers, cache, etc. and non-volatile storage mediums such as read-only memory, a hard-disk drive, a solid-state drive, flash memory, and/or an optical-storage device, among other possibilities.
The microphone array 606 may be a plurality of microphones arranged to detect sound in the environment of the network microphone device 600. Microphone array 606 may include any type of microphone now known or later developed such as a condenser microphone, electret condenser microphone, or a dynamic microphone, among other possibilities. In one example, the microphone array may be arranged to detect audio from one or more directions relative to the network microphone device. The microphone array 606 may be sensitive to a portion of a frequency range. In one example, a first subset of the microphone array 606 may be sensitive to a first frequency range, while a second subset of the microphone array may be sensitive to a second frequency range. The microphone array 606 may further be arranged to capture location information of an audio source (e.g., voice, audible sound) and/or to assist in filtering background noise. Notably, in some embodiments the microphone array may consist of only a single microphone, rather than a plurality of microphones.
The network interface 608 may be configured to facilitate wireless and/or wired communication between various network devices, such as, in reference to
The user interface 610 of the network microphone device 600 may be configured to facilitate user interactions with the network microphone device. In one example, the user interface 608 may include one or more of physical buttons, graphical interfaces provided on touch sensitive screen(s) and/or surface(s), among other possibilities, for a user to directly provide input to the network microphone device 600. The user interface 610 may further include one or more of lights and the speaker(s) 614 to provide visual and/or audio feedback to a user. In one example, the network microphone device 600 may further be configured to playback audio content via the speaker(s) 614.
III. Example Systems and Methods
To execute a voice command to control the media playback system, it is desirable in some instances for the media playback system to receive a voice command and determine an appropriate action for the media playback system to execute based on user identification (or at least based on the user who spoke the voice command). In some embodiments, the media playback system includes one or more of PBDs 532, 534, 536, and 538 and computing device 506 (which is configured as a media playback system server). In some embodiments, the media playback system may include or communicate with a networked microphone system that includes one or more of NMDs 512, 514, and 516 and computing device 504 (which is configured as a networked microphone system server).
Generally, it should be understood that one or more functions described herein may be performed by the networked microphone system individually or in combination with the media playback system. It should be further understood that one or more functions performed by the computing device 506 may be performed by CR 522 and/or one or more of PBDs 532, 534, 536, and 538 of the media playback system.
Examples of voice commands include commands to control any of the media playback system controls discussed previously. For example, in some embodiments, the voice command may be a command for the media playback system to play media content via one or more playback devices of the media playback system. In some embodiments, the voice command may be a command to trigger a time period or window in which to receive additional voice commands associated with the initial voice command. In some embodiments, the voice command may be a command to modify a playback setting for one or more media playback devices of the media playback system. Playback settings may include, for example, playback volume, playback transport controls, music source selection, and grouping, among other possibilities.
Examples of media content include, talk radio, books, audio from television, music stored on a local drive, or music from media sources, among others. Examples of media sources include Pandora® Radio, Spotify®, Slacker®, Radio, Google Play™, and iTunes Radio, among others.
Examples of user identification include identifying a user as a registered user, a guest user, a child, or an unknown user.
Example registered users include one or more users linked or associated with the media playback system by a user profile, and/or voice configuration settings, among other possibilities. Example user profiles may include information about a user's age, location, preferred playback settings, preferred playlists, preferred audio content, access restrictions set on the user, and information identifying the user's voice, user history, among other possibilities. Example information identifying the user's voice includes the tone or frequency of a user's voice, age, gender, and user history, among other information. Example voice configuration settings may include settings that ask a user to provide voice inputs or a series of voice inputs for the media playback system to recognize and associate the user with.
Example guest users include one or more users linked or associated with the media playback system by a registered user's user profile, or a guest profile created by a registered user or a guest user with the registered user's permission. Example guest profiles may include any type of information included in a user profile.
In some embodiments, a guest with his or her own media playback system in his or her own house may have a user profile associated with his or her own media playback system stored in computing device 506, for example. In operation, when that guest arrives at the host's home and tries to use voice commands to control the host's media playback system, the computing device 506 connected to the host's playback system may be able to access user profile settings of the guest, including but not limited to (i) music services that the guest has user accounts with, (ii) the guest's playlists, (iii) whether the host has granted the guest access to control the host's media playback system, and/or (iv) perhaps other user information in the guest's user profile.
A child user may be identified by, for example, information in a user profile if the child is one of the registered users of the media playback system, information in a guest profile, and/or the tone or frequency of the user's voice.
In some embodiments, receiving a voice command includes the media playback system receiving a voice command via one or more of PBDs 532, 534, 536, and 538 and/or computing device 506 (which is configured as a media playback system server). In one example, computing device 506 may convert the voice command to an equivalent text command, and parse the text command to identify a command.
In some embodiments, one or more functions may be performed by the networked microphone system individually or in combination with the media playback system. In some embodiments, receiving a voice command includes the networked microphone system receiving a voice command via one or more of NMDs 512, 514, or 516, and transmitting the voice command to the computing device 506 and/or one or more of PBDs 532, 534, 536, and 538 for further processing. In some embodiments, the computing device 506 may convert the voice command to an equivalent text command, and parse the text command to identify a command. In some embodiments, the networked microphone system may convert the voice command to an equivalent text command and transmit the text command to the computing device 506 and/or one or more of PBDs 532, 534, 536, and 538 to parse the text command and identify a command.
After receiving a voice command, the computing device 506 and/or one or more of PBDs 532, 534, 536, and 538 determines whether the voice command was received from a registered user of the media playback system. In some embodiments, determining whether the voice command was received from a registered user may include the computing device 506 and/or one or more of PBDs 532, 534, 536, and 538 determining whether there is a user profile stored on the media playback system that is associated with the voice command. For example, the computing device 506 and/or one or more of PBDs 532, 534, 536, and 538 may try to match the voice command to information identifying a user's voice that may be included in a user profile stored on the media playback system. In some embodiments, the networked microphone system individually or in combination with the media playback system may determine whether the voice command was received from a registered user of the media playback system by communicating with computing device 506.
In some embodiments, determining whether the voice command was received from a registered user may include the computing device 506 and/or one or more of PBDs 532, 534, 536, and 538 determining whether the voice command matches the voice inputs in the media playback system's voice configuration settings. For example, a user may have previously configured the media playback system to recognize the user's voice by providing a voice input or a series of voice inputs for the media playback system to recognize and associate the user with. The voice input or series of voice inputs may be stored on the computing device 506 and/or one or more of PBDs 532, 534, 536, and 538. In some embodiments, the voice input or series of voice inputs may be stored on the networked microphone system.
In some embodiments, determining whether the voice command was received from a registered user may include the computing device 506, CR 522, and/or one or more of PBDs 532, 534, 536, and 538, individually or in combination, determining a confidence level associated with a voice command received. A confidence level may be determined based on user history, location, individually or in combination with any other information generally found in a user profile.
For example, the media playback system, may receive a first voice command from a registered user in the kitchen and determine a confidence level based on the voice command received. The media playback system may receive the first voice command from any one or more of NMDs 512-513, CR 522, and PBDs 532-538. Further, the media playback system may receive the same voice command from the registered user in another room in the user's house and determine a confidence level based on the voice command received. The media playback system may receive the second voice command from any one or more of NMDs 512-513, CR 522, and PBDs 532-538. The media playback system may then determine a new confidence level based on the received commands from different computing devices (e.g., CR 522), NMDs, and/or PBDs throughout the user's house. As a result, the media playback system may have a greater confidence level that the voice command was received from a registered user.
In another example, the media playback system may receive a voice command from a registered user and determine a confidence level based on user history. In operation, the media playback system may receive the voice command from any one or more of NMDs 512-513, CR 522, and PBDs 532-538. After receiving the voice command, computing device 506, CR 522, and/or one or more of PBDs 532, 534, 536, and 538, individually or in combination, may determine a higher confidence level if the voice command received includes an artist, playlist, genre, or any other information found in a user profile that is typically associated with the registered user. For example, if a registered user typically listens to songs by Michael Jackson, the media playback system may have a greater confidence level that a voice command to play “Thriller” by Michael Jackson was received from a registered user. Many other examples, similar and different from the above, are possible.
In some embodiments, the media playback system may build a confidence level based on a registered user's pattern of voice commands found in a user's profile. For example, the media playback system may receive a voice command from a registered user to play a particular song by Britney Spears, and determine a confidence level based on the received voice command. Every time the media playback system receives the same voice command or similar voice command, such as a command to play another song by Britney Spears, the media playback system may build a higher confidence level and thus, may have a greater confidence level that the voice command was received from a registered user.
Generally, as mentioned previously, it should be understood that one or more functions described herein may be performed by the networked microphone system individually or in combination with the media playback system. It should be further understood that one or more functions performed by the computing device 506 may be performed by CR 522 and/or one or more of PBDs 532, 534, 536, and 538 of the media playback system and/or perhaps one or more of NMDs 512, 514, and 516.
In some embodiments, determining a confidence level includes the media playback system determining a confidence level via computing device 506 (which is configured as a media playback system server), CR 522, and/or one or more of PBDs 532, 534, 536, and 538, individually or in combination with one another. For example, CR 522, and/or one or more of PBDs 532, 534, 536, and 538 may (i) determine a confidence level associated with a received voice command, (ii) determine that the voice command was received from a registered user based on the determined confidence level, and (iii) send an instruction to computing device 506 (which is configured as a media playback system server) to execute the voice command. In another example, CR 522, and/or one or more of PBDs 532, 534, 536, and 538 may (i) determine a confidence level associated with a received voice command, and (ii) send data associated with the confidence level to computing device 506 for further processing. Computing device 506 may then (i) determine that the voice command was received from a registered user based on the determined confidence level, and (ii) send an instruction to execute the voice command to CR 522, and/or one or more of PBDs 532, 534, 536, and 538.
In some embodiments, determining a confidence level includes the media playback system determining a confidence level individually or in combination with the networked microphone system. For example, the media playback system may receive a voice command via CR 522 and/or one or more of PBDs 532, 534, 536, and 538 of the media playback system and/or perhaps one or more of NMDs 512, 514, and 516. In response to the received voice command, the media playback system may send data associated with a confidence level to one or more of NMDs 512, 514, or 516. The networked microphone may then (i) determine a confidence level associated with the received data, and (ii) execute a command or send an instruction to the media playback system to execute a command. In response to determining that the voice command was received from a registered user, the computing device 506 may configure an instruction or a set of instructions for one or more PBDs of the media playback system. The instructions may be based on content from the voice command and information in a user profile for the registered user. Additionally or alternatively, the instructions may be based on content from the voice command and voice configuration settings stored on the computing device 506, one or more of PBDs 532, 534, 536, and 538, or the networked microphone system.
In some embodiments, the content from the voice command may include a command for one or more PBDs of the media playback system to play media content. In some embodiments, based on the command for the media playback system to play media content and information in a user profile for the registered user, computing device 506 may configure an instruction or a set of instructions to cause one or more of PBDs to obtain media content from a preferred media source of a registered user.
In some embodiments, based on the command for the media playback system to play media content and information in a user profile for the registered user, computing device 506 may configure an instruction or a set of instructions to cause the media playback system to play the media content via one or more PBDs of the media playback system.
In some embodiments, based on the command for the media playback system to play media content and information in a user profile for the registered user, the computing device 506 may include instructions to (i) configure the media playback system with one or more of the registered user's preferred playback settings and (ii) cause one or more PBDs to play the media content with the registered user's preferred playback settings. Preferred playback settings may be preferred playback settings stored in a registered user's user profile. Additionally or alternatively, preferred playback settings may be based on user history stored in a registered user's user profile. User history may include commonly used or previously used playback settings by the user to play media content.
In some embodiments, the content from the voice command may include a command for the media playback system to play media content but may not identify a particular listening zone or playback zone of the media playback system. Based on this content and information in a user profile for the registered user, such as user history, the computing device 506 may (i) configure an instruction or a set of instructions to cause the media playback system to play the media content via one or more PBDs within the particular playback zone of the media playback system and (ii) implement the configured instruction or set of instructions to play the media content via the one or more PBDs.
In some embodiments, the content from the voice command may include a command for the media playback system to modify a playback setting. Based on the command for the media playback system to modify a playback setting and information in a user profile for the registered user, the computing device 506 may (i) configure an instruction or a set of instructions to cause the media playback system to modify the playback setting for one or more PBDs of the media playback system and (ii) implement the configured instruction or set of instructions to modify the playback setting via the one or more PBDs.
Some embodiments include the media playback system determining whether the voice command was received from a child. In some embodiments, the computing device 506 may distinguish between an adult and a child based on information in a user profile if the child is one of the registered users of the media playback system. In some embodiments, the computing device 506 may distinguish between an adult and a child based on the tone or frequency of the user's voice.
In some embodiments, determining whether the voice command was received from a child may include the computing device 506, CR 522, and/or one or more of PBDs 532, 534, 536, and 538 (individually or in combination) determining a confidence level associated with a voice command received. As described above, a confidence level may be determined based on user history, location, individually or in combination with any other information generally found in a user profile.
For example, the media playback system may receive a voice command from an NMD or PBD located in a particular room where a child is likely to be (e.g., child's bedroom, playroom, basement, etc). Because the voice command was received from a device (an NMD or PBD) located in a room where a child is likely to be, the media playback system may have a greater confidence level that the voice command was received from a child.
In another example, the media playback system, may receive a voice command for a particular type of content, and based on the type of content, determine a higher confidence level that the voice command was received from a child. For example, if the media playback system receives a voice command to play a song from a cartoon show or movie, the media playback system may have a greater confidence level that the voice command was received from a child. Many other examples, similar and different from the above, are possible. In response to determining that the voice command was received from a child, some embodiments may prevent one or more PBDs from playing given media that may be inappropriate for the child. Some embodiments may prevent the computing device 506 and/or one or more PBDs from modifying a playback setting based on the content of a child's voice command. For example, the computing device 506 and/or one or more PBDs may disregard a child's voice command to increase the volume of one or more PBDs.
Some embodiments include the media playback device taking actions based on determining whether a voice command was received from a guest user instead of a registered user of the media playback system. In some embodiments, computing device 506 may have stored a previously created guest profile that may be associated with a particular guest. In some embodiments, computing device 506 may determine that a voice command was not received from a registered user, and may then ask the registered user if the voice command came from a guest. The registered user may then have the option to prevent the computing device 506 and/or one or more PBDs from executing all or part of the contents of the voice command.
In some embodiments, determining whether the voice command was received from a guest user may include the computing device 506, CR 522, and/or one or more of PBDs 532, 534, 536, and 538 (individually or in combination) determining a confidence level associated with a voice command received. As described above, a confidence level may be determined based on user history, location, individually or in combination with any other information generally found in a user profile.
In response to determining that the voice command was received from a guest user, the computing device 506 may (1) assign a restriction setting for the guest user, (2) configure an instruction for one or more PBDs based on content from the voice command and the assigned restriction setting for the guest user, and (3) send the instruction to one or more PBDs for execution. In some embodiments, assigning a restriction setting for a guest user may include the computing device 506 matching the voice command to a particular guest profile stored on the computing device 506 and/or one or more PBDs. The guest profile may include restriction settings, and information regarding the voice of the particular guest user, such as frequency or tone of the guest's voice, among other information described previously. A restriction setting may be any setting that limits the control of the media playback system.
Some embodiments include the media playback system determining an order of preference to resolve conflicting voice commands received from different users. A conflicting voice commands may be, for example, a voice command received from a user to play a song and a subsequent voice command received from another user to stop playing the song. Other examples are possible, such as a voice command received from a user to increase the volume of one or more PBDs 532, 534, 536, and 538, and a subsequent voice command received from another user to decrease the volume. In particular, the media playback system (via one or more of NMDs 512-516, CR 522, PBDs 532-538, and/or computing device 506) may receive a voice command from a registered user or host to play a song in a playback zone. Subsequently, the media playback system may receive a conflicting voice command from a nonregistered user or guest to stop playing the song in the playback zone. To resolve this conflict, the media playback system may apply an order of preference in which voice commands received from a registered user have a higher priority than a nonregistered user or guest.
In some embodiments, the media playback system may assign an order of preference in which voice commands received from registered guests have a higher priority than nonregistered guests. In some embodiments, voice commands received from one registered guest may have a higher priority than another registered guest. Additionally or alternatively, voice commands received from an adult may have a higher priority than a child.
In another embodiment, controller-issued commands (e.g., commands issued by CR 522 or another computing device configured to control the media playback system) received by the media playback system may have a lower priority than a registered user, but may have a higher priority than a nonregistered user or guest. In some embodiments, some registered guests may have a higher priority than controller-issued commands. Other examples of determining and assigning an order of preference are possible.
Additionally, the media playback system may take actions based on receiving a wakeup word or wakeup phrase, associated with a registered user. A wakeup word or phrase may be a specific word or phrase (e.g., “Hey, Sonos”) stored in a registered user's profile. In some embodiments, different users may configure the media playback system for different wakeup words or phrases. In other embodiments, the media playback system may be configured with the same wakeup word or phrase for all (or any) users.
In some embodiments, a registered user may have a universal wakeup word or phrase that triggers a time period or window for the media playback system to receive additional voice commands associated with the wakeup word or phrase from the registered user, a guest, and/or a nonregistered user. For example, a registered user or host may send a voice command to add songs to a play queue (e.g., “Hey Sonos, let's queue up songs”), which may open a time period or window (e.g., five minutes) during which the registered user can send additional voice commands to add specific songs to the play queue (e.g., “Add Thriller by Michael Jackson”). In another example, a registered user or host may send a voice command (e.g., “Hey Sonos, open control for my house system”) that authorizes all guests in a house to send voice commands to add songs to a play queue, play songs, or change the volume, among other functions for a user-defined or default time period or window, or for a specific period of time (e.g., “Hey Sonos, open control for my house system for the next 4 hours” or “Hey Sonos, open control for my house system from now until Saturday at 2 pm”). In some embodiments, a registered user or host may send a voice command (e.g., “Hey Sonos, restrict control for my living room to authorized guests”) that authorizes only some of the guests to send voice commands for a time period or window to control one or more PBDs 532, 534, 536, and 538 and/or computing device 506 in a playback zone.
In some embodiments, a registered user may have a different wakeup word or phrase for different voice commands that triggers a time period or window for the media playback system to receive additional voice commands associated with the wakeup word or phrase. For example, a registered user or host may have a user-specific wakeup word or phrase to send a voice command to add songs to a play queue (e.g., “Hey Sonos, let's queue up songs” “Yo, Sonos, queue songs,” “Alpha song queue,” etc), and may have a different user-specific wake up word or phrase to authorize guests in a house to control the media playback device (e.g., “Hey Sonos, open access,” “It's party time,” etc).
In some embodiments, a registered user or host may have a user-specific or universal wakeup word or phrase to send a voice command to authorize certain guests in a house to have restricted control of the media playback system for a time period or window. U.S. Patent Pub. No. 2013/0346859 entitled, “Systems, Methods, Apparatus, and Articles of Manufacture to Provide a Crowd-Sourced Playlist with Guest Access,” which is hereby incorporated by reference, provides in more detail some examples for restricted control of the media playback system.
Additionally, a registered user or host may have a user-specific or universal wakeup word or phrase to send a voice command to authorize registered guests in a house to have open control or restricted control of the media playback back system for a time period or window, while preventing nonregistered guests from having control. In some embodiments, a registered user or host may have a user-specific or universal wakeup word or phrase to send a voice command to authorize adults in a house to have open control or restricted control of the media playback system for a time period or window, while preventing children from having control. Many other examples, similar and different from the above, are possible.
In some embodiments, a registered user or host may specify the time period or window for the media playback system to receive additional voice commands. For example, a registered user or host may send a voice command (e.g., Hey, Sonos, open control for my house system for one hour”) that authorizes guests to send additional voice commands to control the media playback system for the specified time period (e.g., one hour). Many other examples, similar and different from the above, are possible.
In some embodiments, a registered user or host may close or key off the time period or window for receiving additional voice commands associated with the initial wakeup word or phrase. For example, if a registered user or host speaks a voice command with a wake up word or phrase that opens a time period or window to receive additional voice commands for an hour, the registered user or host may send another voice command (e.g., “Hey Sonos, queue songs complete”) to key off the one hour time period or window before the one hour time period expires. Many other examples, similar and different from the above, are possible.
In some embodiments, the media playback system may take actions based on receiving a wakeup word or wakeup phrase from a registered guest user. A registered guest user may have wakeup words or phrases stored in a guest profile. In response to determining that a wakeup word or wakeup phrase was received from a guest user, the media playback system may (i) determine whether there is a restriction setting associated with the guest user, (ii) configure an instruction for one or more PBDs based on the wakeup word or phrase and the assigned restriction setting for the guest user, and (iii) send the instruction to one or more PBDs for execution (e.g., to open a time period or window to receive additional voice commands associated with the wake up word command).
In some embodiments, the media playback system may refrain from taking actions based on receiving a wakeup word or phrase from a registered guest user if, for example, the media playback system has already received a voice command with a wakeup word or phrase from a registered user or host, and the time period or window to receive additional commands has not expired.
In some embodiments, the media playback system may take actions based on receiving a wakeup word or wakeup phrase from a registered guest user and subsequently close or key off the time period or window for receiving additional voice commands if the media playback device subsequently receives a voice command from a registered user or host. In some embodiments, the registered guest may close or key off the time period or window before it expires. In other embodiments, an adult may close or key off the time period or window before it expires if the registered guest is a child. Many other examples, similar and different from the above, are possible.
After configuring an instruction or set of instructions for the media playback system, some embodiments may send the instruction or set of instructions to one or more PBDs of the media playback system to execute the instructions. In some embodiments, the media playback system may send the instruction or set of instructions to computing device 506. In some embodiments, the media playback system may send the instruction or set of instructions to the networked microphone system.
Method 700 shown in
In addition, for the method 700 and other processes and methods disclosed herein, the flowchart shows functionality and operation of one possible implementation of some embodiments. In this regard, each block may represent a module, a segment, or a portion of program code, which includes one or more instructions executable by a processor for implementing specific logical functions or steps in the process. The program code may be stored on any type of computer readable medium, for example, such as a storage device including a disk or hard drive. The computer readable medium may include non-transitory computer readable medium, for example, such as tangible, non-transitory computer-readable media that stores data for short periods of time like register memory, processor cache and Random Access Memory (RAM). The computer readable medium may also include non-transitory media, such as secondary or persistent long term storage, like read only memory (ROM), optical or magnetic disks, compact-disc read only memory (CD-ROM), for example. The computer readable media may also be any other volatile or non-volatile storage systems. The computer readable medium may be considered a computer readable storage medium, for example, or a tangible storage device. In addition, for the method 700 and other processes and methods disclosed herein, each block in
Method 700 begins at block 702, which includes receiving a voice command for a media playback system. In some embodiments, receiving a voice command includes the media playback system receiving a voice command via one or more of PBDs 532, 534, 536, and 538 and/or computing device 506 (which is configured as a media playback system server). In one example, the computing device 506 may convert the voice command to an equivalent text command, and parse the text command to identify a command.
In some embodiments, one or more functions may be performed by the networked microphone system individually or in combination with the media playback system. In some embodiments, receiving a voice command includes the networked microphone system receiving a voice command via one or more of NMDs 512, 514, or 516, and transmitting the voice command to computing device 506 and/or one or more of PBDs 532, 534, 536, and 538 for further processing. In some embodiments, computing device 506 and/or one or more of PBDs 532, 534, 536, and 538 may convert the voice command to an equivalent text command, and parse the text command to identify a command. In some embodiments, the networked microphone system may convert the voice command to an equivalent text command and transmit the text command to computing device 506 and/or one or more of PBDs 532, 534, 536, and 538 to parse the text command and identify a command.
Next, method 700 advances to block 704, which includes determining whether the voice command was received from a registered user of the media playback system. In some embodiments, determining whether the voice command was received from a registered user may include computing device 506 determining whether there is a user profile stored on the media playback system that is associated with the voice command. For example, computing device 506 may try to match the voice command to information identifying a user's voice in a user profile.
In some embodiments, determining whether the voice command was received from a registered user may include determining whether the voice command matches the voice inputs stored in the media playback system's voice configuration settings. For example, a user may have previously configured the media playback system to recognize the user's voice by providing a voice input or a series of voice inputs for the media playback system to recognize and associate the user with. Voice configuration settings may be stored on the computing device 506 and/or one or more of PBDs 532, 534, 536, and 538. Alternatively, the computing device 506 may communicate with the networked microphone system to store the voice configuration settings.
In some embodiments, determining whether the voice command was received from a registered user may include determining a confidence level associated with a voice command received. The confidence level may be a confidence level associated with the person who spoke the command, e.g., a confidence level that the command was received from a registered user generally, a confidence level that the command was received from a specific registered user, a confidence level that the command was received from someone other than a registered user, a confidence level that the command was received from a registered guest, a confidence level that the command was received from a child, and/or a confidence level that the command was received from a particular child. The confidence level may also be a confidence level associated with the content of the request, e.g., a confidence level that the request was a request to play “AC/DC” rather than, for example, “Hayseed Dixie,” which are two very different bands with very similar sounding names. The confidence level may be determined based on user history, location, individually or in combination with any other information generally found in a user profile. In operation, determination of the confidence level may be performed by any one or more of CR 522, PBDs 532-538, NMDs 512-516, and/or computing devices 504-508, individually or in combination.
For example, in some embodiments, the media playback system receives a voice command from a registered user in the kitchen and determines a confidence level based on the voice command received. In operation, the media playback device may receive the voice command from any one or more of CR 522, NMDs 512-516, and/or PBDs 532-538. Next, the media playback system receives the same voice command from the registered user in another room in the user's house and determines a confidence level based on the voice command received. The media playback system may then determine a new confidence level based on the received commands from different devices in different rooms throughout the user's house, based at least in part on the room where the voice command was received. As a result, the media playback system may have a greater confidence level that the voice command was received from a registered user.
In another example, the media playback system may receive a voice command from a registered user and determine a confidence level based on user history. In particular, the media playback system may determine a higher confidence level if the voice command received includes an artist, playlist, genre, or any other information found in a user profile that is typically associated with the registered user. For example, if a registered user typically listens to songs by Michael Jackson, the media playback system may have a greater confidence level that the voice command to “Play Thriller” was received from a registered user. Likewise, if the registered user typically listens to songs by Michael Jackson or songs from the 1980's in general, the media playback system may have a greater confidence level that the voice command to “Play Thriller” is a command to play the song “Thriller” by the artist Michael Jackson rather than the song “Thriller” by the band Fall Out Boy. Many other examples, similar and different from the above, are possible.
In some embodiments, the media playback system may build a confidence level based on a registered user's pattern of voice commands found in a user's profile. For example, the media playback system may receive a voice command from a registered user to play a particular song by Britney Spears, and determine a confidence level based on the received voice command. Every time the media playback system receives the same voice command or similar voice command, such as a command to play another song by Britney Spears, the media playback system may build a higher confidence level and may have a greater confidence level that the voice command was received from that registered user.
Finally, method 700 advances to block 706, which includes in response to determining that the voice command was received from a registered user, configuring an instruction for the media playback system based on content from the voice command and information in a user profile for the registered user.
In some embodiments, the content from the voice command may include a command for one or more PBDs of the media playback system to play media content. In some embodiments, based on the command for one or more PBDs to play media content and information in a user profile for the registered user, the computing device 506 may configure an instruction or a set of instructions to cause the media playback system to obtain media or audio content from a preferred media source of a registered user.
In some embodiments, based on the command for the media playback system to play media content and information in a user profile for the registered user, the media playback system may configure an instruction or a set of instructions to cause the media playback system to play the media content via one or more PBDs of the media playback system.
In some embodiments, based on the command for the media playback system to play media content and information in a user profile for the registered user, the computing device 506 may include instructions to (i) configure the media playback system with one or more of the registered user's preferred playback settings and (ii) cause one or more PBDs of the media playback system to play the media content with the registered user's preferred playback settings. Preferred playback settings may be preferred playback settings stored in a registered user's user profile. Additionally or alternatively, preferred playback settings may be based on user history stored in a registered user's user profile. User history may include commonly used or previously used playback settings by the user to play media content.
In some embodiments, the content from the voice command may include a command for one or more PBDs of the media playback system to play media content but may not identify a particular listening zone or playback zone of the media playback system. Based on this content and information in a user profile for the registered user, such as user history, computing device 506 may configure an instruction or a set of instructions to cause the media playback system to play the media content via one or more media playback devices within the particular playback zone of the media playback system.
In some embodiments, the content from the voice command may include a command for the media playback system to modify a playback setting. Based on the command for the media playback system to modify a playback setting and information in a user profile for the registered user, computing device 506 may (i) configure an instruction or a set of instructions to cause the media playback system to modify the playback setting for one or more PBDs of the media playback system, and (ii) implement the configured instruction or set of instructions to modify the playback setting via the one or more PBDs.
Some embodiments include the media playback system determining whether the voice command was received from a child. In some embodiments, the computing device 506 may distinguish between an adult and a child based on information in a user profile if the child is one of the registered users of the media playback system. In some embodiments, the computing device 506 may distinguish between an adult and a child based on the tone or frequency of the user's voice.
In some embodiments, determining whether the voice command was received from a child may include determining a confidence level associated with a received voice command. As described above, a confidence level may be determined based on user history, location, individually or in combination with any other information generally found in a user profile.
For example, the media playback system may receive a voice command via a device (e.g., any of NMDs 512-516 or PBDs 532-538) in a particular room where a child is likely to be (e.g., child's bedroom, playroom, basement, etc). Because the command was received from a device located in a room where a child is likely to be, the media playback system may have a greater confidence level that the voice command was received from a child.
In another example, the media playback system may receive a voice command and determine a confidence level that the command was received from a child based on the content of the voice command. For example, if the media playback system receives a voice command to play a song from a cartoon show or movie, the media playback system may have a greater confidence level that the voice command was received from a child. Many other examples, similar and different from the above, are possible.
In response to determining that the voice command was received from a child, some embodiments may prevent one or more PBDs of the media playback system from playing given media that may be inappropriate for the child. Some embodiments may prevent the computing device 506 and/or one or more PBDs from modifying a playback setting based on the content of a child's voice command. For example, the computing device 506 may disregard a child's voice command to increase the volume of one or more PBDs.
Some embodiments include actions based on determining whether a voice command was received from a guest user instead of a registered user of the media playback system. In some embodiments, computing device 506 may have stored a previously created guest profile that may be associated with a particular guest. In some embodiments, computing device 506 may determine that a voice command was not received from a registered user, and may then ask the registered user if the voice command came from a guest.
In some embodiments, determining whether the voice command was received from a guest user may include the media playback system determining a confidence level associated with a voice command received. As described above, a confidence level may be determined based on user history, location, individually or in combination with any other information generally found in a user profile.
In response to determining that the voice command was received from a guest user, computing device 506 may (1) assign a restriction setting for the guest user, (2) configure an instruction for one or more PBDs based on content from the voice command and the assigned restriction setting for the guest user, and (3) send the instruction to one or more PBDs for execution. In some embodiments, assigning a restriction setting for a guest user may include computing device 506 matching the voice command to a particular guest profile stored on the computing device 506. The guest profile may include restriction settings, and information regarding the voice of the particular guest user, such as frequency or tone of the guest's voice, among other information previously described. A restriction setting may be any setting that limits the control of the media playback system.
Some embodiments include the media playback system applying an order of preference to resolve conflicting voice commands received from different users. Conflicting voice commands may be, for example, a voice command received from a user to play a song and a subsequent voice command received from another user to stop playing the song. Other examples are possible, such as a voice command received from a user to increase the volume of one or more playback devices (e.g., PBDs 532, 534, 536, and 538) and a subsequent voice command received from another user to decrease the volume. In particular, the media playback system may receive a voice command from a registered user or host to play a song in a playback zone. Subsequently, the media playback system may receive a conflicting voice command from a nonregistered user or guest to stop playing the song in the playback zone. To resolve this conflict, the media playback system may apply an order of preference in which voice commands received from a registered user have a higher priority than voice commands from a nonregistered user or guest.
In some embodiments, the media playback system may assign an order of preference in which voice commands received from registered guests have a higher priority than voice commands received from nonregistered guests. In some embodiments, voice commands received from one registered guest may have a higher priority than another registered guest. Additionally or alternatively, voice commands received from an adult may have a higher priority than a child.
In still further embodiments, controller-issued commands received by the media playback system (e.g., commands received from CR 522 or other computing devices configured to control the media playback system, or perhaps commands received from computing device 506) may have a lower priority than a registered user, but may have a higher priority than a nonregistered user or guest. In some embodiments, some registered guest may have a higher priority than controller-issued commands. Other examples of determining and assigning an order of preference are possible.
After configuring an instruction or set of instructions for the media playback system, some embodiments may send the instruction or set of instructions to one or more PBDs of the media playback system to execute the instructions. In some embodiments, the computing device 506 may send the instruction or set of instructions to the networked microphone system.
Method 800 shown in
In addition, for the method 800 and other processes and methods disclosed herein, the flowchart shows functionality and operation of one possible implementation of some embodiments. In this regard, each block may represent a module, a segment, or a portion of program code, which includes one or more instructions executable by one or more processors for implementing specific logical functions or steps in the process. The program code may be stored on any type of computer readable medium, for example, such as a storage device including a disk or hard drive. The computer readable medium may include non-transitory computer readable medium, for example, such as tangible, non-transitory computer-readable media that stores data for short periods of time like register memory, processor cache and Random Access Memory (RAM). The computer readable medium may also include non-transitory media, such as secondary or persistent long term storage, like read only memory (ROM), optical or magnetic disks, compact-disc read only memory (CD-ROM), for example. The computer readable media may also be any other volatile or non-volatile storage systems. The computer readable medium may be considered a computer readable storage medium, for example, or a tangible storage device. In addition, for the method 800 and other processes and methods disclosed herein, each block in
Method 800 begins at block 802, which includes receiving a wakeup word or wakeup phrase associated with a voice command for a media playback system. A wakeup word or phrase, as described above, may be a specific word or phrase (e.g., “Hey, Sonos”) stored in a user profile. In some embodiments, the media playback system, may receive a universal wakeup word or phrase (e.g., “Hey Sonos”) associated with a voice command of a registered user. Additionally or alternatively, the media playback system may receive a universal wakeup word or phrase associated with a voice command of a registered guest user. In some embodiments, the media playback system may be configured for different registered users to have different wake up words or phrases.
In some embodiments, a registered user may have a different, user-specific wakeup word or phrase for different voice commands. For example, the media playback system may receive a wakeup word or phrase to add songs to a play queue (e.g., “Hey Sonos, let's queue up songs” “Yo, Sonos, queue songs,” “Alpha song queue,” etc), and may receive a different user-specific wake up word or phrase to authorize guests in a house to control the media playback device (e.g., “Hey Sonos, open access,” “It's party time,” etc).
Next, method 800 advances to block 804, which includes determining whether the wakeup word associated with the voice command was received from a registered user of the media playback system. In some embodiments, determining whether the wakeup word associated with a voice command was received from a registered user may be similar to determining whether a voice command was received from a registered user described in block 704 for method 700.
Finally, method 800 advances to block 806, which includes in response to determining that the wakeup word associated with the voice command was received from a registered user, configuring an instruction for the media playback system based on the received wakeup word, content from the voice command, and information in a user profile for the registered user.
In some embodiments, the instruction for the media playback system may include an instruction to open a time period or window for the media playback system to receive additional voice commands associated with the received wakeup word from the registered user, a guest, and/or a nonregistered user. For example, in response to determining that the wakeup word to add songs to a play queue was received from a registered user, the media playback system may open a time period (e.g., five minutes) for the registered user to send additional voice commands to add specific songs to the play queue (e.g., “Add Thriller by Michael Jackson”).
In another example, in response to determining that the wakeup word to authorize all guests to control the media playback system was received from a registered user, the media playback system may open a time period (e.g., one hour) to allow all guests in a house to send voice commands to add songs to a play queue, play songs, or change the volume, among other functions for a user-defined or default time period or window.
Next, method 800 advances to block 806, which includes in response to determining that the wakeup word was received from a registered user, determining whether the wakeup word is associated with a restriction setting based on the received wakeup word or phrase, content from the voice command, and information in a user profile for the registered user.
In some embodiments, the media playback system may configure an instruction based on restriction settings in a user profile for the registered user or registered guest user. A wakeup word received from a registered user may be associated with restriction settings for certain guests. For example, a registered user or host may send a voice command (e.g., “Hey Sonos, restrict control for my living room to authorized guests”) that authorizes registered guests to send additional voice commands for a time period or window to control one or more PBDs 532, 534, 536, and 538 and/or computing device 506 in a playback zone, while preventing nonregistered guests from sending additional voice commands. In yet another example, the wake up word received may be associated with restriction settings for a child. Many other examples, similar and different from the above, are possible, including but not limited to the examples described elsewhere herein.
In some embodiments, a wakeup word received from a registered user may be associated with restriction settings that allow certain guests to have restricted control of the media playback system for a time period or window. U.S. Patent Pub. No. 2013/0346859 entitled, “Systems, Methods, Apparatus, and Articles of Manufacture to Provide a Crowd-Sourced Playlist with Guest Access,” which is hereby incorporated by reference, provides in more detail some examples for restricted control of the media playback system.
In some embodiments, in response to determining that a wakeup word or wakeup phrase was received from a guest user, the media playback system may (i) determine whether there is a restriction setting associated with the guest user, (ii) configure an instruction for one or more PBDs based on the wakeup word or phrase and the assigned restriction setting for the guest user, and (iii) send the instruction to one or more PBDs for execution (e.g., to open a time period or window to receive additional voice commands associated with the wake up word command).
In some embodiments, the media playback device, via the one or more PBDs 532, 534, 536, and 538 and/or computing device 506, may refrain from taking actions based on receiving a wakeup word or phrase from a registered guest user if, for example, the media playback system has already received a voice command with a wakeup word or phrase from a registered user or host, and the time period or window to receive additional commands has not expired.
After configuring an instruction or set of instructions for the media playback system, some embodiments may include sending commands or set of commands to one or more PBDs of the media playback system to execute the instructions. In some embodiments, the computing device 506 may send the commands or set of commands to one or more PBDs of the media playback system.
In some embodiments, after configuring an instruction or set of instructions for the media playback system to execute, a registered user or host may close or key off the time period or window for receiving additional voice commands associated with the instruction. For example, if a registered user or host sends a voice command with a wake up word or phrase that opens a time period or window to receive additional voice commands for an hour, the registered user or host may send another voice command (e.g., “Hey Sonos, queue songs complete”) to key off the one hour time period or window before the one hour time period expires. Many other examples, similar and different from the above, are possible.
In some embodiments, the media playback system may take actions based on receiving a wakeup word or wakeup phrase from a registered guest user and subsequently close or key off the time period or window for receiving additional voice commands if the media playback device subsequently receives a voice command from a registered user or host. In some embodiments, the registered guest may close or key off the time period or window before it expires. In other embodiments, an adult may close or key off the time period or window before it expires if the registered guest is a child. Many other examples, similar and different from the above, are possible.
IV. Conclusion
The description above discloses, among other things, various example systems, methods, apparatus, and articles of manufacture including, among other components, firmware and/or software executed on hardware. It is understood that such examples are merely illustrative and should not be considered as limiting. For example, it is contemplated that any or all of the firmware, hardware, and/or software aspects or components can be embodied exclusively in hardware, exclusively in software, exclusively in firmware, or in any combination of hardware, software, and/or firmware. Accordingly, the examples provided are not the only way(s) to implement such systems, methods, apparatus, and/or articles of manufacture.
Additionally, references herein to “embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment can be included in at least one example embodiment of an invention. The appearances of this phrase in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. As such, the embodiments described herein, explicitly and implicitly understood by one skilled in the art, can be combined with other embodiments.
The specification is presented largely in terms of illustrative environments, systems, procedures, steps, logic blocks, processing, and other symbolic representations that directly or indirectly resemble the operations of data processing devices coupled to networks. These process descriptions and representations are typically used by those skilled in the art to most effectively convey the substance of their work to others skilled in the art. Numerous specific details are set forth to provide a thorough understanding of the present disclosure. However, it is understood to those skilled in the art that certain embodiments of the present disclosure can be practiced without certain, specific details. In other instances, well known methods, procedures, components, and circuitry have not been described in detail to avoid unnecessarily obscuring aspects of the embodiments. Accordingly, the scope of the present disclosure is defined by the appended claims rather than the forgoing description of embodiments.
When any of the appended claims are read to cover a purely software and/or firmware implementation, at least one of the elements in at least one example is hereby expressly defined to include a tangible, non-transitory medium such as a memory, DVD, CD, Blu-ray, and so on, storing the software and/or firmware.
The present application is a continuation of U.S. application Ser. No. 15/973,413, filed May 7, 2018, which is a continuation of U.S. application Ser. No. 15/131,776, filed Apr. 18, 2016, which claims priority to (i) U.S. Provisional App. 62/298,433, filed Feb. 22, 2016, titled “Room-corrected Voice Detection,” (ii) U.S. Provisional App. 62/298,439, filed Feb. 22, 2016, titled “Content Mixing,” (iii) U.S. Provisional App. 62/298,425, filed Feb. 22, 2016, titled “Music Service Selection,” (iv) U.S. Provisional App. 62/298,350, filed Feb. 22, 2016, titled “Metadata exchange involving a networked playback system and a networked microphone system,” (v) U.S. Provisional App. 62/298,388, filed Feb. 22, 2016, titled “Handling of loss of pairing between networked devices,” and (vi) U.S. Provisional App. 62/298,393, filed Feb. 22, 2016, titled “Action based on User ID.” The entire contents of all of these applications is incorporated herein by reference. This application also incorporates herein by reference the entire contents of (i) U.S. Provisional App. 62/298,410, filed Feb. 22, 2016, titled “Default Playback Device(s)”; (ii) U.S. Provisional App. 62/298,418, filed Feb. 22, 2016, titled “Audio Response Playback”; and (iii) U.S. Provisional App. 62/312,350, filed Mar. 23, 2016, titled “Voice Control of a Media Playback System.”
Number | Name | Date | Kind |
---|---|---|---|
4741038 | Elko et al. | Apr 1988 | A |
4941187 | Slater | Jul 1990 | A |
4974213 | Siwecki | Nov 1990 | A |
5036538 | Oken et al. | Jul 1991 | A |
5440644 | Farinelli et al. | Aug 1995 | A |
5588065 | Tanaka et al. | Dec 1996 | A |
5740260 | Odom | Apr 1998 | A |
5761320 | Farinelli et al. | Jun 1998 | A |
5923902 | Inagaki | Jul 1999 | A |
5949414 | Namikata et al. | Sep 1999 | A |
6032202 | Lea et al. | Feb 2000 | A |
6088459 | Hobelsberger | Jul 2000 | A |
6256554 | Dilorenzo | Jul 2001 | B1 |
6301603 | Maher et al. | Oct 2001 | B1 |
6311157 | Strong | Oct 2001 | B1 |
6366886 | Dragosh et al. | Apr 2002 | B1 |
6404811 | Cvetko et al. | Jun 2002 | B1 |
6408078 | Hobelsberger | Jun 2002 | B1 |
6469633 | Wachter et al. | Oct 2002 | B1 |
6522886 | Youngs et al. | Feb 2003 | B1 |
6594347 | Calder et al. | Jul 2003 | B1 |
6594630 | Zlokarnik et al. | Jul 2003 | B1 |
6611537 | Edens et al. | Aug 2003 | B1 |
6611604 | Irby et al. | Aug 2003 | B1 |
6631410 | Kowalski et al. | Oct 2003 | B1 |
6757517 | Chang | Jun 2004 | B2 |
6778869 | Champion | Aug 2004 | B2 |
6937977 | Gerson | Aug 2005 | B2 |
7099821 | Visser et al. | Aug 2006 | B2 |
7103542 | Doyle | Sep 2006 | B2 |
7130608 | Hollstrom et al. | Oct 2006 | B2 |
7130616 | Janik | Oct 2006 | B2 |
7143939 | Henzerling | Dec 2006 | B2 |
7174299 | Fujii et al. | Feb 2007 | B2 |
7228275 | Endo et al. | Jun 2007 | B1 |
7236773 | Thomas | Jun 2007 | B2 |
7295548 | Blank et al. | Nov 2007 | B2 |
7356471 | Ito et al. | Apr 2008 | B2 |
7383297 | Atsmon et al. | Jun 2008 | B1 |
7391791 | Balassanian et al. | Jun 2008 | B2 |
7483538 | McCarty et al. | Jan 2009 | B2 |
7571014 | Lambourne et al. | Aug 2009 | B1 |
7577757 | Carter et al. | Aug 2009 | B2 |
7630501 | Blank et al. | Dec 2009 | B2 |
7643894 | Braithwaite et al. | Jan 2010 | B2 |
7657910 | McAulay et al. | Feb 2010 | B1 |
7661107 | Van et al. | Feb 2010 | B1 |
7702508 | Bennett | Apr 2010 | B2 |
7792311 | Holmgren et al. | Sep 2010 | B1 |
7853341 | McCarty et al. | Dec 2010 | B2 |
7961892 | Fedigan | Jun 2011 | B2 |
7987294 | Bryce et al. | Jul 2011 | B2 |
8014423 | Thaler et al. | Sep 2011 | B2 |
8019076 | Lambert | Sep 2011 | B1 |
8032383 | Bhardwaj et al. | Oct 2011 | B1 |
8041565 | Bhardwaj | Oct 2011 | B1 |
8045952 | Qureshey et al. | Oct 2011 | B2 |
8073125 | Zhang et al. | Dec 2011 | B2 |
8073681 | Baldwin et al. | Dec 2011 | B2 |
8085947 | Haulick et al. | Dec 2011 | B2 |
8103009 | McCarty et al. | Jan 2012 | B2 |
8136040 | Fleming | Mar 2012 | B2 |
8165867 | Fish | Apr 2012 | B1 |
8233632 | MacDonald et al. | Jul 2012 | B1 |
8234395 | Millington | Jul 2012 | B2 |
8239206 | Lebeau et al. | Aug 2012 | B1 |
8255224 | Singleton et al. | Aug 2012 | B2 |
8284982 | Bailey | Oct 2012 | B2 |
8290603 | Lambourne | Oct 2012 | B1 |
8340975 | Rosenberger | Dec 2012 | B1 |
8364481 | Strope et al. | Jan 2013 | B2 |
8385557 | Tashev et al. | Feb 2013 | B2 |
8386261 | Mellott et al. | Feb 2013 | B2 |
8386523 | Mody et al. | Feb 2013 | B2 |
8423893 | Ramsay et al. | Apr 2013 | B2 |
8428758 | Naik et al. | Apr 2013 | B2 |
8453058 | Coccaro et al. | May 2013 | B1 |
8473618 | Spear et al. | Jun 2013 | B2 |
8483853 | Lambourne | Jul 2013 | B1 |
8484025 | Moreno et al. | Jul 2013 | B1 |
8588849 | Patterson et al. | Nov 2013 | B2 |
8600443 | Kawaguchi et al. | Dec 2013 | B2 |
8710970 | Oelrich et al. | Apr 2014 | B2 |
8719039 | Sharifi | May 2014 | B1 |
8738925 | Park et al. | May 2014 | B1 |
8775191 | Sharifi et al. | Jul 2014 | B1 |
8831761 | Kemp et al. | Sep 2014 | B2 |
8831957 | Taubman et al. | Sep 2014 | B2 |
8848879 | Coughlan et al. | Sep 2014 | B1 |
8861756 | Zhu et al. | Oct 2014 | B2 |
8874448 | Kauffmann et al. | Oct 2014 | B1 |
8938394 | Faaborg et al. | Jan 2015 | B1 |
8942252 | Balassanian et al. | Jan 2015 | B2 |
8983383 | Haskin | Mar 2015 | B1 |
8983844 | Thomas et al. | Mar 2015 | B1 |
9015049 | Baldwin et al. | Apr 2015 | B2 |
9042556 | Kallai et al. | May 2015 | B2 |
9060224 | List | Jun 2015 | B1 |
9094539 | Noble | Jul 2015 | B1 |
9098467 | Blanksteen et al. | Aug 2015 | B1 |
9124650 | Maharajh et al. | Sep 2015 | B2 |
9124711 | Park et al. | Sep 2015 | B2 |
9148742 | Koulomzin et al. | Sep 2015 | B1 |
9190043 | Krisch et al. | Nov 2015 | B2 |
9208785 | Ben-David et al. | Dec 2015 | B2 |
9215545 | Dublin et al. | Dec 2015 | B2 |
9245527 | Lindahl | Jan 2016 | B2 |
9251793 | Lebeau et al. | Feb 2016 | B2 |
9253572 | Beddingfield, Sr. et al. | Feb 2016 | B2 |
9262612 | Cheyer | Feb 2016 | B2 |
9263042 | Sharifi | Feb 2016 | B1 |
9275637 | Salvador et al. | Mar 2016 | B1 |
9288597 | Carlsson et al. | Mar 2016 | B2 |
9300266 | Grokop | Mar 2016 | B2 |
9304736 | Whiteley et al. | Apr 2016 | B1 |
9307321 | Unruh | Apr 2016 | B1 |
9318107 | Sharifi | Apr 2016 | B1 |
9319816 | Narayanan | Apr 2016 | B1 |
9324322 | Torok et al. | Apr 2016 | B1 |
9335819 | Jaeger et al. | May 2016 | B1 |
9354687 | Bansal et al. | May 2016 | B2 |
9361878 | Boukadakis | Jun 2016 | B2 |
9361885 | Ganong, III et al. | Jun 2016 | B2 |
9368105 | Freed et al. | Jun 2016 | B1 |
9373329 | Strope et al. | Jun 2016 | B2 |
9374634 | Macours | Jun 2016 | B2 |
9386154 | Baciu et al. | Jul 2016 | B2 |
9390708 | Hoffmeister | Jul 2016 | B1 |
9401058 | De La Fuente et al. | Jul 2016 | B2 |
9412392 | Lindahl et al. | Aug 2016 | B2 |
9426567 | Lee et al. | Aug 2016 | B2 |
9431021 | Scalise et al. | Aug 2016 | B1 |
9443516 | Katuri et al. | Sep 2016 | B2 |
9443527 | Watanabe et al. | Sep 2016 | B1 |
9472201 | Sleator | Oct 2016 | B1 |
9472203 | Ayrapetian et al. | Oct 2016 | B1 |
9484030 | Meaney et al. | Nov 2016 | B1 |
9489948 | Chu et al. | Nov 2016 | B1 |
9494683 | Sadek | Nov 2016 | B1 |
9509269 | Rosenberg | Nov 2016 | B1 |
9510101 | Polleros | Nov 2016 | B1 |
9514476 | Kay et al. | Dec 2016 | B2 |
9514752 | Sharifi | Dec 2016 | B2 |
9516081 | Tebbs et al. | Dec 2016 | B2 |
9532139 | Lu et al. | Dec 2016 | B1 |
9536541 | Chen et al. | Jan 2017 | B2 |
9548053 | Basye et al. | Jan 2017 | B1 |
9548066 | Jain et al. | Jan 2017 | B2 |
9552816 | Vanlund et al. | Jan 2017 | B2 |
9554210 | Ayrapetian et al. | Jan 2017 | B1 |
9560441 | McDonough, Jr. et al. | Jan 2017 | B1 |
9576591 | Kim et al. | Feb 2017 | B2 |
9601116 | Casado et al. | Mar 2017 | B2 |
9615170 | Kirsch et al. | Apr 2017 | B2 |
9615171 | O'Neill et al. | Apr 2017 | B1 |
9626695 | Balasubramanian et al. | Apr 2017 | B2 |
9632748 | Faaborg et al. | Apr 2017 | B2 |
9633186 | Ingrassia, Jr. et al. | Apr 2017 | B2 |
9633368 | Greenzeiger et al. | Apr 2017 | B2 |
9633660 | Haughay et al. | Apr 2017 | B2 |
9633661 | Typrin et al. | Apr 2017 | B1 |
9633671 | Giacobello et al. | Apr 2017 | B2 |
9633674 | Sinha et al. | Apr 2017 | B2 |
9640179 | Hart et al. | May 2017 | B1 |
9640183 | Jung et al. | May 2017 | B2 |
9641919 | Poole et al. | May 2017 | B1 |
9646614 | Bellegarda et al. | May 2017 | B2 |
9648564 | Cui et al. | May 2017 | B1 |
9653060 | Hilmes et al. | May 2017 | B1 |
9653075 | Chen et al. | May 2017 | B1 |
9659555 | Hilmes et al. | May 2017 | B1 |
9672821 | Krishnaswamy et al. | Jun 2017 | B2 |
9674587 | Triplett et al. | Jun 2017 | B2 |
9685171 | Yang | Jun 2017 | B1 |
9691378 | Meyers et al. | Jun 2017 | B1 |
9691379 | Mathias et al. | Jun 2017 | B1 |
9697826 | Sainath et al. | Jul 2017 | B2 |
9697828 | Prasad et al. | Jul 2017 | B1 |
9698999 | Mutagi et al. | Jul 2017 | B2 |
9704478 | Vitaladevuni et al. | Jul 2017 | B1 |
9706320 | Starobin et al. | Jul 2017 | B2 |
9721566 | Newendorp et al. | Aug 2017 | B2 |
9721568 | Polansky et al. | Aug 2017 | B1 |
9721570 | Beal et al. | Aug 2017 | B1 |
9728188 | Rosen et al. | Aug 2017 | B1 |
9734822 | Sundaram et al. | Aug 2017 | B1 |
9736578 | Iyengar et al. | Aug 2017 | B2 |
9743204 | Welch et al. | Aug 2017 | B1 |
9743207 | Hartung | Aug 2017 | B1 |
9747011 | Lewis et al. | Aug 2017 | B2 |
9747899 | Pogue et al. | Aug 2017 | B2 |
9747920 | Ayrapetian et al. | Aug 2017 | B2 |
9747926 | Sharifi et al. | Aug 2017 | B2 |
9749760 | Lambourne | Aug 2017 | B2 |
9754605 | Chhetri | Sep 2017 | B1 |
9756422 | Paquier et al. | Sep 2017 | B2 |
9762967 | Clarke et al. | Sep 2017 | B2 |
9767786 | Starobin et al. | Sep 2017 | B2 |
9769420 | Moses | Sep 2017 | B1 |
9779725 | Sun et al. | Oct 2017 | B2 |
9779732 | Lee et al. | Oct 2017 | B2 |
9779734 | Lee | Oct 2017 | B2 |
9779735 | Civelli et al. | Oct 2017 | B2 |
9805733 | Park | Oct 2017 | B2 |
9811314 | Plagge et al. | Nov 2017 | B2 |
9813810 | Nongpiur | Nov 2017 | B1 |
9813812 | Berthelsen et al. | Nov 2017 | B2 |
9818407 | Secker-Walker et al. | Nov 2017 | B1 |
9820036 | Tritschler et al. | Nov 2017 | B1 |
9820039 | Lang | Nov 2017 | B2 |
9826306 | Lang | Nov 2017 | B2 |
9865259 | Typrin et al. | Jan 2018 | B1 |
9865264 | Gelfenbeyn et al. | Jan 2018 | B2 |
9875740 | Kumar et al. | Jan 2018 | B1 |
9881616 | Beckley et al. | Jan 2018 | B2 |
9898250 | Williams et al. | Feb 2018 | B1 |
9899021 | Vitaladevuni et al. | Feb 2018 | B1 |
9900723 | Choisel et al. | Feb 2018 | B1 |
9916839 | Scalise et al. | Mar 2018 | B1 |
9947316 | Millington et al. | Apr 2018 | B2 |
9947333 | David | Apr 2018 | B1 |
9972318 | Kelly et al. | May 2018 | B1 |
9972343 | Thorson et al. | May 2018 | B1 |
9973849 | Zhang et al. | May 2018 | B1 |
9979560 | Kim et al. | May 2018 | B2 |
9992642 | Rapp et al. | Jun 2018 | B1 |
10013381 | Mayman et al. | Jul 2018 | B2 |
10013995 | Lashkari et al. | Jul 2018 | B1 |
10025447 | Dixit et al. | Jul 2018 | B1 |
10026401 | Mutagi et al. | Jul 2018 | B1 |
10048930 | Vega et al. | Aug 2018 | B1 |
10049675 | Haughay | Aug 2018 | B2 |
10051366 | Buoni et al. | Aug 2018 | B1 |
10051600 | Zhong et al. | Aug 2018 | B1 |
10057698 | Drinkwater et al. | Aug 2018 | B2 |
RE47049 | Zhu et al. | Sep 2018 | E |
10068573 | Aykac et al. | Sep 2018 | B1 |
10074369 | Devaraj et al. | Sep 2018 | B2 |
10074371 | Wang et al. | Sep 2018 | B1 |
10079015 | Lockhart et al. | Sep 2018 | B1 |
10089981 | Elangovan et al. | Oct 2018 | B1 |
10108393 | Millington et al. | Oct 2018 | B2 |
10115400 | Wilberding | Oct 2018 | B2 |
10116748 | Farmer et al. | Oct 2018 | B2 |
10127911 | Kim et al. | Nov 2018 | B2 |
10134388 | Lilly | Nov 2018 | B1 |
10134399 | Lang et al. | Nov 2018 | B2 |
10136204 | Poole et al. | Nov 2018 | B1 |
10152969 | Reilly et al. | Dec 2018 | B2 |
10181323 | Beckhardt et al. | Jan 2019 | B2 |
10186265 | Lockhart et al. | Jan 2019 | B1 |
10186266 | Devaraj et al. | Jan 2019 | B1 |
10192546 | Piersol et al. | Jan 2019 | B1 |
10224056 | Torok et al. | Mar 2019 | B1 |
10225651 | Lang | Mar 2019 | B2 |
10229680 | Gillespie et al. | Mar 2019 | B1 |
10241754 | Kadarundalagi Raghuram Doss et al. | Mar 2019 | B1 |
10248376 | Keyser-Allen et al. | Apr 2019 | B2 |
10276161 | Hughes et al. | Apr 2019 | B2 |
10297256 | Reilly et al. | May 2019 | B2 |
10318236 | Pal et al. | Jun 2019 | B1 |
10339917 | Aleksic et al. | Jul 2019 | B2 |
10339957 | Chenier et al. | Jul 2019 | B1 |
10346122 | Morgan | Jul 2019 | B1 |
10354650 | Gruenstein et al. | Jul 2019 | B2 |
10354658 | Wilberding | Jul 2019 | B2 |
10365887 | Mulherkar | Jul 2019 | B1 |
10365889 | Plagge et al. | Jul 2019 | B2 |
10366688 | Gunn et al. | Jul 2019 | B2 |
10366699 | Dharia et al. | Jul 2019 | B1 |
10374816 | Leblang et al. | Aug 2019 | B1 |
10381001 | Gunn et al. | Aug 2019 | B2 |
10381002 | Gunn et al. | Aug 2019 | B2 |
10381003 | Wakisaka et al. | Aug 2019 | B2 |
10388272 | Thomson et al. | Aug 2019 | B1 |
10433058 | Torgerson et al. | Oct 2019 | B1 |
10445057 | Vega et al. | Oct 2019 | B2 |
10445365 | Luke et al. | Oct 2019 | B2 |
10469966 | Lambourne | Nov 2019 | B2 |
10499146 | Lang et al. | Dec 2019 | B2 |
10510340 | Fu et al. | Dec 2019 | B1 |
10511904 | Buoni et al. | Dec 2019 | B2 |
10515625 | Metallinou et al. | Dec 2019 | B1 |
10522146 | Tushinskiy | Dec 2019 | B1 |
10546583 | White et al. | Jan 2020 | B2 |
10565998 | Wilberding | Feb 2020 | B2 |
10573312 | Thomson et al. | Feb 2020 | B1 |
10573321 | Smith et al. | Feb 2020 | B1 |
10580405 | Wang et al. | Mar 2020 | B1 |
10586540 | Smith et al. | Mar 2020 | B1 |
10599287 | Kumar et al. | Mar 2020 | B2 |
10600406 | Shapiro et al. | Mar 2020 | B1 |
10602268 | Soto | Mar 2020 | B1 |
10614807 | Beckhardt et al. | Apr 2020 | B2 |
10621981 | Sereshki | Apr 2020 | B2 |
10622009 | Zhang et al. | Apr 2020 | B1 |
10623811 | Cwik | Apr 2020 | B1 |
10624612 | Sumi et al. | Apr 2020 | B2 |
10643609 | Pogue et al. | May 2020 | B1 |
10645130 | Corbin et al. | May 2020 | B2 |
10672383 | Thomson et al. | Jun 2020 | B1 |
10679625 | Lockhart et al. | Jun 2020 | B1 |
10681460 | Woo et al. | Jun 2020 | B2 |
10685669 | Lan et al. | Jun 2020 | B1 |
10694608 | Baker et al. | Jun 2020 | B2 |
10706843 | Elangovan et al. | Jul 2020 | B1 |
10712997 | Wilberding et al. | Jul 2020 | B2 |
10728196 | Wang | Jul 2020 | B2 |
10740065 | Jarvis et al. | Aug 2020 | B2 |
10748531 | Kim | Aug 2020 | B2 |
10762896 | Yavagal et al. | Sep 2020 | B1 |
10777189 | Fu et al. | Sep 2020 | B1 |
10777203 | Pasko | Sep 2020 | B1 |
10797667 | Fish et al. | Oct 2020 | B2 |
10825471 | Walley et al. | Nov 2020 | B2 |
10837667 | Nelson et al. | Nov 2020 | B2 |
10847143 | Millington et al. | Nov 2020 | B2 |
10847149 | Mok et al. | Nov 2020 | B1 |
10848885 | Lambourne | Nov 2020 | B2 |
RE48371 | Zhu et al. | Dec 2020 | E |
10867596 | Yoneda et al. | Dec 2020 | B2 |
10878811 | Smith et al. | Dec 2020 | B2 |
10878826 | Li et al. | Dec 2020 | B2 |
10897679 | Lambourne | Jan 2021 | B2 |
10911596 | Do et al. | Feb 2021 | B1 |
10943598 | Singh et al. | Mar 2021 | B2 |
10971158 | Patangay et al. | Apr 2021 | B1 |
11127405 | Antos et al. | Sep 2021 | B1 |
11212612 | Lang et al. | Dec 2021 | B2 |
11411763 | MacKay et al. | Aug 2022 | B2 |
20010003173 | Lim | Jun 2001 | A1 |
20010042107 | Palm | Nov 2001 | A1 |
20020022453 | Balog et al. | Feb 2002 | A1 |
20020026442 | Lipscomb et al. | Feb 2002 | A1 |
20020034280 | Infosino | Mar 2002 | A1 |
20020046023 | Fujii et al. | Apr 2002 | A1 |
20020072816 | Shdema et al. | Jun 2002 | A1 |
20020116196 | Tran | Aug 2002 | A1 |
20020124097 | Isely et al. | Sep 2002 | A1 |
20030015354 | Edwards et al. | Jan 2003 | A1 |
20030038848 | Lee et al. | Feb 2003 | A1 |
20030040908 | Yang et al. | Feb 2003 | A1 |
20030070182 | Pierre et al. | Apr 2003 | A1 |
20030070869 | Hlibowicki | Apr 2003 | A1 |
20030072462 | Hlibowicki | Apr 2003 | A1 |
20030095672 | Hobelsberger | May 2003 | A1 |
20030130850 | Badt et al. | Jul 2003 | A1 |
20030157951 | Hasty | Aug 2003 | A1 |
20030235244 | Pessoa et al. | Dec 2003 | A1 |
20040024478 | Hans et al. | Feb 2004 | A1 |
20040093219 | Shin et al. | May 2004 | A1 |
20040105566 | Matsunaga et al. | Jun 2004 | A1 |
20040127241 | Shostak | Jul 2004 | A1 |
20040128135 | Anastasakos et al. | Jul 2004 | A1 |
20040234088 | McCarty et al. | Nov 2004 | A1 |
20050031131 | Browning et al. | Feb 2005 | A1 |
20050031132 | Browning et al. | Feb 2005 | A1 |
20050031133 | Browning et al. | Feb 2005 | A1 |
20050031134 | Leske | Feb 2005 | A1 |
20050031137 | Browning et al. | Feb 2005 | A1 |
20050031138 | Browning et al. | Feb 2005 | A1 |
20050031139 | Browning et al. | Feb 2005 | A1 |
20050031140 | Browning | Feb 2005 | A1 |
20050033582 | Gadd et al. | Feb 2005 | A1 |
20050047606 | Lee et al. | Mar 2005 | A1 |
20050077843 | Benditt | Apr 2005 | A1 |
20050164664 | DiFonzo et al. | Jul 2005 | A1 |
20050195988 | Tashev et al. | Sep 2005 | A1 |
20050201254 | Looney et al. | Sep 2005 | A1 |
20050207584 | Bright | Sep 2005 | A1 |
20050235334 | Togashi et al. | Oct 2005 | A1 |
20050254662 | Blank et al. | Nov 2005 | A1 |
20050268234 | Rossi et al. | Dec 2005 | A1 |
20050283330 | Laraia et al. | Dec 2005 | A1 |
20050283475 | Beranek et al. | Dec 2005 | A1 |
20060004834 | Pyhalammi et al. | Jan 2006 | A1 |
20060023945 | King et al. | Feb 2006 | A1 |
20060041431 | Maes | Feb 2006 | A1 |
20060093128 | Oxford | May 2006 | A1 |
20060104451 | Browning et al. | May 2006 | A1 |
20060147058 | Wang | Jul 2006 | A1 |
20060190269 | Tessel et al. | Aug 2006 | A1 |
20060190968 | Jung et al. | Aug 2006 | A1 |
20060247913 | Huerta et al. | Nov 2006 | A1 |
20060262943 | Oxford | Nov 2006 | A1 |
20070018844 | Sutardja | Jan 2007 | A1 |
20070019815 | Asada et al. | Jan 2007 | A1 |
20070033043 | Hyakumoto | Feb 2007 | A1 |
20070038999 | Millington | Feb 2007 | A1 |
20070060054 | Romesburg | Mar 2007 | A1 |
20070071206 | Gainsboro et al. | Mar 2007 | A1 |
20070071255 | Schobben | Mar 2007 | A1 |
20070076131 | Li et al. | Apr 2007 | A1 |
20070076906 | Takagi et al. | Apr 2007 | A1 |
20070140058 | McIntosh et al. | Jun 2007 | A1 |
20070140521 | Mitobe et al. | Jun 2007 | A1 |
20070142944 | Goldberg et al. | Jun 2007 | A1 |
20070147651 | Mitobe et al. | Jun 2007 | A1 |
20070201639 | Park et al. | Aug 2007 | A1 |
20070254604 | Kim | Nov 2007 | A1 |
20070286426 | Xiang et al. | Dec 2007 | A1 |
20080008333 | Nishikawa et al. | Jan 2008 | A1 |
20080031466 | Buck et al. | Feb 2008 | A1 |
20080037814 | Shau | Feb 2008 | A1 |
20080090537 | Sutardja | Apr 2008 | A1 |
20080090617 | Sutardja | Apr 2008 | A1 |
20080144858 | Khawand et al. | Jun 2008 | A1 |
20080146289 | Korneluk et al. | Jun 2008 | A1 |
20080182518 | Lo | Jul 2008 | A1 |
20080207115 | Lee et al. | Aug 2008 | A1 |
20080208594 | Cross et al. | Aug 2008 | A1 |
20080221897 | Cerra et al. | Sep 2008 | A1 |
20080247530 | Barton et al. | Oct 2008 | A1 |
20080248797 | Freeman et al. | Oct 2008 | A1 |
20080291896 | Tuubel et al. | Nov 2008 | A1 |
20080301729 | Broos et al. | Dec 2008 | A1 |
20090003620 | McKillop et al. | Jan 2009 | A1 |
20090005893 | Sugii et al. | Jan 2009 | A1 |
20090010445 | Matsuo | Jan 2009 | A1 |
20090018828 | Nakadai et al. | Jan 2009 | A1 |
20090043206 | Towfiq et al. | Feb 2009 | A1 |
20090046866 | Feng et al. | Feb 2009 | A1 |
20090052688 | Ishibashi et al. | Feb 2009 | A1 |
20090076821 | Brenner et al. | Mar 2009 | A1 |
20090153289 | Hope et al. | Jun 2009 | A1 |
20090191854 | Beason | Jul 2009 | A1 |
20090197524 | Haff et al. | Aug 2009 | A1 |
20090220107 | Every et al. | Sep 2009 | A1 |
20090228919 | Zott et al. | Sep 2009 | A1 |
20090238377 | Ramakrishnan et al. | Sep 2009 | A1 |
20090238386 | Usher et al. | Sep 2009 | A1 |
20090248397 | Garcia et al. | Oct 2009 | A1 |
20090249222 | Schmidt et al. | Oct 2009 | A1 |
20090264072 | Dai | Oct 2009 | A1 |
20090323907 | Gupta et al. | Dec 2009 | A1 |
20090326949 | Douthitt et al. | Dec 2009 | A1 |
20100014690 | Wolff et al. | Jan 2010 | A1 |
20100023638 | Bowman | Jan 2010 | A1 |
20100035593 | Franco et al. | Feb 2010 | A1 |
20100041443 | Yokota | Feb 2010 | A1 |
20100070922 | DeMaio et al. | Mar 2010 | A1 |
20100075723 | Min et al. | Mar 2010 | A1 |
20100088100 | Lindahl | Apr 2010 | A1 |
20100092004 | Kuze | Apr 2010 | A1 |
20100161335 | Whynot | Jun 2010 | A1 |
20100172516 | Lastrucci | Jul 2010 | A1 |
20100178873 | Lee et al. | Jul 2010 | A1 |
20100179806 | Zhang et al. | Jul 2010 | A1 |
20100179874 | Higgins et al. | Jul 2010 | A1 |
20100185448 | Meisel | Jul 2010 | A1 |
20100211199 | Naik et al. | Aug 2010 | A1 |
20100278351 | Fozunbal et al. | Nov 2010 | A1 |
20110033059 | Bhaskar et al. | Feb 2011 | A1 |
20110035580 | Wang et al. | Feb 2011 | A1 |
20110044461 | Kuech et al. | Feb 2011 | A1 |
20110044489 | Saiki et al. | Feb 2011 | A1 |
20110046952 | Koshinaka | Feb 2011 | A1 |
20110066634 | Phillips et al. | Mar 2011 | A1 |
20110091055 | Leblanc | Apr 2011 | A1 |
20110103615 | Sun | May 2011 | A1 |
20110131032 | Yang et al. | Jun 2011 | A1 |
20110145581 | Malhotra et al. | Jun 2011 | A1 |
20110170707 | Yamada et al. | Jul 2011 | A1 |
20110182436 | Murgia et al. | Jul 2011 | A1 |
20110202924 | Banguero et al. | Aug 2011 | A1 |
20110218656 | Bishop et al. | Sep 2011 | A1 |
20110267985 | Wilkinson et al. | Nov 2011 | A1 |
20110276333 | Wang et al. | Nov 2011 | A1 |
20110280422 | Neumeyer et al. | Nov 2011 | A1 |
20110285808 | Feng et al. | Nov 2011 | A1 |
20110289506 | Trivi et al. | Nov 2011 | A1 |
20110299706 | Sakai | Dec 2011 | A1 |
20120020486 | Fried et al. | Jan 2012 | A1 |
20120022863 | Cho et al. | Jan 2012 | A1 |
20120022864 | Leman et al. | Jan 2012 | A1 |
20120078635 | Rothkopf et al. | Mar 2012 | A1 |
20120086568 | Scott et al. | Apr 2012 | A1 |
20120123268 | Tanaka et al. | May 2012 | A1 |
20120128160 | Kim et al. | May 2012 | A1 |
20120131125 | Seidel et al. | May 2012 | A1 |
20120148075 | Goh et al. | Jun 2012 | A1 |
20120162540 | Ouchi et al. | Jun 2012 | A1 |
20120163603 | Abe et al. | Jun 2012 | A1 |
20120177215 | Bose et al. | Jul 2012 | A1 |
20120183149 | Hiroe | Jul 2012 | A1 |
20120224715 | Kikkeri | Sep 2012 | A1 |
20120245941 | Cheyer | Sep 2012 | A1 |
20120288100 | Cho | Nov 2012 | A1 |
20120297284 | Matthews, III et al. | Nov 2012 | A1 |
20120308044 | Vander Mey et al. | Dec 2012 | A1 |
20120308046 | Muza | Dec 2012 | A1 |
20130006453 | Wang et al. | Jan 2013 | A1 |
20130024018 | Chang et al. | Jan 2013 | A1 |
20130034241 | Pandey et al. | Feb 2013 | A1 |
20130039527 | Jensen et al. | Feb 2013 | A1 |
20130051755 | Brown et al. | Feb 2013 | A1 |
20130058492 | Silzle et al. | Mar 2013 | A1 |
20130066453 | Seefeldt | Mar 2013 | A1 |
20130080146 | Kato et al. | Mar 2013 | A1 |
20130124211 | McDonough | May 2013 | A1 |
20130148821 | Sorensen | Jun 2013 | A1 |
20130170647 | Reilly et al. | Jul 2013 | A1 |
20130179173 | Lee et al. | Jul 2013 | A1 |
20130183944 | Mozer et al. | Jul 2013 | A1 |
20130191119 | Sugiyama | Jul 2013 | A1 |
20130191122 | Mason | Jul 2013 | A1 |
20130198298 | Li et al. | Aug 2013 | A1 |
20130211826 | Mannby | Aug 2013 | A1 |
20130216056 | Thyssen | Aug 2013 | A1 |
20130262101 | Srinivasan | Oct 2013 | A1 |
20130315420 | You | Nov 2013 | A1 |
20130317635 | Bates et al. | Nov 2013 | A1 |
20130322462 | Poulsen | Dec 2013 | A1 |
20130322665 | Bennett et al. | Dec 2013 | A1 |
20130324031 | Loureiro | Dec 2013 | A1 |
20130329896 | Krishnaswamy et al. | Dec 2013 | A1 |
20130331970 | Beckhardt et al. | Dec 2013 | A1 |
20130332165 | Beckley et al. | Dec 2013 | A1 |
20130339028 | Rosner et al. | Dec 2013 | A1 |
20130343567 | Triplett et al. | Dec 2013 | A1 |
20140003611 | Mohammad et al. | Jan 2014 | A1 |
20140003625 | Sheen et al. | Jan 2014 | A1 |
20140003635 | Mohammad et al. | Jan 2014 | A1 |
20140005813 | Reimann | Jan 2014 | A1 |
20140006026 | Lamb et al. | Jan 2014 | A1 |
20140006825 | Shenhav | Jan 2014 | A1 |
20140019743 | DeLuca | Jan 2014 | A1 |
20140034929 | Hamada et al. | Feb 2014 | A1 |
20140046464 | Reimann | Feb 2014 | A1 |
20140064501 | Olsen et al. | Mar 2014 | A1 |
20140073298 | Rossmann | Mar 2014 | A1 |
20140075306 | Rega | Mar 2014 | A1 |
20140075311 | Boettcher et al. | Mar 2014 | A1 |
20140094151 | Klappert et al. | Apr 2014 | A1 |
20140100854 | Chen et al. | Apr 2014 | A1 |
20140109138 | Cannistraro et al. | Apr 2014 | A1 |
20140122075 | Bak et al. | May 2014 | A1 |
20140136195 | Abdossalami et al. | May 2014 | A1 |
20140145168 | Ohsawa et al. | May 2014 | A1 |
20140146983 | Kim et al. | May 2014 | A1 |
20140149118 | Lee et al. | May 2014 | A1 |
20140163978 | Basye et al. | Jun 2014 | A1 |
20140164400 | Kruglick | Jun 2014 | A1 |
20140167931 | Lee et al. | Jun 2014 | A1 |
20140168344 | Shoemake et al. | Jun 2014 | A1 |
20140172899 | Hakkani-Tur et al. | Jun 2014 | A1 |
20140172953 | Blanksteen | Jun 2014 | A1 |
20140181271 | Millington | Jun 2014 | A1 |
20140192986 | Lee et al. | Jul 2014 | A1 |
20140195252 | Gruber et al. | Jul 2014 | A1 |
20140207457 | Biatov et al. | Jul 2014 | A1 |
20140214429 | Pantel | Jul 2014 | A1 |
20140215332 | Lee et al. | Jul 2014 | A1 |
20140219472 | Huang et al. | Aug 2014 | A1 |
20140222436 | Binder et al. | Aug 2014 | A1 |
20140229184 | Shires | Aug 2014 | A1 |
20140244013 | Reilly | Aug 2014 | A1 |
20140244712 | Walters et al. | Aug 2014 | A1 |
20140249817 | Hart et al. | Sep 2014 | A1 |
20140252386 | Ito et al. | Sep 2014 | A1 |
20140254805 | Su et al. | Sep 2014 | A1 |
20140258292 | Thramann et al. | Sep 2014 | A1 |
20140259075 | Chang et al. | Sep 2014 | A1 |
20140269757 | Park et al. | Sep 2014 | A1 |
20140270282 | Tammi et al. | Sep 2014 | A1 |
20140274185 | Luna et al. | Sep 2014 | A1 |
20140274203 | Ganong, III et al. | Sep 2014 | A1 |
20140274218 | Kadiwala et al. | Sep 2014 | A1 |
20140277650 | Zurek et al. | Sep 2014 | A1 |
20140278372 | Nakadai et al. | Sep 2014 | A1 |
20140278445 | Eddington, Jr. | Sep 2014 | A1 |
20140278933 | McMillan | Sep 2014 | A1 |
20140291642 | Watabe et al. | Oct 2014 | A1 |
20140303969 | Inose et al. | Oct 2014 | A1 |
20140310002 | Nitz et al. | Oct 2014 | A1 |
20140310614 | Jones | Oct 2014 | A1 |
20140324203 | Coburn, IV et al. | Oct 2014 | A1 |
20140330896 | Addala et al. | Nov 2014 | A1 |
20140334645 | Yun et al. | Nov 2014 | A1 |
20140340888 | Ishisone et al. | Nov 2014 | A1 |
20140357248 | Tonshal et al. | Dec 2014 | A1 |
20140358535 | Lee et al. | Dec 2014 | A1 |
20140363022 | Dizon et al. | Dec 2014 | A1 |
20140363024 | Apodaca | Dec 2014 | A1 |
20140365227 | Cash et al. | Dec 2014 | A1 |
20140369491 | Kloberdans et al. | Dec 2014 | A1 |
20140372109 | Iyer et al. | Dec 2014 | A1 |
20150006176 | Pogue et al. | Jan 2015 | A1 |
20150006184 | Marti et al. | Jan 2015 | A1 |
20150010169 | Popova et al. | Jan 2015 | A1 |
20150014680 | Yamazaki et al. | Jan 2015 | A1 |
20150016642 | Walsh et al. | Jan 2015 | A1 |
20150018992 | Griffiths et al. | Jan 2015 | A1 |
20150019201 | Schoenbach | Jan 2015 | A1 |
20150019219 | Tzirkel-Hancock et al. | Jan 2015 | A1 |
20150030172 | Gaensler et al. | Jan 2015 | A1 |
20150036831 | Klippel | Feb 2015 | A1 |
20150039303 | Lesso et al. | Feb 2015 | A1 |
20150039317 | Klein et al. | Feb 2015 | A1 |
20150058018 | Georges et al. | Feb 2015 | A1 |
20150063580 | Huang et al. | Mar 2015 | A1 |
20150066479 | Pasupalak et al. | Mar 2015 | A1 |
20150086034 | Lombardi et al. | Mar 2015 | A1 |
20150088500 | Conliffe | Mar 2015 | A1 |
20150091709 | Reichert et al. | Apr 2015 | A1 |
20150092947 | Gossain et al. | Apr 2015 | A1 |
20150104037 | Lee et al. | Apr 2015 | A1 |
20150106085 | Lindahl | Apr 2015 | A1 |
20150110294 | Chen et al. | Apr 2015 | A1 |
20150112672 | Giacobello et al. | Apr 2015 | A1 |
20150124975 | Pontoppidan | May 2015 | A1 |
20150128065 | Torii et al. | May 2015 | A1 |
20150134456 | Baldwin | May 2015 | A1 |
20150154976 | Mutagi | Jun 2015 | A1 |
20150161990 | Sharifi | Jun 2015 | A1 |
20150169279 | Duga | Jun 2015 | A1 |
20150170645 | Di Censo et al. | Jun 2015 | A1 |
20150170665 | Gundeti et al. | Jun 2015 | A1 |
20150172843 | Quan | Jun 2015 | A1 |
20150179181 | Morris et al. | Jun 2015 | A1 |
20150180432 | Gao et al. | Jun 2015 | A1 |
20150181318 | Gautama et al. | Jun 2015 | A1 |
20150189438 | Hampiholi et al. | Jul 2015 | A1 |
20150200454 | Heusdens et al. | Jul 2015 | A1 |
20150200923 | Triplett | Jul 2015 | A1 |
20150201271 | Diethorn et al. | Jul 2015 | A1 |
20150221678 | Yamazaki et al. | Aug 2015 | A1 |
20150222563 | Burns et al. | Aug 2015 | A1 |
20150222987 | Angel, Jr. et al. | Aug 2015 | A1 |
20150228274 | Leppanen et al. | Aug 2015 | A1 |
20150228803 | Koezuka et al. | Aug 2015 | A1 |
20150237406 | Ochoa et al. | Aug 2015 | A1 |
20150243287 | Nakano et al. | Aug 2015 | A1 |
20150245152 | Ding et al. | Aug 2015 | A1 |
20150245154 | Dadu et al. | Aug 2015 | A1 |
20150249889 | Iyer et al. | Sep 2015 | A1 |
20150253292 | Larkin et al. | Sep 2015 | A1 |
20150253960 | Lin et al. | Sep 2015 | A1 |
20150254057 | Klein et al. | Sep 2015 | A1 |
20150263174 | Yamazaki et al. | Sep 2015 | A1 |
20150271593 | Sun et al. | Sep 2015 | A1 |
20150277846 | Yen et al. | Oct 2015 | A1 |
20150280676 | Holman et al. | Oct 2015 | A1 |
20150296299 | Klippel et al. | Oct 2015 | A1 |
20150302856 | Kim et al. | Oct 2015 | A1 |
20150319529 | Klippel | Nov 2015 | A1 |
20150325267 | Lee et al. | Nov 2015 | A1 |
20150331663 | Beckhardt et al. | Nov 2015 | A1 |
20150334471 | Innes et al. | Nov 2015 | A1 |
20150338917 | Steiner et al. | Nov 2015 | A1 |
20150341406 | Rockefeller et al. | Nov 2015 | A1 |
20150346845 | Di Censo et al. | Dec 2015 | A1 |
20150348548 | Piernot et al. | Dec 2015 | A1 |
20150348551 | Gruber et al. | Dec 2015 | A1 |
20150355878 | Corbin | Dec 2015 | A1 |
20150363061 | De Nigris, III et al. | Dec 2015 | A1 |
20150363401 | Chen et al. | Dec 2015 | A1 |
20150370531 | Faaborg | Dec 2015 | A1 |
20150371657 | Gao | Dec 2015 | A1 |
20150371659 | Gao | Dec 2015 | A1 |
20150371664 | Bar-Or et al. | Dec 2015 | A1 |
20150380010 | Srinivasan et al. | Dec 2015 | A1 |
20150382047 | Van Os et al. | Dec 2015 | A1 |
20160007116 | Holman | Jan 2016 | A1 |
20160018873 | Fernald et al. | Jan 2016 | A1 |
20160021458 | Johnson et al. | Jan 2016 | A1 |
20160026428 | Morganstern et al. | Jan 2016 | A1 |
20160029142 | Isaac et al. | Jan 2016 | A1 |
20160035321 | Cho et al. | Feb 2016 | A1 |
20160035337 | Aggarwal et al. | Feb 2016 | A1 |
20160036962 | Rand et al. | Feb 2016 | A1 |
20160042748 | Jain et al. | Feb 2016 | A1 |
20160044151 | Shoemaker et al. | Feb 2016 | A1 |
20160050488 | Matheja et al. | Feb 2016 | A1 |
20160055850 | Nakadai et al. | Feb 2016 | A1 |
20160057522 | Choisel et al. | Feb 2016 | A1 |
20160070526 | Sheen | Mar 2016 | A1 |
20160072804 | Chien et al. | Mar 2016 | A1 |
20160077710 | Lewis et al. | Mar 2016 | A1 |
20160077794 | Kim et al. | Mar 2016 | A1 |
20160086609 | Yue et al. | Mar 2016 | A1 |
20160088036 | Corbin et al. | Mar 2016 | A1 |
20160088392 | Huttunen et al. | Mar 2016 | A1 |
20160093304 | Kim et al. | Mar 2016 | A1 |
20160094718 | Mani et al. | Mar 2016 | A1 |
20160094917 | Wilk et al. | Mar 2016 | A1 |
20160098393 | Hebert | Apr 2016 | A1 |
20160098992 | Renard et al. | Apr 2016 | A1 |
20160103653 | Jang | Apr 2016 | A1 |
20160104480 | Sharifi | Apr 2016 | A1 |
20160111110 | Gautama et al. | Apr 2016 | A1 |
20160125876 | Schroeter et al. | May 2016 | A1 |
20160127780 | Roberts et al. | May 2016 | A1 |
20160133259 | Rubin et al. | May 2016 | A1 |
20160134966 | Fitzgerald et al. | May 2016 | A1 |
20160134982 | Iyer | May 2016 | A1 |
20160140957 | Duta et al. | May 2016 | A1 |
20160148615 | Lee et al. | May 2016 | A1 |
20160154089 | Altman | Jun 2016 | A1 |
20160155442 | Kannan et al. | Jun 2016 | A1 |
20160155443 | Khan et al. | Jun 2016 | A1 |
20160157035 | Russell et al. | Jun 2016 | A1 |
20160162469 | Santos | Jun 2016 | A1 |
20160171976 | Sun et al. | Jun 2016 | A1 |
20160173578 | Sharma et al. | Jun 2016 | A1 |
20160173983 | Berthelsen et al. | Jun 2016 | A1 |
20160180853 | Vanlund et al. | Jun 2016 | A1 |
20160189716 | Lindahl et al. | Jun 2016 | A1 |
20160192099 | Oishi et al. | Jun 2016 | A1 |
20160196499 | Khan et al. | Jul 2016 | A1 |
20160203331 | Khan et al. | Jul 2016 | A1 |
20160210110 | Feldman | Jul 2016 | A1 |
20160212538 | Fullam et al. | Jul 2016 | A1 |
20160216938 | Millington | Jul 2016 | A1 |
20160217789 | Lee et al. | Jul 2016 | A1 |
20160225385 | Hammarqvist | Aug 2016 | A1 |
20160232451 | Scherzer | Aug 2016 | A1 |
20160234204 | Rishi | Aug 2016 | A1 |
20160234615 | Lambourne | Aug 2016 | A1 |
20160239255 | Chavez et al. | Aug 2016 | A1 |
20160240192 | Raghuvir | Aug 2016 | A1 |
20160241976 | Pearson | Aug 2016 | A1 |
20160253050 | Mishra et al. | Sep 2016 | A1 |
20160260431 | Newendorp et al. | Sep 2016 | A1 |
20160283841 | Sainath et al. | Sep 2016 | A1 |
20160302018 | Russell et al. | Oct 2016 | A1 |
20160314782 | Klimanis | Oct 2016 | A1 |
20160316293 | Klimanis | Oct 2016 | A1 |
20160322045 | Hatfield et al. | Nov 2016 | A1 |
20160336519 | Seo et al. | Nov 2016 | A1 |
20160343866 | Koezuka et al. | Nov 2016 | A1 |
20160343949 | Seo et al. | Nov 2016 | A1 |
20160343954 | Seo et al. | Nov 2016 | A1 |
20160345114 | Hanna et al. | Nov 2016 | A1 |
20160352915 | Gautama | Dec 2016 | A1 |
20160353217 | Starobin et al. | Dec 2016 | A1 |
20160353218 | Starobin et al. | Dec 2016 | A1 |
20160357503 | Triplett et al. | Dec 2016 | A1 |
20160364206 | Keyser-Allen et al. | Dec 2016 | A1 |
20160366515 | Mendes et al. | Dec 2016 | A1 |
20160372113 | David et al. | Dec 2016 | A1 |
20160372688 | Seo et al. | Dec 2016 | A1 |
20160373269 | Okubo et al. | Dec 2016 | A1 |
20160373909 | Rasmussen et al. | Dec 2016 | A1 |
20160379634 | Yamamoto et al. | Dec 2016 | A1 |
20170003931 | Dvortsov et al. | Jan 2017 | A1 |
20170012207 | Seo et al. | Jan 2017 | A1 |
20170012232 | Kataishi et al. | Jan 2017 | A1 |
20170019732 | Mendes et al. | Jan 2017 | A1 |
20170025124 | Mixter et al. | Jan 2017 | A1 |
20170025615 | Seo et al. | Jan 2017 | A1 |
20170025630 | Seo et al. | Jan 2017 | A1 |
20170026769 | Patel | Jan 2017 | A1 |
20170032244 | Kurata | Feb 2017 | A1 |
20170034263 | Archambault et al. | Feb 2017 | A1 |
20170039025 | Kielak | Feb 2017 | A1 |
20170040002 | Basson et al. | Feb 2017 | A1 |
20170040018 | Tormey | Feb 2017 | A1 |
20170041724 | Master et al. | Feb 2017 | A1 |
20170053648 | Chi | Feb 2017 | A1 |
20170053650 | Ogawa | Feb 2017 | A1 |
20170060526 | Barton et al. | Mar 2017 | A1 |
20170062734 | Suzuki et al. | Mar 2017 | A1 |
20170070478 | Park et al. | Mar 2017 | A1 |
20170076212 | Shams et al. | Mar 2017 | A1 |
20170076720 | Gopalan et al. | Mar 2017 | A1 |
20170076726 | Bae | Mar 2017 | A1 |
20170078824 | Heo | Mar 2017 | A1 |
20170083285 | Meyers et al. | Mar 2017 | A1 |
20170084277 | Sharifi | Mar 2017 | A1 |
20170084292 | Yoo | Mar 2017 | A1 |
20170084295 | Tsiartas et al. | Mar 2017 | A1 |
20170090864 | Jorgovanovic | Mar 2017 | A1 |
20170092278 | Evermann et al. | Mar 2017 | A1 |
20170092297 | Sainath et al. | Mar 2017 | A1 |
20170092299 | Matsuo | Mar 2017 | A1 |
20170092889 | Seo et al. | Mar 2017 | A1 |
20170092890 | Seo et al. | Mar 2017 | A1 |
20170094215 | Western | Mar 2017 | A1 |
20170103748 | Weissberg et al. | Apr 2017 | A1 |
20170103754 | Higbie et al. | Apr 2017 | A1 |
20170103755 | Jeon et al. | Apr 2017 | A1 |
20170110124 | Boesen et al. | Apr 2017 | A1 |
20170110144 | Sharifi et al. | Apr 2017 | A1 |
20170117497 | Seo et al. | Apr 2017 | A1 |
20170123251 | Nakada et al. | May 2017 | A1 |
20170125037 | Shin | May 2017 | A1 |
20170125456 | Kasahara | May 2017 | A1 |
20170133007 | Drewes | May 2017 | A1 |
20170133011 | Chen et al. | May 2017 | A1 |
20170134872 | Silva et al. | May 2017 | A1 |
20170139720 | Stein | May 2017 | A1 |
20170140449 | Kannan | May 2017 | A1 |
20170140748 | Roberts et al. | May 2017 | A1 |
20170140759 | Kumar et al. | May 2017 | A1 |
20170151930 | Boesen | Jun 2017 | A1 |
20170177585 | Rodger et al. | Jun 2017 | A1 |
20170178662 | Ayrapetian et al. | Jun 2017 | A1 |
20170180561 | Kadiwala et al. | Jun 2017 | A1 |
20170186427 | Wang et al. | Jun 2017 | A1 |
20170188150 | Brunet et al. | Jun 2017 | A1 |
20170188437 | Banta | Jun 2017 | A1 |
20170193999 | Aleksic et al. | Jul 2017 | A1 |
20170206896 | Ko et al. | Jul 2017 | A1 |
20170206900 | Lee et al. | Jul 2017 | A1 |
20170214996 | Yeo | Jul 2017 | A1 |
20170236512 | Williams et al. | Aug 2017 | A1 |
20170236515 | Pinsky et al. | Aug 2017 | A1 |
20170242649 | Jarvis et al. | Aug 2017 | A1 |
20170242651 | Lang et al. | Aug 2017 | A1 |
20170242653 | Lang et al. | Aug 2017 | A1 |
20170242656 | Plagge et al. | Aug 2017 | A1 |
20170242657 | Jarvis et al. | Aug 2017 | A1 |
20170243576 | Millington et al. | Aug 2017 | A1 |
20170243587 | Plagge et al. | Aug 2017 | A1 |
20170245076 | Kusano et al. | Aug 2017 | A1 |
20170255612 | Sarikaya et al. | Sep 2017 | A1 |
20170257686 | Gautama et al. | Sep 2017 | A1 |
20170269975 | Wood et al. | Sep 2017 | A1 |
20170270919 | Parthasarathi et al. | Sep 2017 | A1 |
20170278512 | Pandya et al. | Sep 2017 | A1 |
20170287485 | Civelli et al. | Oct 2017 | A1 |
20170300990 | Tanaka et al. | Oct 2017 | A1 |
20170330565 | Daley et al. | Nov 2017 | A1 |
20170331869 | Bendahan et al. | Nov 2017 | A1 |
20170332168 | Moghimi et al. | Nov 2017 | A1 |
20170346872 | Naik et al. | Nov 2017 | A1 |
20170352357 | Fink | Dec 2017 | A1 |
20170353789 | Kim et al. | Dec 2017 | A1 |
20170357475 | Lee et al. | Dec 2017 | A1 |
20170357478 | Piersol et al. | Dec 2017 | A1 |
20170364371 | Nandi et al. | Dec 2017 | A1 |
20170366393 | Shaker et al. | Dec 2017 | A1 |
20170374454 | Bernardini et al. | Dec 2017 | A1 |
20170374552 | Xia et al. | Dec 2017 | A1 |
20180018964 | Reilly et al. | Jan 2018 | A1 |
20180018965 | Daley | Jan 2018 | A1 |
20180018967 | Lang et al. | Jan 2018 | A1 |
20180020306 | Sheen | Jan 2018 | A1 |
20180025733 | Qian et al. | Jan 2018 | A1 |
20180033428 | Kim et al. | Feb 2018 | A1 |
20180033438 | Toma et al. | Feb 2018 | A1 |
20180040324 | Wilberding | Feb 2018 | A1 |
20180047394 | Tian et al. | Feb 2018 | A1 |
20180053504 | Wang et al. | Feb 2018 | A1 |
20180054506 | Hart et al. | Feb 2018 | A1 |
20180061396 | Srinivasan et al. | Mar 2018 | A1 |
20180061402 | Devaraj et al. | Mar 2018 | A1 |
20180061404 | Devaraj et al. | Mar 2018 | A1 |
20180061419 | Melendo Casado et al. | Mar 2018 | A1 |
20180061420 | Patil et al. | Mar 2018 | A1 |
20180062871 | Jones et al. | Mar 2018 | A1 |
20180084367 | Greff et al. | Mar 2018 | A1 |
20180088900 | Glaser et al. | Mar 2018 | A1 |
20180091898 | Yoon et al. | Mar 2018 | A1 |
20180091913 | Hartung et al. | Mar 2018 | A1 |
20180096683 | James et al. | Apr 2018 | A1 |
20180096696 | Mixter | Apr 2018 | A1 |
20180107446 | Wilberding et al. | Apr 2018 | A1 |
20180108351 | Beckhardt et al. | Apr 2018 | A1 |
20180122372 | Wanderlust | May 2018 | A1 |
20180122378 | Mixter et al. | May 2018 | A1 |
20180130469 | Gruenstein et al. | May 2018 | A1 |
20180132217 | Stirling-Gallacher | May 2018 | A1 |
20180132298 | Birnam et al. | May 2018 | A1 |
20180137861 | Ogawa | May 2018 | A1 |
20180152557 | White et al. | May 2018 | A1 |
20180158454 | Campbell et al. | Jun 2018 | A1 |
20180165055 | Yu et al. | Jun 2018 | A1 |
20180167981 | Jonna et al. | Jun 2018 | A1 |
20180174597 | Lee et al. | Jun 2018 | A1 |
20180182383 | Kim et al. | Jun 2018 | A1 |
20180182390 | Hughes et al. | Jun 2018 | A1 |
20180182397 | Carbune et al. | Jun 2018 | A1 |
20180188948 | Ouyang et al. | Jul 2018 | A1 |
20180190274 | Kirazci et al. | Jul 2018 | A1 |
20180190285 | Heckman et al. | Jul 2018 | A1 |
20180197533 | Lyon et al. | Jul 2018 | A1 |
20180199146 | Sheen | Jul 2018 | A1 |
20180204569 | Nadkar et al. | Jul 2018 | A1 |
20180205963 | Matei et al. | Jul 2018 | A1 |
20180210698 | Park et al. | Jul 2018 | A1 |
20180211665 | Park et al. | Jul 2018 | A1 |
20180218747 | Moghimi et al. | Aug 2018 | A1 |
20180219976 | Decenzo et al. | Aug 2018 | A1 |
20180225933 | Park et al. | Aug 2018 | A1 |
20180228006 | Baker et al. | Aug 2018 | A1 |
20180233130 | Kaskari et al. | Aug 2018 | A1 |
20180233136 | Torok et al. | Aug 2018 | A1 |
20180233137 | Torok et al. | Aug 2018 | A1 |
20180233139 | Finkelstein et al. | Aug 2018 | A1 |
20180233141 | Solomon et al. | Aug 2018 | A1 |
20180233142 | Koishida et al. | Aug 2018 | A1 |
20180233150 | Gruenstein et al. | Aug 2018 | A1 |
20180234765 | Torok et al. | Aug 2018 | A1 |
20180260680 | Finkelstein et al. | Sep 2018 | A1 |
20180261213 | Arik et al. | Sep 2018 | A1 |
20180262793 | Lau et al. | Sep 2018 | A1 |
20180262831 | Matheja et al. | Sep 2018 | A1 |
20180270565 | Ganeshkumar | Sep 2018 | A1 |
20180277107 | Kim | Sep 2018 | A1 |
20180277113 | Hartung et al. | Sep 2018 | A1 |
20180277119 | Baba et al. | Sep 2018 | A1 |
20180277133 | Deetz et al. | Sep 2018 | A1 |
20180286394 | Li et al. | Oct 2018 | A1 |
20180286414 | Ravindran et al. | Oct 2018 | A1 |
20180293221 | Finkelstein et al. | Oct 2018 | A1 |
20180293484 | Wang et al. | Oct 2018 | A1 |
20180308470 | Park et al. | Oct 2018 | A1 |
20180314552 | Kim et al. | Nov 2018 | A1 |
20180322891 | Van Den Oord et al. | Nov 2018 | A1 |
20180324756 | Ryu et al. | Nov 2018 | A1 |
20180330727 | Tulli | Nov 2018 | A1 |
20180335903 | Coffman et al. | Nov 2018 | A1 |
20180336274 | Choudhury et al. | Nov 2018 | A1 |
20180349093 | McCarty et al. | Dec 2018 | A1 |
20180356962 | Corbin | Dec 2018 | A1 |
20180358009 | Daley et al. | Dec 2018 | A1 |
20180365567 | Kolavennu et al. | Dec 2018 | A1 |
20180367944 | Heo et al. | Dec 2018 | A1 |
20190012141 | Piersol et al. | Jan 2019 | A1 |
20190013019 | Lawrence | Jan 2019 | A1 |
20190014592 | Hampel et al. | Jan 2019 | A1 |
20190019112 | Gelfenbeyn et al. | Jan 2019 | A1 |
20190033446 | Bultan et al. | Jan 2019 | A1 |
20190042187 | Truong et al. | Feb 2019 | A1 |
20190043488 | Bocklet et al. | Feb 2019 | A1 |
20190043492 | Lang | Feb 2019 | A1 |
20190051298 | Lee et al. | Feb 2019 | A1 |
20190066672 | Wood et al. | Feb 2019 | A1 |
20190066687 | Wood et al. | Feb 2019 | A1 |
20190074025 | Lashkari et al. | Mar 2019 | A1 |
20190079724 | Feuz et al. | Mar 2019 | A1 |
20190081507 | Ide | Mar 2019 | A1 |
20190081810 | Jung | Mar 2019 | A1 |
20190082255 | Tajiri et al. | Mar 2019 | A1 |
20190087455 | He et al. | Mar 2019 | A1 |
20190088261 | Lang et al. | Mar 2019 | A1 |
20190090056 | Rexach et al. | Mar 2019 | A1 |
20190096408 | Li et al. | Mar 2019 | A1 |
20190098400 | Buoni et al. | Mar 2019 | A1 |
20190104119 | Giorgi et al. | Apr 2019 | A1 |
20190104373 | Wodrich et al. | Apr 2019 | A1 |
20190108839 | Reilly et al. | Apr 2019 | A1 |
20190115011 | Khellah et al. | Apr 2019 | A1 |
20190130906 | Kobayashi et al. | May 2019 | A1 |
20190163153 | Price et al. | May 2019 | A1 |
20190172452 | Smith et al. | Jun 2019 | A1 |
20190173687 | MacKay et al. | Jun 2019 | A1 |
20190179607 | Thangarathnam et al. | Jun 2019 | A1 |
20190179611 | Wojogbe et al. | Jun 2019 | A1 |
20190182072 | Roe et al. | Jun 2019 | A1 |
20190188328 | Oyenan et al. | Jun 2019 | A1 |
20190189117 | Kumar | Jun 2019 | A1 |
20190206391 | Busch et al. | Jul 2019 | A1 |
20190206405 | Gillespie et al. | Jul 2019 | A1 |
20190206412 | Li et al. | Jul 2019 | A1 |
20190219976 | Giorgi et al. | Jul 2019 | A1 |
20190220246 | Orr et al. | Jul 2019 | A1 |
20190221206 | Chen et al. | Jul 2019 | A1 |
20190237067 | Friedman et al. | Aug 2019 | A1 |
20190239008 | Lambourne | Aug 2019 | A1 |
20190239009 | Lambourne | Aug 2019 | A1 |
20190243603 | Keyser-Allen et al. | Aug 2019 | A1 |
20190243606 | Jayakumar et al. | Aug 2019 | A1 |
20190244608 | Choi et al. | Aug 2019 | A1 |
20190251960 | Maker et al. | Aug 2019 | A1 |
20190281397 | Lambourne | Sep 2019 | A1 |
20190287546 | Ganeshkumar | Sep 2019 | A1 |
20190288970 | Siddiq | Sep 2019 | A1 |
20190289367 | Siddiq | Sep 2019 | A1 |
20190295542 | Huang et al. | Sep 2019 | A1 |
20190295555 | Wilberding | Sep 2019 | A1 |
20190295556 | Wilberding | Sep 2019 | A1 |
20190295563 | Kamdar et al. | Sep 2019 | A1 |
20190297388 | Panchaksharaiah et al. | Sep 2019 | A1 |
20190304443 | Bhagwan | Oct 2019 | A1 |
20190311710 | Eraslan et al. | Oct 2019 | A1 |
20190311712 | Firik et al. | Oct 2019 | A1 |
20190311720 | Pasko | Oct 2019 | A1 |
20190317606 | Jain et al. | Oct 2019 | A1 |
20190342962 | Chang et al. | Nov 2019 | A1 |
20190347063 | Liu et al. | Nov 2019 | A1 |
20190348044 | Chun et al. | Nov 2019 | A1 |
20190362714 | Mori et al. | Nov 2019 | A1 |
20190364375 | Soto et al. | Nov 2019 | A1 |
20190364422 | Zhuo | Nov 2019 | A1 |
20190371310 | Fox et al. | Dec 2019 | A1 |
20200007987 | Woo et al. | Jan 2020 | A1 |
20200034492 | Verbeke et al. | Jan 2020 | A1 |
20200051554 | Kim et al. | Feb 2020 | A1 |
20200074990 | Kim et al. | Mar 2020 | A1 |
20200090647 | Kurtz | Mar 2020 | A1 |
20200092687 | Devaraj et al. | Mar 2020 | A1 |
20200098354 | Lin et al. | Mar 2020 | A1 |
20200098379 | Tai et al. | Mar 2020 | A1 |
20200105245 | Gupta et al. | Apr 2020 | A1 |
20200105256 | Fainberg et al. | Apr 2020 | A1 |
20200105264 | Jang et al. | Apr 2020 | A1 |
20200125162 | D'Amato et al. | Apr 2020 | A1 |
20200135224 | Bromand et al. | Apr 2020 | A1 |
20200152206 | Shen et al. | May 2020 | A1 |
20200175989 | Lockhart et al. | Jun 2020 | A1 |
20200184964 | Myers et al. | Jun 2020 | A1 |
20200184980 | Wilberding | Jun 2020 | A1 |
20200193973 | Tolomei et al. | Jun 2020 | A1 |
20200211539 | Lee | Jul 2020 | A1 |
20200211550 | Pan et al. | Jul 2020 | A1 |
20200211556 | Mixter et al. | Jul 2020 | A1 |
20200213729 | Soto | Jul 2020 | A1 |
20200216089 | Garcia et al. | Jul 2020 | A1 |
20200234709 | Kunitake | Jul 2020 | A1 |
20200251107 | Wang et al. | Aug 2020 | A1 |
20200265838 | Lee et al. | Aug 2020 | A1 |
20200310751 | Anand et al. | Oct 2020 | A1 |
20200336846 | Rohde et al. | Oct 2020 | A1 |
20200366477 | Brown et al. | Nov 2020 | A1 |
20200395006 | Smith et al. | Dec 2020 | A1 |
20200395010 | Smith et al. | Dec 2020 | A1 |
20200395013 | Smith et al. | Dec 2020 | A1 |
20200409652 | Wilberding et al. | Dec 2020 | A1 |
20210035561 | D'Amato et al. | Feb 2021 | A1 |
20210035572 | D'Amato et al. | Feb 2021 | A1 |
20210067867 | Kagoshima | Mar 2021 | A1 |
20210118429 | Shan | Apr 2021 | A1 |
20210166680 | Jung et al. | Jun 2021 | A1 |
20210183366 | Reinspach et al. | Jun 2021 | A1 |
20210280185 | Tan et al. | Sep 2021 | A1 |
Number | Date | Country |
---|---|---|
2017100486 | Jun 2017 | AU |
2017100581 | Jun 2017 | AU |
101310558 | Nov 2008 | CN |
101480039 | Jul 2009 | CN |
101661753 | Mar 2010 | CN |
101686282 | Mar 2010 | CN |
101907983 | Dec 2010 | CN |
102123188 | Jul 2011 | CN |
102256098 | Nov 2011 | CN |
102567468 | Jul 2012 | CN |
103052001 | Apr 2013 | CN |
103181192 | Jun 2013 | CN |
103210663 | Jul 2013 | CN |
103546616 | Jan 2014 | CN |
103811007 | May 2014 | CN |
104010251 | Aug 2014 | CN |
104035743 | Sep 2014 | CN |
104053088 | Sep 2014 | CN |
104092936 | Oct 2014 | CN |
104104769 | Oct 2014 | CN |
104115224 | Oct 2014 | CN |
104282305 | Jan 2015 | CN |
104520927 | Apr 2015 | CN |
104538030 | Apr 2015 | CN |
104575504 | Apr 2015 | CN |
104635539 | May 2015 | CN |
104865550 | Aug 2015 | CN |
105187907 | Dec 2015 | CN |
105204357 | Dec 2015 | CN |
105206281 | Dec 2015 | CN |
105284076 | Jan 2016 | CN |
105493442 | Apr 2016 | CN |
105679318 | Jun 2016 | CN |
106028223 | Oct 2016 | CN |
106375902 | Feb 2017 | CN |
106531165 | Mar 2017 | CN |
106708403 | May 2017 | CN |
107004410 | Aug 2017 | CN |
107919123 | Apr 2018 | CN |
108028047 | May 2018 | CN |
108028048 | May 2018 | CN |
109712626 | May 2019 | CN |
1349146 | Oct 2003 | EP |
1389853 | Feb 2004 | EP |
2051542 | Apr 2009 | EP |
2166737 | Mar 2010 | EP |
2683147 | Jan 2014 | EP |
2986034 | Feb 2016 | EP |
3128767 | Feb 2017 | EP |
3133595 | Feb 2017 | EP |
2351021 | Sep 2017 | EP |
3270377 | Jan 2018 | EP |
3285502 | Feb 2018 | EP |
H0883091 | Mar 1996 | JP |
2003223188 | Aug 2003 | JP |
2005242134 | Sep 2005 | JP |
2005250867 | Sep 2005 | JP |
2006092482 | Apr 2006 | JP |
2007013400 | Jan 2007 | JP |
2007142595 | Jun 2007 | JP |
2008079256 | Apr 2008 | JP |
2008158868 | Jul 2008 | JP |
2015161551 | Sep 2015 | JP |
2015527768 | Sep 2015 | JP |
2016095383 | May 2016 | JP |
2017072857 | Apr 2017 | JP |
2017129860 | Jul 2017 | JP |
2017227912 | Dec 2017 | JP |
2018055259 | Apr 2018 | JP |
20100036351 | Apr 2010 | KR |
100966415 | Jun 2010 | KR |
20100111071 | Oct 2010 | KR |
20130050987 | May 2013 | KR |
20140005410 | Jan 2014 | KR |
20140035310 | Mar 2014 | KR |
20140054643 | May 2014 | KR |
20140111859 | Sep 2014 | KR |
20140112900 | Sep 2014 | KR |
200153994 | Jul 2001 | WO |
03054854 | Jul 2003 | WO |
2003093950 | Nov 2003 | WO |
2008048599 | Apr 2008 | WO |
2012166386 | Dec 2012 | WO |
2013184792 | Dec 2013 | WO |
2014064531 | May 2014 | WO |
2014159581 | Oct 2014 | WO |
2015017303 | Feb 2015 | WO |
2015105788 | Jul 2015 | WO |
2015131024 | Sep 2015 | WO |
2015178950 | Nov 2015 | WO |
2016014142 | Jan 2016 | WO |
2016014686 | Jan 2016 | WO |
2016022926 | Feb 2016 | WO |
2016033364 | Mar 2016 | WO |
2016057268 | Apr 2016 | WO |
2016085775 | Jun 2016 | WO |
2016136062 | Sep 2016 | WO |
2016165067 | Oct 2016 | WO |
2016171956 | Oct 2016 | WO |
2016200593 | Dec 2016 | WO |
2017039632 | Mar 2017 | WO |
2017058654 | Apr 2017 | WO |
2017138934 | Aug 2017 | WO |
2017147075 | Aug 2017 | WO |
2017147936 | Sep 2017 | WO |
2018027142 | Feb 2018 | WO |
2018067404 | Apr 2018 | WO |
20190052 | Jan 2019 | WO |
Entry |
---|
US 9,299,346 B1, 03/2016, Hart et al. (withdrawn) |
Japanese Patent Office, Office Action and Translation dated Apr. 27, 2021, issued in connection with Japanese Patent Application No. 2020-518400, 10 pages. |
Japanese Patent Office, Office Action and Translation dated Jul. 6, 2021, issued in connection with Japanese Patent Application No. 2019-073349, 6 pages. |
Joseph Szurley et al, “Efficient computation of microphone utility in a wireless acoustic sensor network with multi-channel Wiener filter based noise reduction”, 2012 IEEE International Conference on Acoustics, Speech and Signal Processing, Kyoto, Japan, Mar. 25-30, 2012, pp. 2657-2660, XP032227701, DOI: 10.1109/ICASSP .2012.6288463 ISBN: 978-1-4673-0045-2. |
Korean Patent Office, Korean Examination Report and Translation dated Apr. 26, 2021, issued in connection with Korean Application No. 10-2021-7008937, 15 pages. |
Korean Patent Office, Korean Office Action and Translation dated Jan. 4, 2021, issued in connection with Korean Application No. 10-2020-7034425, 14 pages. |
Non-Final Office Action dated Jul. 12, 2021, issued in connection with U.S. Appl. No. 17/008,104, filed Aug. 31, 2020, 6 pages. |
Non-Final Office Action dated Jun. 18, 2021, issued in connection with U.S. Appl. No. 17/236,559, filed Apr. 21, 2021, 9 pages. |
Non-Final Office Action dated Apr. 21, 2021, issued in connection with U.S. Appl. No. 16/109,375, filed Aug. 22, 2018, 9 pages. |
Non-Final Office Action dated Apr. 23, 2021, issued in connection with U.S. Appl. No. 16/660,197, filed Oct. 22, 2019, 9 pages. |
Non-Final Office Action dated Jun. 25, 2021, issued in connection with U.S. Appl. No. 16/213,570, filed Dec. 7, 2018, 11 pages. |
Non-Final Office Action dated Jul. 8, 2021, issued in connection with U.S. Appl. No. 16/813,643, filed Mar. 9, 2020, 12 pages. |
Non-Final Office Action dated Jul. 9, 2021, issued in connection with U.S. Appl. No. 16/806,747, filed Mar. 2, 2020, 18 pages. |
Non-Final Office Action dated Feb. 11, 2021, issued in connection with U.S. Appl. No. 16/876,493, filed May 18, 2020, 16 pages. |
Non-Final Office Action dated Mar. 11, 2021, issued in connection with U.S. Appl. No. 16/834,483, filed Mar. 30, 2020, 11 pages. |
Non-Final Office Action dated Apr. 12, 2021, issued in connection with U.S. Appl. No. 16/528,224, filed Jul. 31, 2019, 9 pages. |
Non-Final Office Action dated Jun. 23, 2021, issued in connection with U.S. Appl. No. 16/439,032, filed Jun. 12, 2019, 13 pages. |
Non-Final Office Action dated Mar. 29, 2021, issued in connection with U.S. Appl. No. 16/528,265, filed Jul. 31, 2019, 18 pages. |
Non-Final Office Action dated Jan. 6, 2021, issued in connection with U.S. Appl. No. 16/439,046, filed Jun. 12, 2019, 13 pages. |
Non-Final Office Action dated Apr. 9, 2021, issued in connection with U.S. Appl. No. 16/780,483, filed Feb. 3, 2020, 45 pages. |
Non-Final Office Action dated Feb. 9, 2021, issued in connection with U.S. Appl. No. 16/806,747, filed Mar. 2, 2020, 16 pages. |
Notice of Allowance dated Mar. 31, 2021, issued in connection with U.S. Appl. No. 16/813,643, filed Mar. 9, 2020, 11 pages. |
Notice of Allowance dated Jun. 1, 2021, issued in connection with U.S. Appl. No. 16/219,702, filed Dec. 13, 2018, 8 pages. |
Notice of Allowance dated Jun. 1, 2021, issued in connection with U.S. Appl. No. 16/685,135, filed Nov. 15, 2019, 10 pages. |
Notice of Allowance dated Feb. 10, 2021, issued in connection with U.S. Appl. No. 16/138,111, filed Sep. 21, 2018, 8 pages. |
Notice of Allowance dated May 12, 2021, issued in connection with U.S. Appl. No. 16/402,617, filed May 3, 2019, 8 pages. |
Notice of Allowance dated Jan. 13, 2021, issued in connection with U.S. Appl. No. 16/539,843, filed Aug. 13, 2019, 5 pages. |
Notice of Allowance dated Jan. 14, 2021, issued in connection with U.S. Appl. No. 17/087,423, filed Nov. 2, 2020, 8 pages. |
Notice of Allowance dated Apr. 16, 2021, issued in connection with U.S. Appl. No. 16/798,967, filed Feb. 24, 2020, 16 pages. |
Notice of Allowance dated Feb. 17, 2021, issued in connection with U.S. Appl. No. 16/715,984, filed Dec. 16, 2019, 8 pages. |
Notice of Allowance dated Mar. 18, 2021, issued in connection with U.S. Appl. No. 16/177,185, filed Oct. 31, 2018, 8 pages. |
Notice of Allowance dated Mar. 19, 2021, issued in connection with U.S. Appl. No. 17/157,686, filed Jan. 25, 2021, 11 pages. |
Notice of Allowance dated Apr. 21, 2021, issued in connection with U.S. Appl. No. 16/145,275, filed Sep. 28, 2018, 8 pages. |
Notice of Allowance dated Jan. 21, 2021, issued in connection with U.S. Appl. No. 16/600,644, filed Oct. 14, 2019, 7 pages. |
Notice of Allowance dated Jun. 23, 2021, issued in connection with U.S. Appl. No. 16/814,844, filed Mar. 10, 2020, 8 pages. |
Notice of Allowance dated May 26, 2021, issued in connection with U.S. Appl. No. 16/927,670, filed Jul. 13, 2020, 10 pages. |
Notice of Allowance dated May 28, 2021, issued in connection with U.S. Appl. No. 16/524,306, filed Jul. 29, 2019, 9 pages. |
Notice of Allowance dated Jan. 29, 2021, issued in connection with U.S. Appl. No. 16/290,599, filed Mar. 1, 2019, 9 pages. |
Notice of Allowance dated Mar. 29, 2021, issued in connection with U.S. Appl. No. 16/600,949, filed Oct. 14, 2019, 9 pages. |
Notice of Allowance dated Jun. 3, 2021, issued in connection with U.S. Appl. No. 16/876,493, filed May 18, 2020, 7 pages. |
Notice of Allowance dated Jun. 4, 2021, issued in connection with U.S. Appl. No. 16/528,265, filed Jul. 31, 2019, 17 pages. |
Notice of Allowance dated Jun. 7, 2021, issued in connection with U.S. Appl. No. 16/528,224, filed Jul. 31, 2019, 9 pages. |
Pre-Appeal Brief Decision dated Jun. 2, 2021, issued in connection with U.S. Appl. No. 16/213,570, filed Dec. 7, 2018, 2 pages. |
Preinterview First Office Action dated Jan. 8, 2021, issued in connection with U.S. Appl. No. 16/798,967, filed Feb. 24, 2020, 4 pages. |
Tsung-Hsien Wen et al: “A Network-based End-to-End Trainable Task-oriented Dialogue System”, CORR ARXIV, vol. 1604.04562v1, Apr. 15, 2016, pp. 1-11, XP055396370, Stroudsburg, PA, USA. |
Notice of Allowance dated Nov. 14, 2018, issued in connection with U.S. Appl. No. 15/297,627, filed Oct. 19, 2016, 5 pages. |
Notice of Allowance dated Dec. 15, 2017, issued in connection with U.S. Appl. No. 15/223,218, filed Jul. 29, 2016, 7 pages. |
Notice of Allowance dated Jan. 15, 2020, issued in connection with U.S. Appl. No. 16/439,009, filed Jun. 12, 2019, 9 pages. |
Notice of Allowance dated Mar. 15, 2019, issued in connection with U.S. Appl. No. 15/804,776, filed Nov. 6, 2017, 9 pages. |
Notice of Allowance dated Oct. 15, 2019, issued in connection with U.S. Appl. No. 16/437,437, filed Jun. 11, 2019, 9 pages. |
Notice of Allowance dated Oct. 15, 2020, issued in connection with U.S. Appl. No. 16/715,713, filed Dec. 16, 2019, 9 pages. |
Notice of Allowance dated Aug. 16, 2017, issued in connection with U.S. Appl. No. 15/098,892, filed Apr. 14, 2016, 9 pages. |
Notice of Allowance dated Aug. 17, 2017, issued in connection with U.S. Appl. No. 15/131,244, filed Apr. 18, 2016, 9 pages. |
Notice of Allowance dated Jul. 17, 2019, issued in connection with U.S. Appl. No. 15/718,911, filed Sep. 28, 2017, 5 pages. |
Notice of Allowance dated Jun. 17, 2020, issued in connection with U.S. Appl. No. 16/141,875, filed Sep. 25, 2018, 6 pages. |
Notice of Allowance dated Sep. 17, 2018, issued in connection with U.S. Appl. No. 15/211,689, filed Jul. 15, 2016, 6 pages. |
Notice of Allowance dated Apr. 18, 2019, issued in connection with U.S. Appl. No. 16/173,797, filed Oct. 29, 2018, 9 pages. |
Notice of Allowance dated Dec. 18, 2019, issued in connection with U.S. Appl. No. 16/434,426, filed Jun. 7, 2019, 13 pages. |
Notice of Allowance dated Feb. 18, 2020, issued in connection with U.S. Appl. No. 16/022,662, filed Jun. 28, 2018, 8 pages. |
Notice of Allowance dated Jul. 18, 2019, issued in connection with U.S. Appl. No. 15/438,749, filed Feb. 21, 2017, 9 pages. |
Notice of Allowance dated Jul. 18, 2019, issued in connection with U.S. Appl. No. 15/721,141, filed Sep. 29, 2017, 8 pages. |
Notice of Allowance dated Aug. 19, 2020, issued in connection with U.S. Appl. No. 16/271,560, filed Feb. 8, 2019, 9 pages. |
Notice of Allowance dated Dec. 19, 2018, issued in connection with U.S. Appl. No. 15/818,051, filed Nov. 20, 2017, 9 pages. |
Notice of Allowance dated Jul. 19, 2018, issued in connection with U.S. Appl. No. 15/681,937, filed Aug. 21, 2017, 7 pages. |
Notice of Allowance dated Aug. 2, 2019, issued in connection with U.S. Appl. No. 16/102,650, filed Aug. 13, 2018, 5 pages. |
Notice of Allowance dated Dec. 2, 2020, issued in connection with U.S. Appl. No. 15/989,715, filed May 25, 2018, 11 pages. |
Notice of Allowance dated Sep. 2, 2020, issued in connection with U.S. Appl. No. 16/214,711, filed Dec. 10, 2018, 9 pages. |
Notice of Allowance dated Jul. 20, 2020, issued in connection with U.S. Appl. No. 15/984,073, filed May 18, 2018, 12 pages. |
Notice of Allowance dated Mar. 20, 2018, issued in connection with U.S. Appl. No. 15/784,952, filed Oct. 16, 2017, 7 pages. |
Notice of Allowance dated Sep. 20, 2018, issued in connection with U.S. Appl. No. 15/946,599, filed Apr. 5, 2018, 7 pages. |
Notice of Allowance dated Feb. 21, 2020, issued in connection with U.S. Appl. No. 16/416,752, filed May 20, 2019, 6 pages. |
Notice of Allowance dated Jan. 21, 2020, issued in connection with U.S. Appl. No. 16/672,764, filed Nov. 4, 2019, 10 pages. |
Notice of Allowance dated Oct. 21, 2019, issued in connection with U.S. Appl. No. 15/946,585, filed Apr. 5, 2018, 5 pages. |
Notice of Allowance dated Aug. 22, 2017, issued in connection with U.S. Appl. No. 15/273,679, filed Sep. 22, 2016, 5 pages. |
Notice of Allowance dated Jan. 22, 2018, issued in connection with U.S. Appl. No. 15/178,180, filed Jun. 9, 2016, 9 pages. |
Notice of Allowance dated Jul. 22, 2020, issued in connection with U.S. Appl. No. 16/131,409, filed Sep. 14, 2018, 13 pages. |
Notice of Allowance dated Jul. 22, 2020, issued in connection with U.S. Appl. No. 16/790,621, filed Feb. 13, 2020, 10 pages. |
Notice of Allowance dated Apr. 24, 2019, issued in connection with U.S. Appl. No. 16/154,469, filed Oct. 8, 2018, 5 pages. |
Notice of Allowance dated Aug. 26, 2020, issued in connection with U.S. Appl. No. 15/948,541, filed Apr. 9, 2018, 9 pages. |
Notice of Allowance dated Apr. 27, 2020, issued in connection with U.S. Appl. No. 16/700,607, filed Dec. 2, 2019, 10 pages. |
Notice of Allowance dated Mar. 27, 2019, issued in connection with U.S. Appl. No. 16/214,666, filed Dec. 10, 2018, 6 pages. |
Notice of Allowance dated Mar. 28, 2018, issued in connection with U.S. Appl. No. 15/699,982, filed Sep. 8, 2017, 17 pages. |
Notice of Allowance dated Dec. 29, 2017, issued in connection with U.S. Appl. No. 15/131,776, filed Apr. 18, 2016, 13 pages. |
Notice of Allowance dated Jun. 29, 2020, issued in connection with U.S. Appl. No. 16/216,357, filed Dec. 11, 2018, 8 pages. |
Notice of Allowance dated May 29, 2020, issued in connection with U.S. Appl. No. 16/148,879, filed Oct. 1, 2018, 6 pages. |
Notice of Allowance dated Apr. 3, 2019, issued in connection with U.S. Appl. No. 16/160,107, filed Oct. 15, 2018, 7 pages. |
Notice of Allowance dated Jul. 30, 2018, issued in connection with U.S. Appl. No. 15/098,718, filed Apr. 14, 2016, 5 pages. |
Notice of Allowance dated Jul. 30, 2019, issued in connection with U.S. Appl. No. 15/131,254, filed Apr. 18, 2016, 9 pages. |
Notice of Allowance dated Mar. 30, 2020, issued in connection with U.S. Appl. No. 15/973,413, filed May 7, 2018, 5 pages. |
Notice of Allowance dated Nov. 30, 2018, issued in connection with U.S. Appl. No. 15/438,725, filed Feb. 21, 2017, 5 pages. |
Notice of Allowance dated Oct. 30, 2019, issued in connection with U.S. Appl. No. 16/131,392, filed Sep. 14, 2018, 9 pages. |
Notice of Allowance dated Oct. 30, 2020, issued in connection with U.S. Appl. No. 16/528,016, filed Jul. 31, 2019, 10 pages. |
Notice of Allowance dated May 31, 2019, issued in connection with U.S. Appl. No. 15/717,621, filed Sep. 27, 2017, 9 pages. |
Notice of Allowance dated Mar. 4, 2020, issued in connection with U.S. Appl. No. 16/444,975, filed Jun. 18, 2019, 10 pages. |
Notice of Allowance dated Feb. 5, 2020, issued in connection with U.S. Appl. No. 16/178,122, filed Nov. 1, 2018, 9 pages. |
Notice of Allowance dated Oct. 5, 2018, issued in connection with U.S. Appl. No. 15/211,748, filed Jul. 15, 2018, 10 pages. |
Notice of Allowance dated Feb. 6, 2019, issued in connection with U.S. Appl. No. 16/102,153, filed Aug. 13, 2018, 9 pages. |
Notice of Allowance dated Feb. 6, 2020, issued in connection with U.S. Appl. No. 16/227,308, filed Dec. 20, 2018, 7 pages. |
Notice of Allowance dated Apr. 7, 2020, issued in connection with U.S. Appl. No. 15/098,760, filed Apr. 14, 2016, 7 pages. |
Notice of Allowance dated Apr. 7, 2020, issued in connection with U.S. Appl. No. 16/147,710, filed Sep. 29, 2018, 15 pages. |
Notice of Allowance dated Jun. 7, 2019, issued in connection with U.S. Appl. No. 16/102,153, filed Aug. 13, 2018, 9 pages. |
Notice of Allowance dated Aug. 9, 2018, issued in connection with U.S. Appl. No. 15/229,868, filed Aug. 5, 2016, 11 pages. |
Notice of Allowance dated Mar. 9, 2018, issued in connection with U.S. Appl. No. 15/584,782, filed May 2, 2017, 8 pages. |
Optimizing Siri on HomePod in Far-Field Settings. Audio Software Engineering and Siri Speech Team, Machine Learning Journal vol. 1, Issue 12. https://machinelearning.apple.com/2018/12/03/optimizing-siri-on-homepod-in-far-field-settings.html. Dec. 2018, 18 pages. |
Palm, Inc., “Handbook for the Palm VII Handheld,” May 2000, 311 pages. |
Preinterview First Office Action dated Aug. 5, 2019, issued in connection with U.S. Appl. No. 16/434,426, filed Jun. 7, 2019, 4 pages. |
Preinterview First Office Action dated Mar. 25, 2020, issued in connection with U.S. Appl. No. 16/109,375, filed Aug. 22, 2018, 6 pages. |
Preinterview First Office Action dated Sep. 30, 2019, issued in connection with U.S. Appl. No. 15/989,715, filed May 25, 2018, 4 pages. |
Preinterview First Office Action dated May 7, 2020, issued in connection with U.S. Appl. No. 16/213,570, filed Dec. 7, 2018, 5 pages. |
Presentations at WinHEC 2000, May 2000, 138 pages. |
Restriction Requirement dated Aug. 14, 2019, issued in connection with U.S. Appl. No. 16/214,711, filed Dec. 10, 2018, 5 pages. |
Restriction Requirement dated Aug. 9, 2018, issued in connection with U.S. Appl. No. 15/717,621, filed Sep. 27, 2017, 8 pages. |
Rottondi et al., “An Overview on Networked Music Performance Technologies,” IEEE Access, vol. 4, pp. 8823-8843, 2016, DOI: 10.1109/ACCESS.2016.2628440, 21 pages. |
Souden et al. “An Integrated Solution for Online Multichannel Noise Tracking and Reduction.” IEEE Transactions on Audio, Speech, and Language Processing, vol. 19. No. 7, Sep. 7, 2011, 11 pages. |
Souden et al. “Gaussian Model-Based Multichannel Speech Presence Probability” IEEE Transactions on Audio, Speech, and Language Processing, vol. 18, No. 5, Jul. 5, 2010, 6pages. |
Souden et al. “On Optimal Frequency-Domain Multichannel Linear Filtering for Noise Reduction.” IEEE Transactions on Audio, Speech, and Language Processing, vol. 18, No. 2, Feb. 2010, 17pages. |
Steven J. Nowlan and Geoffrey E. Hinton “Simplifying Neural Networks by Soft Weight-Sharing” Neural Computation 4, 1992, 21 pages. |
Tsiami et al. “Experiments in acoustic source localization using sparse arrays in adverse indoors environments”, 2014 22nd European Signal Processing Conference, Sep. 1, 2014, 5 pages. |
Tweet: “How to start using Google app voice commands to make your life easier Share This Story shop @Bullet”, Jan. 21, 2016, https://bgr.com/2016/01/21/best-ok-google-voice-commands/, 3 page. |
Ullrich et al. “Soft Weight-Sharing for Neural Network Compression.” ICLR 2017, 16 pages. |
U.S. Appl. No. 60/490,768, filed Jul. 28, 2003, entitled “Method for synchronizing audio playback between multiple networked devices,” 13 pages. |
U.S. Appl. No. 60/825,407, filed Sep. 12, 2006, entitled “Controlling and manipulating groupings in a multi-zone music or media system,” 82 pages. |
UPnP; “Universal Plug and Play Device Architecture,” Jun. 8, 2000; version 1.0; Microsoft Corporation; pp. 1-54. |
Vacher at al. “Recognition of voice commands by multisource ASR and noise cancellation in a smart home environment” Signal Processing Conference 2012 Proceedings of the 20th European, IEEE, Aug. 27, 2012, 5 pages. |
Vacher et al. “Speech Recognition in a Smart Home: Some Experiments for Telemonitoring,” 2009 Proceedings of the 5th Conference on Speech Technology and Human-Computer Dialogoue, Constant, 2009, 10 pages. |
“S Voice or Google Now?”; https://web.archive.org/web/20160807040123/lowdown.carphonewarehouse.com/news/s-voice-or-google-now/ . . . , Apr. 28, 2015; 4 pages. |
Wung et al. “Robust Acoustic Echo Cancellation in the Short-Time Fourier Transform Domain Using Adaptive Crossband Filters” IEEE International Conference on Acoustic, Speech and Signal Processing ICASSP, 2014, p. 1300-1304. |
Xiao et al. “A Learning-Based Approach to Direction of Arrival Estimation in Noisy and Reverberant Environments,” 2015 IEEE International Conference on Acoustics, Speech and Signal Processing, Apr. 19, 2015, 5 pages. |
Yamaha DME 64 Owner's Manual; copyright 2004, 80 pages. |
Yamaha DME Designer 3.0 Owner's Manual; Copyright 2008, 501 pages. |
Yamaha DME Designer 3.5 setup manual guide; copyright 2004, 16 pages. |
Yamaha DME Designer 3.5 User Manual; Copyright 2004, 507 pages. |
Non-Final Office Action dated Jan. 10, 2018, issued in connection with U.S. Appl. No. 15/438,725, filed Feb. 21, 2017, 15 pages. |
Non-Final Office Action dated Sep. 10, 2018, issued in connection with U.S. Appl. No. 15/670,361, filed Aug. 7, 2017, 17 pages. |
Non-Final Office Action dated Oct. 11, 2019, issued in connection with U.S. Appl. No. 16/177,185, filed Oct. 31, 2018, 14 pages. |
Non-Final Office Action dated Sep. 11, 2020, issued in connection with U.S. Appl. No. 15/989,715, filed May 25, 2018, 8 pages. |
Non-Final Office Action dated Sep. 11, 2020, issued in connection with U.S. Appl. No. 16/219,702, filed Dec. 13, 2018, 9 pages. |
Non-Final Office Action dated Dec. 12, 2016, issued in connection with U.S. Appl. No. 15/098,718, filed Apr. 14, 2016, 11 pages. |
Non-Final Office Action dated Feb. 12, 2019, issued in connection with U.S. Appl. No. 15/670,361, filed Aug. 7, 2017, 13 pages. |
Non-Final Office Action dated Jan. 13, 2017, issued in connection with U.S. Appl. No. 15/098,805, filed Apr. 14, 2016, 11 pages. |
Non-Final Office Action dated Nov. 13, 2018, issued in connection with U.S. Appl. No. 15/717,621, filed Sep. 27, 2017, 23 pages. |
Non-Final Office Action dated Nov. 13, 2018, issued in connection with U.S. Appl. No. 16/160,107, filed Oct. 15, 2018, 8 pages. |
Non-Final Office Action dated Nov. 13, 2019, issued in connection with U.S. Appl. No. 15/984,073, filed May 18, 2018, 18 pages. |
Non-Final Office Action dated May 14, 2020, issued in connection with U.S. Appl. No. 15/948,541, filed Apr. 9, 2018, 8 pages. |
Non-Final Office Action dated Sep. 14, 2017, issued in connection with U.S. Appl. No. 15/178,180, filed Jun. 9, 2016, 16 pages. |
Non-Final Office Action dated Sep. 14, 2018, issued in connection with U.S. Appl. No. 15/959,907, filed Apr. 23, 2018, 15 pages. |
Non-Final Office Action dated Apr. 15, 2020, issued in connection with U.S. Appl. No. 16/138,111, filed Sep. 21, 2018, 15 pages. |
Non-Final Office Action dated Dec. 15, 2020, issued in connection with U.S. Appl. No. 17/087,423, filed Nov. 2, 2020, 7 pages. |
Non-Final Office Action dated Jan. 15, 2019, issued in connection with U.S. Appl. No. 16/173,797, filed Oct. 29, 2018, 6 pages. |
Non-Final Office Action dated Nov. 15, 2019, issued in connection with U.S. Appl. No. 16/153,530, filed Oct. 5, 2018, 17 pages. |
Non-Final Office Action dated Mar. 16, 2018, issued in connection with U.S. Appl. No. 15/681,937, filed Aug. 21, 2017, 5 pages. |
Non-Final Office Action dated Oct. 16, 2018, issued in connection with U.S. Appl. No. 15/131,254, filed Apr. 18, 2016, 16 pages. |
Non-Final Office Action dated Sep. 17, 2020, issued in connection with U.S. Appl. No. 16/600,949, filed Oct. 14, 2019, 29 pages. |
Non-Final Office Action dated Apr. 18, 2018, issued in connection with U.S. Appl. No. 15/811,468, filed Nov. 13, 2017, 14 pages. |
Non-Final Office Action dated Jan. 18, 2019, issued in connection with U.S. Appl. No. 15/721,141, filed Sep. 29, 2017, 18 pages. |
Non-Final Office Action dated Oct. 18, 2019, issued in connection with U.S. Appl. No. 15/098,760, filed Apr. 14, 2016, 27 pages. |
Non-Final Office Action dated Sep. 18, 2019, issued in connection with U.S. Appl. No. 16/179,779, filed Nov. 2, 2018, 14 pages. |
Non-Final Office Action dated Apr. 19, 2017, issued in connection with U.S. Appl. No. 15/131,776, filed Apr. 18, 2016, 12 pages. |
Non-Final Office Action dated Dec. 19, 2019, issued in connection with U.S. Appl. No. 16/147,710, filed Sep. 29, 2018, 10 pages. |
Non-Final Office Action dated Feb. 19, 2020, issued in connection with U.S. Appl. No. 16/148,879, filed Oct. 1, 2018, 15 pages. |
Non-Final Office Action dated Sep. 2, 2020, issued in connection with U.S. Appl. No. 16/290,599, filed Mar. 1, 2019, 17 pages. |
Non-Final Office Action dated Feb. 20, 2018, issued in connection with U.S. Appl. No. 15/211,748, filed Jul. 15, 2016, 31 pages. |
Non-Final Office Action dated Jun. 20, 2019, issued in connection with U.S. Appl. No. 15/946,585, filed Apr. 5, 2018, 10 pages. |
Non-Final Office Action dated Aug. 21, 2019, issued in connection with U.S. Appl. No. 16/192,126, filed Nov. 15, 2018, 8 pages. |
Non-Final Office Action dated Feb. 21, 2019, issued in connection with U.S. Appl. No. 16/214,666, filed Dec. 10, 2018, 12 pages. |
Non-Final Office Action dated Jan. 21, 2020, issued in connection with U.S. Appl. No. 16/214,711, filed Dec. 10, 2018, 9 pages. |
Non-Final Office Action dated Jan. 21, 2020, issued in connection with U.S. Appl. No. 16/598,125, filed Oct. 10, 2019, 25 pages. |
Non-Final Office Action dated Oct. 21, 2019, issued in connection with U.S. Appl. No. 15/973,413, filed May 7, 2018, 10 pages. |
Non-Final Office Action dated Jul. 22, 2020, issued in connection with U.S. Appl. No. 16/145,275, filed Sep. 28, 2018, 11 pages. |
Non-Final Office Action dated May 22, 2018, issued in connection with U.S. Appl. No. 15/946,599, filed Apr. 5, 2018, 19 pages. |
Non-Final Office Action dated Sep. 22, 2020, issued in connection with U.S. Appl. No. 16/539,843, filed Aug. 13, 2019, 7 pages. |
Non-Final Office Action dated May 23, 2019, issued in connection with U.S. Appl. No. 16/154,071, filed Oct. 8, 2018, 36 pages. |
Non-Final Office Action dated Nov. 23, 2020, issued in connection with U.S. Appl. No. 16/524,306, filed Jul. 29, 2019, 14 pages. |
Non-Final Office Action dated Sep. 23, 2020, issued in connection with U.S. Appl. No. 16/177,185, filed Oct. 31, 2018, 17 pages. |
Non-Final Office Action dated Aug. 24, 2017, issued in connection with U.S. Appl. No. 15/297,627, filed Oct. 19, 2016, 13 pages. |
Non-Final Office Action dated Jul. 24, 2019, issued in connection with U.S. Appl. No. 16/439,009, filed Jun. 12, 2019, 26 pages. |
Non-Final Office Action dated Jul. 25, 2017, issued in connection with U.S. Appl. No. 15/273,679, filed Jul. 22, 2016, 11 pages. |
Non-Final Office Action dated Dec. 26, 2018, issued in connection with U.S. Appl. No. 16/154,469, filed Oct. 8, 2018, 7 pages. |
Non-Final Office Action dated Jan. 26, 2017, issued in connection with U.S. Appl. No. 15/098,867, filed Apr. 14, 2016, 16 pages. |
Non-Final Office Action dated Oct. 26, 2017, issued in connection with U.S. Appl. No. 15/438,744, filed Feb. 21, 2017, 12 pages. |
Non-Final Office Action dated Jun. 27, 2018, issued in connection with U.S. Appl. No. 15/438,749, filed Feb. 21, 2017, 16 pages. |
Non-Final Office Action dated Jun. 27, 2019, issued in connection with U.S. Appl. No. 16/437,437, filed Jun. 11, 2019, 8 pages. |
Final Office Action dated Nov. 10, 2020, issued in connection with U.S. Appl. No. 16/600,644, filed Oct. 14, 2019, 19 pages. |
Final Office Action dated Apr. 11, 2019, issued in connection with U.S. Appl. No. 15/131,254, filed Apr. 18, 2016, 17 pages. |
Final Office Action dated Aug. 11, 2017, issued in connection with U.S. Appl. No. 15/131,776, filed Apr. 18, 2016, 7 pages. |
Final Office Action dated Dec. 11, 2019, issued in connection with U.S. Appl. No. 16/227,308, filed Dec. 20, 2018, 10 pages. |
Final Office Action dated Sep. 11, 2019, issued in connection with U.S. Appl. No. 16/178,122, filed Nov. 1, 2018, 13 pages. |
Final Office Action dated Apr. 13, 2018, issued in connection with U.S. Appl. No. 15/131,254, filed Apr. 18, 2016, 18 pages. |
Final Office Action dated Apr. 13, 2018, issued in connection with U.S. Appl. No. 15/438,744, filed Feb. 21, 2017, 20 pages. |
Final Office Action dated May 13, 2020, issued in connection with U.S. Appl. No. 16/153,530, filed Oct. 5, 2018, 20 pages. |
Final Office Action dated Jun. 15, 2017, issued in connection with U.S. Appl. No. 15/098,718, filed Apr. 14, 2016, 15 pages. |
Final Office Action dated Oct. 15, 2018, issued in connection with U.S. Appl. No. 15/804,776, filed Nov. 6, 2017, 18 pages. |
Final Office Action dated Oct. 15, 2020, issued in connection with U.S. Appl. No. 16/109,375, filed Aug. 22, 2018, 9 pages. |
Final Office Action dated Oct. 16, 2018, issued in connection with U.S. Appl. No. 15/438,725, filed Feb. 21, 2017, 10 pages. |
Final Office Action dated May 18, 2020, issued in connection with U.S. Appl. No. 16/177,185, filed Oct. 31, 2018, 16 pages. |
Final Office Action dated Feb. 21, 2018, issued in connection with U.S. Appl. No. 15/297,627, filed Oct. 19, 2016, 12 pages. |
Final Office Action dated May 21, 2020, issued in connection with U.S. Appl. No. 15/989,715, filed May 25, 2018, 21 pages. |
Final Office Action dated Jun. 22, 2020, issued in connection with U.S. Appl. No. 16/179,779, filed Nov. 2, 2018, 16 pages. |
Final Office Action dated Mar. 23, 2020, issued in connection with U.S. Appl. No. 16/145,275, filed Sep. 28, 2018, 11 pages. |
Final Office Action dated Feb. 24, 2020, issued in connection with U.S. Appl. No. 15/936,177, filed Mar. 26, 2018, 20 pages. |
Final Office Action dated Apr. 26, 2019, issued in connection with U.S. Appl. No. 15/721,141, filed Sep. 29, 2017, 20 pages. |
Final Office Action dated Apr. 30, 2019, issued in connection with U.S. Appl. No. 15/098,760, filed Apr. 14, 2016, 6 pages. |
Final Office Action dated Feb. 5, 2019, issued in connection with U.S. Appl. No. 15/438,749, filed Feb. 21, 2017, 17 pages. |
Final Office Action dated Feb. 7, 2020, issued in connection with U.S. Appl. No. 15/948,541, filed Apr. 9, 2018, 8 pages. |
Final Office Action dated Sep. 8, 2020, issued in connection with U.S. Appl. No. 16/213,570, filed Dec. 7, 2018, 12 pages. |
Fiorenza Arisio et al. “Deliverable 1.1 User Study, analysis of requirements and definition of the application task,” May 31, 2012, http://dirha.fbk.eu/sites/dirha.fbk.eu/files/docs/DIRHA_D1.1., 31 pages. |
First Action Interview Office Action dated Aug. 14, 2019, issued in connection with U.S. Appl. No. 16/227,308, filed Dec. 20, 2018, 4 pages. |
First Action Interview Office Action dated Jun. 15, 2020, issued in connection with U.S. Appl. No. 16/213,570, filed Dec. 7, 2018, 4 pages. |
First Action Interview Office Action dated Jun. 2, 2020, issued in connection with U.S. Appl. No. 16/109,375, filed Aug. 22, 2018, 10 pages. |
First Action Interview Office Action dated Jan. 22, 2020, issued in connection with U.S. Appl. No. 15/989,715, filed May 25, 2018, 3 pages. |
First Action Interview Office Action dated Jul. 5, 2019, issued in connection with U.S. Appl. No. 16/227,308, filed Dec. 20, 2018, 4 pages. |
Freiberger, Karl, “Development and Evaluation of Source Localization Algorithms for Coincident Microphone Arrays,” Diploma Thesis, Apr. 1, 2010, 106 pages. |
Giacobello et al. “A Sparse Nonuniformly Partitioned Multidelay Filter for Acoustic Echo Cancellation,” 2013, IEEE Workshop on Applications of Signal Processing to Audio and Acoustics, Oct. 2013, New Paltz, NY, 4 pages. |
Giacobello et al. “Tuning Methodology for Speech Enhancement Algorithms using a Simulated Conversational Database and Perceptual Objective Measures,” 2014, 4th Joint Workshop on Hands-free Speech Communication and Microphone Arrays HSCMA, 2014, 5 pages. |
Han et al. “Deep Compression: Compressing Deep Neural Networks with Pruning, Trained Quantization and Huffman Coding.” ICLR 2016, Feb. 15, 2016, 14 pages. |
Helwani et al “Source-domain adaptive filtering for MIMO systems with application to acoustic echo cancellation”, Acoustics Speech and Signal Processing, 2010 IEEE International Conference, Mar. 14, 2010, 4 pages. |
Hirano et al. “A Noise-Robust Stochastic Gradient Algorithm with an Adaptive Step-Size Suitable for Mobile Hands-Free Telephones,” 1995, International Conference on Acoustics, Speech, and Signal Processing, vol. 2, 4 pages. |
International Bureau, International Preliminary Report on Patentability, dated Apr. 11, 2019, issued in connection with International Application No. PCT/US2017/0054063, filed on Sep. 28, 2017, 9 pages. |
International Bureau, International Preliminary Report on Patentability, dated Feb. 20, 2020, issued in connection with International Application No. PCT/US2018/045397, filed on Aug. 6, 2018, 8 pages. |
International Bureau, International Preliminary Report on Patentability, dated Apr. 23, 2019, issued in connection with International Application No. PCT/US2017/057220, filed on Oct. 18, 2017, 7 pages. |
International Bureau, International Preliminary Report on Patentability, dated Mar. 31, 2020, issued in connection with International Application No. PCT/US2018053123, filed on Sep. 27, 2018, 12 pages. |
International Bureau, International Preliminary Report on Patentability, dated Mar. 31, 2020, issued in connection with International Application No. PCT/US2018053472, filed on Sep. 28, 2018, 8 pages. |
International Bureau, International Preliminary Report on Patentability, dated Mar. 31, 2020, issued in connection with International Application No. PCT/US2018053517, filed on Sep. 28, 2018, 10 pages. |
International Bureau, International Preliminary Report on Patentability, dated Sep. 7, 2018, issued in connection with International Application No. PCT/US2017/018728, filed on Feb. 21, 2017, 8 pages. |
International Bureau, International Preliminary Report on Patentability, dated Sep. 7, 2018, issued in connection with International Application No. PCT/US2017/018739, filed on Feb. 21, 2017, 7 pages. |
International Bureau, International Search Report and Written Opinion dated Nov. 10, 2020, issued in connection with International Application No. PCT/US2020/044250, filed on Jul. 30, 2020, 15 pages. |
International Bureau, International Search Report and Written Opinion dated Jul. 14, 2020, issued in connection with International Application No. PCT/US2020/017150, filed on Feb. 7, 2020, 27 pages. |
International Bureau, International Search Report and Written Opinion dated Nov. 18, 2019, issued in connection with International Application No. PCT/US2019052841, filed on Sep. 25, 2019, 12 pages. |
International Bureau, International Search Report and Written Opinion dated Mar. 2, 2020, issued in connection with International Application No. PCT/US2019064907, filed on Dec. 6, 2019, 11 pages. |
International Bureau, International Search Report and Written Opinion dated Dec. 20, 2019, issued in connection with International Application No. PCT/US2019052654, filed on Sep. 24, 2019, 11 pages. |
International Bureau, International Search Report and Written Opinion dated Sep. 21, 2020, issued in connection with International Application No. PCT/US2020/037229, filed on Jun. 11, 2020, 17 pages. |
International Bureau, International Search Report and Written Opinion dated Dec. 6, 2019, issued in connection with International Application No. PCT/US2019050852, filed on Sep. 12, 2019, 10 pages. |
International Bureau, International Search Report and Written Opinion dated Apr. 8, 2020, issued in connection with International Application No. PCT/US2019/067576, filed on Dec. 19, 2019, 12 pages. |
International Searching Authority, International Search Report and Written Opinion dated Dec. 19, 2018, in connection with International Application No. PCT/US2018/053517, 13 pages. |
International Searching Authority, International Search Report and Written Opinion dated Nov. 22, 2017, issued in connection with International Application No. PCT/US2017/054063, filed on Sep. 28, 2017, 11 pages. |
International Searching Authority, International Search Report and Written Opinion dated Jan. 23, 2018, issued in connection with International Application No. PCT/US2017/57220, filed on Oct. 18, 2017, 8 pages. |
International Searching Authority, International Search Report and Written Opinion dated May 23, 2017, issued in connection with International Application No. PCT/US2017/018739, Filed on Feb. 21, 2017, 10 pages. |
International Searching Authority, International Search Report and Written Opinion dated Oct. 23, 2017, issued in connection with International Application No. PCT/US2017/042170, filed on Jul. 14, 2017, 15 pages. |
International Searching Authority, International Search Report and Written Opinion dated Oct. 24, 2017, issued in connection with International Application No. PCT/US2017/042227, filed on Jul. 14, 2017, 16 pages. |
International Searching Authority, International Search Report and Written Opinion dated May 30, 2017, issued in connection with International Application No. PCT/US2017/018728, Filed on Feb. 21, 2017, 11 pages. |
Japanese Patent Office, English Translation of Office Action dated Aug. 27, 2020, issued in connection with Japanese Application No. 2019-073349, 6 pages. |
Japanese Patent Office, English Translation of Office Action dated Jul. 30, 2020, issued in connection with Japanese Application No. 2019-517281, 26 pages. |
Japanese Patent Office, Non-Final Office Action and Translation dated Nov. 5, 2019, issued in connection with Japanese Patent Application No. 2019-517281, 6 pages. |
Japanese Patent Office, Office Action and Translation dated Nov. 17, 2020, issued in connection with Japanese Patent Application No. 2019-145039, 7 pages. |
Japanese Patent Office, Office Action and Translation dated Aug. 27, 2020, issued in connection with Japanese Patent Application No. 2019-073349, 6 pages. |
Japanese Patent Office, Office Action and Translation dated Jul. 30, 2020, issued in connection with Japanese Patent Application No. 2019-517281, 6 pages. |
Japanese Patent Office, Office Action and Translation dated Jul. 6, 2020, issued in connection with Japanese Patent Application No. 2019-073348, 10 pages. |
Japanese Patent Office, Office Action and Translation dated Oct. 8, 2019, issued in connection with Japanese Patent Application No. 2019-521032, 5 pages. |
Japanese Patent Office, Office Action Translation dated Nov. 5, 2019, issued in connection with Japanese Patent Application No. 2019-517281, 2 pages. |
Japanese Patent Office, Office Action Translation dated Oct. 8, 2019, issued in connection with Japanese Patent Application No. 2019-521032, 8 pages. |
Jo et al., “Synchronized One-to-many Media Streaming with Adaptive Playout Control,” Proceedings of SPIE, 2002, pp. 71-82, vol. 4861. |
Johnson, “Implementing Neural Networks into Modern Technology,” IJCNN'99. International Joint Conference on Neural Networks . Proceedings [Cat. No. 99CH36339], Washington, DC, USA, 1999, pp. 1028-1032, vol. 2, doi: 10.1109/IJCNN.1999.831096. [retrieved on Jun. 22, 2020]. |
Jones, Stephen, “Dell Digital Audio Receiver: Digital upgrade for your analog stereo,” Analog Stereo, Jun. 24, 2000 http://www.reviewsonline.com/articles/961906864.htm retrieved Jun. 18, 2014, 2 pages. |
Jose Alvarez and Mathieu Salzmann “Compression-aware Training of Deep Networks” 31st Conference on Neural Information Processing Systems, Nov. 13, 2017, 12pages. |
Korean Patent Office, Korean Office Action and Translation dated Aug. 16, 2019, issued in connection with Korean Application No. 10-2018-7027452, 14 pages. |
Korean Patent Office, Korean Office Action and Translation dated Apr. 2, 2020, issued in connection with Korean Application No. 10-2020-7008486, 12 pages. |
Korean Patent Office, Korean Office Action and Translation dated Mar. 25, 2020, issued in connection with Korean Application No. 10-2019-7012192, 14 pages. |
Korean Patent Office, Korean Office Action and Translation dated Aug. 26, 2020, issued in connection with Korean Application No. 10-2019-7027640, 16 pages. |
Korean Patent Office, Korean Office Action and Translation dated Mar. 30, 2020, issued in connection with Korean Application No. 10-2020-7004425, 5 pages. |
Korean Patent Office, Korean Office Action and Translation dated Sep. 9, 2019, issued in connection with Korean Application No. 10-2018-7027451, 21 pages. |
Korean Patent Office, Korean Office Action dated May 8, 2019, issued in connection with Korean Application No. 10-2018-7027451, 7 pages. |
Korean Patent Office, Korean Office Action dated May 8, 2019, issued in connection with Korean Application No. 10-2018-7027452, 5 pages. |
Louderback, Jim, “Affordable Audio Receiver Furnishes Homes With MP3,” TechTV Vault. Jun. 28, 2000 retrieved Jul. 10, 2014, 2 pages. |
Maja Taseska and Emanual A.P. Habets, “MMSE-Based Blind Source Extraction in Diffuse Noise Fields Using a Complex Coherence-Based a Priori Sap Estimator.” International Workshop on Acoustic Signal Enhancement 2012, Sep. 4-6, 2012, 4pages. |
Morales-Cordovilla et al. “Room Localization for Distant Speech Recognition,” Proceedings of Interspeech 2014, Sep. 14, 2014, 4 pages. |
Newman, Jared. “Chromecast Audio's multi-room support has arrived,” Dec. 11, 2015, https://www.pcworld.com/article/3014204/customer-electronic/chromcase-audio-s-multi-room-support-has . . . , 1 page. |
Ngo et al. “Incorporating the Conditional Speech Presence Probability in Multi-Channel Wiener Filter Based Noise Reduction in Hearing Aids.” EURASIP Journal on Advances in Signal Processing vol. 2009, Jun. 2, 2009, 11 pages. |
Non-Final Office Action dated Dec. 21, 2020, issued in connection with U.S. Appl. No. 16/153,530, filed Oct. 5, 2018, 22 pages. |
Non-Final Office Action dated Dec. 9, 2020, issued in connection with U.S. Appl. No. 16/271,550, filed Feb. 8, 2019, 35 pages. |
Non-Final Office Action dated Jun. 1, 2017, issued in connection with U.S. Appl. No. 15/223,218, filed Jul. 29, 2016, 7 pages. |
Non-Final Office Action dated Nov. 2, 2017, issued in connection with U.S. Appl. No. 15/584,782, filed May 2, 2017, 11 pages. |
Non-Final Office Action dated Nov. 3, 2017, issued in connection with U.S. Appl. No. 15/438,741, filed Feb. 21, 2017, 11 pages. |
Non-Final Office Action dated Nov. 4, 2019, issued in connection with U.S. Appl. No. 16/022,662, filed Jun. 28, 2018, 16 pages. |
Non-Final Office Action dated Sep. 5, 2019, issued in connection with U.S. Appl. No. 16/416,752, filed May 20, 2019, 14 pages. |
Non-Final Office Action dated Feb. 7, 2017, issued in connection with U.S. Appl. No. 15/131,244, filed Apr. 18, 2016, 12 pages. |
Non-Final Office Action dated Feb. 8, 2017, issued in connection with U.S. Appl. No. 15/098,892, filed Apr. 14, 2016, 17 pages. |
Non-Final Office Action dated Mar. 9, 2017, issued in connection with U.S. Appl. No. 15/098,760, filed Apr. 14, 2016, 13 pages. |
Non-Final Office Action dated Oct. 9, 2019, issued in connection with U.S. Appl. No. 15/936,177, filed Mar. 26, 2018, 16 pages. |
Non-Final Office Action dated Jul. 1, 2020, issued in connection with U.S. Appl. No. 16/138,111, filed Sep. 21, 2018, 14 pages. |
Non-Final Office Action dated Jan. 10, 2018, issued in connection with U.S. Appl. No. 15/098,718, filed Apr. 14, 2016, 15 pages. |
Non-Final Office Action dated Jan. 10, 2018, issued in connection with U.S. Appl. No. 15/229,868, filed Aug. 5, 2016, 13 pages. |
Advisory Action dated Apr. 23, 2021, issued in connection with U.S. Appl. No. 16/219,702, filed Dec. 13, 2018, 3 pages. |
Anonymous: “What are the function of 4 Microphones on iPhone 6S/6S+?”, ETrade Supply, Dec. 24, 2015, XP055646381, Retrieved from the Internet: URL:https://www.etradesupply.com/blog/4-microphones-iphone-6s6s-for/ [retrieved on Nov. 26, 2019]. |
Australian Patent Office, Australian Examination Report Action dated Apr. 7, 2021, issued in connection with Australian Application No. 2019333058, 2 pages. |
Australian Patent Office, Examination Report dated Jun. 28, 2021, issued in connection with Australian Patent Application No. 2019395022, 2 pages. |
Canadian Patent Office, Canadian Examination Report dated Mar. 9, 2021, issued in connection with Canadian Application No. 3067776, 5 pages. |
Chinese Patent Office, Chinese Office Action and Translation dated Jul. 2, 2021, issued in connection with Chinese Application No. 201880077216.4, 22 pages. |
Chinese Patent Office, Chinese Office Action and Translation dated Mar. 30, 2021, issued in connection with Chinese Application No. 202010302650.7, 15 pages. |
Chinese Patent Office, First Office Action and Translation dated May 27, 2021, issued in connection with Chinese Application No. 201880026360.5, 15 pages. |
Chinese Patent Office, First Office Action and Translation dated Dec. 28, 2020, issued in connection with Chinese Application No. 201880072203.8, 11 pages. |
European Patent Office, European EPC Article 94.3 mailed on Feb. 23, 2021, issued in connection with European Application No. 17200837.7, 8 pages. |
European Patent Office, European EPC Article 94.3 mailed on Feb. 26, 2021, issued in connection with European Application No. 18789515.6, 8 pages. |
European Patent Office, European Office Action dated Jan. 21, 2021, issued in connection with European Application No. 17792272.1, 7 pages. |
Final Office Action dated Feb. 10, 2021, issued in connection with U.S. Appl. No. 16/219,702, filed Dec. 13, 2018, 9 pages. |
Final Office Action dated Feb. 10, 2021, issued in connection with U.S. Appl. No. 16/402,617, filed May 3, 2019, 13 pages. |
Final Office Action dated Jul. 15, 2021, issued in connection with U.S. Appl. No. 16/153,530, filed Oct. 5, 2018, 22 pages. |
Final Office Action dated Jun. 15, 2021, issued in connection with U.S. Appl. No. 16/819,755, filed Mar. 16, 2020, 12 pages. |
Final Office Action dated Feb. 22, 2021, issued in connection with U.S. Appl. No. 15/936,177, filed Mar. 26, 2018, 20 pages. |
Final Office Action dated Feb. 22, 2021, issued in connection with U.S. Appl. No. 16/213,570, filed Dec. 7, 2018, 12 pages. |
Final Office Action dated Jun. 4, 2021, issued in connection with U.S. Appl. No. 16/168,389, filed Oct. 23, 2018, 38 pages. |
Final Office Action dated Jun. 8, 2021, issued in connection with U.S. Appl. No. 16/271,550, filed Feb. 8, 2019, 41 pages. |
First Action Interview Office Action dated Mar. 8, 2021, issued in connection with U.S. Appl. No. 16/798,967, filed Feb. 24, 2020, 4 pages. |
Hans Speidel: “Chatbot Training: How to use training data to provide fully automated customer support”, Jun. 29, 2017, pp. 1-3, XP055473185, Retrieved from the Internet: URL:https://www.crowdguru.de/wp-content/uploads/Case-Study-Chatbot-training-How-to-use-training-data-to-provide-fully-automated-customer-support.pdf [retrieved on May 7, 2018]. |
Indian Patent Office, Examination Report dated May 24, 2021, issued in connection with Indian Patent Application No. 201847035595, 6 pages. |
Indian Patent Office, Examination Report dated Feb. 25, 2021, issued in connection with Indian Patent Application No. 201847035625, 6 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Apr. 1, 2021, issued in connection with International Application No. PCT/US2019/052129, filed on Sep. 20, 2019, 13 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Jul. 1, 2021, issued in connection with International Application No. PCT/US2019/067576, filed on Dec. 19, 2019, 8 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Dec. 10, 2020, issued in connection with International Application No. PCT/US2019/033945, filed on May 25, 2018, 7 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Apr. 15, 2021, issued in connection with International Application No. PCT/US2019/054332, filed on Oct. 2, 2019, 9 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Mar. 25, 2021, issued in connection with International Application No. PCT/US2019/050852, filed on Sep. 12, 2019, 8 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Jan. 7, 2021, issued in connection with International Application No. PCT/US2019/039828, filed on Jun. 28, 2019, 11 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Apr. 8, 2021, issued in connection with International Application No. PCT/US2019/052654, filed on Sep. 24, 2019, 7 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Apr. 8, 2021, issued in connection with International Application No. PCT/US2019/052841, filed on Sep. 25, 2019, 8 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Apr. 8, 2021, issued in connection with International Application No. PCT/US2019/053253, filed on Sep. 26, 2019, 10 pages. |
International Bureau, International Preliminary Report on Patentability, dated Jun. 17, 2021, issued in connection with International Application No. PCT/US2019/064907, filed on Dec. 6, 2019, 8 pages. |
International Bureau, International Preliminary Report on Patentability, dated Mar. 2, 2021, issued in connection with International Application No. PCT/US2019/048558, filed on Aug. 28, 2019, 8 pages. |
International Bureau, International Search Report and Written Opinion dated Dec. 11, 2019, issued in connection with International Application No. PCT/US2019/052129, filed on Sep. 20, 2019, 18 pages. |
International Bureau, International Search Report and Written Opinion dated Nov. 13, 2018, issued in connection with International Application No. PCT/US2018/045397, filed on Aug. 6, 2018, 11 pages. |
International Bureau, International Search Report and Written Opinion dated Jan. 14, 2019, issued in connection with International Application No. PCT/US2018053472, filed on Sep. 28, 2018, 10 pages. |
International Bureau, International Search Report and Written Opinion dated Jul. 17, 2019, issued in connection with International Application No. PCT/US2019/032934, filed on May 17, 2019, 17 pages. |
International Bureau, International Search Report and Written Opinion dated Nov. 18, 2019, issued in connection with International Application No. PCT/US2019/048558, filed on Aug. 28, 2019, 11 pages. |
International Bureau, International Search Report and Written Opinion dated Apr. 23, 2021, issued in connection with International Application No. PCT/US2021/070007, filed on Jan. 6, 2021, 11 pages. |
International Bureau, International Search Report and Written Opinion dated Sep. 27, 2019, issued in connection with International Application No. PCT/US2019/039828, filed on Jun. 28, 2019, 13 pages. |
International Bureau, International Search Report and Written Opinion dated Nov. 29, 2019, issued in connection with International Application No. PCT/US2019/053523, filed on Sep. 29, 2019, 14 pages. |
International Bureau, International Search Report and Written Opinion dated Sep. 4, 2019, issued in connection with International Application No. PCT/US2019/033945, filed on May 24, 2019, 8 pages. |
Japanese Patent Office, Decision of Refusal and Translation dated Jun. 8, 2021, issued in connection with Japanese Patent Application No. 2019-073348, 5 pages. |
Japanese Patent Office, English Translation of Office Action dated Nov. 17, 2020, issued in connection with Japanese Application No. 2019-145039, 5 pages. |
Japanese Patent Office, Notice of Reasons for Refusal and Translation dated Jun. 22, 2021, issued in connection with Japanese Patent Application No. 2020-517935, 4 pages. |
Japanese Patent Office, Office Action and Translation dated Mar. 16, 2021, issued in connection with Japanese Patent Application No. 2020-506725, 7 pages. |
Japanese Patent Office, Office Action and Translation dated Apr. 20, 2021, issued in connection with Japanese Patent Application No. 2020-513852, 9 pages. |
Japanese Patent Office, Office Action and Translation dated Feb. 24, 2021, issued in connection with Japanese Patent Application No. 2019-517281, 4 pages. |
Advisory Action dated Jun. 10, 2020, issued in connection with U.S. Appl. No. 15/936,177, filed Mar. 26, 2018, 4 pages. |
Advisory Action dated Apr. 24, 2020, issued in connection with U.S. Appl. No. 15/948,541, filed Apr. 9, 2018, 4 pages. |
Advisory Action dated Jun. 28, 2018, issued in connection with U.S. Appl. No. 15/438,744, filed Feb. 21, 2017, 3 pages. |
Advisory Action dated Dec. 31, 2018, issued in connection with U.S. Appl. No. 15/804,776, filed Nov. 6, 2017, 4 pages. |
Advisory Action dated Jun. 9, 2020, issued in connection with U.S. Appl. No. 16/145,275, filed Sep. 28, 2018, 3 pages. |
Anonymous,. S Voice or Google Now—The Lowdown. Apr. 28, 2015, 9 pages. [online], [retrieved on Nov. 29, 2017]. Retrieved from the Internet (URL:http://web.archive.org/web/20160807040123/http://lowdown.carphonewarehouse.com/news/s-voice-or-google-now/29958/). |
AudioTron Quick Start Guide, Version 1.0, Mar. 2001, 24 pages. |
AudioTron Reference Manual, Version 3.0, May 2002, 70 pages. |
AudioTron Setup Guide, Version 3.0, May 2002, 38 pages. |
Australian Patent Office, Australian Examination Report Action dated Apr. 14, 2020, issued in connection with Australian Application No. 2019202257, 3 pages. |
Australian Patent Office, Australian Examination Report Action dated Oct. 3, 2019, issued in connection with Australian Application No. 2018230932, 3 pages. |
Australian Patent Office, Australian Examination Report Action dated Aug. 7, 2020, issued in connection with Australian Application No. 2019236722, 4 pages. |
Australian Patent Office, Examination Report dated Oct. 30, 2018, issued in connection with Australian Application No. 2017222436, 3 pages. |
“Automatic Parameter Tying in Neural Networks” ICLR 2018, 14 pages. |
Bertrand et al. “Adaptive Distributed Noise Reduction for Speech Enhancement in Wireless Acoustic Sensor Networks” Jan. 2010, 4 pages. |
Bluetooth. “Specification of the Bluetooth System: The ad hoc Scatternet for affordable and highly functional wireless connectivity,” Core, Version 1.0 A, Jul. 26, 1999, 1068 pages. |
Bluetooth. “Specification of the Bluetooth System: Wireless connections made easy,” Core, Version 1.0 B, Dec. 1, 1999, 1076 pages. |
Canadian Patent Office, Canadian Office Action dated Nov. 14, 2018, issued in connection with Canadian Application No. 3015491, 3 pages. |
Chinese Patent Office, First Office Action and Translation dated Mar. 20, 2019, issued in connection with Chinese Application No. 201780025028.2, 18 pages. |
Chinese Patent Office, First Office Action and Translation dated Mar. 27, 2019, issued in connection with Chinese Application No. 201780025029.7, 9 pages. |
Chinese Patent Office, First Office Action and Translation dated Nov. 5, 2019, issued in connection with Chinese Application No. 201780072651.3, 19 pages. |
Chinese Patent Office, First Office Action dated Feb. 28, 2020, issued in connection with Chinese Application No. 201780061543.6, 29 pages. |
Chinese Patent Office, Second Office Action and Translation dated May 11, 2020, issued in connection with Chinese Application No. 201780061543.6, 17 pages. |
Chinese Patent Office, Second Office Action and Translation dated Jul. 18, 2019, issued in connection with Chinese Application No. 201780025029.7, 14 pages. |
Chinese Patent Office, Second Office Action and Translation dated Sep. 23, 2019, issued in connection with Chinese Application No. 201780025028.2, 15 pages. |
Chinese Patent Office, Second Office Action and Translation dated Mar. 31, 2020, issued in connection with Chinese Application No. 201780072651.3, 17 pages. |
Chinese Patent Office, Third Office Action and Translation dated Sep. 16, 2019, issued in connection with Chinese Application No. 201780025029.7, 14 pages. |
Chinese Patent Office, Third Office Action and Translation dated Aug. 5, 2020, issued in connection with Chinese Application No. 201780072651.3, 10 pages. |
Chinese Patent Office, Translation of Office Action dated Jul. 18, 2019, issued in connection with Chinese Application No. 201780025029.7, 8 pages. |
Cipriani,. The complete list of OK, Google commands—CNET. Jul. 1, 2016, 5 pages. [online], [retrieved on Jan. 15, 2020]. Retrieved from the Internet: (URL:https://web.archive.org/web/20160803230926/https://www.cnet.com/how-to/complete-list-of-ok-google--commands/). |
Corrected Notice of Allowability dated Mar. 8, 2017, issued in connection with U.S. Appl. No. 15/229,855, filed Aug. 5, 2016, 6 pages. |
Dell, Inc. “Dell Digital Audio Receiver: Reference Guide,” Jun. 2000, 70 pages. |
Dell, Inc. “Start Here,” Jun. 2000, 2 pages. |
“Denon 2003-2004 Product Catalog,” Denon, 2003-2004, 44 pages. |
European Patent Office, European Extended Search Report dated Nov. 25, 2020, issued in connection with European Application No. 20185599.6, 9 pages. |
European Patent Office, European Extended Search Report dated Feb. 3, 2020, issued in connection with European Application No. 19197116.7, 9 pages. |
European Patent Office, European Extended Search Report dated Jan. 3, 2019, issued in connection with European Application No. 177570702, 8 pages. |
European Patent Office, European Extended Search Report dated Jan. 3, 2019, issued in connection with European Application No. 17757075.1, 9 pages. |
European Patent Office, European Extended Search Report dated Oct. 30, 2017, issued in connection with EP Application No. 17174435.2, 11 pages. |
European Patent Office, European Extended Search Report dated Aug. 6, 2020, issued in connection with European Application No. 20166332.5, 10 pages. |
European Patent Office, European Office Action dated Jul. 1, 2020, issued in connection with European Application No. 17757075.1, 7 pages. |
European Patent Office, European Office Action dated Jan. 14, 2020, issued in connection with European Application No. 17757070.2, 7 pages. |
European Patent Office, European Office Action dated Jan. 22, 2019, issued in connection with European Application No. 17174435.2, 9 pages. |
European Patent Office, European Office Action dated Sep. 23, 2020, issued in connection with European Application No. 18788976.1, 7 pages. |
European Patent Office, European Office Action dated Oct. 26, 2020, issued in connection with European Application No. 18760101.8, 4 pages. |
European Patent Office, European Office Action dated Aug. 30, 2019, issued in connection with European Application No. 17781608.9, 6 pages. |
European Patent Office, European Office Action dated Sep. 9, 2020, issued in connection with European Application No. 18792656.3, 10 pages. |
European Patent Office, Summons to Attend Oral Proceedings mailed on Dec. 20, 2019, issued in connection with European Application No. 17174435.2, 13 pages. |
Fadilpasic,“Cortana can now be the default PDA on your Android”, IT Pro Portal: Accessed via WayBack Machine; http://web.archive.org/web/20171129124915/https://www.itproportal.com/2015/08/11/cortana-can-now-be- . . . , Aug. 11, 2015, 6 pages. |
Final Office Action dated Oct. 6, 2017, issued in connection with U.S. Appl. No. 15/098,760, filed Apr. 14, 2016, 25 pages. |
Non-Final Office Action dated Jun. 27, 2019, issued in connection with U.S. Appl. No. 16/437,476, filed Jun. 11, 2019, 8 pages. |
Non-Final Office Action dated Mar. 27, 2020, issued in connection with U.S. Appl. No. 16/790,621, filed Feb. 13, 2020, 8 pages. |
Non-Final Office Action dated May 27, 2020, issued in connection with U.S. Appl. No. 16/715,713, filed Dec. 16, 2019, 14 pages. |
Non-Final Office Action dated Oct. 27, 2020, issued in connection with U.S. Appl. No. 16/213,570, filed Dec. 7, 2018, 13 pages. |
Non-Final Office Action dated Oct. 27, 2020, issued in connection with U.S. Appl. No. 16/715,984, filed Dec. 16, 2019, 14 pages. |
Non-Final Office Action dated Oct. 27, 2020, issued in connection with U.S. Appl. No. 16/819,755, filed Mar. 16, 2020, 8 pages. |
Non-Final Office Action dated Oct. 28, 2019, issued in connection with U.S. Appl. No. 16/145,275, filed Sep. 28, 2018, 11 pages. |
Non-Final Office Action dated Mar. 29, 2019, issued in connection with U.S. Appl. No. 16/102,650, filed Aug. 13, 2018, 11 pages. |
Non-Final Office Action dated Sep. 29, 2020, issued in connection with U.S. Appl. No. 16/402,617, filed May 3, 2019, 12 pages. |
Non-Final Office Action dated Dec. 3, 2020, issued in connection with U.S. Appl. No. 16/145,275, filed Sep. 28, 2018, 11 pages. |
Non-Final Office Action dated Jul. 3, 2019, issued in connection with U.S. Appl. No. 15/948,541, filed Apr. 9, 2018, 7 pages. |
Non-Final Office Action dated May 3, 2019, issued in connection with U.S. Appl. No. 16/178,122, filed Nov. 1, 2018, 14 pages. |
Non-Final Office Action dated Oct. 3, 2018, issued in connection with U.S. Appl. No. 16/102,153, filed Aug. 13, 2018, 20 pages. |
Non-Final Office Action dated Apr. 30, 2019, issued in connection with U.S. Appl. No. 15/718,521, filed Sep. 28, 2017, 39 pages. |
Non-Final Office Action dated Jun. 30, 2017, issued in connection with U.S. Appl. No. 15/277,810, filed Sep. 27, 2016, 13 pages. |
Non-Final Office Action dated Apr. 4, 2019, issued in connection with U.S. Appl. No. 15/718,911, filed Sep. 28, 2017, 21 pages. |
Non-Final Office Action dated Aug. 4, 2020, issued in connection with U.S. Appl. No. 16/600,644, filed Oct. 14, 2019, 30 pages. |
Non-Final Office Action dated Jan. 4, 2019, issued in connection with U.S. Appl. No. 15/948,541, filed Apr. 9, 2018, 6 pages. |
Non-Final Office Action dated Apr. 6, 2020, issued in connection with U.S. Appl. No. 16/424,825, filed May 29, 2019, 22 pages. |
Non-Final Office Action dated Feb. 6, 2018, issued in connection with U.S. Appl. No. 15/211,689, filed Jul. 15, 2016, 32 pages. |
Non-Final Office Action dated Feb. 6, 2018, issued in connection with U.S. Appl. No. 15/237,133, filed Aug. 15, 2016, 6 pages. |
Non-Final Office Action dated Mar. 6, 2020, issued in connection with U.S. Appl. No. 16/141,875, filed Sep. 25, 2018, 8 pages. |
Non-Final Office Action dated Sep. 6, 2017, issued in connection with U.S. Appl. No. 15/131,254, filed Apr. 18, 2016, 13 pages. |
Non-Final Office Action dated Sep. 6, 2018, issued in connection with U.S. Appl. No. 15/098,760, filed Apr. 14, 2016, 29 pages. |
Non-Final Office Action dated Sep. 8, 2020, issued in connection with U.S. Appl. No. 15/936,177, filed Mar. 26, 2018, 19 pages. |
Non-Final Office Action dated Apr. 9, 2018, issued in connection with U.S. Appl. No. 15/804,776, filed Nov. 6, 2017, 18 pages. |
Non-Final Office Action dated May 9, 2018, issued in connection with U.S. Appl. No. 15/818,051, filed Nov. 20, 2017, 22 pages. |
Non-Final Office Action dated Sep. 9, 2020, issued in connection with U.S. Appl. No. 16/168,389, filed Oct. 23, 2018, 29 pages. |
Notice of Allowance dated Dec. 2, 2019, issued in connection with U.S. Appl. No. 15/718,521, filed Sep. 28, 2017, 15 pages. |
Notice of Allowance dated Dec. 4, 2017, issued in connection with U.S. Appl. No. 15/277,810, filed Sep. 27, 2016, 5 pages. |
Notice of Allowance dated Jul. 5, 2018, issued in connection with U.S. Appl. No. 15/237,133, filed Aug. 15, 2016, 5 pages. |
Notice of Allowance dated Jul. 9, 2018, issued in connection with U.S. Appl. No. 15/438,741, filed Feb. 21, 2017, 5 pages. |
Notice of Allowance dated Apr. 1, 2019, issued in connection with U.S. Appl. No. 15/935,966, filed Mar. 26, 2018, 5 pages. |
Notice of Allowance dated Aug. 1, 2018, issued in connection with U.S. Appl. No. 15/297,627, filed Oct. 19, 2016, 9 pages. |
Notice of Allowance dated Aug. 10, 2020, issued in connection with U.S. Appl. No. 16/424,825, filed May 29, 2019, 9 pages. |
Notice of Allowance dated Apr. 11, 2018, issued in connection with U.S. Appl. No. 15/719,454, filed Sep. 28, 2017, 15 pages. |
Notice of Allowance dated Oct. 11, 2019, issued in connection with U.S. Appl. No. 16/437,476, filed Jun. 11, 2019, 9 pages. |
Notice of Allowance dated Sep. 11, 2019, issued in connection with U.S. Appl. No. 16/154,071, filed Oct. 8, 2018, 5 pages. |
Notice of Allowance dated Dec. 12, 2018, issued in connection with U.S. Appl. No. 15/811,468, filed Nov. 13, 2017, 9 pages. |
Notice of Allowance dated Jul. 12, 2017, issued in connection with U.S. Appl. No. 15/098,805, filed Apr. 14, 2016, 8 pages. |
Notice of Allowance dated Jun. 12, 2019, issued in connection with U.S. Appl. No. 15/670,361, filed Aug. 7, 2017, 7 pages. |
Notice of Allowance dated Sep. 12, 2018, issued in connection with U.S. Appl. No. 15/438,744, filed Feb. 21, 2017, 15 pages. |
Notice of Allowance dated Dec. 13, 2017, issued in connection with U.S. Appl. No. 15/784,952, filed Oct. 16, 2017, 9 pages. |
Notice of Allowance dated Feb. 13, 2019, issued in connection with U.S. Appl. No. 15/959,907, filed Apr. 23, 2018, 10 pages. |
Notice of Allowance dated Jan. 13, 2020, issued in connection with U.S. Appl. No. 16/192,126, filed Nov. 15, 2018, 6 pages. |
Notice of Allowance dated Nov. 13, 2020, issued in connection with U.S. Appl. No. 16/131,409, filed Sep. 14, 2018, 11 pages. |
Notice of Allowance dated Aug. 14, 2017, issued in connection with U.S. Appl. No. 15/098,867, filed Apr. 14, 2016, 10 pages. |
Notice of Allowance dated Aug. 14, 2020, issued in connection with U.S. Appl. No. 16/598,125, filed Oct. 10, 2019, 5 pages. |
Notice of Allowance dated Feb. 14, 2017, issued in connection with U.S. Appl. No. 15/229,855, filed Aug. 5, 2016, 11 pages. |
Notice of Allowance dated Jun. 14, 2017, issued in connection with U.S. Appl. No. 15/282,554, filed Sep. 30, 2016, 11 pages. |
Advisory Action dated Aug. 13, 2021, issued in connection with U.S. Appl. No. 16/271,550, filed Feb. 8, 2019, 4 pages. |
Advisory Action dated Feb. 28, 2022, issued in connection with U.S. Appl. No. 16/813,643, filed Mar. 9, 2020, 3 pages. |
Advisory Action dated Sep. 8, 2021, issued in connection with U.S. Appl. No. 16/168,389, filed Oct. 23, 2018, 4 pages. |
Andra et al. Contextual Keyword Spotting in Lecture Video With Deep Convolutional Neural Network. 2017 International Conference on Advanced Computer Science and Information Systems, IEEE, Oct. 28, 2017, 6 pages. |
Audhkhasi Kartik et al. End-to-end ASR-free keyword search from speech. 2017 IEEE International Conference on Acoustics, Speech and Signal Processing, Mar. 5, 2017, 7 pages. |
Australian Patent Office, Australian Examination Report Action dated May 19, 2022, issued in connection with Australian Application No. 2021212112, 2 pages. |
Australian Patent Office, Australian Examination Report Action dated Mar. 4, 2022, issued in connection with Australian Application No. 2021202786, 2 pages. |
Canadian Patent Office, Canadian Examination Report dated Dec. 1, 2021, issued in connection with Canadian Application No. 3096442, 4 pages. |
Canadian Patent Office, Canadian Examination Report dated Nov. 2, 2021, issued in connection with Canadian Application No. 3067776, 4 pages. |
Canadian Patent Office, Canadian Examination Report dated Oct. 26, 2021, issued in connection with Canadian Application No. 3072492, 3 pages. |
Canadian Patent Office, Canadian Examination Report dated Mar. 29, 2022, issued in connection with Canadian Application No. 3111322, 3 pages. |
Canadian Patent Office, Canadian Examination Report dated Jun. 7, 2022, issued in connection with Canadian Application No. 3105494, 5 pages. |
Chinese Patent Office, First Office Action and Translation dated Jun. 1, 2021, issued in connection with Chinese Application No. 201980089721.5, 21 pages. |
Chinese Patent Office, First Office Action and Translation dated Dec. 1, 2021, issued in connection with Chinese Application No. 201780077204.7, 11 pages. |
Chinese Patent Office, First Office Action and Translation dated Dec. 20, 2021, issued in connection with Chinese Application No. 202010302650.7, 10 pages. |
Chinese Patent Office, Second Office Action and Translation dated Mar. 3, 2022, issued in connection with Chinese Application No. 201880077216.4, 11 pages. |
Chung et al. Empirical Evaluation of Gated Recurrent Neural Network on Sequence Modeling. Dec. 11, 2014, 9 pages. |
Co-pending U.S. Application No. 202117236559, inventor Millington; Nicholas A.J., filed on Apr. 21, 2021. |
Couke et al. Efficient Keyword Spotting using Dilated Convolutions and Gating, arXiv:1811.07684v2, Feb. 18, 2019, 5 pages. |
European Patent Office, Decision to Refuse European Patent Application dated May 30, 2022, issued in connection with European Application No. 17200837.7, 4 pages. |
European Patent Office, European EPC Article 94.3 mailed on Mar. 11, 2022, issued in connection with European Application No. 19731415.6, 7 pages. |
European Patent Office, European EPC Article 94.3 mailed on Nov. 11, 2021, issued in connection with European Application No. 19784172.9, 5 pages. |
European Patent Office, European EPC Article 94.3 mailed on May 2, 2022, issued in connection with European Application No. 20185599.6, 7 pages. |
European Patent Office, European EPC Article 94.3 mailed on Jun. 21, 2022, issued in connection with European Application No. 19780508.8, 5 pages. |
European Patent Office, European EPC Article 94.3 mailed on Mar. 3, 2022, issued in connection with European Application No. 19740292.8, 10 pages. |
European Patent Office, European EPC Article 94.3 mailed on Jun. 30, 2022, issued in connection with European Application No. 19765953.5, 4 pages. |
European Patent Office, European Extended Search Report dated Oct. 7, 2021, issued in connection with European Application No. 21193616.6, 9 pages. |
European Patent Office, European Extended Search Report dated Apr. 22, 2022, issued in connection with European Application No. 21195031.6, 14 pages. |
European Patent Office, European Extended Search Report dated Jun. 23, 2022, issued in connection with European Application No. 22153180.9, 6 pages. |
European Patent Office, European Extended Search Report dated Jun. 30, 2022, issued in connection with European Application No. 21212763.3, 9 pages. |
European Patent Office, European Extended Search Report dated Jul. 8, 2022, issued in connection with European Application No. 22153523.0, 9 pages. |
European Patent Office, European Search Report dated Mar. 1, 2022, issued in connection with European Application No. 21180778.9, 9 pages. |
European Patent Office, Examination Report dated Jul. 15, 2021, issued in connection with European Patent Application No. 19729968.8, 7 pages. |
European Patent Office, Extended Search Report dated Aug. 13, 2021, issued in connection with European Patent Application No. 21164130.3, 11 pages. |
European Patent Office, Extended Search Report dated May 16, 2018, issued in connection with European Patent Application No. 17200837.7, 11 pages. |
European Patent Office, Extended Search Report dated Jul. 25, 2019, issued in connection with European Patent Application No. 18306501.0, 14 pages. |
European Patent Office, Extended Search Report dated May 29, 2020, issued in connection with European Patent Application No. 19209389.6, 8 pages. |
European Patent Office, Summons to Attend Oral Proceedings mailed on Jul. 15, 2022, issued in connection with European Application No. 17792272.1, 11 pages. |
European Patent Office, Summons to Attend Oral Proceedings mailed on Feb. 4, 2022, issued in connection with European Application No. 17757075.1, 10 pages. |
European Patent Office, Summons to Attend Oral Proceedings mailed on Dec. 9, 2021, issued in connection with European Application No. 17200837.7, 10 pages. |
Final Office Action dated Jul. 23, 2021, issued in connection with U.S. Appl. No. 16/439,046, filed Jun. 12, 2019, 12 pages. |
Final Office Action dated Jun. 1, 2022, issued in connection with U.S. Appl. No. 16/806,747, filed Mar. 2, 2020, 20 pages. |
Final Office Action dated Aug. 17, 2022, issued in connection with U.S. Appl. No. 16/179,779, filed Nov. 2, 2018, 26 pages. |
Final Office Action dated Dec. 17, 2021, issued in connection with U.S. Appl. No. 16/813,643, filed Mar. 9, 2020, 12 pages. |
Final Office Action dated Mar. 21, 2022, issued in connection with U.S. Appl. No. 16/153,530, filed Oct. 5, 2018, 23 pages. |
Final Office Action dated Aug. 22, 2022, issued in connection with U.S. Appl. No. 16/168,389, filed Oct. 23, 2018, 37 pages. |
Final Office Action dated Jul. 27, 2022, issued in connection with U.S. Appl. No. 16/989,350, filed Aug. 10, 2020, 15 pages. |
Final Office Action dated Nov. 29, 2021, issued in connection with U.S. Appl. No. 17/236,559, filed Apr. 21, 2021, 11 pages. |
Final Office Action dated Oct. 4, 2021, issued in connection with U.S. Appl. No. 16/806,747, filed Mar. 2, 2020, 17 pages. |
Final Office Action dated Jun. 7, 2022, issued in connection with U.S. Appl. No. 16/736,725, filed Jan. 7, 2020, 14 pages. |
Snips: How to Snips—Assistant creation & Installation, Jun. 26, 2017, 6 pages. |
Speidel, Hans. Chatbot Training: How to use training data to provide fully automated customer support. Retrieved from the Internet: URL: https://www.crowdguru.de/wp-content/uploads/Case-Study-Chatbox-training-How-to-use-training-data-to-provide-fully-automated-customer-support.pdf. Jun. 29, 2017, 4 pages. |
Stemmer et al. Speech Recognition and Understanding on Hardware-Accelerated DSP. Proceedings of Interspeech 2017: Show & Tell Contribution, Aug. 20, 2017, 2 pages. |
Tsung-Hsien Wen et al: “A Network-based End-to-End Trainable Task-oriented Dialogue System”, CORR (ARXIV), vol. 1604.04562v1, Apr. 15, 2016 (Apr. 15, 2016), pp. 1-11. |
Wen et al. A Network-based End-to-End Trainable Task-oriented Dialogue System, CORR (ARXIV), Apr. 15, 2016, 11 pages. |
Wikipedia. “The Wayback Machine”, Speech recognition software for Linux, Sep. 22, 2016, 4 pages. [retrieved on Mar. 28, 2022], Retrieved from the Internet: URL: https://web.archive.org/web/20160922151304/https://en.wikipedia.org/wiki/Speech_recognition_software_for_Linux. |
Wu et al. End-to-End Recurrent Entity Network for Entity-Value Independent Goal-Oriented Dialog Learning. DSTC6—Dialog System Technology Challenges, Dec. 10, 2017, 5 pages. |
Xiaoguang et al. “Robust Small-Footprint Keyword Spotting Using Sequence-To-Sequence Model with Connectionist Temporal Classifier”, 2019 IEEE, Sep. 28, 2019, 5 pages. |
Xu et al. An End-to-end Approach for Handling Unknown Slot Values in Dialogue State Tracking. ARXIV.org, Cornell University Library, May 3, 2018, 10 pages. |
Zaykovskiy, Dmitry. Survey of the Speech Recognition Techniques for Mobile Devices. Proceedings of Specom 2006, Jun. 25, 2006, 6 pages. |
Non-Final Office Action dated Oct. 28, 2021, issued in connection with U.S. Appl. No. 17/247,736, filed Dec. 21, 2020, 12 pages. |
Non-Final Office Action dated Nov. 29, 2021, issued in connection with U.S. Appl. No. 16/989,350, filed Aug. 10, 2020, 15 pages. |
Non-Final Office Action dated Jan. 4, 2022, issued in connection with U.S. Appl. No. 16/879,549, filed May 20, 2020, 14 pages. |
Non-Final Office Action dated Nov. 5, 2021, issued in connection with U.S. Appl. No. 16/153,530, filed Oct. 5, 2018, 21 pages. |
Non-Final Office Action dated Dec. 7, 2021, issued in connection with U.S. Appl. No. 16/168,389, filed Oct. 23, 2018, 36 pages. |
Non-Final Office Action dated Jan. 7, 2022, issued in connection with U.S. Appl. No. 17/135,123, filed Dec. 28, 2020, 16 pages. |
Non-Final Office Action dated Mar. 7, 2022, issued in connection with U.S. Appl. No. 16/812,758, filed Mar. 9, 2020, 18 pages. |
Non-Final Office Action dated Feb. 8, 2022, issued in connection with U.S. Appl. No. 16/806,747, filed Mar. 2, 2020, 17 pages. |
Notice of Allowance dated Aug. 10, 2021, issued in connection with U.S. Appl. No. 17/157,686, filed Jan. 25, 2021, 9 pages. |
Notice of Allowance dated Aug. 2, 2021, issued in connection with U.S. Appl. No. 16/660,197, filed Oct. 22, 2019, 7 pages. |
Notice of Allowance dated Aug. 4, 2021, issued in connection with U.S. Appl. No. 16/780,483, filed Feb. 3, 2020, 5 pages. |
Notice of Allowance dated Feb. 1, 2022, issued in connection with U.S. Appl. No. 16/439,046, filed Jun. 12, 2019, 9 pages. |
Notice of Allowance dated Mar. 1, 2022, issued in connection with U.S. Appl. No. 16/879,549, filed May 20, 2020, 9 pages. |
Notice of Allowance dated Sep. 1, 2021, issued in connection with U.S. Appl. No. 15/936,177, filed Mar. 26, 2018, 22 pages. |
Notice of Allowance dated Jun. 10, 2022, issued in connection with U.S. Appl. No. 16/879,549, filed May 20, 2020, 8 pages. |
Notice of Allowance dated May 11, 2022, issued in connection with U.S. Appl. No. 17/135,123, filed Dec. 28, 2020, 8 pages. |
Notice of Allowance dated May 11, 2022, issued in connection with U.S. Appl. No. 17/145,667, filed Jan. 11, 2021, 7 pages. |
Notice of Allowance dated Aug. 12, 2021, issued in connection with U.S. Appl. No. 16/819,755, filed Mar. 16, 2020, 6 pages. |
Notice of Allowance dated Jul. 12, 2022, issued in connection with U.S. Appl. No. 16/907,953, filed Jun. 22, 2020, 8 pages. |
Notice of Allowance dated Jul. 12, 2022, issued in connection with U.S. Appl. No. 17/391,404, filed Aug. 2, 2021, 13 pages. |
Notice of Allowance dated Apr. 13, 2022, issued in connection with U.S. Appl. No. 17/236,559, filed Apr. 21, 2021, 7 pages. |
Notice of Allowance dated Dec. 13, 2021, issued in connection with U.S. Appl. No. 16/879,553, filed May 20, 2020, 15 pages. |
Notice of Allowance dated Jan. 14, 2022, issued in connection with U.S. Appl. No. 16/966,397, filed Jul. 30, 2020, 5 pages. |
Notice of Allowance dated Aug. 15, 2022, issued in connection with U.S. Appl. No. 17/101,949, filed Nov. 23, 2020, 11 pages. |
Notice of Allowance dated Oct. 15, 2021, issued in connection with U.S. Appl. No. 16/213,570, filed Dec. 7, 2018, 8 pages. |
Notice of Allowance dated Sep. 15, 2021, issued in connection with U.S. Appl. No. 16/685,135, filed Nov. 15, 2019, 10 pages. |
Notice of Allowance dated Aug. 17, 2022, issued in connection with U.S. Appl. No. 17/135,347, filed Dec. 28, 2020, 14 pages. |
Notice of Allowance dated Jul. 18, 2022, issued in connection with U.S. Appl. No. 17/222,151, filed Apr. 5, 2021, 5 pages. |
Notice of Allowance Dec. 2, 2021, issued in connection with U.S. Appl. No. 16/841,116, filed Apr. 6, 2020, 5 pages. |
Notice of Allowance dated Jun. 20, 2022, issued in connection with U.S. Appl. No. 16/947,895, filed Aug. 24, 2020, 7 pages. |
Notice of Allowance dated Oct. 20, 2021, issued in connection with U.S. Appl. No. 16/439,032, filed Jun. 12, 2019, 8 pages. |
Notice of Allowance dated Dec. 21, 2021, issued in connection with U.S. Appl. No. 16/271,550, filed Feb. 8, 2019, 11 pages. |
Notice of Allowance dated Nov. 22, 2021, issued in connection with U.S. Appl. No. 16/834,483, filed Mar. 30, 2020, 10 pages. |
Notice of Allowance dated Aug. 23, 2021, issued in connection with U.S. Appl. No. 16/109,375, filed Aug. 22, 2018, 10 pages. |
Notice of Allowance dated Mar. 24, 2022, issued in connection with U.S. Appl. No. 16/378,516, filed Apr. 8, 2019, 7 pages. |
Notice of Allowance dated Oct. 25, 2021, issued in connection with U.S. Appl. No. 16/723,909, filed Dec. 20, 2019, 11 pages. |
Notice of Allowance dated Jun. 27, 2022, issued in connection with U.S. Appl. No. 16/812,758, filed Mar. 9, 2020, 16 pages. |
Notice of Allowance dated Jul. 29, 2022, issued in connection with U.S. Appl. No. 17/236,559, filed Apr. 21, 2021, 6 pages. |
Notice of Allowance dated Sep. 29, 2021, issued in connection with U.S. Appl. No. 16/876,493, filed May 18, 2020, 5 pages. |
Notice of Allowance dated Mar. 3, 2022, issued in connection with U.S. Appl. No. 16/679,538, filed Nov. 11, 2019, 7 pages. |
Notice of Allowance dated Apr. 8, 2022, issued in connection with U.S. Appl. No. 16/813,643, filed Mar. 9, 2020, 7 pages. |
Notice of Allowance dated Nov. 8, 2021, issued in connection with U.S. Appl. No. 17/008,104, filed Aug. 31, 2020, 9 pages. |
Notice of Allowance dated Dec. 9, 2021, issued in connection with U.S. Appl. No. 16/845,946, filed Apr. 10, 2020, 10 pages. |
Notice of Allowance dated Feb. 9, 2022, issued in connection with U.S. Appl. No. 17/247,736, filed Dec. 21, 2020, 8 pages. |
Oord et al. WaveNet: A Generative Model for Raw Audio. Arxiv.org, Cornell University Library, Sep. 12, 2016, 15 pages. |
Parada et al. Contextual Information Improves OOV Detection in Speech. Proceedings of the 2010 Annual Conference of the North American Chapter of the Association for Computational Linguistics, Jun. 2, 2010, 9 pages. |
Pre-Appeal Brief Decision dated Jan. 18, 2022, issued in connection with U.S. Appl. No. 16/806,747, filed Mar. 2, 2020, 2 pages. |
Renato De Mori. Spoken Language Understanding: A Survey. Automatic Speech Recognition & Understanding, 2007. IEEE, Dec. 1, 2007, 56 pages. |
Rybakov et al. Streaming keyword spotting on mobile devices, arXiv:2005.06720v2, Jul. 29, 2020, 5 pages. |
Shan et al. Attention-based End-to-End Models for Small-Footprint Keyword Spotting, arXiv:1803.10916v1, Mar. 29, 2018, 5 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Aug. 10, 2021, issued in connection with International Application No. PCT/US2020/017150, filed on Feb. 7, 2020, 20 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Mar. 10, 2020, issued in connection with International Application No. PCT/US2018/050050, filed on Sep. 7, 2018, 7 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Jan. 15, 2019, issued in connection with International Application No. PCT/US2017/042170, filed on Jul. 14, 2017, 7 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Jan. 15, 2019, issued in connection with International Application No. PCT/US2017/042227, filed on Jul. 14, 2017, 7 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Aug. 27, 2019, issued in connection with International Application No. PCT/US2018/019010, filed on Feb. 21, 2018, 9 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Mar. 31, 2020, issued in connection with International Application No. PCT/US2018/053517, filed on Sep. 28, 2018, 10 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Feb. 5, 2019, issued in connection with International Application No. PCT/US2017/045521, filed on Aug. 4, 2017, 7 pages. |
International Bureau, International Preliminary Report on Patentability and Written Opinion, dated Feb. 5, 2019, issued in connection with International Application No. PCT/US2017/045551, filed on Aug. 4, 2017, 9 pages. |
International Bureau, International Preliminary Report on Patentability, dated Jul. 21, 2022, issued in connection with International Application No. PCT/US2021/070007, filed on Jan. 6, 2021, 8 pages. |
International Bureau, International Preliminary Report on Patentability, dated Apr. 26, 2022, issued in connection with International Application No. PCT/US2020/056632, filed on Oct. 21, 2020, 7 pages. |
International Bureau, International Search Report and Written Opinion dated Nov. 14, 2017, issued in connection with International Application No. PCT/US2017/045521, filed on Aug. 4, 2017, 10 pages. |
International Bureau, International Search Report and Written Opinion dated Mar. 2, 2020, issued in connection with International Application No. PCT/US2019/064907, filed on Dec. 6, 2019, 9 pages. |
International Bureau, International Search Report and Written Opinion dated Oct. 22, 2020, issued in connection with International Application No. PCT/US2020/044282, filed on Jul. 30, 2020, 15 pages. |
International Bureau, International Search Report and Written Opinion dated Jul. 24, 2018, issued in connection with International Application No. PCT/US2018/019010, filed on Feb. 21, 2018, 12 pages. |
International Bureau, International Search Report and Written Opinion, dated Feb. 27, 2019, issued in connection with International Application No. PCT/US2018/053123, filed on Sep. 27, 2018, 16 pages. |
International Bureau, International Search Report and Written Opinion dated Aug. 6, 2020, issued in connection with International Application No. PCT/FR2019/000081, filed on May 24, 2019, 12 pages. |
International Bureau, International Search Report and Written Opinion dated Dec. 6, 2018, issued in connection with International Application No. PCT/US2018/050050, filed on Sep. 7, 2018, 9 pages. |
International Bureau, International Search Report and Written Opinion dated Oct. 6, 2017, issued in connection with International Application No. PCT/US2017/045551, filed on Aug. 4, 2017, 12 pages. |
International Searching Authority, International Search Report and Written Opinion dated Feb. 8, 2021, issued in connection with International Application No. PCT/EP2020/082243, filed on Nov. 16, 2020, 10 pages. |
International Searching Authority, International Search Report and Written Opinion dated Feb. 12, 2021, issued in connection with International Application No. PCT/US2020/056632, filed on Oct. 21, 2020, 10 pages. |
International Searching Authority, International Search Report and Written Opinion dated Apr. 23, 2021, issued in connection with International Application No. PCT/US2020/066231, filed on Dec. 18, 2020, 9 pages. |
Japanese Patent Office, Decision of Refusal and Translation dated Jul. 26, 2022, issued in connection with Japanese Patent Application No. 2020-513852, 10 pages. |
Japanese Patent Office, Notice of Reasons for Refusal and Translation dated Nov. 28, 2021, issued in connection with Japanese Patent Application No. 2020-550102, 9 pages. |
Japanese Patent Office, Office Action dated Dec. 7, 2021, issued in connection with Japanese Patent Application No. 2020-513852, 6 pages. |
Ketabdar et al. Detection of Out-of-Vocabulary Words in Posterior Based ASR. Proceedings of Interspeech 2007, Aug. 27, 2007, 4 pages. |
Kim et al. Character-Aware Neural Language Models. Retrieved from the Internet: URL: https://arxiv.org/pdf/1508.06615v3.pdf, Oct. 16, 2015, 9 pages. |
Korean Patent Office, Korean Examination Report and Translation dated Apr. 19, 2022, issued in connection with Korean Application No. 10-2021-7008937, 14 pages. |
Korean Patent Office, Korean Examination Report and Translation dated Nov. 25, 2021, issued in connection with Korean Application No. 10-2021-7008937, 14 pages. |
Korean Patent Office, Korean Examination Report and Translation dated Jul. 26, 2022, issued in connection with Korean Application No. 10-2022-7016656, 17 pages. |
Korean Patent Office, Korean Examination Report and Translation dated Dec. 27, 2021, issued in connection with Korean Application No. 10-2021-7008937, 22 pages. |
Korean Patent Office, Korean Office Action and Translation dated Oct. 14, 2021, issued in connection with Korean Application No. 10-2020-7011843, 29 pages. |
Lei et al. Accurate and Compact Large Vocabulary Speech Recognition on Mobile Devices. Interspeech 2013, Aug. 25, 2013, 4 pages. |
Lengerich et al. An End-to-End Architecture for Keyword Spotting and Voice Activity Detection, arXiv:1611.09405v1, Nov. 28, 2016, 5 pages. |
Matrix—The Ultimate Development Board Sep. 14, 2019 Matrix—The Ultimate Development Board Sep. 14, 2019 https-//web.archive.org/web/20190914035838/https-//www.matrix.one/ , 1 page. |
Mesaros et al. Detection and Classification of Acoustic Scenes and Events: Outcome of the DCASE 2016 Challenge. IEEE/ACM Transactions on Audio, Speech, and Language Processing. Feb. 2018, 16 pages. |
Molina et al., “Maximum Entropy-Based Reinforcement Learning Using a Confidence Measure in Speech Recognition for Telephone Speech,” in IEEE Transactions on Audio, Speech, and Language Processing, vol. 18, No. 5, pp. 1041-1052, Jul. 2010, doi: 10.1109/TASL.2009.2032618. [Retrieved online] URLhttps://ieeexplore.ieee.org/document/5247099?partnum=5247099&searchProductType=IEEE%20Journals%20Transactions. |
Non-Final Office Action dated Jul. 22, 2021, issued in connection with U.S. Appl. No. 16/179,779, filed Nov. 2, 2018, 19 pages. |
Non-Final Office Action dated Aug. 11, 2021, issued in connection with U.S. Appl. No. 16/841,116, filed Apr. 6, 2020, 9 pages. |
Non-Final Office Action dated Feb. 11, 2022, issued in connection with U.S. Appl. No. 17/145,667, filed Jan. 11, 2021, 9 pages. |
Non-Final Office Action dated Oct. 13, 2021, issued in connection with U.S. Appl. No. 16/679,538, filed Nov. 11, 2019, 8 pages. |
Non-Final Office Action dated Aug. 15, 2022, issued in connection with U.S. Appl. No. 17/448,015, filed Sep. 17, 2021, 12 pages. |
Non-Final Office Action dated Sep. 16, 2021, issued in connection with U.S. Appl. No. 16/879,553, filed May 20, 2020, 24 pages. |
Non-Final Office Action dated Aug. 17, 2021, issued in connection with U.S. Appl. No. 17/236,559, filed Apr. 21, 2021, 10 pages. |
Non-Final Office Action dated Aug. 18, 2021, issued in connection with U.S. Appl. No. 16/845,946, filed Apr. 10, 2020, 14 pages. |
Non-Final Office Action dated Sep. 2, 2021, issued in connection with U.S. Appl. No. 16/947,895, filed Aug. 24, 2020, 16 pages. |
Non-Final Office Action dated Mar. 23, 2022, issued in connection with U.S. Appl. No. 16/907,953, filed Jun. 22, 2020, 7 pages. |
Non-Final Office Action dated May 24, 2022, issued in connection with U.S. Appl. No. 17/101,949, filed Nov. 23, 2020, 10 pages. |
Non-Final Office Action dated Oct. 26, 2021, issued in connection with U.S. Appl. No. 16/736,725, filed Jan. 7, 2020, 12 pages. |
Non-Final Office Action dated Mar. 28, 2022, issued in connection with U.S. Appl. No. 17/222,151, filed Apr. 5, 2021, 5 pages. |
Non-Final Office Action dated Oct. 28, 2021, issued in connection with U.S. Appl. No. 16/378,516, filed Apr. 8, 2019, 10 pages. |
Number | Date | Country | |
---|---|---|---|
20210026595 A1 | Jan 2021 | US |
Number | Date | Country | |
---|---|---|---|
62298433 | Feb 2016 | US | |
62298439 | Feb 2016 | US | |
62298425 | Feb 2016 | US | |
62298350 | Feb 2016 | US | |
62298388 | Feb 2016 | US | |
62298393 | Feb 2016 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15973413 | May 2018 | US |
Child | 16989805 | US | |
Parent | 15131776 | Apr 2016 | US |
Child | 15973413 | US |