Playback device pairing

Information

  • Patent Grant
  • 10555082
  • Patent Number
    10,555,082
  • Date Filed
    Tuesday, September 11, 2018
    6 years ago
  • Date Issued
    Tuesday, February 4, 2020
    4 years ago
Abstract
Technology for grouping, consolidating, and pairing individual playback devices with network capability (players) to stimulate a multi-channel listening environment is disclosed. An example method includes receiving an audio signal containing a range of audio frequencies; amplifying, according to a gain parameter, the audio signal to be reproduced by at least one speaker. The example method includes automatically increasing the gain parameter to a higher gain parameter responsive to a determination that no more than a subset of the range of audio frequencies is to be reproduced by the at least one speaker. The example method includes amplifying, according to the higher gain parameter, the audio signal containing no more than the subset of the range of audio frequencies to be reproduced by the at least one speaker.
Description
BACKGROUND

The presently described technology is directed towards technology for use in the area of consumer electronics. In particular, certain embodiments are directed to multi-channel pairing in a media system.


Music is very much a part of our everyday lives. And thanks to the advancement of technology, music content is now more accessible than ever. The same can be said of other types of media, such as television, movies, and other audio and video content. In fact, now a user can even access the content over the Internet through an online store, an Internet radio station, online music service, online movie service, and the like, in addition to the more traditional means of accessing audio and video content.


The demand for such audio and video content continues to surge. Given the high demand over the years, technology used to access and play such content has likewise improved. Even still, technology used in accessing the content and the playback of such content can be significantly improved or developed in ways that the market or end users may not anticipate.


SUMMARY

The embodiments described herein include, but are not limited to, various devices, systems, methods, and computer program products. This section is for the purpose of summarizing some aspects of certain embodiments. Simplifications or omissions in this section as well as in the abstract or the title of this description may be made to avoid obscuring the purpose of this section, the abstract and the title. Such simplifications or omissions are not intended to limit the scope of the various inventions described herein.


In brief summary, the embodiments described herein provide technology for grouping, consolidating, and pairing individual playback devices to create or enhance a multi-channel listening environment. Particularly, the embodiments described herein enable two or more playback devices to be paired, such that multi-channel audio is achieved or enhanced. Such embodiments may be used to produce stereo sound or other audio environments suitable for audio content encoded with more than two channels, such as for certain kinds of television, movies, and music.


For example, an apparatus according to an embodiment comprises a network interface, a plurality of speaker drivers, an amplifier, and a processor. The network interface receives audio data over a network. The amplifier powers the plurality of speaker drivers. The processor processes the audio data to be output through the plurality of speaker drivers. The processor further configures a first equalization of the output from the plurality of speaker drivers in accordance with a first type of pairing and configuring a second equalization of the output from the plurality of speaker drivers in accordance with a second type of pairing.


In another example, a method according to an embodiment comprises receiving audio data over a network and processing the audio data to be output through a plurality of speaker drivers. The method further includes configuring a first equalization of the output from the plurality of speaker drivers in accordance with a first type of pairing and configuring a second equalization of the output from the plurality of speaker drivers in accordance with a second type of pairing.


One of the objects, features, and advantages of the present invention is to achieve or enhance a multi-channel listening environment. Many other objects, features, and advantages of the present invention will become apparent upon examining the following detailed description of an embodiment thereof, taken in conjunction with the attached drawings.





BRIEF DESCRIPTION OF THE DRAWINGS

These and other features, aspects, and advantages of the presently described technology will become better understood by a person skilled in the art with regard to the following description, appended claims, and accompanying drawings where:



FIG. 1 shows an illustrative configuration in which certain embodiments may be practiced;



FIG. 2A shows an illustrative functional block diagram of a player in accordance with certain embodiments;



FIG. 2B shows an example of a controller that may be used to remotely control one or more players of FIG. 2A;



FIG. 2C shows an example of a controller that may be used to remotely control one or more players of FIG. 2A;



FIG. 2D shows an example internal functional block diagram of a controller in accordance with certain embodiments;



FIG. 3A provides an illustration of a zone scene configuration;



FIG. 3B shows that a user defines multiple groups to be gathered at the same time;



FIG. 4 shows an example user interface that may be displayed on a controller or a computer of FIG. 1;



FIG. 5A shows an example user interface to allow a user to form a scene;



FIG. 5B shows another example user interface to allow a user to form a scene;



FIG. 5C shows an example user interface to allow a user to adjust a volume level of the zone players in a zone scene individually or collectively;



FIG. 6 shows a flowchart or process of providing a player theme or a zone scene for a plurality of players, where one or more of the players are placed in a zone;



FIG. 7 shows an illustrative configuration in which an audio source is played back on two players in accordance to an embodiment;



FIG. 8 shows an illustrative configuration of a pairing amongst multiple players in accordance to an embodiment;



FIG. 9 shows a flowchart or process of grouping a plurality of audio products to play separated sound tracks in synchronization to simulate a multi-channel listening environment; and



FIGS. 10A to 10F show example snapshots of a controller used in certain embodiments.





In addition, the drawings are for the purpose of illustrating certain embodiments, but it is understood that the inventions are not limited to the arrangements and instrumentality shown in the drawings.


DETAILED DESCRIPTION
I. Overview

The embodiments described herein relate to multi-channel pairing in a media system. The embodiments are particularly useful in pairing two or more playback devices together to create or enhance multi-channel audio reproduction, like stereo, surround sound, or some other multi-channel environment. The embodiments will also find utility in connection with any system for which multi-channel pairing is desired.


In an embodiment, two playback devices that are each configured to output a plurality of audio channels independent of each other are selectively paired, such that subsequent to pairing, one playback device is configured to output a first subset of the plurality of audio channels and the other playback device is configured to output a second subset of the plurality of audio channels. The first and second subsets are different. For instance, each playback device is configured to operate in a two-channel mode or stereo mode prior to being paired (e.g., each playback device is configured to play both right and left channel audio). Subsequent to pairing, one playback device is reconfigured to output a first channel (e.g., right channel audio and not left channel audio) and the other playback device is reconfigured to output a second channel (e.g., left channel audio and not right channel audio), which is different from the first channel.


In another embodiment, a collection of three or more playback devices that are each configured to output a plurality of audio channels independent of another playback device in the collection are selectively paired, such that subsequent to pairing, each of the playback devices is configured to output a different audio channel from the collection. This embodiment is particularly useful in a television or movie theater type setting where a particular playback device of the multiple playback devices is configured to output in two-channel or stereo mode at one time (e.g., when playing a song), and subsequent to pairing, is configured to output as a front-right channel, a front-center channel, a front-left channel, a rear-right channel, a rear-left channel, a subwoofer channel, and so on (e.g., when watching television or a movie or listening to music that contains more than two-channels).


In another embodiment, one of the paired playback devices processes the data of an audio item to separate the data into channels, each of the channels representing a single-sound track, for example. The playback device sends the separated channel(s) to the other, respective playback device(s). The playback devices play their distinctive channels in synchrony, thus creating a multi-channel listening environment. Alternatively, each of the paired playback devices processes the data of an audio item, or a portion of the data, and plays only those one or more channels designated for the respective playback device.


In another embodiment, two or more playback devices may be grouped into a single or consolidated playback device and the consolidated playback device may be paired with one or more playback devices. For instance, two playback devices may be grouped into a first consolidated playback device and two additional playback devices may be grouped into a second consolidated playback device. The first and second consolidated playback devices may be paired, for example. In certain embodiments, each playback device of a consolidated playback device is put into consolidated mode, which may result in a changed equalization for one or more speaker drivers of any particular playback device. Further, one or more additional playback devices may be added to a consolidated playback device.


In certain embodiments, a playback device that is configured to output an audio channel is paired with one or more additional playback devices, such that the playback device is reconfigured to output a different audio channel. For instance, the playback device might be configured to output a right channel for stereo mode, but subsequent to being paired with one or more additional playback devices, might be reconfigured to output a rear, right channel for theater mode. The playback device may be paired to one or more other playback devices.


In certain embodiments, a playback device that is configured to output a plurality of audio channels is paired with one or more additional playback devices, such that the playback device is configured to output a subset of the plurality of audio channels relative to the one or more additional playback devices. For instance, the playback device might be configured to output in two-channel or stereo mode, but subsequent to being paired with one or more playback devices might be configured to output a right or left channel. The playback device may be paired to one or more other playback devices.


In certain embodiments, a playback device comprises a network interface, one or more speaker drivers, an amplifier, and a processor. The network interface receives audio data over a network. The amplifier powers the speaker drivers. The processor processes the audio data to be output through the speaker drivers. The processor further configures a first equalization of the output from the speaker drivers in accordance with a first type of pairing and configuring a second equalization of the output from the speaker drivers in accordance with a second type of pairing. The playback device may operate in any of: non paired mode, paired mode, consolidated mode, and grouped mode.


In certain embodiments, a controller is configured to, among other things, pair two or more playback devices to establish a multi-channel audio environment. That is, through the controller, a user can select which playback devices to pair. Once programmed, the playback devices may operate in paired mode until disengaged, for example. In some embodiments, the controller is wirelessly coupled to the one or more playback devices. In other embodiments, the controller is wired to the one or more playback devices.


According to certain embodiments, the action of pairing two or more playback devices is triggered based on a command from a user via a control interface (e.g., a manual command creates a pair) or responsive to an event (e.g., an automatic command creates a pair). Example events include a detection in the change of audio content (e.g., the audio content goes from having two-channel content to three or more channel content, and vice-versa), a detection of a certain time, a detection of a certain kind of entertainment (e.g., detecting that the user is watching television versus just listening to music), or any other event that is programmed to create a pair amongst playback devices. The event detection might occur by a controller, one of the playback devices, or some other device, for example.


According to certain embodiments, in an attempt to optimize the multi-channel pairing, the configuration of a playback device may include any of: changing the equalization of the playback device by changing the equalization of one or more specific speaker drivers and optimizing the synchronization between paired devices. Examples of changing the equalization are described more below.


These embodiments and many additional embodiments are described more below. Further, the detailed description is presented largely in terms of illustrative environments, systems, procedures, steps, logic blocks, processing, and other symbolic representations that directly or indirectly resemble the operations of data processing devices coupled to networks. These process descriptions and representations are typically used by those skilled in the art to most effectively convey the substance of their work to others skilled in the art. Numerous specific details are set forth in order to provide a thorough understanding of the present invention. However, it is understood to those skilled in the art that certain embodiments of the present invention may be practiced without certain, specific details. In other instances, well known methods, procedures, components, and circuitry have not been described in detail to avoid unnecessarily obscuring aspects of the embodiments.


Reference herein to “embodiment” means that a particular feature, structure, or characteristic described in connection with the embodiment can be included in at least one embodiment of the invention. The appearances of this phrase in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. The embodiments described herein, explicitly and implicitly understood by one skilled in the art, may be combined with other embodiments.


II. Example Environment

Referring now to the drawings, in which like numerals may refer to like parts throughout the several views. FIG. 1 shows an exemplary configuration 100 in which certain embodiments may be practiced. The configuration 100 may represent, but not be limited to, a part of a residential home, a business building, or a complex with multiple zones. There are a number of multimedia players of which three examples 102, 104 and 106 are shown as audio devices. Each of the audio devices may be installed or provided in one particular area or zone and hence referred to as a zone player herein. It is understood that a zone can comprise more than one zone player.


As used herein, unless explicitly stated otherwise, an audio source or audio sources are generally in digital format and can be transported or streamed over a data network. To facilitate the understanding of the example environment of FIG. 1, it is assumed that the configuration 100 represents a home. Though, it is understood that this technology is not limited to its place of application. Referring back to FIG. 1, the zone players 102 and 104 may be located in one or two of the bedrooms while the zone player 106 may be installed or positioned in a living room. All of the zone players 102, 104, and 106 are coupled directly or indirectly to a data network 108. In addition, a computing device 110 is shown to be coupled on the network 108. In reality, any other device such as a home gateway device, a storage device, or an MP3 player may be coupled to the network 108 as well.


The network 108 may be a wired network, a wireless network or a combination of both. In one example, all devices including the zone players 102, 104, and 106 are coupled to the network 108 by wireless means based on an industry standard such as IEEE 802.11. In yet another example, all devices including the zone players 102, 104, and 106 are part of a local area network that communicates with a wide area network (e.g., the Internet). In still another example, all devices including the zone players 102, 104 and 106 and a controller 142 forms an ad-hoc network and may be specifically named, e.g., a household identifier: Smith Family, to be differentiated from a similar neighboring setup with a household identifier, e.g., Kallai Family.


Many devices on the network 108 are configured to download and store audio sources. For example, the computing device 110 can download audio sources, such as music or audio associated with videos, from the Internet (e.g., the “cloud”) or some other source and store the downloaded audio sources locally for sharing with other devices on the Internet or the network 108. The computing device 110 or any of the zone players 102, 104, and 106 can also be configured to receive streaming audio. Shown as a stereo system, the device 112 is configured to receive an analog audio source (e.g., from broadcasting) or retrieve a digital audio source (e.g., from a compact disk). The analog audio sources can be converted to digital audio sources. In accordance with certain embodiments, the various audio sources may be shared among the devices on the network 108.


Two or more zone players (e.g., any two or more of the zone players 102, 104, and 106) may be grouped together to form a new zone group. Any combinations of zone players and an existing zone group may be grouped together. In one instance, a new zone group is formed by adding one zone player to another zone player or an existing zone group.


In certain embodiments, there are two or more zone players in one environment (e.g., a living room in a house). Instead of grouping these two zone players to play back the same audio source in synchrony, these two zone players may be configured to play two separate sounds in left and right channels. In other words, the stereo effects of a sound are reproduced or enhanced through these two zone players, one for the left sound and the other for the right sound. Likewise, for a 3-channel (or 2.1 sound effects) sound, three such zone players may be reconfigured as if there are three speakers: left and right speakers and a subwoofer to form a stereo sound. The details of the reconfiguring the zone players and operating these audio products are described more below. Similar configurations with multiple channels (greater than 3, such as 4, 5, 6, 7, 9 channels and so on) also apply. For example, configurations that use more than two channels may be useful in television and theater type settings, where video content such as in the form of television and movies is played together with audio content that contains more than two channels. Further, certain music might similarly be encoded with more than two channel sound.


In certain embodiments, two or more zone players may be consolidated to form a single, consolidated zone player. The consolidated zone player may further be paired with a single zone player or yet another consolidated zone player. A consolidated zone player may comprise one or more individual playback devices. Each playback device of a consolidated playback device is preferably set in a consolidated mode.


According to some embodiments, one can continue to do any of: group, consolidate, and pair until a desired configuration is complete. The actions of grouping, consolidation, and pairing are preferably performed through a control interface and not by physically connecting and re-connecting speaker wire, for example, to individual, discrete speakers to create different configurations. As such, certain embodiments described herein provide a more flexible and dynamic platform through which sound reproduction can be offered to the end-user.


It is understood that the technology described herein is not limited to its place of application. For example, it is understood that zones and zone players, and the embodiments described herein, may also be used in vehicles, on water craft, airplanes, amphitheaters, outdoors, along the streets in a village or city, and so on, in addition to homes, offices, gyms, schools, hospitals, hotels, movie theaters, malls, stores, casinos, museum, entertainment parks, or any other place where audio content is played. As such, it will be appreciated that the embodiments described herein may be used in connection with any system or application for which multi-channel pairing is desired.


III. Example Playback Devices

Referring now to FIG. 2A, there is shown an exemplary functional block diagram of a zone player 200 in accordance with an embodiment. The zone player 200 includes a network interface 202, a processor 204, a memory 206, an audio processing circuit 210, a module 212, optionally, an audio amplifier 214 that may be internal or external, and optionally, a speaker unit 218 connected to the audio amplifier 214. The network interface 202 facilitates a data flow between a data network (i.e., the data network 108 of FIG. 1) and the zone player 200 and typically executes a special set of rules (i.e., a protocol) to send data back and forth. One of the common protocols used in the Internet is TCP/IP (Transmission Control Protocol/Internet Protocol). In general, a network interface 202 manages the assembling of an audio source or file into smaller packets that are to be transmitted over the data network or reassembles received packets into the original source or file. In addition, the network interface 202 handles the address part of each packet so that it gets to the right destination or intercepts packets destined for the zone player 200. Accordingly, in certain embodiments, each of the packets includes an IP-based source address as well as an IP-based destination address.


The network interface 202 may include one or both of a wireless interface 216 and a wired interface 217. The wireless interface 216, also referred to as an RF interface, provides network interface functions by a wireless means for the zone player 200 to communicate with other devices in accordance with a communication protocol (such as the wireless standard IEEE 802.11a, 802.11b, 802.11g, 802.11n, or 802.15.1). The wired interface 217 provides network interface functions by a wired means (e.g., an Ethernet cable). In one embodiment, a zone player includes both of the interfaces 216 and 217, and other zone players include only a RF or wired interface. Thus these other zone players communicate with other devices on a network or retrieve audio sources via the zone player. The processor 204 is configured to control the operation of other parts in the zone player 200. The memory 206 may be loaded with one or more software modules that can be executed by the processor 204 to achieve desired tasks. According to one embodiment, a software module implementing an embodiment, such as described herein, is executed, the processor 204 operates in accordance with the software module in reference to a saved zone group configuration characterizing a zone group created by a user, the zone player 200 is caused to retrieve an audio source from another zone player or a device on the network and synchronize the players in the zone group to play back the audio source as desired. According to another embodiment, a software module implementing an embodiment described herein creates a pair between two or more zone players to create a desired multi-channel audio environment.


According to one embodiment, the memory 206 is used to save one or more saved zone configuration files that may be retrieved for modification at any time. Typically, a saved zone group configuration file is transmitted to a controller (e.g., the controlling device 140 or 142 of FIG. 1, a computer, a portable device, or a TV) when a user operates the controlling device. The zone group configuration provides an interactive user interface so that various manipulations or control of the zone players may be performed.


In certain embodiments, the audio processing circuit 210 resembles the circuitry in an audio playback device and includes one or more digital-to-analog converters (DAC), an audio preprocessing part, an audio enhancement part or a digital signal processor and others. In operation, when an audio source is retrieved via the network interface 202, the audio source is processed in the audio processing circuit 210 to produce analog audio signals. The processed analog audio signals are then provided to the audio amplifier 214 for playback on speakers. In addition, the audio processing circuit 210 may include necessary circuitry to process analog signals as inputs to produce digital signals for sharing with other devices on a network.


Depending on an exact implementation, the module 212 may be implemented as a combination of hardware and software. In one embodiment, the module 212 is used to save a scene. The audio amplifier 214 is typically an analog circuit that powers the provided analog audio signals to drive one or more speakers.


It is understood that zone player 200 is an example of a playback device. Examples of playback devices include those zone players that are commercially offered for sale by Sonos, Inc. of Santa Barbara, Calif. They currently include a ZonePlayer 90, ZonePlayer 120, and Sonos S5. The ZonePlayer 90 is an example zone player without a built-in amplifier, whereas the ZonePlayer 120 is an example zone player with a built-in amplifier. The S5 is an example zone player with a built-in amplifier and speakers. In particular, the S5 is a five-driver speaker system that includes two tweeters, two mid-range drivers, and one subwoofer. When playing audio content via the S5, the left audio data of a track is sent out of the left tweeter and left mid-range driver, the right audio data of a track is sent out of the right tweeter and the right mid-range driver, and mono bass is sent out of the subwoofer. Further, both mid-range drivers and both tweeters have the same equalization (or substantially the same equalization). That is, they are both sent the same frequencies, just from different channels of audio. While the S5 is an example of a zone player with speakers, it is understood that a zone player with speakers is not limited to one with a certain number of speakers (e.g., five speakers as in the S5), but rather can contain one or more speakers. Further, a zone player may be part of another device, which might even serve a primary purpose different than audio.


IV. Example Controller

Referring now to FIG. 2B, there is shown an exemplary controller 240, which may correspond to the controlling device 140 or 142 of FIG. 1. The controller 240 may be used to facilitate the control of multi-media applications, automation and others in a complex. In particular, the controller 240 is configured to facilitate a selection of a plurality of audio sources available on the network, controlling operations of one or more zone players (e.g., the zone player 200) through a RF interface corresponding to the wireless interface 216 of FIG. 2A. According to one embodiment, the wireless means is based on an industry standard (e.g., infrared, radio, wireless standard IEEE 802.11a, 802.11b 802.11g, 802.11n, or 802.15.1). When a particular audio source is being played in the zone player 200, a picture, if there is any, associated with the audio source may be transmitted from the zone player 200 to the controller 240 for display. In one embodiment, the controller 240 is used to synchronize audio playback of more than one zone player by grouping the zone players in a group. In another embodiment, the controller 240 is used to control the volume of each of the zone players in a zone group individually or together.


In an embodiment, the controller 240 is used to create a pairing between two or more playback devices to create or enhance a multi-channel listening environment. For example, the controller 240 may be used to select and pair two or more playback devices. In addition, the controller 240 may be used to turn pairing on or off. The controller 240 may also be used to consolidate playback devices, and further to set a particular playback device in consolidated mode. Accordingly, in some embodiments, the controller 240 provides a flexible mechanism for dynamically configuring a multi-channel audio environment. In some instances, the pairing creates a multi-channel listening environment. In some instances, the pairing enhances a multi-channel listening environment by increasing the separation between devices. For example, two individual playback devices, which are positioned at a distance from each other, may provide more channel separation to the listener than the audio coming from only a single device.


The user interface for the controller 240 includes a screen 242 (e.g., a LCD screen) and a set of functional buttons as follows: a “zones” button 244, a “back” button 246, a “music” button 248, a scroll wheel 250, “ok” button 252, a set of transport control buttons 254, a mute button 262, a volume up/down button 264, a set of soft buttons 266 corresponding to the labels 268 displayed on the screen 242.


The screen 242 displays various screen menus in response to a user's selection. In one embodiment, the “zones” button 244 activates a zone management screen or “Zone Menu”, which is described in more details below. The “back” button 246 may lead to different actions depending on the current screen. In one embodiment, the “back” button triggers the current screen display to go back to a previous one. In another embodiment, the ‘back” button negates the user's erroneous selection. The “music” button 248 activates a music menu, which allows the selection of an audio source (e.g., a song) to be added to a zone player's music queue for playback.


The scroll wheel 250 is used for selecting an item within a list, whenever a list is presented on the screen 242. When the items in the list are too many to be accommodated in one screen display, a scroll indicator such as a scroll bar or a scroll arrow is displayed beside the list. When the scroll indicator is displayed, a user may rotate the scroll wheel 250 to either choose a displayed item or display a hidden item in the list. The “OK” button 252 is used to confirm the user selection on the screen 242.


There are three transport buttons 254, which are used to control the effect of the currently playing song. For example, the functions of the transport buttons may include play/pause and forward/rewind a song, move forward to a next song track, or move backward to a previous track. According to one embodiment, pressing one of the volume control buttons such as the mute button 262 or the volume up/down button 264 activates a volume panel. In addition, there are three soft buttons 266 that can be activated in accordance with the labels 268 on the screen 242. It is understood that, in a multi-zone system, there may be multiple audio sources being played respectively in more than one zone players. The music transport functions described herein shall apply selectively to one of the sources when a corresponding one of the zone players or zone groups is selected.



FIG. 2C shows an exemplary controller 260 which may correspond to the controlling device 140 or 142 of FIG. 1. The controller 260 is provided with a touch screen that allows a user to interact with the controller, for example, to navigate a playlist of many items, to control operations of one or more players. In one embodiment as it will be further shown in FIGS. 10A to 10F, a user may interact with the controller to make a multi-channel audio environment, such as create a stereo pair for example, and may even be used to separate the multi-channel audio environment, such as disengage a stereo pair. It should be noted that other network-enabled portable devices such as an iPhone, iPad or any other smart phone or network-enabled device may be used as a controller to interact or control multiple zone players in an environment (e.g., a networked computer such as a PC or Mac may also be used as a controller). According to one embodiment, an application may be downloaded into a network enabled device. Such an application may implement most of the functions discussed above for the controller 240 using a navigating mechanism or touch screen in the device. Those skilled in the art will appreciate the flexibility of such an application and its ability to be ported to a new type of portable device given the detailed description herein.



FIG. 2D illustrates an internal functional block diagram of an exemplary controller 270, which may correspond to the controller 240 of FIG. 2B, a computing device, smart phone, or any other communicative device. The screen 272 on the controller 270 may be an LCD screen. The screen 272 communicates with and is commanded by a screen driver 274 that is controlled by a microcontroller (e.g., a processor) 276. The memory 282 may be loaded with one or more application modules 284 that can be executed by the microcontroller 276 with or without a user input via the user interface 278 to achieve desired tasks. In one embodiment, an application module is configured to facilitate grouping a number of selected zone players into a zone group and synchronizing the zone players for one audio source. In another embodiment, an application module is configured to control together the audio sounds (e.g., volume) of the zone players in a zone group. In operation, when the microcontroller 276 executes one or more of the application modules 284, the screen driver 274 generates control signals to drive the screen 272 to display an application specific user interface accordingly, more of which will be described below.


The controller 270 includes a network interface 280 referred to as a RF interface 280 that facilitates wireless communication with a zone player via a corresponding RF interface thereof. In one embodiment, the commands such as volume control and audio playback synchronization are sent via the RF interfaces. In another embodiment, a saved zone group configuration is transmitted between a zone player and a controller via the RF interfaces. The controller 270 may control one or more zone players, such as 102, 104 and 106 of FIG. 1. Nevertheless, there may be more than one controller, each preferably in a zone (e.g., a room or rooms nearby each other) and configured to control any one and all of the zone players.


In one embodiment, a user creates a zone group including at least two zone players from the controller 240 that sends signals or data to one of the zone players. As all the zone players are coupled on a network, the received signals in one zone player can cause other zone players in the group to be synchronized so that all the zone players in the group play back an identical audio source or a list of identical audio sources in a timely synchronized manner such that no (or substantially no) audible delays or hiccups could be heard. Similarly, when a user increases the audio volume of the group from the controller, the signals or data of increasing the audio volume for the group are sent to one of the zone players and causes other zone players in the group to be increased together in volume and in scale.


According to one implementation, an application module is loaded in memory 282 for zone group management. When a predetermined key (e.g. the “zones” button 244) is activated on the controller 240, the application module is executed in the microcontroller 276. The input interface 278 coupled to and controlled by the microcontroller 276 receives inputs from a user. A “Zone Menu” is then displayed on the screen 272. The user may start grouping zone players into a zone group by activating a “Link Zones” or “Add Zone” soft button, or de-grouping a zone group by activating an “Unlink Zones” or “Drop Zone” button. The detail of the zone group manipulation will be further discussed below.


As described above, the input interface 278 includes a number of function buttons as well as a screen graphical user interface. It should be pointed out that the controller 240 in FIG. 2B is not the only controlling device that may practice the embodiments. Other devices that provide the equivalent control functions (e.g., a computing device, a hand-held device) may also be configured to practice the present invention. In the above description, unless otherwise specifically described, it is clear that keys or buttons are generally referred to as either the physical buttons or soft buttons, enabling a user to enter a command or data.


One mechanism for ‘joining’ zone players together for music playback is to link a number of zone players together to form a group. To link a number of zone players together, a user may manually link each zone player or room one after the other. For example, there is a multi-zone system that includes the following zones:

    • Bathroom
    • Bedroom
    • Den
    • Dining Room
    • Family Room
    • Foyer


If a user wishes to link five of the six zone players using the current mechanism, the user may start with a single zone and then manually link each zone to that zone. This mechanism may be sometimes quite time consuming. According to one embodiment, a set of zones can be dynamically linked together using one command. Using what is referred to herein as a theme or a zone scene, zones can be configured in a particular scene (e.g., morning, afternoon, or garden), where a predefined zone grouping and setting of attributes for the grouping are automatically effectuated.


For instance, a “Morning” zone scene/configuration command would link the Bedroom, Den and Dining Room together in one action. Without this single command, the user would need to manually and individually link each zone. FIG. 3A provides an illustration of one zone scene, where the left column shows the starting zone grouping—all zones are separate, the column on the right shows the effects of grouping the zones to make a group of 3 zones named after “Morning”.


Expanding this idea further, a Zone Scene can be set to create multiple sets of linked zones. For example, a scene creates 3 separate groups of zones, the downstairs zones would be linked together, the upstairs zones would be linked together in their own group, and the outside zones (in this case the patio) would move into a group of its own.


In one embodiment as shown in FIG. 3B, a user defines multiple groups to be gathered at the same time. For example: an “Evening Scene” is desired to link the following zones:

    • Group 1
      • Bedroom
      • Den
      • Dining Room
    • Group 2
      • Garage
      • Garden


        where Bathroom, Family Room and Foyer should be separated from any group if they were part of a group before the Zone Scene was invoked.


A feature of certain embodiments is that that zones do not need to be separated before a zone scene is invoked. In one embodiment, a command is provided and links all zones in one step, if invoked. The command is in a form of a zone scene. After linking the appropriate zones, a zone scene command could apply the following attributes:

    • Set volumes levels in each zones (each zone can have a different volume)
    • Mute/Unmute zones.
    • Select and play specific music in the zones.
    • Set the play mode of the music (Shuffle, Repeat, Shuffle-repeat)
    • Set the music playback equalization of each zone (e.g., bass treble).


A further extension of this embodiment is to trigger a zone scene command as an alarm clock function. For instance the zone scene is set to apply at 8:00 am. It could link appropriate zones automatically, set specific music to play and then stop the music after a defined duration. Although a single zone may be assigned to an alarm, a scene set as an alarm clock provides a synchronized alarm, allowing any zones linked in the scene to play a predefined audio (e.g., a favorable song, a predefined playlist) at a specific time or for a specific duration. If, for any reason, the scheduled music failed to be played (e.g., an empty playlist, no connection to a share, failed UPnP, no Internet connection for an Internet Radio station), a backup buzzer will sound. This buzzer will be a sound file that is stored in a zone player.



FIG. 4 shows an exemplary user interface 400 that may be displayed on a controller 142 or a computer 110 of FIG. 1. The interface 400 shows a list of items that may be set up by a user to cause a scene to function at a specific time. In the embodiment shown in FIG. 4, the list of items includes “Alarm”, “Time”, “Zone”, “Music”, “Frequency” and “Alarm length”. “Alarm” can be set on or off. When “Alarm” is set on, “Time” is a specific time to set off the alarm. “Zone” shows which zone players are being set to play a specified audio at the specific time. “Music” shows what to be played when the specific time arrives. “Frequency” allows the user to define a frequency of the alarm. “Alarm length” defines how long the audio is to be played. It should be noted that the user interface 400 is provided herein to show some of the functions associated with setting up an alarm. Depending on an exact implementation, other functions, such as time zone, daylight savings, time synchronization, and time/date format for display may also be provided.


According to one embodiment, each zone player in a scene may be set up for different alarms. For example, a “Morning” scene includes three zone players, each in a bedroom, a den, and a dining room. After selecting the scene, the user may set up an alarm for the scene as whole. As a result, each of the zone players will be activated at a specific time.



FIG. 5A shows a user interface 500 to allow a user to form a scene. The panel on the left shows the available zones in a household. The panel on the right shows the zones that have been selected and be grouped as part of this scene. Depending on an exact implementation of a user interface, Add/Remove buttons may be provided to move zones between the panels, or zones may be dragged along between panels.



FIG. 5B shows another user interface 520 to allow a user to form a scene. The user interface 520 that may be displayed on a controller or a computing device, lists available zones in a system. A checkbox is provided next to each of the zones so that a user may check in the zones to be associated with the scene.



FIG. 5C shows a user interface 510 to allow a user to adjust a volume level of the zone players in a zone scene individually or collectively. As shown in the user interface 510, the ‘Volumes . . . ’ button (shown as sliders, other forms are possible) allows the user to affect the volumes of the associated zone players when a zone scene is invoked. In one embodiment, the zone players can be set to retain whatever volume that they currently have when the scene is invoked. Additionally the user can decide if the volumes should be unmuted or muted when the scene is invoked.


V. Providing Example Player Themes or Zone Scenes


FIG. 6 shows a flowchart or process 600 of providing a player theme or a zone scene for a plurality of players, where one or more of the players are placed in a zone. The process 600 is presented in accordance with one embodiment of the present invention and may be implemented in a module to be located in the memory 282 of FIG. 2C.


The process 600 is initiated when a user decides to proceed with a zone scene at 602. The process 600 then moves to 604 where it allows a user to decide which zone players to be associated with the scene. For example, there are ten players in a household, and the scene is named after “Morning”. The user may be given an interface to select four of the ten players to be associated with the scene. At 606, the scene is saved. The scene may be saved in any one of the members in the scene. In the example of FIG. 1, the scene is saved in one of the zone players and displayed on the controller 142. In operation, a set of data pertaining to the scene includes a plurality of parameters. In one embodiment, the parameters include, but may not be limited to, identifiers (e.g., IP address) of the associated players and a playlist. The parameters may also include volume/tone settings for the associated players in the scene. The user may go back to 602 to configure another scene if desired.


Given a saved scene, a user may activate the scene at any time or set up a timer to activate the scene at 610. The process 600 can continue when a saved scene is activated at 610. At 612, upon the activation of a saved scene, the process 600 checks the status of the players associated with the scene. The status of the players means that each of the players shall be in condition to react in a synchronized manner. In one embodiment, the interconnections of the players are checked to make sure that the players communicate among themselves and/or with a controller if there is such a controller in the scene.


It is assumed that all players associated with the scene are in good condition. At 614, commands are executed with the parameters (e.g., pertaining to a playlist and volumes). In one embodiment, data including the parameters is transported from a member (e.g., a controller) to other members in the scene so that the players are caused to synchronize an operation configured in the scene. The operation may cause all players to play back a song in identical or different volumes or to play back a pre-stored file.


VI. Example Multi-Channel Environments


FIG. 7 shows an example configuration in which an audio source is played back on two players 702 and 704, according to an example embodiment. These two players 702 and 704 may be located in and around one place (e.g., a hall, room, or nearby rooms) and are designated to play two sound tracks respectively. For example, an audio source may have left and right sound channels or tracks (e.g., stereo sound). Instead of grouping the players 702 and 704 to play back the audio source together in synchrony, where each player 702 and 704 plays the same audio content at substantially the same time, the players 702 and 704 can be paired to play different channels of the audio source in synchrony. As a result of pairing, the stereo sound effects can be simulated or enhanced via two players 702 and 704 versus one player or none of the players, for example.


In certain embodiments, each player of players 702 and 704 includes a network interface, one or more speaker drivers (two or more speaker drivers in some instances, such as when the player can play in stereo mode absent pairing), an amplifier, and a processor, such as shown in FIG. 2A. The network interface receives audio data over a network. One or more amplifiers power the speaker drivers. The processor processes the audio data to be output through the speaker drivers. The processor may further configure a first equalization of the output from the speaker drivers in accordance with a first type of pairing and configuring a second equalization of the output from the speaker drivers in accordance with a second type of pairing.


In an embodiment, the two players 702 and 704 are configured to output a plurality of audio channels independent of each other. For example, each player 702 and 704 may be configured to output audio content in stereo independently from each other. Subsequent to pairing, one playback device (e.g., player 702) is configured to output a first subset of the plurality of audio channels and the other playback device (e.g., player 704) is configured to output a second subset of the plurality of audio channels. The first and second subsets are different. In this example, subsequent to pairing players 702 and 704, player 702 might play the right channel and player 704 might play the left channel. In another example, player 702 might play the right channel plus a center channel (e.g., in television or theater mode) and player 704 might play the left channel plus the center channel. Even in the latter example, the first and second subsets are different in that player 702 is playing channels Right+Center and player 704 is playing channels Left+Center. In yet another embodiment, subsequent to pairing, player 702 might play all channels except certain bass frequencies, which may be played via player 704, thereby using player 704 as a subwoofer.


In another embodiment, a collection of three or more playback devices (e.g., players 702, 704, and one or more additional players) are each configured to output a plurality of audio channels independent of another playback device in the collection. Subsequent to pairing, each of the playback devices is configured to output a generally different audio channel(s) from the collection. This embodiment is particularly useful in a television or movie theater setting where a particular playback device of the multiple playback devices is configured to output in two-channel or stereo mode at one time (e.g., when playing a song), and subsequent to pairing, is configured to output as a front-right channel, a front-center channel, a front-left channel, a rear-right channel, a rear-left channel, and so on (e.g., when watching a movie or television).


In another embodiment, one of the paired playback devices (e.g., player 702 or player 704) processes the data of the audio item, essentially separating the data into channels, each of the channels representing a single-sound track, for example, and being played back in one of the playback devices, thus creating or enhancing a multi-channel listening environment. In an alternative embodiment, both playback devices (e.g., players 702 and 704) may receive and process the data of the audio item and each playback device may output only the audio content designated for the respective player. For example, player 702 might receive both left and right channel audio, but only play the left channel, whereas player 704 might also receive both left and right channel audio, but only play the right channel.


In another embodiment, two or more playback devices (e.g., players 702 or 704) may be grouped into a single or consolidated playback device and the consolidated playback device (e.g., consolidated player 702+704) may be paired with one or more playback devices. For instance, two playback devices maybe grouped into a first consolidated playback device and two additional playback devices maybe grouped into a second consolidated playback device. Then, the first and second consolidated playback devices may be paired to create or enhance a multi-channel listening environment.


In certain embodiments, a playback device (e.g., either player 702 or 704) that is configured to output an audio channel is paired with one or more additional playback devices, such that the playback device is configured to output a different audio channel than previously configured. For instance, the playback device might be configured to output a right channel in stereo mode, but subsequent to being paired with one or more additional playback devices, might be configured to output a rear, right channel in theater mode. The playback device may be paired to one or more other playback devices.


In certain embodiments, a playback device (e.g., either player 702 or 704) that is configured to output a plurality of audio channels is paired with one or more additional playback devices, such that the playback device is configured to output a subset of the plurality of audio channels relative to the one or more additional playback devices. For instance, the playback device might be configured to output in two-channel or stereo mode, but subsequent to being paired with one or more playback devices might be configured to output a right or left channel. The playback device may be paired to one or more other playback devices.


According to certain embodiments, the action of pairing two or more playback devices is triggered based on a command from a user via a control interface (e.g., a manual command) or responsive to an event (e.g., an automatic command). For example, using a controller, a user can create a pairing between two or more playback devices or disengage the pairing between two or more playback devices. In another example, pairing may be triggered by the audio content itself, a signal received from a source device, or some other predefined event, such that pairing occurs when the event is detected by the controller or playback device, for example. In addition, another device might be programmed to detect the event and provide a pairing signal to the controller and/or playback devices.


Further, it is understood that going from a configuration of no pairing (unpaired or non paired) to a configuration of pairing or from one kind of pairing (e.g., a pairing used in a type of stereo mode or theater mode) to a different kind of pairing (e.g., another pairing used in a type of stereo mode or theater mode) are all various types of “pairing” that can occur according to certain embodiments. In addition, disengaging a pairing between multiple playback devices might go from pairing to no pairing or from pairing of a first kind back to pairing of a previous kind, for example.


In one example, a first type of pairing might include “no pairing” with another playback device and a second type of pairing might include pairing with one or more additional playback devices. In a second example, a first type of pairing might include pairing with a second playback device and a second type of pairing might include pairing with a plurality of playback devices. In a third example, a first type of pairing might include reproducing two channel sound via the speaker drivers and a second type of pairing comprises reproducing no more than one channel of the two channel sound via the speaker drivers. In a fourth example, a first type of pairing might comprise reproducing a first audio channel via the speaker drivers and the second type of pairing might include reproducing a second audio channel via the speaker drivers. In a fifth example, a first type of pairing might include reproducing the audio content via the speaker drivers in stereo mode and a second type of pairing might include reproducing the audio content via the speaker drivers in theater mode. In a sixth example, a first type of pairing might include reproducing the audio content via the speaker drivers and a second type of pairing comprises reproducing the audio content via the speaker drivers when in consolidated mode. It is understood that various variations and modifications may be made to the examples described just above with the attainment of some or all of the advantages of the technology described herein.


According to certain embodiments, the configuration of a playback device may include any of: changing the equalization of the playback device by changing the equalization of one or more specific speaker drivers and optimizing the synchronization between paired devices. Changing the equalization of the playback device might include any of: turning on or off (or effectively muting) one or more specific speaker drivers, changing the channel output of one or more speaker drivers, changing the frequency response of one or more specific speaker drivers, changing the amplifier gain of any particular speaker driver, changing the amplifier gain of the playback device as a whole.


In certain embodiments, changing the equalization of a playback device (e.g., changing the equalization of one or more speaker drivers of the playback device) may affect frequency dependent parameters. Examples might include the adjustment of the strength of frequencies within the audio data, a phase adjustment, and time-delay adjustment. In addition, a particular equalization may use a first type of pass filter, such as one that attenuates high, middle, or low frequencies, for example, while allowing other frequencies to pass unfiltered (or substantially unfiltered). Filters might also be different kinds or of a different order (e.g., first order filter, second order filter, third order filter, fourth order filter, and so on). For example, a first equalization of a playback device might include using a first type of pass filter to modify the output based on a first type of pairing and a second equalization of the playback device might include using a second type of pass filter to modify the output based on the second type of pairing. In this example, the first and second type of pass filters have one or different properties and/or behaviors, thus changing the equalization and sonic behavior of the device.


By way of illustration, when two S5 devices are paired to create a stereo pair, for example, one S5 device may be configured as the “left” and the other S5 device may be configured as the “right.” In one embodiment, the user may determine which is left or right. In this configuration, for example, the left and right audio data may be sent to both S5 devices, but the left audio data of the track is played out of the S5 device configured as left and the right audio data of a track is played out of the S5 device configured as right. In addition, the equalization of each S5 device is changed in an attempt to reduce or eliminate certain constructive or destructive interference. For example, one tweeter on each S5 device may be turned off or substantially muted. In certain embodiments, the crossover frequency to each driver may even be changed from a previous configuration so that two or more drivers are not necessarily outputting the exact same audio data, otherwise constructive and/or destructive interference may occur. In certain embodiments, the amplifier gain is adjusted for a particular speaker driver and/or for the playback device as a whole.


In operation, according to certain embodiments, a controller 706 (e.g., a controller 142 of FIG. 1 or 240 of FIG. 2B or a portable device) is used to initiate the operation. Through a user interface, the controller 706 causes a player 702 to retrieve the audio source, provided the audio source is on a network 708 (e.g., the Internet or a local area network). Similarly, the controller 706 may also cause a designated device (e.g., another networked device) to establish a communication session with the player 702 to deliver the requested audio source. In any case, either one or both of the players 702 and 704 may have access to the data representing the audio source.


In certain embodiments, a module in the player 702 is activated to process the data. According to one embodiment, the right and left sound tracks are separated. One sound track is retained locally in one player and the other sound track is pushed or uploaded to the other device (e.g., via an ad-hoc network). When the right and left sound tracks are played back simultaneously or substantially simultaneously, the stereo sound effect can be appreciated.


In another embodiment, several tracks are separated, such as in television or theater mode. For example, the tracks may be separated into a center channel, right front channel, left front channel, right rear channel, left rear channel, and so on. Accordingly, one or more sound tracks may be retained locally in one player and the other sound tracks are pushed or uploaded to the other devices.


In yet another embodiment, one player might process the data and retain one or more tracks locally, while the remaining data is sent onto another player. The receiving player may then process the data and retain one or more tracks locally and send any remaining data onto another player. This process, or one like it, may continue until all of the tracks are retained locally by corresponding player devices.


In yet another embodiment, each player might receive and process the data and play only the channel or channels that are designated for that player.


In certain embodiments, it is important to maintain good synchronization, especially when pairing two or more independently clocked playback devices so that the multi-channel audio content is played back as it was originally intended. According to an embodiment, a message may be initiated from one device to another that is also activated to send back an acknowledgement. Upon receiving the acknowledgement, the time delay in transporting data from one device to another can be measured. The time delay will be considered when synchronizing the two players to play back the two separated sound tracks. In certain embodiments, if sending a packet (e.g., a packet in accordance with SNTP protocol) to a playback device and receiving a response takes more than fifteen milliseconds, for example, the timing information contained within that packet, such as clock information, is discarded. If sending and receiving a packet is less than fifteen milliseconds, then the information from the packet is used to adjust playback, if so necessary.


Additional details of synchronizing operations of two or more independently clocked players are provided in commonly assigned U.S. application Ser. No. 10/816,217, filed Apr. 1, 2004, entitled “System and Method For Synchronizing Operations Among A Plurality Of Independently Clocked Digital Data Processing Devices” which is hereby incorporated by reference.



FIG. 8 shows an example configuration of a pairing amongst multiple players 802, 804, 806, 808, 810, and 812 in a theater-like environment, in accordance to an embodiment. Player 802 may operate as a front-left channel, player 804 may operate as a center channel, player 806 may operate as a front-right channel, player 808 may operate as a subwoofer, player 810 may operate as a rear, right channel, and player 812 may operate as a rear, right channel. In this example, the players 802, 804, 806, 808, 810, and 812 are wirelessly coupled over network 815 so as to receive and transmit data over a wireless network, and obtain power from power outlets in the wall or through some other power source (e.g., a battery). Players 802, 804, 806, 808, 810, and 812 may be wired, if so configured in an alternate embodiment. Controller 814 may be a network-enabled device, examples of which include a smart phone, tablet computer, laptop computer, desktop computer, or a television.


In one embodiment, a designated player, such as player 804, receives multi-channel audio content from a source 816. Source 816 might include audio and/or video content downloaded or streamed from the Internet, a DVD or Blu-Ray player, or from some other source of audio and/or video content. Player 804 separates the multi-channel audio and sends respective audio channels to its playback owner. For example, if a particular audio channel is designated for the front, right speaker, then that content is wirelessly directed from player 804 to player 802, and so on. Players 802, 804, 806, 808, 810, and 812 play the audio content synchronously, so as to create a multi-channel listening environment. Moreover, if source 816 provides video content along with audio content, then the audio content is preferably played in synchrony with the video content.


In another embodiment, each player of players 802, 804, 806, 808, 810, and 812 may separate out its own one or more channels for playback. That is, either all audio content, or a portion thereof, is sent to each player (e.g., from source 816 or another playback device) and the player itself obtains its own data for playback.


In addition, players 802, 804, 806, 808, 810, and 812 may be reconfigured to operate in many different configurations, such as described above. For example, players 802 and 806 may be paired to operate in stereo mode, while the other players remain in sleep mode or turned off (player 808 may remain on in any particular configuration, if so desired and configured, because it is operating as a subwoofer). In another example, players 802 and 810 may be consolidated and output left channel audio, while players 806 and 812 may be consolidated and output right channel audio. In yet another example, some of players 802, 804, 806, 808, 810, and 812 are consolidated into a single player and paired with additional playback devices, such as in an adjacent room. In a further example, players 802, 804, 806, 808, 810, and 812 are grouped and not paired, when the audio content is music (versus movie content, for example). These are just some configuration examples. Many other configurations are possible using the teachings described herein.



FIG. 9 shows a flowchart or process 900 of grouping a plurality of audio products to play separated sound tracks in synchronization to simulate a multi-channel listening environment. The process 900 is presented in accordance with certain embodiments and may be implemented in a module to be located in the memory 282 of FIG. 2D. To facilitate the description of process 900, a listening environment of stereo sound with left and right channels is described. Those skilled in the art can appreciate that the description can be equally applied to other forms of multi-channel listening environment (e.g., three, five, seven channel environments).


Typically, there is a plurality of players being controlled by one or more controllers, where these players are disposed in various locations. For example, there are five players in a house; three of them are respectively disposed in three rooms while two players are disposed in a larger room. Accordingly, these two players would be candidates to be paired to simulate a stereo listening environment, instead of just playing synchronized audio from both in a grouped fashion. In another example, there are four players in a large space or adjacent spaces, two pairs of the players may be paired to simulate a stereo listening environment, in which two players in one consolidated pair can be grouped to play back one (left) sound track and the other two in the other consolidated pair can be grouped to play back one (right) sound track.


In any case, two groups of players or two players are decided to be paired at 902. If no players are paired, the process 900 will not be activated. It is assumed that two players from a group of players being controlled by a controller are selected to be paired at 902. The process 900 proceeds.


At 904, a user may decide which player is to play back which sound track. Depending on the location of the user or listener(s) with respect to the selected players, it is assumed that a player or unit A is chosen to play back a left sound track and another player or unit B is chosen to play back a right sound track. In an alternative embodiment, the players themselves (or the controller) may automatically determine which unit is configured to play the right channel and which unit is configured to play the left channel without input from the user.


According to one embodiment, a time delay in transporting data between the two units A and B is measured at 906. This time delay may facilitate sound synchronization between the two units as one of the units will receive a processed sound track from the other. The user may continue to operate on a controller to select a title (e.g., an audio source or an item from a playlist) for playback on the two units at 910.


Once the title is determined at 912, the data for the title is accessed. Depending on where the data is located, the controller may be configured to cause one of the two units to obtain or stream in the data. In one embodiment, the controller or unit A initiates a request to a remotely-networked device providing or storing the data. Assuming an authentication procedure, if any, completes successfully, the remote device starts to upload the data to the unit A. Likewise, if the data is locally stored in the unit A, the data can be accessed locally without requesting the same from the network. As the data is being received or accessed in the unit A, a processing module is activated in the unit A to process the data, essentially separating the data into two streams of sound tracks at 914. In an alternative embodiment, each unit may receive and process the data, essentially separating the data into a stream to be played by the respective unit.


At 916, one of the streams is uploaded from the unit A to unit B via a local network (e.g., the ad-hoc network formed by all the players being controlled by the controller). As the streams are being distributed, the two units are configured to play back the streams respectively, each reproducing the sound of a single sound track at 918. Together, in synchrony, the two units create a stereo sound listening environment.


It should be noted that the delay time, if noticeable, may be incorporated into the unit A to delay the consumption of the stream by the delay time to synchronize with the unit B. Alternatively, a non-selected player may be used to process a streaming data of the title and configured to supply two streams to the pair of players, thus equalizing the delay time that would be otherwise experienced by the unit B.



FIGS. 10A-10F show illustrative screenshots of a controller for creating a stereo pair in accordance with certain embodiments. The screenshots are from a computing device (e.g., a tablet computer, laptop, or desktop) used as a controller. Those skilled in the art can appreciate that FIGS. 10A-10F may be readily modified to be used in a portable device with network capability, such as, for example, iPhone or iTouch or other smart phone or other network-enabled devices. Additionally, the controller might exist as part of a player itself or directly/indirectly coupled to the player, and therefore such screenshots may be modified accordingly—such a controller need not have network capability as the player will have network connectivity.



FIG. 10A shows a graphic interface 1000 that may be displayed on a controller when a user desires to create a stereo pair with two players in a system. It is understood that the system may include two or more players. If a stereo pair is desired, such as discussed with respect to the example of FIGS. 10A-10F, then any two players (one or both of which may be a consolidated player) in the system may be paired. However, if pairing more than two players is desired, such as creating an environment which is capable of playing more than two channel audio data, then the graphic interface 1000 may include an additional option or options. For example, an option might include “Make a Movie Surround Sound Pairing,” “Make a Music Surround Sound Pairing,” or “Make a Dolby Pro Logic Pairing.” Any descriptive language may be used to appropriately indicate to the user the type of pairing that can be created. Upon selecting an option, a setup wizard on the controller may help the user appropriately configure the system such that multi-channel discrete audio may be effectively realized by the system.


Turning back to FIG. 10A, the interface 1000 allows a user to initiate a stereo pair with a zone player named “ZPSS-Black.” In certain embodiments, the system recognizes that ZPSS-Black is part of a particular zone (e.g., kitchen, family room, bedroom, and so on). The system may allow the user to pair ZPSS-Black with another player in the same zone only, or alternatively, the system may allow the user to pair ZPSS-Black with another player in a different zone (such as an adjacent zone). Pairing players in different zones may be particularly useful when an open space is divided into two or more zones (e.g., an open space might include a kitchen and family room, for example).


Additionally, the system may be programmed such that pairing players from different zones creates another zone to reflect the players in paired mode (e.g., a single kitchen-family room zone during paired operation might originate from a kitchen zone and a family room zone during non-paired operation). In such an embodiment, a user may be able to switch between zones or dynamically create new zones.


In certain embodiments, if another similar player is available to be paired, then the screenshot of FIG. 10B may be displayed. If the user wishes to continue with creating a pair, then the user may select “OK.” If not, then the user may select “Cancel.” In another embodiment, a different player (e.g., a player that is not an S5) may be paired together. That is, different types of players may be paired, if the players are so designed to be paired. To accommodate the differences in player type, the equalization of one or more players may be adjusted accordingly to compensate for things like the number and size of speaker drivers used in one player versus the other player. In yet another embodiment, a list of the players in the system may be displayed (not shown), from which the user selects two or more players to make the stereo pair. The list of players may be automatically determined by the system based on a player's particular location within a home, room, or configuration with other players within a room, for example.


Turning now to FIG. 10C, in this example, it is assumed that the user may select a zone player named “ZPSS-White” to be paired with “ZPSS-Black” to create a stereo pair. If so desired, the user may select “OK” to proceed with the pairing. Otherwise, the user may select “Cancel.” In certain embodiments, ZPSS-White may be in the same zone as ZPSS-Black. In other embodiments, ZPSS-White may be in a different zone as ZPSS-Black.


Upon selecting “OK” in FIG. 10C, a screenshot like that of FIG. 10D may be displayed to the user, thereby instructing the user to press the mute button (or some other designated button) on the “LEFT” player of the stereo pair. Further, a light on the players may flash to further indicate that each of the players is a possibility for left channel pairing. Upon selection of the left player, FIG. 10E may be displayed to inform the user that a pair has been created along with a name for the pair, if so desired. Responsively, the system will play the left channel audio from the user designated player and will automatically play the right channel audio from the other player. FIG. 10F provides an example screenshot to allow the user to separate the stereo pair if so desired.


In an alternative embodiment, the creation of a stereo pair may be an option for a particular zone or a number of zones (e.g., a household of zones). For example, an option like “Create a Stereo Pair” may exist such that upon selection, a setup wizard may launch asking the user to press a flashing mute button (or some other designated button) on whichever speaker the user wanted to be the left speaker in the zone, a portion of zones, or all of the zones. In one embodiment, flashing would occur for all of the same speaker types. In another embodiment, flashing would occur for all speaker types that are capable of being paired. After choosing the left speaker, the wizard screen would ask the user to do the same for the right speaker. Preferably, only the speakers that are capable of being paired as the right speaker are flashing so as to appropriately narrow the choices for the user.


Additionally, in one embodiment and as shown in FIG. 3A or 3B, a graphic display is provided to show to the user all the players in a system and how they are grouped or named. A nickname for the stereo pair in the display 1040 may be highlighted and would be further displayed in FIG. 3A if FIG. 3A is modified after the stereo pair is complete


A similar graphic interface may be used to create a pair in an environment having more than two channels. For example, in a home theater environment, the system may list more than two separate players from which the user can create a pairing by selecting which player is to operate as the front right, center, front left, rear right, and rear left. A subwoofer may also be added to the list, so that it can be integrated into the multi-channel pairing by the user.


As an example, similar to what is described in the various embodiments above with respect to creating a stereo pair, the system may flash an indicator light on all relevant players and a setup wizard may ask the user to select the “front-left,” then the “front-right,” then the “front-center,” then the “rear-left,” then the “rear-right,” and so on until all of the players are appropriately paired. Preferably, only the speakers that are capable of being paired as the next speaker are flashing so as to appropriately narrow the choices for the user.


VII. Conclusion

The components, elements, and/or functionality of the systems discussed above may be implemented alone or in combination in various forms in hardware, firmware, and/or as a set of instructions in software, for example. Certain embodiments may be provided as a set of instructions residing on a computer-readable medium, such as a memory, hard disk, CD-ROM, DVD, and/or EPROM, for execution on a processing device, such as a controller and/or playback device.


Various inventions have been described in sufficient detail with a certain degree of particularity. It is understood to those skilled in the art that the present disclosure of embodiments has been made by way of examples only and that numerous changes in the arrangement and combination of parts may be resorted without departing from the spirit and scope of the invention as claimed. While the embodiments discussed herein may appear to include some limitations as to the presentation of the information units, in terms of the format and arrangement, the embodiments have applicability well beyond such embodiment, which can be appreciated by those skilled in the art. Accordingly, the scope of the present invention is defined by the appended claims rather than the forgoing description of embodiments.

Claims
  • 1. A first zone player comprising: a network interface that is configured to provide an interconnection with at least one data network;a clock that is configured to provide a clock time of the first zone player;at least one processor;a tangible, non-transitory computer-readable medium;a first set of program instructions stored on the tangible, non-transitory computer-readable medium that, when executed by the at least one processor, cause the first zone player to perform functions comprising: receiving a request to enter into a synchrony group that will play multi-channel audio content according to a first scheme for synchronizing playback with at least one other zone player that is communicatively coupled with the first zone player over a local area network (LAN); andin response to receiving the request to enter into the synchrony group, entering into the synchrony group and beginning to operate in a grouped mode in which the first zone player is configured to play back each of a plurality of audio channels included in the multi-channel audio content in accordance with the first scheme for synchronizing playback of the multi-channel audio content with the at least one other zone player that is also configured to play back each of the plurality of audio channels; anda second set of program instructions stored on the non-transitory computer-readable medium that, when executed by the at least one processor, cause the first zone player to perform functions comprising: receiving a request to enter into a multi-channel pairing that will play multi-channel audio content according to a second scheme for synchronizing playback with at least one other zone player; andin response to receiving the request to enter into the multi-channel pairing, entering into the multi-channel pairing with the at least one other zone player and beginning to operate in a paired mode in which the first zone player is configured to play back a given subset of a plurality of audio channels included in the multi-channel audio content in accordance with the second scheme for synchronizing playback with the at least one other zone player that is configured to play back a different subset of the plurality of audio channels, wherein the second scheme for synchronizing playback differs from the first scheme for synchronizing playback.
  • 2. The first zone player of claim 1, wherein receiving the request to enter into the synchrony group comprises: receiving the request to enter into the synchrony group from one of (a) a controller device that is communicatively coupled with the first zone player over the LAN or (b) the at least one other zone player.
  • 3. The first zone player of claim 1, wherein receiving the request to enter into the multi-channel pairing comprises: receiving the request to enter into the multi-channel pairing from one of (a) a controller device that is communicatively coupled with the first zone player over the LAN or (b) the at least one other zone player.
  • 4. The first zone player of claim 1, wherein the plurality of audio channels comprises a left channel and a right channel, wherein the given subset of the plurality of audio channels consists of the left channel, and wherein the different subset of the plurality of audio channels consists of the right channel.
  • 5. The first zone player of claim 1, wherein the plurality of audio channels comprises a center channel, a front-left channel, a front-right channel, a rear-left channel, and a rear-right channel, wherein the given subset of the plurality of audio channels consists of the center channel, the front-left channel, and the front-right channel, and wherein the different subset of the plurality of audio channels consists of one of the rear-left channel and the rear-right channel.
  • 6. The first zone player of claim 1, wherein: the first scheme for synchronizing playback involves using a first time delay to facilitate synchronous audio playback between the first zone player and the at least one other zone player; andthe second scheme for synchronizing playback involves using a second time delay that differs from the first time delay to facilitate synchronous audio playback between the first zone player and the at least one other zone player.
  • 7. The first zone player of claim 1, wherein: the first scheme for synchronizing playback involves transmitting, to the at least one other zone player over the LAN, audio data representing each of the plurality of audio channels included in the multi-channel audio content; andthe second scheme for synchronizing playback involves transmitting, to the at least one other zone player over the LAN, audio data representing only the different subset of the plurality of audio channels included in the multi-channel audio content.
  • 8. The first zone player of claim 1, wherein: the first scheme for synchronizing playback involves receiving, from the at least one other zone player over the LAN, audio data representing each of the plurality of audio channels included in the multi-channel audio content; andthe second scheme for synchronizing playback involves receiving, from the at least one other zone player over the LAN, audio data representing only the given subset of the plurality of audio channels included in the multi-channel audio content.
  • 9. The first zone player of claim 1, wherein: the first scheme for synchronizing playback involves synchronizing playback of the multi-channel audio content without reference to any video content; andthe second scheme for synchronizing playback involves synchronizing playback of the multi-channel audio content in a manner that also synchronizes playback of the multi-channel audio content with video content that corresponds to the multi-channel audio content.
  • 10. The first zone player of claim 1, wherein the second set of program instructions further comprise program instructions stored on the non-transitory computer-readable medium that, when executed by the at least one processor, cause the first zone player, while operating in the paired mode, to perform functions comprising: separating multi-channel audio content into a first set of audio data representing the given subset of the plurality of audio channels included in the multi-channel audio content and a second set of audio data representing the different subset of the plurality of audio channels included in the multi-channel audio content.
  • 11. The first zone player of claim 1, wherein entering into the multi-channel pairing with the at least one other zone player and beginning to operate in the paired mode comprises: changing an equalization of the first zone player.
  • 12. The first zone player of claim 11, wherein changing the equalization of the first zone player comprises: changing a channel output of one or more speaker drivers of the first zone player.
  • 13. The first zone player of claim 11, wherein changing the equalization of the first zone player comprises: changing an amplifier gain of one or more amplifiers connected to one or more speaker drivers of the first zone player.
  • 14. The first zone player of claim 11, wherein changing the equalization of the first zone player comprises: changing one or more of a frequency parameter, a time delay, or a phase associated with audio data that is to be output by the first zone player.
  • 15. The first zone player of claim 14, wherein changing a frequency parameter associated with the audio data to be output by the first zone player comprises: beginning to apply a given type of filter to the audio data.
  • 16. The first zone player of claim 1, further comprising a third set of program instructions stored on the tangible, non-transitory computer-readable medium that, when executed by the at least one processor, cause the first zone player to perform functions comprising: operating in a standalone mode in which the first zone player is configured to individually play back each of a plurality of audio channels included in multi-channel audio content rather than in synchrony with another zone player.
  • 17. The first zone player of claim 1, wherein the second set of program instructions further comprise program instructions stored on the non-transitory computer-readable medium that, when executed by the at least one processor, cause the first zone player to perform additional functions comprising: in response to receiving the request to enter into the multi-channel pairing, identifying the given subset of the plurality of audio channels that is to be played back by the first zone player and the different subset of the plurality of audio channels that is to be played back by the at least one other zone player.
  • 18. The first zone player of claim 1, wherein receiving the request to enter into the synchrony group comprises receiving a request to enter into the synchrony group with at least one other zone player that is located in a different room from the first zone player; and wherein receiving the request to enter into the multi-channel pairing comprises receiving a request to enter into the multi-channel pairing with at least one other zone player that is located in the same room as the first zone player.
  • 19. Tangible, non-transitory computer-readable storage media comprising software that, when executed by a first zone player, causes the first zone player to perform functions comprising: receiving a request to enter into a synchrony group that will play multi-channel audio content according to a first scheme for synchronizing playback with at least one other zone player that is communicatively coupled with the first zone player over a local area network (LAN);in response to receiving the request to enter into the synchrony group, entering into the synchrony group and beginning to operate in a grouped mode in which the first zone player is configured to play back each of a plurality of audio channels included in the multi-channel audio content in accordance with the first scheme for synchronizing playback of the multi-channel audio content with the at least one other zone player that is also configured to play back each of the plurality of audio channels; andreceiving a request to enter into a multi-channel pairing that will play multi-channel audio content according to a second scheme for synchronizing playback with at least one other zone player; andin response to receiving the request to enter into the multi-channel pairing, entering into the multi-channel pairing with the at least one other zone player and beginning to operate in a paired mode in which the first zone player is configured to play back a first subset of a plurality of audio channels included in the multi-channel audio content in accordance with the second scheme for synchronizing playback with the at least one other zone player that is configured to play back a second subset of the plurality of audio channels, wherein the second scheme for synchronizing playback differs from the first scheme for synchronizing playback.
  • 20. The tangible, non-transitory computer-readable storage media of claim 19, wherein: the first scheme for synchronizing playback involves using a first time delay to facilitate synchronous audio playback between the first zone player and the at least one other zone player; andthe second scheme for synchronizing playback involves using a second time delay that differs from the first time delay to facilitate synchronous audio playback between the first zone player and the at least one other zone player.
  • 21. The tangible, non-transitory computer-readable storage media of claim 19, wherein: the first scheme for synchronizing playback involves transmitting, to the at least one other zone player over the LAN, audio data representing each of the plurality of audio channels included in the multi-channel audio content; andthe second scheme for synchronizing playback involves transmitting, to the at least one other zone player over the LAN, audio data representing only the second subset of the plurality of audio channels included in the multi-channel audio content.
  • 22. The tangible, non-transitory computer-readable storage media of claim 19, wherein: the first scheme for synchronizing playback involves receiving, from the at least one other zone player over the LAN, audio data representing each of the plurality of audio channels included in the multi-channel audio content; andthe second scheme for synchronizing playback involves receiving, from the at least one other zone player over the LAN, audio data representing only the first subset of the plurality of audio channels included in the multi-channel audio content.
  • 23. The tangible, non-transitory computer-readable storage media of claim 19, wherein: the first scheme for synchronizing playback involves synchronizing playback of the multi-channel audio content without reference to any video content; andthe second scheme for synchronizing playback involves synchronizing playback of the multi-channel audio content in a manner that also synchronizes playback of the multi-channel audio content with video content that corresponds to the multi-channel audio content.
  • 24. The tangible, non-transitory computer-readable storage media of claim 19, wherein entering into the multi-channel pairing with the at least one other zone player and beginning to operate in the paired mode comprises changing an equalization of the first zone player, and wherein changing the equalization of the first zone player comprises one or more of: (i) changing a channel output of one or more speaker drivers of the first zone player, (ii) changing an amplifier gain of one or more amplifiers connected to one or more speaker drivers of the first zone player, or (iii) changing one or more of a frequency parameter, a time delay, or a phase associated with audio data that is to be output by the first zone player.
  • 25. A method carried out by a first zone player, the method comprising: receiving a request to enter into a configuration in which the first zone player is to play back multi-channel audio content in synchrony with at least one other zone player that is communicatively coupled with the first zone player over a local area network (LAN);if the request comprises a request to enter into a synchrony group that will play multi-channel audio content according to a first scheme for synchronizing playback with the at least one other zone player, responsively entering into the synchrony group and beginning to operate in a grouped mode in which the first zone player is configured to play back each of a plurality of audio channels included in the multi-channel audio content in accordance with the first scheme for synchronizing playback of the multi-channel audio content with the at least one other zone player that is also configured to play back each of the plurality of audio channels; andif the request comprises a request to enter into a multi-channel pairing that will play multi-channel audio content according to a second scheme for synchronizing playback with the at least one other zone player, responsively entering into the multi-channel pairing with the at least one other zone player and beginning to operate in a paired mode in which the first zone player is configured to play back a first subset of a plurality of audio channels included in the multi-channel audio content in accordance with the second scheme for synchronizing playback with the at least one other zone player that is configured to play back a second subset of the plurality of audio channels, wherein the second scheme for synchronizing playback differs from the first scheme for synchronizing playback.
  • 26. The method of claim 25, wherein: the first scheme for synchronizing playback involves using a first time delay to facilitate synchronous audio playback between the first zone player and the at least one other zone player; andthe second scheme for synchronizing playback involves using a second time delay that differs from the first time delay to facilitate synchronous audio playback between the first zone player and the at least one other zone player.
  • 27. The method of claim 25, wherein: the first scheme for synchronizing playback involves transmitting, to the at least one other zone player over the LAN, audio data representing each of the plurality of audio channels included in the multi-channel audio content; andthe second scheme for synchronizing playback involves transmitting, to the at least one other zone player over the LAN, audio data representing only the second subset of the plurality of audio channels included in the multi-channel audio content.
  • 28. The method of claim 25, wherein: the first scheme for synchronizing playback involves receiving, from the at least one other zone player over the LAN, audio data representing each of the plurality of audio channels included in the multi-channel audio content; andthe second scheme for synchronizing playback involves receiving, from the at least one other zone player over the LAN, audio data representing only the first subset of the plurality of audio channels included in the multi-channel audio content.
  • 29. The method of claim 25, wherein: the first scheme for synchronizing playback involves synchronizing playback of the multi-channel audio content without reference to any video content; andthe second scheme for synchronizing playback involves synchronizing playback of the multi-channel audio content in a manner that also synchronizes playback of the multi-channel audio content with video content that corresponds to the multi-channel audio content.
  • 30. The method of claim 25, wherein entering into the multi-channel pairing with the at least one other zone player and beginning to operate in the paired mode comprises changing an equalization of the first zone player, and wherein changing the equalization of the first zone player comprises one or more of: (i) changing a channel output of one or more speaker drivers of the first zone player, (ii) changing an amplifier gain of one or more amplifiers connected to one or more speaker drivers of the first zone player, or (iii) changing one or more of a frequency parameter, a time delay, or a phase associated with audio data that is to be output by the first zone player.
CROSS-REFERENCE TO RELATED APPLICATIONS

The present application is a continuation of U.S. application Ser. No. 14/629,937 titled “Gain Based on Play Responsibility” filed on Feb. 24, 2015, and; U.S. application Ser. No. 14/629,937 is a continuation of U.S. application Ser. No. 14/299,847 titled “Multi-channel pairing in a media system” filed on Jun. 9, 2014, and issued on Dec. 22, 2015, as U.S. Pat. No. 9,219,959; U.S. application Ser. No. 14/299,847 is a continuation of U.S. application Ser. No. 13/083,499 titled “Multi-channel pairing in a media system” filed on Apr. 8, 2011, and issued on Jul. 22, 2014, as U.S. Pat. No. 8,788,080; U.S. application Ser. No. 13/083,499 is a continuation-in-part of U.S. application Ser. No. 13/013,740 titled “Controlling and grouping in a multi-zone media system” filed on Jan. 25, 2011 and issued on Dec. 1, 2015, as U.S. Pat. No. 9,202,509; U.S. application Ser. No. 13/013,740 is a continuation-in-part of U.S. application Ser. No. 11/853,790 titled “Controlling and manipulating groupings in a multi-zone media system,” filed Sep. 11, 2007, and issued on Jul. 9, 2013, as U.S. Pat. No. 8,483,853; U.S. application Ser. No. 11/853,790 claims priority to U.S. Provisional App. 60/825,407 titled “Controlling and manipulating groupings in a multi-zone music or media system” filed on Sep. 12, 2006. The entire contents of the Ser. No. 14/629,937; Ser. No. 14/299,847; Ser. No. 13/083,499; Ser. No. 13/013,740; Ser. No. 11/853,790; and 60/825,407 applications are incorporated herein by reference.

US Referenced Citations (1055)
Number Name Date Kind
3956591 Gates, Jr. May 1976 A
4105974 Rogers Aug 1978 A
D260764 Castagna et al. Sep 1981 S
4296278 Cullison et al. Oct 1981 A
4306114 Callahan Dec 1981 A
4382158 Ohshita et al. May 1983 A
4509211 Robbins Apr 1985 A
D279779 Taylor Jul 1985 S
4530091 Crockett Jul 1985 A
4696037 Fierens Sep 1987 A
4701629 Citroen Oct 1987 A
4712105 Koehler Dec 1987 A
D293671 Beaumont Jan 1988 S
4731814 Becker et al. Mar 1988 A
4816989 Finn et al. Mar 1989 A
4824059 Butler Apr 1989 A
D301037 Matsuda May 1989 S
4845751 Schwab Jul 1989 A
D304443 Grinyer et al. Nov 1989 S
D313023 Kolenda et al. Dec 1990 S
D313398 Gilchrist Jan 1991 S
D313600 Weber Jan 1991 S
4994908 Kuban et al. Feb 1991 A
4995778 Bruessel Feb 1991 A
D320598 Auerbach et al. Oct 1991 S
D322609 Patton Dec 1991 S
5086385 Launey et al. Feb 1992 A
D326450 Watanabe May 1992 S
D327060 Wachob et al. Jun 1992 S
5151922 Weiss Sep 1992 A
5153579 Fisch et al. Oct 1992 A
D331388 Dahnert et al. Dec 1992 S
5182552 Paynting Jan 1993 A
D333135 Wachob et al. Feb 1993 S
5185680 Kakubo Feb 1993 A
5197099 Hirasawa Mar 1993 A
5237327 Saitoh et al. Aug 1993 A
5239458 Suzuki Aug 1993 A
5272757 Scofield et al. Dec 1993 A
5299266 Lumsden Mar 1994 A
D350531 Tsuji Sep 1994 S
D350962 Reardon et al. Sep 1994 S
5361381 Short Nov 1994 A
5372441 Louis Dec 1994 A
D354059 Hendricks Jan 1995 S
D354751 Hersh et al. Jan 1995 S
D356093 McCauley et al. Mar 1995 S
D356312 Althans Mar 1995 S
D357024 Tokiyama et al. Apr 1995 S
5406634 Anderson Apr 1995 A
5430485 Lankford et al. Jul 1995 A
5440644 Farinelli et al. Aug 1995 A
D362446 Gasiorek et al. Sep 1995 S
5457448 Totsuka et al. Oct 1995 A
D363933 Starck Nov 1995 S
5467342 Logston et al. Nov 1995 A
D364877 Tokiyama et al. Dec 1995 S
D364878 Green et al. Dec 1995 S
D365102 Gioscia Dec 1995 S
D366044 Hara et al. Jan 1996 S
5481251 Buys et al. Jan 1996 A
5491839 Schotz Feb 1996 A
5515345 Barreira et al. May 1996 A
5519641 Beers et al. May 1996 A
5533021 Branstad et al. Jul 1996 A
D372716 Thorne Aug 1996 S
5553147 Pineau Sep 1996 A
5553222 Milne et al. Sep 1996 A
5553314 Grube et al. Sep 1996 A
D377651 Biasotti et al. Jan 1997 S
5596696 Tindell et al. Jan 1997 A
5602992 Danneels Feb 1997 A
5623483 Agrawal et al. Apr 1997 A
5625350 Fukatsu et al. Apr 1997 A
D379816 Laituri et al. Jun 1997 S
5640388 Woodhead et al. Jun 1997 A
5642171 Baumgartner et al. Jun 1997 A
D380752 Hanson Jul 1997 S
5652749 Davenport et al. Jul 1997 A
D382271 Akwiwu Aug 1997 S
5661665 Glass et al. Aug 1997 A
5668884 Clair, Jr. et al. Sep 1997 A
5673323 Schotz et al. Sep 1997 A
D384940 Kono et al. Oct 1997 S
D387352 Kaneko et al. Dec 1997 S
5696896 Badovinatz et al. Dec 1997 A
D388792 Nykerk Jan 1998 S
D389143 Wicks Jan 1998 S
D392641 Fenner Mar 1998 S
5726989 Dokic Mar 1998 A
D393628 Ledbetter et al. Apr 1998 S
5740235 Lester et al. Apr 1998 A
5742623 Nuber et al. Apr 1998 A
D394659 Biasotti et al. May 1998 S
5751819 Dorrough May 1998 A
5761320 Farinelli et al. Jun 1998 A
5774016 Ketterer Jun 1998 A
D395889 Gerba et al. Jul 1998 S
5787249 Badovinatz et al. Jul 1998 A
5790543 Cloutier Aug 1998 A
D397996 Smith Sep 1998 S
5808662 Kinney et al. Sep 1998 A
5812201 Yoo Sep 1998 A
5815689 Shaw et al. Sep 1998 A
5818948 Gulick Oct 1998 A
D401587 Rudolph Nov 1998 S
5832024 Schotz et al. Nov 1998 A
5848152 Slipy et al. Dec 1998 A
5852722 Hamilton Dec 1998 A
D404741 Schumaker et al. Jan 1999 S
D405071 Gambaro Feb 1999 S
5867691 Shiraishi Feb 1999 A
5875233 Cox Feb 1999 A
5875354 Charlton et al. Feb 1999 A
D406847 Gerba et al. Mar 1999 S
D407071 Keating Mar 1999 S
5887143 Saito et al. Mar 1999 A
5905768 Maturi et al. May 1999 A
D410927 Yamagishi Jun 1999 S
5910991 Farrar Jun 1999 A
D412337 Hamano Jul 1999 S
5923869 Kashiwagi et al. Jul 1999 A
5923902 Inagaki Jul 1999 A
5946343 Schotz et al. Aug 1999 A
5956025 Goulden et al. Sep 1999 A
5956088 Shen et al. Sep 1999 A
5960006 Maturi et al. Sep 1999 A
D415496 Gerba et al. Oct 1999 S
D416021 Godette et al. Nov 1999 S
5984512 Jones et al. Nov 1999 A
5987611 Freund Nov 1999 A
5990884 Douma et al. Nov 1999 A
5991307 Komuro et al. Nov 1999 A
5999906 Mercs et al. Dec 1999 A
6009457 Moller Dec 1999 A
6018376 Nakatani Jan 2000 A
D420006 Tonino Feb 2000 S
6026150 Frank et al. Feb 2000 A
6029196 Lenz Feb 2000 A
6031818 Lo et al. Feb 2000 A
6032202 Lea et al. Feb 2000 A
6038614 Chan et al. Mar 2000 A
6046550 Ference et al. Apr 2000 A
6061457 Stockhamer May 2000 A
6078725 Tanaka Jun 2000 A
6081266 Sciammarella Jun 2000 A
6088063 Shiba Jul 2000 A
D429246 Holma Aug 2000 S
D430143 Renk Aug 2000 S
6101195 Lyons et al. Aug 2000 A
6108485 Kim Aug 2000 A
6108686 Williams, Jr. Aug 2000 A
6119239 Fujii Sep 2000 A
6122668 Teng et al. Sep 2000 A
6122749 Gulick Sep 2000 A
D431552 Backs et al. Oct 2000 S
D432525 Beecroft Oct 2000 S
6127941 Van Ryzin Oct 2000 A
6128318 Sato Oct 2000 A
6148205 Cotton Nov 2000 A
6157957 Berthaud Dec 2000 A
6163647 Terashima et al. Dec 2000 A
6169725 Gibbs et al. Jan 2001 B1
6175872 Neumann et al. Jan 2001 B1
6181383 Fox et al. Jan 2001 B1
6185737 Northcutt et al. Feb 2001 B1
6195435 Kitamura Feb 2001 B1
6195436 Scibora et al. Feb 2001 B1
6199169 Voth Mar 2001 B1
6212282 Mershon Apr 2001 B1
6246701 Slattery Jun 2001 B1
6253293 Rao et al. Jun 2001 B1
D444475 Levey et al. Jul 2001 S
6255961 Van et al. Jul 2001 B1
6256554 DiLorenzo Jul 2001 B1
6269406 Dutcher et al. Jul 2001 B1
6301012 White et al. Oct 2001 B1
6308207 Tseng et al. Oct 2001 B1
6310652 Li et al. Oct 2001 B1
6313879 Kubo et al. Nov 2001 B1
6321252 Bhola et al. Nov 2001 B1
6324586 Johnson Nov 2001 B1
D452520 Gotham et al. Dec 2001 S
6332147 Moran et al. Dec 2001 B1
6343028 Kuwaoka Jan 2002 B1
6349285 Liu et al. Feb 2002 B1
6349339 Williams Feb 2002 B1
6351821 Voth Feb 2002 B1
6353172 Fay et al. Mar 2002 B1
6356871 Hemkumar et al. Mar 2002 B1
6404811 Cvetko et al. Jun 2002 B1
6418150 Staats Jul 2002 B1
6430353 Honda et al. Aug 2002 B1
6442443 Fujii et al. Aug 2002 B1
D462339 Allen et al. Sep 2002 S
D462340 Allen et al. Sep 2002 S
D462945 Skulley Sep 2002 S
6449642 Bourke-Dunphy et al. Sep 2002 B2
6449653 Klemets et al. Sep 2002 B2
6456783 Ando et al. Sep 2002 B1
6463474 Fuh et al. Oct 2002 B1
6466832 Zuqert et al. Oct 2002 B1
6469633 Wachter Oct 2002 B1
D466108 Glodava et al. Nov 2002 S
6487296 Allen et al. Nov 2002 B1
6493832 Itakura et al. Dec 2002 B1
D468297 Ikeda Jan 2003 S
6522886 Youngs et al. Feb 2003 B1
6526325 Sussman et al. Feb 2003 B1
6535121 Mathney et al. Mar 2003 B2
D474763 Tozaki et al. May 2003 S
D475993 Meyer Jun 2003 S
D476643 Yamagishi Jul 2003 S
D477310 Moransais Jul 2003 S
6587127 Leeke et al. Jul 2003 B1
6598172 Vandeusen et al. Jul 2003 B1
D478051 Sagawa Aug 2003 S
D478069 Beck et al. Aug 2003 S
D478896 Summers Aug 2003 S
6604023 Brown et al. Aug 2003 B1
6611537 Edens et al. Aug 2003 B1
D479520 De Sep 2003 S
D481056 Kawasaki et al. Oct 2003 S
6631410 Kowalski et al. Oct 2003 B1
6636269 Baldwin Oct 2003 B1
6653899 Organvidez et al. Nov 2003 B2
6654720 Graham et al. Nov 2003 B1
6654956 Trinh et al. Nov 2003 B1
6658091 Naidoo et al. Dec 2003 B1
6674803 Kesselring Jan 2004 B1
6684060 Curtin Jan 2004 B1
D486145 Kaminski et al. Feb 2004 S
6687664 Sussman et al. Feb 2004 B1
6703940 Allen et al. Mar 2004 B1
6704421 Kitamura Mar 2004 B1
6741961 Lim May 2004 B2
D491925 Griesau et al. Jun 2004 S
6757517 Chang Jun 2004 B2
D493148 Shibata et al. Jul 2004 S
6763274 Gilbert Jul 2004 B1
D495333 Borsboom Aug 2004 S
6778073 Lutter et al. Aug 2004 B2
6778493 Ishii Aug 2004 B1
6778869 Champion Aug 2004 B2
D496003 Spira Sep 2004 S
D496005 Wang Sep 2004 S
D496335 Spira Sep 2004 S
D497363 Olson et al. Oct 2004 S
6803964 Post et al. Oct 2004 B1
6809635 Kaaresoja Oct 2004 B1
D499086 Polito Nov 2004 S
6816510 Banerjee Nov 2004 B1
6816818 Wolf et al. Nov 2004 B2
6823225 Sass Nov 2004 B1
6826283 Wheeler et al. Nov 2004 B1
D499395 Hsu Dec 2004 S
D499718 Chen Dec 2004 S
D500015 Gubbe Dec 2004 S
6836788 Kim et al. Dec 2004 B2
6839752 Miller et al. Jan 2005 B1
D501477 Hall Feb 2005 S
6859460 Chen Feb 2005 B1
6859538 Voltz Feb 2005 B1
6870934 Krochmal et al. Mar 2005 B2
6873862 Reshefsky Mar 2005 B2
6882335 Saarinen Apr 2005 B2
D504872 Uehara et al. May 2005 S
D504885 Zhang et al. May 2005 S
6889207 Slemmer et al. May 2005 B2
6898642 Chafle et al. May 2005 B2
6901439 Bonasia et al. May 2005 B1
D506463 Daniels Jun 2005 S
6907458 Tomassetti et al. Jun 2005 B2
6912610 Spencer Jun 2005 B2
6915347 Hanko et al. Jul 2005 B2
6916980 Ishida et al. Jul 2005 B2
6917592 Ramankutty et al. Jul 2005 B1
6919771 Nakajima Jul 2005 B2
6920373 Xi et al. Jul 2005 B2
6931134 Waller, Jr. et al. Aug 2005 B1
6931557 Togawa Aug 2005 B2
6934766 Russell Aug 2005 B1
6937988 Hemkumar et al. Aug 2005 B1
6970482 Kim Nov 2005 B2
6985694 De Bonet et al. Jan 2006 B1
6987767 Saito Jan 2006 B2
6987947 Richenstein et al. Jan 2006 B2
D515072 Lee Feb 2006 S
D515557 Okuley Feb 2006 S
7007106 Flood et al. Feb 2006 B1
7020791 Aweya et al. Mar 2006 B1
D518475 Yang et al. Apr 2006 S
7043477 Mercer et al. May 2006 B2
7043651 Aweya et al. May 2006 B2
7046677 Monta et al. May 2006 B2
7047308 Deshpande May 2006 B2
7054888 Lachapelle et al. May 2006 B2
7058889 Trovato et al. Jun 2006 B2
7068596 Mou Jun 2006 B1
D524296 Kita Jul 2006 S
7072477 Kincaid Jul 2006 B1
D527375 Flora et al. Aug 2006 S
7092528 Patrick et al. Aug 2006 B2
7092694 Griep et al. Aug 2006 B2
7096169 Crutchfield et al. Aug 2006 B2
7113999 Pestoni et al. Sep 2006 B2
7115017 Laursen et al. Oct 2006 B1
7120168 Zimmermann Oct 2006 B2
7130316 Kovacevic Oct 2006 B2
7130368 Aweya et al. Oct 2006 B1
7130608 Hollstrom et al. Oct 2006 B2
7130616 Janik Oct 2006 B2
7136934 Carter et al. Nov 2006 B2
7139981 Mayer et al. Nov 2006 B2
7143141 Morgan et al. Nov 2006 B1
7143939 Henzerling Dec 2006 B2
7146260 Preston et al. Dec 2006 B2
7158488 Fujimori Jan 2007 B2
7161939 Israel et al. Jan 2007 B2
7162315 Gilbert Jan 2007 B2
7171010 Martin et al. Jan 2007 B2
7185090 Kowalski et al. Feb 2007 B2
7187947 White et al. Mar 2007 B1
7197148 Nourse et al. Mar 2007 B2
7206367 Moore Apr 2007 B1
7206618 Latto et al. Apr 2007 B2
7206967 Marti et al. Apr 2007 B1
7209795 Sullivan et al. Apr 2007 B2
7218708 Berezowski et al. May 2007 B2
7236739 Chang Jun 2007 B2
7236773 Thomas Jun 2007 B2
7257398 Ukita et al. Aug 2007 B1
7260616 Cook Aug 2007 B1
7263110 Fujishiro Aug 2007 B2
7269338 Janevski Sep 2007 B2
7277547 Delker et al. Oct 2007 B1
7286652 Azriel et al. Oct 2007 B1
7289631 Ishidoshiro Oct 2007 B2
7293060 Komsi Nov 2007 B2
7295548 Blank et al. Nov 2007 B2
7302468 Wijeratne Nov 2007 B2
7305694 Commons et al. Dec 2007 B2
7308188 Namatame Dec 2007 B2
7310334 Fitzgerald et al. Dec 2007 B1
7312785 Tsuk et al. Dec 2007 B2
7313384 Meenan et al. Dec 2007 B1
7313593 Pulito et al. Dec 2007 B1
7319764 Reid et al. Jan 2008 B1
7324857 Goddard Jan 2008 B2
7330875 Parasnis et al. Feb 2008 B1
7333519 Sullivan et al. Feb 2008 B2
7346332 McCarty et al. Mar 2008 B2
7356011 Waters et al. Apr 2008 B1
7359006 Xiang et al. Apr 2008 B1
7366206 Lockridge et al. Apr 2008 B2
7372846 Zwack May 2008 B2
7391791 Balassanian et al. Jun 2008 B2
7392102 Sullivan et al. Jun 2008 B2
7392481 Gewickey et al. Jun 2008 B2
7400644 Sakamoto et al. Jul 2008 B2
7412499 Chang et al. Aug 2008 B2
7424267 Eisenbach Sep 2008 B2
7428310 Park Sep 2008 B2
7430181 Hong Sep 2008 B1
7457948 Bilicksa et al. Nov 2008 B1
7472058 Tseng et al. Dec 2008 B2
7474677 Trott Jan 2009 B2
7483538 McCarty et al. Jan 2009 B2
7483540 Rabinowitz et al. Jan 2009 B2
7483958 Elabbady et al. Jan 2009 B1
7490044 Kulkarni Feb 2009 B2
7492912 Chung et al. Feb 2009 B2
7505889 Salmonsen et al. Mar 2009 B2
7509181 Champion Mar 2009 B2
7519188 Berardi et al. Apr 2009 B2
7519667 Capps Apr 2009 B1
7539551 Komura May 2009 B2
7548744 Oesterling et al. Jun 2009 B2
7548851 Lau et al. Jun 2009 B1
7558224 Surazski et al. Jul 2009 B1
7558635 Thiel et al. Jul 2009 B1
7561932 Holmes et al. Jul 2009 B1
7571014 Lambourne et al. Aug 2009 B1
7574274 Holmes Aug 2009 B2
7599685 Goldberg et al. Oct 2009 B2
7606174 Ochi et al. Oct 2009 B2
7620468 Shimizu Nov 2009 B2
7626952 Slemmer et al. Dec 2009 B2
7627825 Kakuda Dec 2009 B2
7630500 Beckman et al. Dec 2009 B1
7630501 Blank et al. Dec 2009 B2
7631119 Moore et al. Dec 2009 B2
7643894 Braithwaite et al. Jan 2010 B2
7653344 Feldman et al. Jan 2010 B1
7657224 Goldberg et al. Feb 2010 B2
7657644 Zheng Feb 2010 B1
7657910 McAulay et al. Feb 2010 B1
7665115 Gallo et al. Feb 2010 B2
7668990 Krzyzanowski et al. Feb 2010 B2
7669113 Moore et al. Feb 2010 B1
7669219 Scott Feb 2010 B2
7672470 Lee Mar 2010 B2
7675943 Mosig et al. Mar 2010 B2
7676044 Sasaki et al. Mar 2010 B2
7676142 Hung Mar 2010 B1
7688306 Wehrenberg et al. Mar 2010 B2
7689304 Sasaki Mar 2010 B2
7689305 Kreifeldt et al. Mar 2010 B2
7702279 Ko et al. Apr 2010 B2
7702403 Gladwin et al. Apr 2010 B1
7710941 Rietschel et al. May 2010 B2
7711774 Rothschild May 2010 B1
7720096 Klemets May 2010 B2
7721032 Bushell et al. May 2010 B2
7742740 Goldberg et al. Jun 2010 B2
7742832 Feldman et al. Jun 2010 B1
7743009 Hangartner et al. Jun 2010 B2
7746906 Jinzaki et al. Jun 2010 B2
7761176 Ben-Yaacov et al. Jul 2010 B2
7765315 Batson et al. Jul 2010 B2
RE41608 Blair et al. Aug 2010 E
7792311 Holmgren et al. Sep 2010 B1
7793206 Lim et al. Sep 2010 B2
7804972 Melanson Sep 2010 B2
7805210 Cucos et al. Sep 2010 B2
7817960 Tan et al. Oct 2010 B2
7827259 Heller et al. Nov 2010 B2
7831054 Ball et al. Nov 2010 B2
7835689 Goldberg et al. Nov 2010 B2
7849181 Slemmer et al. Dec 2010 B2
7853341 McCarty Dec 2010 B2
7865137 Goldberg et al. Jan 2011 B2
7882234 Watanabe et al. Feb 2011 B2
7885622 Krampf et al. Feb 2011 B2
7907819 Ando et al. Mar 2011 B2
7916877 Goldberg et al. Mar 2011 B2
7917082 Goldberg et al. Mar 2011 B2
7933418 Morishima Apr 2011 B2
7934239 Dagman Apr 2011 B1
7945636 Nelson et al. May 2011 B2
7945708 Ohkita May 2011 B2
7958441 Heller et al. Jun 2011 B2
7962482 Handman et al. Jun 2011 B2
7966388 Pugaczewski et al. Jun 2011 B1
7975051 Saint Clair et al. Jul 2011 B2
7987294 Bryce et al. Jul 2011 B2
7995732 Koch et al. Aug 2011 B2
7996566 Sylvain et al. Aug 2011 B1
7996588 Subbiah et al. Aug 2011 B2
8014423 Thaler et al. Sep 2011 B2
8015306 Bowman Sep 2011 B2
8020023 Millington et al. Sep 2011 B2
8023663 Goldberg Sep 2011 B2
8028038 Weel Sep 2011 B2
8028323 Weel Sep 2011 B2
8041062 Cohen et al. Oct 2011 B2
8045721 Burgan et al. Oct 2011 B2
8045952 Qureshey et al. Oct 2011 B2
8050203 Jacobsen et al. Nov 2011 B2
8050652 Qureshey et al. Nov 2011 B2
8054987 Seydoux Nov 2011 B2
8055364 Champion Nov 2011 B2
8063698 Howard Nov 2011 B2
8074253 Nathan Dec 2011 B1
8086287 Mooney et al. Dec 2011 B2
8086752 Millington Dec 2011 B2
8090317 Burge et al. Jan 2012 B2
8103009 McCarty et al. Jan 2012 B2
8111132 Allen et al. Feb 2012 B2
8112032 Ko et al. Feb 2012 B2
8116476 Inohara Feb 2012 B2
8126172 Horbach et al. Feb 2012 B2
8131389 Hardwick et al. Mar 2012 B1
8131390 Braithwaite et al. Mar 2012 B2
8135141 Shiba Mar 2012 B2
8139774 Berardi et al. Mar 2012 B2
8144883 Pdersen et al. Mar 2012 B2
8148622 Rothkopf et al. Apr 2012 B2
8150079 Maeda et al. Apr 2012 B2
8160281 Kim et al. Apr 2012 B2
8169938 Duchscher et al. May 2012 B2
8170222 Dunko May 2012 B2
8170260 Reining et al. May 2012 B2
8175292 Aylward et al. May 2012 B2
8175297 Ho et al. May 2012 B1
8185674 Moore et al. May 2012 B2
8189824 Strauss et al. May 2012 B2
8194874 Starobin et al. Jun 2012 B2
8204890 Gogan Jun 2012 B1
8208653 Eo et al. Jun 2012 B2
8214447 Deslippe et al. Jul 2012 B2
8214740 Johnson Jul 2012 B2
8214873 Weel Jul 2012 B2
8218790 Bull et al. Jul 2012 B2
8229125 Short Jul 2012 B2
8230099 Weel Jul 2012 B2
8233029 Yoshida et al. Jul 2012 B2
8233632 MacDonald et al. Jul 2012 B1
8233635 Shiba Jul 2012 B2
8233648 Sorek et al. Jul 2012 B2
8234395 Millington et al. Jul 2012 B2
8238578 Aylward Aug 2012 B2
8239559 Rajapakse Aug 2012 B2
8239748 Moore et al. Aug 2012 B1
8243961 Morrill Aug 2012 B1
8265310 Berardi et al. Sep 2012 B2
8270631 Kusunoki Sep 2012 B2
8279709 Choisel et al. Oct 2012 B2
8281001 Busam et al. Oct 2012 B2
8285404 Kekki Oct 2012 B1
8290185 Kim Oct 2012 B2
8290603 Lambourne Oct 2012 B1
8300845 Zurek et al. Oct 2012 B2
8306235 Mahowald Nov 2012 B2
8311226 Lorgeoux et al. Nov 2012 B2
8315555 Ko et al. Nov 2012 B2
8316147 Batson et al. Nov 2012 B2
8325931 Howard et al. Dec 2012 B2
8325935 Rutschman Dec 2012 B2
8331585 Hagen et al. Dec 2012 B2
8340330 Yoon et al. Dec 2012 B2
8345709 Nitzpon et al. Jan 2013 B2
8364295 Beckmann et al. Jan 2013 B2
8370678 Millington et al. Feb 2013 B2
8374595 Chien et al. Feb 2013 B2
8391501 Khawand et al. Mar 2013 B2
8407623 Kerr et al. Mar 2013 B2
8411883 Matsumoto Apr 2013 B2
8423659 Millington Apr 2013 B2
8423893 Ramsay et al. Apr 2013 B2
8432851 Xu et al. Apr 2013 B2
8433076 Zurek et al. Apr 2013 B2
8442239 Bruelle-Drews et al. May 2013 B2
8452020 Gregg et al. May 2013 B2
8457334 Yoon et al. Jun 2013 B2
8463184 Dua Jun 2013 B2
8463875 Katz et al. Jun 2013 B2
8473844 Kreifeldt et al. Jun 2013 B2
8477958 Moeller et al. Jul 2013 B2
8483853 Lambourne Jul 2013 B1
8498726 Kim et al. Jul 2013 B2
8509211 Trotter et al. Aug 2013 B2
8520870 Sato et al. Aug 2013 B2
8565455 Worrell et al. Oct 2013 B2
8577045 Gibbs Nov 2013 B2
8577048 Chaikin et al. Nov 2013 B2
8588432 Simon Nov 2013 B1
8588949 Lambourne et al. Nov 2013 B2
8600075 Lim Dec 2013 B2
8600084 Garrett Dec 2013 B1
8601394 Sheehan et al. Dec 2013 B2
8611559 Sanders Dec 2013 B2
8615091 Terwal Dec 2013 B2
8620006 Berardi et al. Dec 2013 B2
8639830 Bowman Jan 2014 B2
8654995 Silber et al. Feb 2014 B2
8672744 Gronkowski et al. Mar 2014 B1
8683009 Ng et al. Mar 2014 B2
8688431 Lyons et al. Apr 2014 B2
8700730 Rowe Apr 2014 B2
8731206 Park May 2014 B1
8750282 Gelter et al. Jun 2014 B2
8751026 Sato et al. Jun 2014 B2
8762565 Togashi et al. Jun 2014 B2
8775546 Millington Jul 2014 B2
8788080 Kallai et al. Jul 2014 B1
8818538 Sakata Aug 2014 B2
8819554 Basso et al. Aug 2014 B2
8843224 Holmgren et al. Sep 2014 B2
8843228 Lambourne Sep 2014 B2
8843586 Pantos et al. Sep 2014 B2
8855319 Liu et al. Oct 2014 B2
8861739 Ojanpera Oct 2014 B2
8879761 Johnson et al. Nov 2014 B2
8885851 Westenbroek Nov 2014 B2
8886347 Lambourne Nov 2014 B2
8904066 Moore et al. Dec 2014 B2
8914559 Kalayjian et al. Dec 2014 B2
8917877 Haaff et al. Dec 2014 B2
8923997 Kallai et al. Dec 2014 B2
8930006 Haatainen Jan 2015 B2
8934647 Joyce et al. Jan 2015 B2
8934655 Breen et al. Jan 2015 B2
8942252 Balassanian et al. Jan 2015 B2
8942395 Lissaman et al. Jan 2015 B2
8954177 Sanders Feb 2015 B2
8965544 Ramsay Feb 2015 B2
8965546 Visser et al. Feb 2015 B2
8966394 Gates et al. Feb 2015 B2
8977974 Kraut Mar 2015 B2
8984442 Pirnack et al. Mar 2015 B2
9020153 Britt, Jr. Apr 2015 B2
9042556 Kallai et al. May 2015 B2
9112622 Miyata et al. Aug 2015 B2
9137602 Mayman et al. Sep 2015 B2
9160965 Redmann et al. Oct 2015 B2
9195258 Millington Nov 2015 B2
9219959 Kallai et al. Dec 2015 B2
9226073 Ramos et al. Dec 2015 B2
9245514 Donaldson Jan 2016 B2
9325286 Yang Apr 2016 B1
9524098 Griffiths et al. Dec 2016 B2
20010001160 Shoff et al. May 2001 A1
20010009604 Ando et al. Jul 2001 A1
20010020193 Teramachi et al. Sep 2001 A1
20010022823 Renaud Sep 2001 A1
20010027498 Van De Meulenhof et al. Oct 2001 A1
20010032188 Miyabe et al. Oct 2001 A1
20010042107 Palm Nov 2001 A1
20010043456 Atkinson Nov 2001 A1
20010046235 Trevitt et al. Nov 2001 A1
20010047377 Sincaglia et al. Nov 2001 A1
20010050991 Eves Dec 2001 A1
20020002039 Qureshey et al. Jan 2002 A1
20020002562 Moran et al. Jan 2002 A1
20020002565 Ohyama Jan 2002 A1
20020003548 Krusche et al. Jan 2002 A1
20020022453 Balog et al. Feb 2002 A1
20020026442 Lipscomb et al. Feb 2002 A1
20020034374 Barton Mar 2002 A1
20020042844 Chiazzese Apr 2002 A1
20020049843 Barone et al. Apr 2002 A1
20020062406 Chang et al. May 2002 A1
20020065926 Hackney et al. May 2002 A1
20020067909 Iivonen Jun 2002 A1
20020072816 Shdema Jun 2002 A1
20020072817 Champion Jun 2002 A1
20020073228 Cognet et al. Jun 2002 A1
20020078161 Cheng Jun 2002 A1
20020078293 Kou et al. Jun 2002 A1
20020080783 Fujimori Jun 2002 A1
20020090914 Kang et al. Jul 2002 A1
20020093478 Yeh Jul 2002 A1
20020095460 Benson Jul 2002 A1
20020098878 Mooney et al. Jul 2002 A1
20020101357 Gharapetian Aug 2002 A1
20020103635 Mesarovic et al. Aug 2002 A1
20020109710 Holtz et al. Aug 2002 A1
20020112244 Liou et al. Aug 2002 A1
20020114354 Sinha et al. Aug 2002 A1
20020114359 Ibaraki et al. Aug 2002 A1
20020124097 Isely et al. Sep 2002 A1
20020129156 Yoshikawa Sep 2002 A1
20020131398 Taylor Sep 2002 A1
20020131761 Kawasaki et al. Sep 2002 A1
20020136335 Liou et al. Sep 2002 A1
20020137505 Eiche et al. Sep 2002 A1
20020143547 Fay et al. Oct 2002 A1
20020143998 Rajagopal et al. Oct 2002 A1
20020150053 Gray et al. Oct 2002 A1
20020159596 Durand et al. Oct 2002 A1
20020163361 Parkin Nov 2002 A1
20020165721 Chang et al. Nov 2002 A1
20020165921 Sapieyevski Nov 2002 A1
20020168938 Chang Nov 2002 A1
20020173273 Spurgat et al. Nov 2002 A1
20020177411 Yajima et al. Nov 2002 A1
20020181355 Shikunami et al. Dec 2002 A1
20020184310 Traversat et al. Dec 2002 A1
20020188762 Tomassetti et al. Dec 2002 A1
20020194309 Carter et al. Dec 2002 A1
20020196951 Tsai Dec 2002 A1
20030002609 Faller et al. Jan 2003 A1
20030002689 Folio Jan 2003 A1
20030008616 Anderson Jan 2003 A1
20030014486 May Jan 2003 A1
20030018797 Dunning et al. Jan 2003 A1
20030020763 Mayer et al. Jan 2003 A1
20030023741 Tomassetti et al. Jan 2003 A1
20030031333 Cohen et al. Feb 2003 A1
20030035072 Hagg Feb 2003 A1
20030035444 Zwack Feb 2003 A1
20030041173 Hoyle Feb 2003 A1
20030041174 Wen et al. Feb 2003 A1
20030043856 Lakaniemi et al. Mar 2003 A1
20030043924 Haddad et al. Mar 2003 A1
20030055892 Huitema et al. Mar 2003 A1
20030061428 Garney et al. Mar 2003 A1
20030063755 Nourse et al. Apr 2003 A1
20030066094 Van Der Schaar et al. Apr 2003 A1
20030067437 McClintock et al. Apr 2003 A1
20030073432 Meade Apr 2003 A1
20030091322 Van May 2003 A1
20030097478 King May 2003 A1
20030099212 Anjum et al. May 2003 A1
20030099221 Rhee May 2003 A1
20030101253 Saito et al. May 2003 A1
20030103088 Dresti et al. Jun 2003 A1
20030110329 Higaki et al. Jun 2003 A1
20030126211 Anttila et al. Jul 2003 A1
20030135822 Evans Jul 2003 A1
20030157951 Hasty Aug 2003 A1
20030161479 Yang et al. Aug 2003 A1
20030167335 Alexander Sep 2003 A1
20030172123 Polan et al. Sep 2003 A1
20030177889 Koseki et al. Sep 2003 A1
20030179780 Walker et al. Sep 2003 A1
20030185400 Yoshizawa et al. Oct 2003 A1
20030195964 Mane Oct 2003 A1
20030198254 Sullivan et al. Oct 2003 A1
20030198255 Sullivan et al. Oct 2003 A1
20030198257 Sullivan et al. Oct 2003 A1
20030200001 Goddard Oct 2003 A1
20030204273 Dinker et al. Oct 2003 A1
20030204509 Dinker et al. Oct 2003 A1
20030210796 McCarty et al. Nov 2003 A1
20030212802 Rector et al. Nov 2003 A1
20030219007 Barrack et al. Nov 2003 A1
20030227478 Chatfield Dec 2003 A1
20030229900 Reisman Dec 2003 A1
20030231208 Hanon et al. Dec 2003 A1
20030231871 Ushimaru Dec 2003 A1
20030235304 Evans et al. Dec 2003 A1
20040001106 Deutscher et al. Jan 2004 A1
20040001484 Ozguner Jan 2004 A1
20040001591 Mani et al. Jan 2004 A1
20040008852 Also et al. Jan 2004 A1
20040010727 Fujinami Jan 2004 A1
20040012620 Buhler et al. Jan 2004 A1
20040014426 Moore Jan 2004 A1
20040015252 Aiso et al. Jan 2004 A1
20040019497 Volk et al. Jan 2004 A1
20040019807 Freund et al. Jan 2004 A1
20040019911 Gates et al. Jan 2004 A1
20040023697 Komura Feb 2004 A1
20040024478 Hans et al. Feb 2004 A1
20040024925 Cypher et al. Feb 2004 A1
20040027166 Mangum et al. Feb 2004 A1
20040032348 Lai et al. Feb 2004 A1
20040032421 Williamson et al. Feb 2004 A1
20040037433 Chen Feb 2004 A1
20040041836 Zaner et al. Mar 2004 A1
20040042629 Mellone et al. Mar 2004 A1
20040044742 Evron et al. Mar 2004 A1
20040048569 Kawamura Mar 2004 A1
20040059842 Hanson et al. Mar 2004 A1
20040059965 Marshall et al. Mar 2004 A1
20040066736 Kroeger Apr 2004 A1
20040071299 Yoshino Apr 2004 A1
20040075767 Neuman et al. Apr 2004 A1
20040078383 Mercer et al. Apr 2004 A1
20040080671 Siemens et al. Apr 2004 A1
20040093096 Huang et al. May 2004 A1
20040098754 Vella et al. May 2004 A1
20040111473 Lysenko et al. Jun 2004 A1
20040114771 Vaughan et al. Jun 2004 A1
20040117044 Konetski Jun 2004 A1
20040117462 Bodin et al. Jun 2004 A1
20040128701 Kaneko et al. Jul 2004 A1
20040131192 Metcalf Jul 2004 A1
20040133689 Vasisht Jul 2004 A1
20040143368 May et al. Jul 2004 A1
20040143852 Meyers Jul 2004 A1
20040147224 Lee Jul 2004 A1
20040148237 Bittmann et al. Jul 2004 A1
20040168081 Ladas et al. Aug 2004 A1
20040170383 Mazur Sep 2004 A1
20040171346 Lin Sep 2004 A1
20040177167 Iwamura et al. Sep 2004 A1
20040179554 Tsao Sep 2004 A1
20040183827 Putterman et al. Sep 2004 A1
20040185773 Gerber et al. Sep 2004 A1
20040203354 Yue Oct 2004 A1
20040203378 Powers Oct 2004 A1
20040203590 Shteyn Oct 2004 A1
20040208158 Fellman et al. Oct 2004 A1
20040213230 Douskalis et al. Oct 2004 A1
20040220687 Klotz et al. Nov 2004 A1
20040223622 Lindemann Nov 2004 A1
20040224638 Fadell et al. Nov 2004 A1
20040225389 Ledoux et al. Nov 2004 A1
20040228367 Mosig et al. Nov 2004 A1
20040248601 Chang Dec 2004 A1
20040249490 Sakai Dec 2004 A1
20040249965 Huggins et al. Dec 2004 A1
20040249982 Arnold et al. Dec 2004 A1
20040252400 Blank et al. Dec 2004 A1
20040253969 Nguyen et al. Dec 2004 A1
20040264717 Fujita et al. Dec 2004 A1
20050002535 Liu et al. Jan 2005 A1
20050010691 Oyadomari et al. Jan 2005 A1
20050011388 Kouznetsov Jan 2005 A1
20050013394 Rausch et al. Jan 2005 A1
20050015551 Eames et al. Jan 2005 A1
20050021470 Martin et al. Jan 2005 A1
20050021590 Debique et al. Jan 2005 A1
20050027821 Alexander et al. Feb 2005 A1
20050031135 Devantier et al. Feb 2005 A1
20050047605 Lee et al. Mar 2005 A1
20050058149 Howe Mar 2005 A1
20050060435 Xue et al. Mar 2005 A1
20050062637 El Zabadani et al. Mar 2005 A1
20050069153 Hall et al. Mar 2005 A1
20050081213 Suzuoki et al. Apr 2005 A1
20050100174 Howard et al. May 2005 A1
20050105052 McCormick et al. May 2005 A1
20050114538 Rose May 2005 A1
20050120128 Willes et al. Jun 2005 A1
20050125222 Brown et al. Jun 2005 A1
20050125357 Saadat et al. Jun 2005 A1
20050131558 Braithwaite et al. Jun 2005 A1
20050144284 Ludwig et al. Jun 2005 A1
20050147261 Yeh Jul 2005 A1
20050154766 Huang et al. Jul 2005 A1
20050159833 Giaimo et al. Jul 2005 A1
20050160270 Goldberg et al. Jul 2005 A1
20050166135 Burke et al. Jul 2005 A1
20050168630 Yamada et al. Aug 2005 A1
20050177256 Shintani et al. Aug 2005 A1
20050177643 Xu Aug 2005 A1
20050181348 Carey et al. Aug 2005 A1
20050195205 Abrams, Jr. Sep 2005 A1
20050195823 Chen et al. Sep 2005 A1
20050195999 Takemura et al. Sep 2005 A1
20050197725 Alexander et al. Sep 2005 A1
20050198574 Lamkin et al. Sep 2005 A1
20050201549 Dedieu et al. Sep 2005 A1
20050216556 Manion et al. Sep 2005 A1
20050254505 Chang et al. Nov 2005 A1
20050262217 Nonaka et al. Nov 2005 A1
20050266798 Moloney et al. Dec 2005 A1
20050266826 Vlad Dec 2005 A1
20050281255 Davies et al. Dec 2005 A1
20050283820 Richards et al. Dec 2005 A1
20050288805 Moore et al. Dec 2005 A1
20050289224 Deslippe et al. Dec 2005 A1
20050289244 Sahu et al. Dec 2005 A1
20060041616 Ludwig et al. Feb 2006 A1
20060041639 Lamkin et al. Feb 2006 A1
20060045281 Korneluk Mar 2006 A1
20060072489 Toyoshima Apr 2006 A1
20060095516 Wijeratne May 2006 A1
20060098936 Ikeda et al. May 2006 A1
20060119497 Miller et al. Jun 2006 A1
20060143236 Wu Jun 2006 A1
20060149402 Chung Jul 2006 A1
20060155721 Grunwald et al. Jul 2006 A1
20060173844 Zhang et al. Aug 2006 A1
20060179160 Uehara et al. Aug 2006 A1
20060193454 Abou-Chakra et al. Aug 2006 A1
20060193482 Harvey et al. Aug 2006 A1
20060199538 Eisenbach Sep 2006 A1
20060205349 Passier et al. Sep 2006 A1
20060222186 Paige et al. Oct 2006 A1
20060227985 Kawanami Oct 2006 A1
20060229752 Chung Oct 2006 A1
20060259649 Hsieh et al. Nov 2006 A1
20060270395 Dhawan Nov 2006 A1
20060294569 Chung Dec 2006 A1
20070003067 Gierl et al. Jan 2007 A1
20070003075 Cooper et al. Jan 2007 A1
20070022207 Millington et al. Jan 2007 A1
20070038999 Millington Feb 2007 A1
20070043847 Carter et al. Feb 2007 A1
20070047712 Gross et al. Mar 2007 A1
20070048713 Plastina et al. Mar 2007 A1
20070054680 Mo et al. Mar 2007 A1
20070071255 Schobben Mar 2007 A1
20070087686 Holm Apr 2007 A1
20070142022 Madonna et al. Jun 2007 A1
20070142944 Goldberg et al. Jun 2007 A1
20070143493 Mullig et al. Jun 2007 A1
20070169115 Ko et al. Jul 2007 A1
20070180137 Rajapakse Aug 2007 A1
20070189544 Rosenberg Aug 2007 A1
20070192156 Gauger Aug 2007 A1
20070206829 Weinans et al. Sep 2007 A1
20070220150 Garg Sep 2007 A1
20070223725 Neumann et al. Sep 2007 A1
20070249295 Ukita et al. Oct 2007 A1
20070265031 Koizumi et al. Nov 2007 A1
20070271388 Bowra et al. Nov 2007 A1
20070288610 Saint et al. Dec 2007 A1
20070299778 Haveson et al. Dec 2007 A1
20080002836 Moeller et al. Jan 2008 A1
20080007649 Bennett Jan 2008 A1
20080007650 Bennett Jan 2008 A1
20080007651 Bennett Jan 2008 A1
20080018785 Bennett Jan 2008 A1
20080022320 Ver Steeg Jan 2008 A1
20080025535 Rajapakse Jan 2008 A1
20080045140 Korhonen Feb 2008 A1
20080065232 Igoe Mar 2008 A1
20080066094 Igoe Mar 2008 A1
20080066120 Igoe Mar 2008 A1
20080072816 Riess et al. Mar 2008 A1
20080075295 Mayman et al. Mar 2008 A1
20080077261 Baudino Mar 2008 A1
20080077619 Gilley et al. Mar 2008 A1
20080077620 Gilley et al. Mar 2008 A1
20080086318 Gilley et al. Apr 2008 A1
20080091771 Allen et al. Apr 2008 A1
20080092204 Bryce et al. Apr 2008 A1
20080109852 Kretz et al. May 2008 A1
20080120429 Millington et al. May 2008 A1
20080126943 Parasnis et al. May 2008 A1
20080144861 Melanson et al. Jun 2008 A1
20080144864 Huon Jun 2008 A1
20080146289 Korneluk et al. Jun 2008 A1
20080152165 Zacchi Jun 2008 A1
20080159545 Takumai et al. Jul 2008 A1
20080162668 Miller Jul 2008 A1
20080189272 Powers et al. Aug 2008 A1
20080205070 Osada Aug 2008 A1
20080212786 Park Sep 2008 A1
20080215169 Debettencourt et al. Sep 2008 A1
20080242222 Bryce et al. Oct 2008 A1
20080247554 Caffrey Oct 2008 A1
20080263010 Roychoudhuri et al. Oct 2008 A1
20080291863 Agren Nov 2008 A1
20080303947 Ohnishi et al. Dec 2008 A1
20090011798 Yamada Jan 2009 A1
20090017868 Ueda et al. Jan 2009 A1
20090031336 Chavez et al. Jan 2009 A1
20090060219 Inohara Mar 2009 A1
20090070434 Himmelstein Mar 2009 A1
20090089327 Kalaboukis et al. Apr 2009 A1
20090097672 Buil et al. Apr 2009 A1
20090100189 Bahren et al. Apr 2009 A1
20090124289 Nishida May 2009 A1
20090157905 Davis Jun 2009 A1
20090164655 Pettersson et al. Jun 2009 A1
20090169030 Inohara Jul 2009 A1
20090180632 Goldberg et al. Jul 2009 A1
20090193345 Wensley et al. Jul 2009 A1
20090222115 Malcolm et al. Sep 2009 A1
20090228919 Zott et al. Sep 2009 A1
20090232326 Gordon et al. Sep 2009 A1
20090251604 Iyer Oct 2009 A1
20100004983 Dickerson et al. Jan 2010 A1
20100010651 Kirkeby et al. Jan 2010 A1
20100031366 Knight et al. Feb 2010 A1
20100049835 Ko et al. Feb 2010 A1
20100052843 Cannistraro Mar 2010 A1
20100067716 Katayama Mar 2010 A1
20100087089 Struthers et al. Apr 2010 A1
20100142735 Yoon et al. Jun 2010 A1
20100153097 Hotho et al. Jun 2010 A1
20100228740 Cannistraro et al. Sep 2010 A1
20100272270 Chaikin et al. Oct 2010 A1
20100284389 Ramsay et al. Nov 2010 A1
20100290643 Mihelich et al. Nov 2010 A1
20100299639 Ramsay et al. Nov 2010 A1
20110001632 Hohorst Jan 2011 A1
20110002487 Panther et al. Jan 2011 A1
20110044476 Burlingame et al. Feb 2011 A1
20110066943 Brillon et al. Mar 2011 A1
20110110533 Choi et al. May 2011 A1
20110170710 Son Jul 2011 A1
20110228944 Croghan et al. Sep 2011 A1
20110299696 Holmgren et al. Dec 2011 A1
20110316768 McRae Dec 2011 A1
20120029671 Millington et al. Feb 2012 A1
20120030366 Collart et al. Feb 2012 A1
20120047435 Holladay et al. Feb 2012 A1
20120051558 Kim et al. Mar 2012 A1
20120051567 Castor-Perry Mar 2012 A1
20120060046 Millington Mar 2012 A1
20120127831 Gicklhorn et al. May 2012 A1
20120129446 Ko et al. May 2012 A1
20120148075 Goh et al. Jun 2012 A1
20120185771 Rothkopf et al. Jul 2012 A1
20120192071 Millington Jul 2012 A1
20120207290 Moyers et al. Aug 2012 A1
20120237054 Eo et al. Sep 2012 A1
20120263325 Freeman et al. Oct 2012 A1
20120281058 Laney et al. Nov 2012 A1
20120290621 Heitz, III et al. Nov 2012 A1
20130010970 Hegarty et al. Jan 2013 A1
20130018960 Knysz et al. Jan 2013 A1
20130028443 Pance et al. Jan 2013 A1
20130031475 Maor et al. Jan 2013 A1
20130038726 Kim Feb 2013 A1
20130041954 Kim et al. Feb 2013 A1
20130047084 Sanders et al. Feb 2013 A1
20130051572 Goh et al. Feb 2013 A1
20130052940 Brillhart et al. Feb 2013 A1
20130070093 Rivera et al. Mar 2013 A1
20130080599 Ko et al. Mar 2013 A1
20130094670 Millington Apr 2013 A1
20130124664 Fonseca, Jr. et al. May 2013 A1
20130129122 Johnson et al. May 2013 A1
20130132837 Mead et al. May 2013 A1
20130159126 Elkady Jun 2013 A1
20130167029 Friesen et al. Jun 2013 A1
20130174100 Seymour et al. Jul 2013 A1
20130174223 Dykeman et al. Jul 2013 A1
20130179163 Herbig et al. Jul 2013 A1
20130191454 Oliver et al. Jul 2013 A1
20130197682 Millington Aug 2013 A1
20130208911 Millington Aug 2013 A1
20130208921 Millington Aug 2013 A1
20130226323 Millington Aug 2013 A1
20130230175 Bech et al. Sep 2013 A1
20130232416 Millington Sep 2013 A1
20130236029 Millington Sep 2013 A1
20130243199 Kallai et al. Sep 2013 A1
20130253679 Lambourne Sep 2013 A1
20130253934 Parekh et al. Sep 2013 A1
20130259254 Xiang et al. Oct 2013 A1
20130279706 Marti Oct 2013 A1
20130287186 Quady Oct 2013 A1
20130290504 Quady Oct 2013 A1
20130293345 Lambourne Nov 2013 A1
20140006483 Garmark et al. Jan 2014 A1
20140016784 Sen et al. Jan 2014 A1
20140016786 Sen Jan 2014 A1
20140016802 Sen Jan 2014 A1
20140023196 Xiang et al. Jan 2014 A1
20140037097 Labosco Feb 2014 A1
20140064501 Olsen et al. Mar 2014 A1
20140075308 Sanders et al. Mar 2014 A1
20140075311 Boettcher et al. Mar 2014 A1
20140079242 Nguyen et al. Mar 2014 A1
20140108929 Garmark et al. Apr 2014 A1
20140112481 Li et al. Apr 2014 A1
20140123005 Forstall et al. May 2014 A1
20140140530 Gomes-Casseres et al. May 2014 A1
20140161265 Chaikin et al. Jun 2014 A1
20140181569 Millington et al. Jun 2014 A1
20140219456 Morrell et al. Aug 2014 A1
20140226823 Sen et al. Aug 2014 A1
20140242913 Pang Aug 2014 A1
20140256260 Ueda et al. Sep 2014 A1
20140267148 Luna et al. Sep 2014 A1
20140270202 Ivanov et al. Sep 2014 A1
20140273859 Luna et al. Sep 2014 A1
20140279889 Luna Sep 2014 A1
20140285313 Luna et al. Sep 2014 A1
20140286496 Luna et al. Sep 2014 A1
20140294200 Baumgarte et al. Oct 2014 A1
20140298174 Ikonomov Oct 2014 A1
20140323036 Daley et al. Oct 2014 A1
20140344689 Scott et al. Nov 2014 A1
20140355768 Sen et al. Dec 2014 A1
20140355794 Morrell et al. Dec 2014 A1
20140378056 Liu Dec 2014 A1
20150019670 Redmann Jan 2015 A1
20150026613 Kwon et al. Jan 2015 A1
20150032844 Tarr et al. Jan 2015 A1
20150043736 Olsen et al. Feb 2015 A1
20150049248 Wang et al. Feb 2015 A1
20150063610 Mossner Mar 2015 A1
20150074527 Sevigny et al. Mar 2015 A1
20150074528 Sakalowsky et al. Mar 2015 A1
20150098576 Sundaresan et al. Apr 2015 A1
20150139210 Marin et al. May 2015 A1
20150146886 Baumgarte May 2015 A1
20150201274 Ellner et al. Jul 2015 A1
20150256954 Carlsson et al. Sep 2015 A1
20150281866 Williams et al. Oct 2015 A1
20150286360 Wachter et al. Oct 2015 A1
20150304288 Balasaygun et al. Oct 2015 A1
20150365987 Weel Dec 2015 A1
20170188152 Watson et al. Jun 2017 A1
Foreign Referenced Citations (76)
Number Date Country
2320451 Mar 2001 CA
1598767 Mar 2005 CN
101095372 Dec 2007 CN
101292500 Oct 2008 CN
101785182 Jul 2010 CN
0251584 Jan 1988 EP
0672985 Sep 1995 EP
0772374 May 1997 EP
1111527 Jun 2001 EP
1122931 Aug 2001 EP
1133896 Aug 2002 EP
1312188 May 2003 EP
1389853 Feb 2004 EP
2713281 Apr 2004 EP
1517464 Mar 2005 EP
0895427 Jan 2006 EP
1416687 Aug 2006 EP
1410686 Mar 2008 EP
2043381 Apr 2009 EP
2161950 Mar 2010 EP
1825713 Oct 2012 EP
0742674 Apr 2014 EP
2591617 Jun 2014 EP
2860992 Apr 2015 EP
2284327 May 1995 GB
2338374 Dec 1999 GB
2379533 Mar 2003 GB
2486183 Jun 2012 GB
63269633 Nov 1988 JP
07-210129 Aug 1995 JP
2000149391 May 2000 JP
2001034951 Feb 2001 JP
2002111817 Apr 2002 JP
2002123267 Apr 2002 JP
2002358241 Dec 2002 JP
2003037585 Feb 2003 JP
2003506765 Feb 2003 JP
2003101958 Apr 2003 JP
2003169089 Jun 2003 JP
2004193868 Jul 2004 JP
2005108427 Apr 2005 JP
2005136457 May 2005 JP
2007241652 Sep 2007 JP
2007288405 Nov 2007 JP
2009506603 Feb 2009 JP
2009135750 Jun 2009 JP
2009218888 Sep 2009 JP
2009535708 Oct 2009 JP
2009538006 Oct 2009 JP
2011010183 Jan 2011 JP
2011130496 Jun 2011 JP
2011176581 Sep 2011 JP
20060030713 Apr 2006 KR
439027 Jun 2001 TW
199525313 Sep 1995 WO
9709756 Mar 1997 WO
1999023560 May 1999 WO
199961985 Dec 1999 WO
0019693 Apr 2000 WO
0110125 Feb 2001 WO
200153994 Jul 2001 WO
02073851 Sep 2002 WO
03093950 Nov 2003 WO
2005013047 Feb 2005 WO
2007023120 Mar 2007 WO
2007127485 Nov 2007 WO
2007131555 Nov 2007 WO
2007135581 Nov 2007 WO
2008082350 Jul 2008 WO
2008114389 Sep 2008 WO
2012050927 Apr 2012 WO
2012137190 Oct 2012 WO
2013012582 Jan 2013 WO
2014004182 Jan 2014 WO
2014149533 Sep 2014 WO
2015024881 Feb 2015 WO
Non-Patent Literature Citations (629)
Entry
Notice of Allowance dated Sep. 6, 2013, issued in connection with U.S. Appl. No. 13/619,237, filed Sep. 14, 2012, 10 pages.
Notice of Allowance dated Sep. 6, 2016, issued in connection with U.S. Appl. No. 15/134,767, filed Apr. 21, 2016, 7 pages.
Notice of Allowance dated Apr. 7, 2016, issued in connection with U.S. Appl. No. 13/533,105, filed Jun. 26, 2012, 40 pages.
Notice of Allowance dated Oct. 7, 2015, issued in connection with U.S. Appl. No. 14/184,526, filed Feb. 19, 2014, 7 pages.
Notice of Allowance dated Oct. 9, 2015, issued in connection with U.S. Appl. No. 13/435,739, filed Mar. 30, 2012, 4 pages.
Notice of Allowance dated Sep. 9, 2016, issued in connection with U.S. Appl. No. 15/134,761, filed Apr. 21, 2016, 7 pages.
Notice of Allowance dated Mar. 1, 2018, issued in connection with U.S. Appl. No. 14/619,813, filed Feb. 11, 2015, 7 pages.
Notice of Allowance dated Aug. 10, 2015, issued in connection with U.S. Appl. No. 13/848,904, filed Mar. 22, 2013, 9 pages.
Notice of Allowance dated Jul. 10, 2015, issued in connection with U.S. Appl. No. 13/013,740, filed Jan. 25, 2011, 7 pages.
Notice of Allowance dated Mar. 10, 2016, issued in connection with U.S. Appl. No. 14/937,523, filed Nov. 10, 2015, 5 pages.
Notice of Allowance dated Nov. 10, 2011, issued in connection with U.S. Appl. No. 11/906,702, filed Oct. 2, 2007, 17 pages.
Notice of Allowance dated Sep. 10, 2014, issued in connection with U.S. Appl. No. 13/892,230, filed May 10, 2013, 5 pages.
Notice of Allowance dated Sep. 10, 2018, issued in connection with U.S. Appl. No. 14/629,937, filed Feb. 24, 2015, 7 pages.
Notice of Allowance dated Apr. 11, 2016, issued in connection with U.S. Appl. No. 13/864,247, filed Apr. 17, 2013, 21 pages.
Notice of Allowance dated Jan. 11, 2016, issued in connection with U.S. Appl. No. 14/564,544, filed Dec. 9, 2014, 5 pages.
Notice of Allowance dated Jul. 11, 2017, issued in connection with U.S. Appl. No. 14/825,961, filed Aug. 13, 2015, 5 pages.
Notice of Allowance dated Aug. 12, 2015, issued in connection with U.S. Appl. No. 13/435,739, filed Mar. 30, 2012, 27 pages.
Notice of Allowance dated Jun. 12, 2014, issued in connection with U.S. Appl. No. 13/896,829, filed May 17, 2013, 5 pages.
Notice of Allowance dated Jul. 13, 2015, issued in connection with U.S. Appl. No. 14/184,526, filed Feb. 19, 2014, 22 pages.
Notice of Allowance dated May 13, 2015, issued in connection with U.S. Appl. No. 14/299,847, filed Jun. 9, 2014, 10 pages.
Notice of Allowance dated Nov. 13, 2013, issued in connection with U.S. Appl. No. 13/724,048, filed Dec. 21, 2012, 7 pages.
Notice of Allowance dated Nov. 13, 2017, issued in connection with U.S. Appl. No. 14/563,515, filed Dec. 8, 2014, 11 pages.
Notice of Allowance dated Oct. 13, 2015, issued in connection with U.S. Appl. No. 13/864,251, filed Apr. 17, 2013, 7 pages.
Notice of Allowance dated Jun. 14, 2012, issued in connection with U.S. Appl. No. 12/035,112, filed Feb. 21, 2008, 9 pages.
Notice of Allowance dated Jul. 15, 2015, issued in connection with U.S. Appl. No. 13/705,174, filed Dec. 5, 2012, 18 pages.
Notice of Allowance dated Mar. 15, 2016, issued in connection with U.S. Appl. No. 14/937,571, filed Nov. 10, 2015, 5 pages.
Notice of Allowance dated Jun. 16, 2009, issued in connection with U.S. Appl. No. 10/861,653, filed Jun. 5, 2004, 11 pages.
Notice of Allowance dated May 16, 2017, issued in connection with U.S. Appl. No. 15/228,685, filed Aug. 4, 2016, 10 pages.
Notice of Allowance dated Jul. 17, 2015, issued in connection with U.S. Appl. No. 13/864,251, filed Apr. 17, 2013, 20 pages.
Notice of Allowance dated Aug. 19, 2016, issued in connection with U.S. Appl. No. 14/619,813, filed Feb. 11, 2015, 9 pages.
Notice of Allowance dated May 19, 2015, issued in connection with U.S. Appl. No. 13/907,666, filed May 31, 2013, 7 pages.
Notice of Allowance dated Oct. 19, 2016, issued in connection with U.S. Appl. No. 14/290,493, filed May 29, 2014, 14 pages.
Notice of Allowance dated Jan. 20, 2016, issued in connection with U.S. Appl. No. 14/465,457, filed Aug. 21, 2014, 10 pages.
Notice of Allowance dated Oct. 21, 2015, issued in connection with U.S. Appl. No. 14/174,244, filed Feb. 6, 2014, 5 pages.
Notice of Allowance dated Oct. 21, 2015, issued in connection with U.S. Appl. No. 14/174,253, filed Feb. 6, 2014, 6 pages.
Notice of Allowance dated Sep. 21, 2015, issued in connection with U.S. Appl. No. 13/297,000, filed Nov. 15, 2011, 11 pages.
Notice of Allowance dated Jan. 22, 2015, issued in connection with U.S. Appl. No. 13/630,565, filed Sep. 28, 2012, 7 pages.
Notice of Allowance dated Sep. 22, 2015, issued in connection with U.S. Appl. No. 13/888,203, filed May 6, 2013, 7 pages.
Notice of Allowance dated May 24, 2017, issued in connection with U.S. Appl. No. 14/806,070, filed Jul. 22, 2015, 5 pages.
Notice of Allowance dated Oct. 24, 2016, issued in connection with U.S. Appl. No. 15/134,761, filed Apr. 21, 2016, 7 pages.
Notice of Allowance dated Oct. 24, 2016, issued in connection with U.S. Appl. No. 15/134,767, filed Apr. 21, 2016, 7 pages.
Notice of Allowance dated Sep. 24, 2015, issued in connection with U.S. Appl. No. 13/705,174, filed Dec. 5, 2012, 7 pages.
Notice of Allowance dated Sep. 24, 2015, issued in connection with U.S. Appl. No. 14/184,935, filed Feb. 20, 2014, 7 pages.
Notice of Allowance dated Aug. 25, 2017, issued in connection with U.S. Appl. No. 14/505,966, filed Oct. 3, 2014, 5 pages.
Notice of Allowance dated Sep. 25, 2014, issued in connection with U.S. Appl. No. 14/176,808, filed Feb. 10, 2014, 5 pages.
Notice of Allowance dated Aug. 27, 2015, issued in connection with U.S. Appl. No. 13/705,177, filed Dec. 5, 2012, 34 pages.
Notice of Allowance dated Aug. 27, 2015, issued in connection with U.S. Appl. No. 14/505,027, filed Oct. 2, 2014, 18 pages.
Notice of Allowance dated Dec. 27, 2011, issued in connection with U.S. Appl. No. 10/816,217, filed Apr. 1, 2004, 15 pages.
Notice of Allowance dated Oct. 17, 2015, issued in connection with U.S. Appl. No. 14/299,847, filed Jun. 9, 2014, 5 pages.
1otice of Allowance dated Oct. 28, 2014, issued in connection withU.S. Appl. No. 13/896,037, filed May 16, 2013, 7 pages.
Non-Final Office Action dated Jan. 3, 2017, issued in connection with U.S. Appl. No. 14/808,397, filed Jul. 24, 2015, 11 pages.
Non-Final Office Action dated Jun. 3, 2015, issued in connection with U.S. Appl. No. 14/564,544, filed Dec. 9, 2014, 7 pages.
Non-Final Office Action dated Nov. 3, 2016, issued in connection with U.S. Appl. No. 14/184,528, filed Feb. 19, 2014, 17 pages.
Non-Final Office Action dated Jan. 4, 2017, issued in connection with U.S. Appl. No. 14/825,961, filed Aug. 13, 2015, 11 pages.
Non-Final Office Action dated Jun. 4, 2015, issued in connection with U.S. Appl. No. 13/871,795, filed Apr. 26, 2013, 16 pages.
Non-Final Office Action dated Mar. 4, 2015, issued in connection with U.S. Appl. No. 13/435,776, filed Mar. 30, 2012, 16 pages.
Non-Final Office Action dated Oct. 4, 2016, issued in connection with U.S. Appl. No. 15/089,758, filed Apr. 4, 2016, 9 pages.
Non-Final Office Action dated Oct. 5, 2016, issued in connection with U.S. Appl. No. 13/864,250, filed Apr. 17, 2013, 10 pages.
Non-Final Office Action dated Oct. 5, 2016, issued in connection with U.S. Appl. No. 13/864,252, filed Apr. 17, 2013, 11 pages.
Non-Final Office Action dated Oct. 6, 2016, issued in connection with U.S. Appl. No. 15/088,678, filed Apr. 1, 2016, 9 pages.
Non-Final Office Action dated Jul. 7, 2015, issued in connection with U.S. Appl. No. 14/174,244, filed Feb. 6, 2014, 9 pages.
Non-Final Office Action dated Oct. 7, 2016, issued in connection with U.S. Appl. No. 15/156,392, filed May 17, 2016, 8 pages.
Non-Final Office Action dated Mar. 8, 2011, issued in connection with U.S. Appl. No. 11/853,790, filed Sep. 11, 2007, 10 pages.
Non-Final Office Action dated Mar. 8, 2016, issued in connection with U.S. Appl. No. 13/848,921, filed Mar. 22, 2013, 13 pages.
Non-Final Office Action dated Aug. 9, 2016, issued in connection with U.S. Appl. No. 13/871,795, filed Apr. 26, 2013, 31 pages.
Non-Final Office Action dated May 9, 2014, issued in connection with U.S. Appl. No. 13/892,230, filed May 10, 2013, 10 pages.
Non-Final Office Action dated Feb. 10, 2016, issued in connection with U.S. Appl. No. 14/937,571, filed Nov. 10, 2015, 9 pages.
Non-Final Office Action dated Mar. 10, 2011, issued in connection with U.S. Appl. No. 12/035,112, filed Feb. 21, 2008, 12 pages.
Non-Final Office Action dated May 10, 2016, issued in connection with U.S. Appl. No. 14/504,812, filed Oct. 2, 2014, 22 pages.
Non-Final Office Action dated Nov. 10, 2016, issued in connection with U.S. Appl. No. 15/243,355, filed Aug. 22, 2016, 11 pages.
Non-Final Office Action dated Jun. 11, 2018, issued in connection with U.S. Appl. No. 15/405,931, filed Jan. 13, 2017, 14 pages.
Non-Final Office Action dated Dec. 12, 2016, issued in connection with U.S. Appl. No. 15/343,019, filed Nov. 3, 2016, 8 pages.
Non-Final Office Action dated Jun. 12, 2015, issued in connection with U.S. Appl. No. 13/848,932, filed Mar. 22, 2013, 16 pages.
Non-Final Office Action dated Mar. 12, 2015, issued in connection with U.S. Appl. No. 13/705,174, filed Dec. 5, 2012, 13 pages.
Non-Final Office Action dated Oct. 12, 2016, issued in connection with U.S. Appl. No. 14/505,966, filed Oct. 3, 2014, 10 pages.
Non-Final Office Action dated Feb. 13, 2014, issued in connection with U.S. Appl. No. 13/896,037, filed May 16, 2013, 10 pages.
Non-Final Office Action dated Feb. 13, 2015, issued in connection with U.S. Appl. No. 13/013,740, filed Jan. 25, 2011, 14 pages.
Non-Final Office Action dated Jan. 13, 2016, issued in connection with U.S. Appl. No. 14/184,528, filed Feb. 19, 2014, 14 pages.
Non-Final Office Action dated Jun. 13, 2016, issued in connection with U.S. Appl. No. 14/620,937, filed Feb. 12, 2015, 12 pages.
Non-Final Office Action dated Jun. 13, 2016, issued in connection with U.S. Appl. No. 15/134,761, filed Apr. 21, 2016, 10 pages.
Non-Final Office Action dated Mar. 13, 2015, issued in connection with U.S. Appl. No. 13/705,177, filed Dec. 5, 2012, 15 pages.
Non-Final Office Action dated May 14, 2018, issued in connection with U.S. Appl. No. 15/228,812, filed Aug. 4, 2016, 15 pages.
Non-Final Office Action dated Dec. 15, 2016, issued in connection with U.S. Appl. No. 13/458,558, filed Apr. 27, 2012, 12 pages.
Non-Final Office Action dated Jul. 15, 2015, issued in connection with U.S. Appl. No. 14/174,253, filed Feb. 6, 2014, 9 pages.
Non-Final Office Action dated Nov. 16, 2016, issued in connection with U.S. Appl. No. 15/228,639, filed Aug. 4, 2016, 15 pages.
Non-Final Office Action dated Dec. 17, 2015, issued in connection with U.S. Appl. No. 13/458,558, filed Apr. 27, 2012, 10 pages.
Non-Final Office Action dated Nov. 17, 2014, issued in connection with U.S. Appl. No. 13/864,247, filed Apr. 17, 2013, 11 pages.
Non-Final Office Action dated Nov. 17, 2016, issued in connection with U.S. Appl. No. 14/620,937, filed Feb. 12, 2015, 14 pages.
Non-Final Office Action dated Feb. 18, 2009, issued in connection with U.S. Appl. No. 10/861,653, filed Jun. 5, 2004, 18 pages.
Non-Final Office Action dated Nov. 18, 2014, issued in connection with U.S. Appl. No. 13/435,739, filed Mar. 30, 2012, 10 pages.
Non-Final Office Action dated Jan. 19, 2018, issued in connection with U.S. Appl. No. 14/629,937, filed Feb. 24, 2015, 14 pages.
Non-Final Office Action dated Jun. 19, 2015, issued in connection with U.S. Appl. No. 13/533,105, filed Jun. 26, 2012, 38 pages.
Non-Final Office Action dated Nov. 19, 2014, issued in connection with U.S. Appl. No. 13/848,921, filed Mar. 22, 2013, 9 pages.
Non-Final Office Action dated Aug. 20, 2009, issued in connection with U.S. Appl. No. 11/906,702, filed Oct. 2, 2007, 27 pages.
Non-Final Office Action dated Oct. 20, 2016, issued in connection with U.S. Appl. No. 14/563,515, filed Dec. 8, 2014, 10 pages.
Non-Final Office Action dated Sep. 21, 2016, issued in connection with U.S. Appl. No. 15/080,591, filed Mar. 25, 2016, 9 pages.
Non-Final Office Action dated Sep. 21, 2016, issued in connection with U.S. Appl. No. 15/080,716, filed Mar. 25, 2016, 8 pages.
Non-Final Office Action dated Sep. 21, 2016, issued in connection with U.S. Appl. No. 15/088,283, filed Apr. 1, 2016, 9 pages.
Non-Final Office Action dated Sep. 21, 2016, issued in connection with U.S. Appl. No. 15/088,532, filed Apr. 1, 2016, 9 pages.
Non-Final Office Action dated Aug. 22, 2018, issued in connection with U.S. Appl. No. 15/487,686, filed Apr. 14, 2017, 13 pages.
LinkSys by Cisco, Wireless Home Audio Player, Wireless-N Music Extender DMP100 Quick Installation Guide, Copyright 2009, 32 pages.
LinkSys by Cisco, Wireless Home Audio Player, Wireless-N Music Extender DMP100 User Guide, Copyright 2008, 65 pages.
Linux SDK for UPnP Devices v. 1.2 (Sep. 6, 2002) (101 pages).
Liu et al., “A synchronization control scheme for real-time streaming multimedia applications,” Packet Video, 2003, 10 pages, vol. 2003.
Liu et al., “Adaptive Delay Concealment for Internet Voice Applications with Packet-Based Time-Scale Modification,” Information Technologies 2000, pp. 91-102.
Louderback, Jim, “Affordable Audio Receiver Furnishes Homes With MP3,” TechTV Vault. Jun. 28, 2000 retrieved Jul. 10, 2014, 2 pages.
Maniactools, “Identify Duplicate Files by Sound,” Sep. 28, 2010, http://www.maniactools.com/soft/music-duplicate-remover/identify-duplicate-files-by-sound.shtml.
MediaRenderer:1 Device Template Version 1.01 for UPnP, Version 1.0 (Jun. 25, 2002) (12 pages).
MediaServer:1 Device Template Version 1.01 for UPnP, Version 1.0 (Jun. 25, 2002) (12 pages).
Microsoft, Universal Plug and Play (UPnP) Client Support (“Microsoft UPnP”) (Aug. 2001) (D+M_0402007-24) (18 pages).
Microsoft Window's XP Reviewer's Guide (Aug. 2001) (D+M_0402225-85) (61 pages).
“Microsoft Windows XP File and Printer Share with Microsoft Windows” Microsoft Windows XP Technical Article, 2003, 65 pages.
Mills David L., “Network Time Protocol (Version 3) Specification, Implementation and Analysis,” Network Working Group, Mar. 1992, 7 pages.
Mills, David L., “Precision Synchronization of Computer Network Clocks,” ACM SIGCOMM Computer Communication Review, 1994, pp. 28-43, vol. 24, No. 2.
“Model MRC44 Four Zone—Four Source Audio/Video Controller/Amplifier System,” Xantech Corporation, 2002, 52 pages.
Motorola, “Simplefi, Wireless Digital Audio Receiver, Installation and User Guide,” Dec. 31, 2001, 111 pages.
“SMPTE Made Simple: A Time Code Tutor by Timeline,” 1996, 46 pages.
Network Time Protocol (NTP), RFC 1305 (Mar. 1992) (D+M_0397417-536) (120 pages).
“NexSys Software v.3 Manual,” Crest Audio, Inc., 1997, 76 pages.
Niederst, Jennifer “O'Reilly Web Design in a Nutshell,” Second Edition, Sep. 2001, 678 pages.
Nilsson, M., “ID3 Tag Version 2,” Mar. 26, 1998, 28 pages.
Non-Final Office Action dated May 1, 2014, issued in connection with U.S. Appl. No. 14/184,522, filed Feb. 19, 2014, 31 pages.
Non-Final Office Action dated Dec. 5, 2013, issued in connection with U.S. Appl. No. 13/827,653, filed Mar. 14, 2013, 28 pages.
Non-Final Office Action dated Jan. 5, 2012, issued in connection with U.S. Appl. No. 13/298,090, filed Nov. 16, 2011, 40 pages.
Non-Final Office Action dated May 6, 2014, issued in connection with U.S. Appl. No. 13/705,176, filed Dec. 5, 2012, 23 pages.
Non-Final Office Action dated Jan. 7, 2014, issued in connection with U.S. Appl. No. 13/896,829, filed May 17, 2013, 11 pages.
Non-Final Office Action dated Sep. 7, 2016, issued in connection with U.S. Appl. No. 13/864,248, filed Apr. 17, 2013, 12 pages.
Non-final Office Action dated Apr. 10, 2013, issued in connection with U.S. Appl. No. 13/619,237, filed Sep. 14, 2012, 10 pages.
Non-Final Office Action dated Feb. 10, 2014, issued in connection with U.S. Appl. No. 13/083,499, filed Apr. 8, 2011, 12 pages.
Non-Final Office Action dated May 12, 2014, issued in connection with U.S. Appl. No. 14/184,528, filed Feb. 19, 2014, 23 pages.
Non-Final Office Action dated May 14, 2014, issued in connection with U.S. Appl. No. 13/848,932, filed Mar. 22, 2013, 14 pages.
Non-Final Office Action dated Jun. 17, 2014, issued in connection with U.S. Appl. No. 14/176,808, filed Feb. 10, 2014, 6 pages.
Non-Final Office Action dated Dec. 18, 2013, issued in connection with U.S. Appl. No. 13/907,666, filed May 31, 2013, 12 pages.
Non-Final Office Action dated Jan. 18, 2008, issued in connection with U.S. Appl. No. 10/816,217, filed Apr. 1, 2004, 38 pages.
Non-Final Office Action dated Apr. 19, 2010, in connection with U.S. Appl. No. 11/801,468, filed May 9, 2007, 16 pages.
Non-Final Office Action dated Mar. 19, 2013, issued in connection with U.S. Appl. No. 13/724,048, filed Dec. 21, 2012, 9 pages.
Non-Final Office Action dated Jun. 21, 2011, issued in connection with U.S. Appl. No. 10/816,217, filed Apr. 1, 2004, 13 pages.
Non-Final Office Action dated Jan. 22, 2009, issued in connection with U.S. Appl. No. 10/816,217, filed Apr. 1, 2004, 18 pages.
Non-Final Office Action dated Jul. 23, 2014, issued in connection with U.S. Appl. No. 14/256,434, filed Apr. 18, 2014, 12 pages.
Non-Final Office Action dated Jul. 25, 2014, issued in connection with U.S. Appl. No. 14/184,526, filed Feb. 19, 2014, 9 pages.
Non-Final Office Action dated Jul. 25, 2014, issued in connection with U.S. Appl. No. 14/184,935, filed Feb. 20, 2014, 11 pages.
Non-Final Office Action dated Jun. 25, 2010, issued in connection with U.S. Appl. No. 10/816,217, filed Apr. 1, 2004, 17 pages.
Non-Final Office Action dated Nov. 25, 2013, issued in connection with U.S. Appl. No. 13/533,105, filed Jun. 26, 2012, 19 pages.
Non-Final Office Action dated May 27, 2014, issued in connection with U.S. Appl. No. 14/186,850, filed Feb. 21, 2014, 13 pages.
Non-Final Office Action dated Feb. 29, 2012, issued in connection with U.S. Appl. No. 13/297,000, filed Nov. 15, 2011, 10 pages.
Non-Final Office Action dated Nov. 29, 2010, issued in connection with U.S. Appl. No. 11/801,468, filed May 9, 2007, 17 pages.
Non-Final Office Action dated Jul. 30, 2013 issued in connection with U.S. Appl. No. 13/724,048, filed Dec. 21, 2012, 7 pages.
Non-Final Office Action dated Jul. 31, 2014, issued in connection with U.S. Appl. No. 13/533,105, filed Jun. 26, 2012, 31 pages.
Non-Final Office Action dated Dec. 1, 2014, issued in connection with U.S. Appl. No. 14/516,867, filed Oct. 17, 2014, 11 pages.
Non-Final Office Action dated Jun. 1, 2016, issued in connection with U.S. Appl. No. 14/184,522, filed Feb. 19, 2014, 21 pages.
Breebaart et al., “Multi-Channel Goes Mobile: MPEG Surround Binaural Rendering,” AES 29th International Conference, Sep. 2-4, 2006, pp. 1-13.
Bretl W.E., et al., MPEG2 Tutorial [online], 2000 [retrieved on Jan. 13, 2009] Retrieved from the Internet:(http://www.bretl.com/mpeghtml/MPEGindex.htm), pp. 1-23.
Buerk et al., “AVTransport:1 Service Template Version 1.01,” Contributing Members of the UPnP Forum, Jun. 25, 2002, 67 pages.
Canadian Intellectual Property Office, Canadian Office Action dated Apr. 4, 2016, issued in connection with Canadian Patent Application No. 2,842,342, 5 pages.
Canadian Intellectual Property Office, Canadian Office Action dated Sep. 14, 2015, issued in connection with Canadian Patent Application No. 2,842,342, 2 pages.
Canadian Patent Office, Canadian Office Action dated Aug. 30, 2017, issued in connection with CA Application No. 2947275, 5 pages.
Canadian Patent Office, Office Action dated Apr. 10, 2015, issued in connection with Canadian Patent Application No. 2,832,542, 3 pages.
Cen et al., “A Distributed Real-Time MPEG Video Audio Player,” Department of Computer Science and Engineering, Oregon Graduate Institute of Science and Technology, 1995, 12 pages.
Chakrabarti et al., “A Remotely Controlled Bluetooth Enabled Environment,” IEEE, 2004, pp. 77-81.
Change Notification: Agere Systems WaveLan Multimode Reference Design (D2 to D3), AVAGO0042, Agere Systems, Nov. 2004, 2 pages.
Chinese Patent Office, First Office Action dated Oct. 12, 2018, issued in connection with Chinese Application No. 201610804134.8, 10 pages.
Chinese Patent Office, Office Action dated Jul. 5, 2016, issued in connection with Chinese Patent Application No. 201380044380.2, 25 pages.
Chinese Patent Office, Office Action dated Nov. 27, 2015, issued in connection with Chinese Patent Application No. 201280028038.9, 26 pages.
Connection Manager: 1 Service Template Version 1.01 for UPnP, Version 1.0 (Jun. 25, 2002) (25 pages).
ContentDirectory:1 Service Template Version 1.01 for UPnP, Version 1.0 (Jun. 25, 2002) (89 pages).
Corrected Notice of Allowance dated Mar. 12, 2015, issued in connection with U.S. Appl. No. 13/630,565, filed Sep. 28, 2012, 4 pages.
Corrected Notice of Allowance dated Aug. 19, 2015, issued in connection with U.S. Appl. No. 13/907,666, filed May 31, 2013, 2 pages.
Corrected Notice of Allowance dated Oct. 30, 2015, issued in connection with U.S. Appl. No. 13/013,740, filed Jan. 25, 2011, 2 pages.
Corrected Notice of Allowance dated Dec. 6, 2017, issued in connection with U.S. Appl. No. 15/228,685, filed Aug. 4, 2016, 5 pages.
Creative, “Connecting Bluetooth Devices with Creative D200,” http://support.creative.com/kb/ShowArticle.aspx?url=http://ask.creative.com:80/SRVS/CGI-BIN/WEBCGI.EXE/,/?St=106,E=0000000000396859016,K=9377,Sxi=8,VARSET=ws:http://us.creative.com,case=63350>, available on Nov. 28, 2011, 2 pages.
Crown PIP Manual available for sale at least 2004, 68 pages.
Dannenberg et al., “A. System Supporting Flexible Distributed Real-Time Music Processing,” Proceedings of the 2001 International Computer Music Conference, 2001, 4 pages.
Dannenberg, Roger B., “Remote Access to Interactive Media,” Proceedings of the SPIE 1785, 1993, pp. 230-237.
Day, Rebecca, “Going Elan!” Primedia Inc., 2003, 4 pages.
Deep-Sleep Implementation in WL60011 for IEEE 802.11b Applications, AVAGO0020, Agere Systems, Jul. 2004, 22 pages.
Dell, Inc. “Dell Digital Audio Receiver: Reference Guide,” Jun. 2000, 70 pages.
Dell, Inc. “Start Here,” Jun. 2000, 2 pages.
“Denon 2003-2004 Product Catalog,” Denon, 2003-2004, 44 pages.
Denon AV Surround Receiver AVR-1604/684 User's Manual, 2004, 128 pages.
Denon AV Surround Receiver AVR-5800 Operating Instructions, Copyright 2000, 67 pages.
Designing a UPnP AV MediaServer, Nelson Kidd (2003) (SONDM000115062-116) (55 pages).
Dorwaldt, Carl, “EASE 4.1 Tutorial,” Renkus-Heinz, Inc., 2004, 417 pages.
“DP-0206 Digital Signal Processor,” TOA Electronics, Inc., 2001, pp. 1-12.
Dynaudio Acoustics Air Series, http://www.soundonsound.com/sos/sep02/articles/dynaudioair.asp, 2002, 4 pages.
European Patent Office, European Extended Search Report dated Mar. 7, 2016, issued in connection with EP Application No. 13810340.3, 9 pages.
European Patent Office, European Extended Search Report dated Feb. 28, 2014, issued in connection with EP Application No. 13184747.7, 8 pages.
European Patent Office, European Extended Search Report dated Mar. 31, 2015, issued in connection with EP Application No. 14181454.1, 9 pages.
European Patent Office, European Search Report dated Jul. 5, 2016, issued in connection with European Patent Application No. 16156935.5, 9 pages.
European Patent Office, Examination Report dated Mar. 22, 2016, issued in connection with European Patent Application No. EP14181454.1, 6 pages.
European Patent Office, Examination Report dated Oct. 24, 2016, issued in connection with European Patent Application No. 13808623.6, 4 pages.
European Patent Office, Extended European Search Report dated Jul. 5, 2016, issued in connection with European Patent Application No. 16156940.5, 7 pages.
Falcone, John, “Sonos BU150 Digital Music System review,” CNET, CNET [online] Jul. 27, 2009 [retrieved on Mar. 16, 2016], 11 pages Retrieved from the Internet: URL:http://www.cnet.com/products/sonos-bu150-digital-music-system/.
Faller, Christof, “Coding of Spatial Audio Compatible with Different Playback Formats,” Audio Engineering Society Convention Paper (Presented at the 117th Convention), Oct. 28-31, 2004, 12 pages.
File History of Re-Examination U.S. Appl. No. 90/013,423.
Final Office Action dated Jun. 5, 2014, issued in connection with U.S. Appl. No. 13/907,666, filed May 31, 2013, 12 pages.
Final Office Action dated Jul. 13, 2009, issued in connection with U.S. Appl. No. 10/816,217, filed Apr. 1, 2004, 16 pages.
Final Office Action dated Sep. 13, 2012, issued in connection with U.S. Appl. No. 13/297,000, filed Nov. 15, 2011, 17 pages.
Final Office Action dated Nov. 18, 2015, issued in connection with U.S. Appl. No. 13/533,105, filed Jun. 26, 2012, 56 pages.
Final Office Action dated Oct. 21, 2011, issued in connection with U.S. Appl. No. 10/816,217, filed Apr. 1, 2004, 19 pages.
Final Office Action dated Jul. 23, 2014, issued in connection with U.S. Appl. No. 13/896,037, filed May 16, 2013, 12 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Amended Invalidity Contentions Exhibit 7: Defendants' Invalidity Contentions for U.S. Pat. No. 9,195,258 filed Sep. 27, 2016, 418 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Amended Invalidity Contentions Exhibit 8: Defendants' Invalidity Contentions for U.S. Pat. No. 9,202,509 filed Sep. 27, 2016, 331 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Amended Invalidity Contentions Exhibit 9: Defendants' Invalidity Contentions for U.S. Pat. No. 9,213,357 filed Sep. 27, 2016, 251 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Initial Invalidity Contentions Exhibit 1: Defendants' Invalidity Contentions for U.S. Pat. No. 7,571,014 filed Apr. 15, 2016, 161 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Initial Invalidity Contentions Exhibit 10: Defendants' Invalidity Contentions for U.S. Pat. No. 9,213,357 filed Apr. 15, 2016, 244 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Initial Invalidity Contentions Exhibit 11: Defendants' Invalidity Contentions for U.S. Pat. No. 9,219,959 filed Apr. 15, 2016, 172 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Initial Invalidity Contentions Exhibit 12: Defendants' Invalidity Contentions for Design U.S. Pat. No. D. 559,197 filed Apr. 15, 2016, 36 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Initial Invalidity Contentions Exhibit 2: Defendants' Invalidity Contentions for U.S. Pat. No. 8,588,949 filed Apr. 15, 2016, 112 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Initial Invalidity Contentions Exhibit 3: Defendants' Invalidity Contentions for U.S. Pat. No. 8,843,224 filed Apr. 15, 2016, 118 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Initial Invalidity Contentions Exhibit 4: Defendants' Invalidity Contentions for U.S. Pat. No. 8,938,312 filed Apr. 15, 2016, 217 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Initial Invalidity Contentions Exhibit 5: Defendants' Invalidity Contentions for U.S. Pat. No. 8,938,637 filed Apr. 15, 2016, 177 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Initial Invalidity Contentions Exhibit 6: Defendants' Invalidity Contentions for U.S. Pat. No. 9,042,556 filed Apr. 15, 2016, 86 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Initial Invalidity Contentions Exhibit 7: Defendants' Invalidity Contentions for U.S. Pat. No. 9,130,771 filed Apr. 15, 2016, 203 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Initial Invalidity Contentions Exhibit 8: Defendants' Invalidity Contentions for U.S. Pat. No. 9,195,258 filed Apr. 15, 2016, 400 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Initial Invalidity Contentions Exhibit 9: Defendants' Invalidity Contentions for U.S. Pat. No. 9,202,509 filed Apr. 15, 2016, 163 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Preliminary Identification of Prior Art References, provided Jul. 29, 2016, 5 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendants' Brief in Support of their Motion for Leave to Amend their Answer to Add the Defense of Inequitable Conduct, provided Oct. 12, 2016, 24 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendants' Opposition to Sonos's Motion to Strike Defendants' New Amended Answer Submitted with their Reply, provided Oct. 3, 2016, 15 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Exhibit A: Defendants' Second Amended Answer to Plaintiffs' Third Amended Complaint, provided Oct. 12, 2016, 43 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Exhibit B: Defendants' Second Amended Answer to Plaintiffs' Third Amended Complaint, provided Oct. 12, 2016, 43 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Opening Brief in Support of Defendants' Motion for Leave to Amend Their Answer to Add the Defense of Inequitable Conduct, provided Aug. 1, 2016, 11 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Order, provided Oct. 7, 2016, 2 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Plaintiff's Opposition to Defendants' Motion for Leave to Amend Their Answer to Add the Defense of Inequitable Conduct, provided Aug. 26, 2016, 25 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Redlined Exhibit B: Defendants' First Amended Answer to Plaintiffs' Third Amended Complaint, provided Aug. 1, 2016, 27 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Reply Brief in Support of Defendants' Motion for Leave to Amend their Answer to Add the Defense of Inequitable Conduct, provided Nov. 10, 2016, 16 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Reply Brief in Support of Defendants' Motion for Leave to Amend their Answer to Add the Defense of Inequitable Conduct, provided Sep. 9, 2016, 16 pages.
Sonos, Inc. v. D&M Holdings (No. 14-1330-RGA), DI 206-1, Transcript of 101 Hearing (Nov. 28, 2016) (28 pages).
Sonos, Inc. v. D&M Holdings (No. 14-1330-RGA), DI 207, Public Joint Claim Construction Brief (Nov. 30, 2016) (88 pages).
Sonos, Inc. v. D&M Holdings (No. 14-1330-RGA), DI 214, D&M Post-Markman Letter (Dec. 22, 2016) (13 pages).
Sonos, Inc. v. D&M Holdings (No. 14-1330-RGA), DI 215, Sonos Post-Markman Letter (Dec. 22, 2016) (15 pages).
Sonos, Inc. v. D&M Holdings (No. 14-1330-RGA), DI 219, Claim Construction Opinion (Jan. 12, 2017) (24 pages).
Sonos, Inc. v. D&M Holdings (No. 14-1330-RGA), DI 221, Claim Construction Order (Jan. 18 2017) (2 pages).
Sonos, Inc. v. D&M Holdings (No. 14-1330-RGA), Markman Hearing Transcript (Dec. 14, 2016) (69 pages).
Sonos Multi-Room Music System User Guide, Version: 091001, 2009, 299 pages.
Sonos Play:3 Product Guide; copyright 2004-2011; 2 pages.
Sonos Play:3 Product Guide; copyright 2004-2012; 14 pages.
Sonos Play:3 Product Guide; copyright 2004-2013; 15 pages.
Sonos Play:3 Teardown; https://www.ifixit.com/Teardown/Sonos+Play%3A3+Teardown/12475; 11 pages.
Sony: AIR-SA 50R Wireless Speaker, Copyright 2009, 2 pages.
Sony: Altus Quick Setup Guide ALT-SA32PC, Copyright 2009, 2 pages.
Sony: BD/DVD Home Theatre System Operating Instructions for BDV-E300, E301 and E801, Copyright 2009, 115 pages.
Sony: BD/DVD Home Theatre System Operating Instructions for BDV-IT1000/BDV-IS1000, Copyright 2008, 159 pages.
Sony: Blu-ray Disc/DVD Home Theatre System Operating Instructions for BDV-IZ1000W, Copyright 2010, 88 pages.
Sony: DVD Home Theatre System Operating Instructions for DAV-DZ380W/DZ680W/DZ880W, Copyright 2009, 136 pages.
Sony: DVD Home Theatre System Operating Instructions for DAV-DZ870W, Copyright 2008, 128 pages.
Sony Ericsson MS500 User Guide, Copyright 2009, 2 pages.
Sony: Home Theatre System Operating Instructions for HT-IS100, Copyright 2008, 168 pages.
Sony: HT-IS100, 5.1 Channel Audio System, last updated Nov. 2009, 2 pages.
Sony: Multi Channel AV Receiver Operating Instructions, 2007, 80 pages.
Sony: Multi Channel AV Receiver Operating Instructions for STR-DN1000, Copyright 2009, 136 pages.
Notice of Allowance dated Jul. 29, 2015, issued in connection with U.S. Appl. No. 13/359,976, filed Jan. 27, 2012, 28 pages.
Notice of Allowance dated Jul. 29, 2015, issued in connection with U.S. Appl. No. 14/186,850, filed Feb. 21, 2014, 9 pages.
Notice of Allowance dated Aug. 30, 2016, issued in connection with U.S. Appl. No. 14/290,493, filed May 29, 2014, 7 pages.
Notice of Allowance dated Jul. 30, 2015, issued in connection with U.S. Appl. No. 13/705,178, filed Dec. 5, 2012, 18 pages.
Notice of Allowance dated Aug. 5, 2015, issued in connection with U.S. Appl. No. 13/435,776, filed Mar. 30, 2012, 26 pages.
Notice of Allowance dated Jul. 6, 2015, issued in connection with U.S. Appl. No. 13/297,000, filed Nov. 15, 2011, 24 pages.
Notice of Allowance dated Apr. 7, 2017, issued in connection with U.S. Appl. No. 14/629,937, filed Feb. 24, 2015, 8 pages.
Notice of Incomplete Re-Exam Request dated May 25, 2017, issued in connection with U.S. Appl. No. 90/013,959 , filed Apr. 1, 2016, 10 pages.
Notice of Intent to Issue Re-Examination Certificate dated Mar. 24, 2017, issued in connection with U.S. Appl. No. 90/013,859, filed Nov. 4, 2016, 10 pages.
Nutzel et al., “Sharing Systems for Future HiFi Systems,” IEEE, 2004, 9 pages.
Office Action in Ex Parte Reexamination mailed on Oct. 20, 2017, issued in connection with Reexamination U.S. Appl. No. 90/013,959, filed Jun. 16, 2017, 50 pages.
Palm, Inc., “Handbook for the Palm VII Handheld,” May 2000, 311 pages.
Parasound Zpre2 Zone Preamplifier with PTZI Remote Control, 2005, 16 pages.
Park et al., “Group Synchronization in MultiCast Media Communications,” Proceedings of the 5th Research on Multicast Technology Workshop, 2003, 5 pages.
Pillai et al., “A Method to Improve the Robustness of MPEG Video Applications over Wireless Networks,” Kent Ridge Digital Labs, 2000, 15 pages.
Polycom Conference Composer User Guide, copyright 2001, 29 pages.
Postel, J., “User Datagram Protocol,” RFC: 768, USC/Information Sciences Institute, Aug. 1980, 3 pages.
Preinterview First Office Action dated Jun. 8, 2016, issued in connection with U.S. Appl. No. 14/619,813, filed Feb. 11, 2015, 4 pages.
Pre-Interview First Office Action dated Mar. 10, 2015, issued in connection with U.S. Appl. No. 14/505,027, filed Oct. 2, 2014, 4 pages.
Presentations at WinHEC 2000, May 2000, 138 pages.
PRISMIQ, Inc., “PRISMIQ Media Player User Guide,” 2003, 44 pages.
Proficient Audio Systems M6 Quick Start Guide, 2011, 5 pages.
Proficient Audio Systems: Proficient Editor Advanced Programming Guide, 2007, 40 pages.
Programming Interface for WL54040 Dual-Band Wireless Transceiver, AVAGO0066, Agere Systems, May 2004, 16 pages.
Radio Shack, “Auto-Sensing 4-Way Audio/Video Selector Switch,” 2004, 1 page.
RadioShack, Pro-2053 Scanner, 2002 Catalog, part 1, 100 pages.
RadioShack, Pro-2053 Scanner, 2002 Catalog, part 2, 100 pages.
RadioShack, Pro-2053 Scanner, 2002 Catalog, part 3, 100 pages.
RadioShack, Pro-2053 Scanner, 2002 Catalog, part 4, 100 pages.
RadioShack, Pro-2053 Scanner, 2002 Catalog, part 5, 46 pages.
Rane: DragNet software; available for sale at least 2006.
Rangan et al., “Feedback Techniques for Continuity and Synchronization in Multimedia Information Retrieval,” ACM Transactions on Information Systems, 1995, pp. 145-176, vol. 13, No. 2.
Real Time Control Protocol (RTCP) and Realtime Transfer Protocol (RTP), RFC 1889 (Jan. 1996) (D+M_0397810-84) (75 pages).
Realtime Streaming Protocol (RTSP), RFC 2326 (Apr. 1998) (D+M_0397945-8036) (92 pages).
Realtime Transport Protocol (RTP), RFC 3550 (Jul. 2003) (D+M_0398235-323) (89 pages).
Re-Exam Final Office Action dated Aug. 5, 2015, issued in connection with U.S. Appl. No. 90/013,423, filed Jan. 5, 2015, 25 pages.
Reexam Non-Final Office Action dated Oct. 17, 2016, issued in connection with U.S. Appl. No. 90/013,756, filed May 25, 2016, 31 pages.
Re-Exam Non-Final Office Action dated Apr. 22, 2015, issued in connection with U.S. Appl. No. 90/013,423, filed Jan. 5, 2015, 16 pages.
Reid, Mark, “Multimedia conferencing over ISDN and IP networks using ITU-T H-series recommendations: architecture, control and coordination,” Computer Networks, 1999, pp. 225-235, vol. 31.
RenderingControl:1 Service Template Version 1.01 for UPnP, Version 1.0, (Jun. 25, 2002) (SONDM000115187-249) (63 pages).
Renewed Request for Ex Parte Re-Examination, U.S. Appl. No. 90/013,959, filed Jun. 16, 2017, 126 pages.
Renkus Heinz Manual; available for sale at least 2004, 6 pages.
Request for Ex Parte Reexamination submitted in U.S. Pat. No. 9,213,357 on May 22, 2017, 85 pages.
“Residential Distributed Audio Wiring Practices,” Leviton Network Solutions, 2001, 13 pages.
Ritchie et al., “MediaServer:1 Device Template Version 1.01,” Contributing Members of the UPnP Forum, Jun. 25, 2002, 12 pages.
Ritchie et al., “UPnP AV Architecture:1, Version 1.0,” Contributing Members of the UPnP Forum, Jun. 25, 2002, 22 pages.
Ritchie, John, “MediaRenderer:1 Device Template Version 1.01,” Contributing Members of the UPnP Forum, Jun. 25, 2002, 12 pages.
Roland Corporation, “Roland announces BA-55 Portable PA System,” press release, Apr. 6, 2011, 2 pages.
Rothermel et al., “An Adaptive Protocol for Synchronizing Media Streams,” Institute of Parallel and Distributed High-Performance Systems (IPVR), 1997, 26 pages.
Rothermel et al., “An Adaptive Stream Synchronization Protocol,” 5th International Workshop on Network and Operating System Support for Digital Audio and Video, 1995, 13 pages.
Sony: STR-DN1000, Audio Video Receiver, last updated Aug. 2009, 2 pages.
Sony: Wireless Surround Kit Operating Instructions for WHAT-SA2, Copyright 2010, 56 pages.
Taylor, Marilou, “Long Island Sound,” Audio Video Interiors, Apr. 2000, 8 pages.
Third Party Request for Ex Parte Re-Examination, U.S. Appl. No. 90/013,859, filed Nov. 4, 2016, 424 pages.
TOA Corporation, Digital Processor DP-0206 DACsys2000 Version 2.00 Software Instruction Manual, Copyright 2001, 67 pages.
Understanding Universal Plug and Play, Microsoft White Paper (Jun. 2000) (D+M_0402074-118) (45 pages).
U.S. Appl. No. 60/490,768, filed Jul. 28, 2003, entitled “Method for synchronizing audio playback between multiple networked devices,” 13 pages.
U.S. Appl. No. 60/825,407, filed Sep. 12, 2006, entitled “Controlling and manipulating groupings in a multi-zone music or media system,” 82 pages.
Universal Plug and Play Device Architecture V. 1.0, (Jun. 8, 2000) (54 pages).
Universal Plug and Play in Windows XP, Tom Fout. Microsoft Corporation (Jul. 2001) (D+M_0402041-73) (33 pages).
Universal Plug and Play (“UPnP”) AV Architecture:1 for UPnP, Version 1.0, (Jun. 25, 2002) (D+M_0298151-72) (22 pages).
Universal Plug and Play Vendor's Implementation Guide (Jan. 5, 2000) (7 pages).
UPnP AV Architecture:0.83 (Jun. 12, 2002) (SONDM000115483-504) (22 pages).
UPnP Design by Example, A Software Developers Guide to Universal Plug and Play Michael Jeronimo and JackWeast, Intel Press (D+M_0401307-818) (Apr. 2003) (511 pages).
ZX135: Installation Manual,LA Audio, Apr. 2003, 44 pages.
U.S. Appl. No. 13/083,499, filed Apr. 8, 2011, “Multi-Channel Pairing in a Media System.”
WANCommonInterfaceConfig:1 Service Template Version 1.01 for UPnP, Ver. 1.0 (Nov. 12, 2001) (D+M_0401820-43) (24 pages).
WANIPConnection:1 Service Template Version 1.01 for UPnP Ver. 1.0 (Nov. 12, 2001) (D+M_0401844-917) (74 pages).
WANPPPConnection:1 Service Template Version 1.01 for UPnP, Version 1.0 (Nov. 12, 2001) (D+M_0401918-2006) (89 pages).
WaveLan High-Speed Multimode Chip Set, AVAGO0003, Agere Systems, Feb. 2003, 4 pages.
WaveLan High-Speed Multimode Chip Set, AVAGO0005, Agere Systems, Feb. 2003, 4 pages.
WaveLAN Wireless Integration Developer Kit (WI-DK) for Access Point Developers, AVAGO0054, Agere Systems, Jul. 2003, 2 pages.
WaveLAN Wireless Integration-Developer Kit (WI-DK) Hardware Control Function (HCF), AVAGO0052, Agere Systems, Jul. 2003, 2 pages.
“Welcome. You're watching Apple TV.” Apple TV 1st Generation Setup Guide, Apr. 8, 2008 Retrieved Oct. 14, 2014, 40 pages.
“Welcome. You're watching Apple TV.” Apple TV 2nd Generation Setup Guide, Mar. 10, 2011 Retrieved Oct. 16, 2014, 36 pages.
“Welcome. You're watching Apple TV.” Apple TV 3rd Generation Setup Guide, Mar. 16, 2012 Retrieved Oct. 16, 2014, 36 pages.
WI-DK Release 2 WaveLan Embedded Drivers for VxWorks and Linux, AVAGO0056, Agere Systems, Jul. 2003, 2 pages.
WI-DK Release 2 WaveLan END Reference Driver for VxWorks, AVAGO0044, Agere Systems, Jul. 2003, 4 pages.
WI-DK Release 2 WaveLan LKM Reference Drivers for Linux, AVAGO0048, Agere Systems, Jul. 2003, 4 pages.
Windows Media Connect Device Compatibility Specification (Apr. 12, 2004) (16 pages).
WPA Reauthentication Rates, AVAGO0063, Agere Systems, Feb. 2004, 3 pages.
Yamaha DME 32 manual: copyright 2001.
Yamaha DME 64 Owner's Manual; copyright 2004, 80 pages.
Yamaha DME Designer 3.5 setup manual guide; copyright 2004, 16 pages.
Yamaha DME Designer 3.5 User Manual; Copyright 2004, 507 pages.
Yamaha DME Designer software manual: Copyright 2004, 482 pages.
“Symantec pcAnywhere User's Guide,” v 10.5.1, 1995-2002, 154 pages.
“Systemline Modular Installation Guide, Multiroom System,” Systemline, 2003, pp. 1-22.
Final Office Action dated Mar. 27, 2014, issued in connection with U.S. Appl. No. 13/533,105, filed Jun. 26, 2012, 29 pages.
Final Office Action dated Jan. 28, 2011, issued in connection with U.S. Appl. No. 10/816,217, filed Apr. 1, 2004, 21 pages.
Final Office Action dated Jun. 30, 2008, issued in connection with U.S. Appl. No. 10/816,217, filed Apr. 1, 2004, 30 pages.
Final Office Action dated Jul. 1, 2016, issued in connection with U.S. Appl. No. 13/458,558, filed Apr. 27, 2012, 11 pages.
Final Office Action dated Jul. 2, 2015, issued in connection with U.S. Appl. No. 13/458,558, filed Apr. 27, 2012, 11 pages.
Final Office Action dated Aug. 3, 2015, issued in connection with U.S. Appl. No. 13/848,921, filed Mar. 22, 2013, 13 pages.
Final Office Action dated Dec. 3, 2014, issued in connection with U.S. Appl. No. 14/184,528, filed Feb. 19, 2014, 12 pages.
Final Office Action dated Jul. 3, 2012, issued in connection with U.S. Appl. No. 13/298,090, filed Nov. 16, 2011, 46 pages.
Final Office Action dated Jun. 3, 2016, issued in connection with U.S. Appl. No. 13/705,176, filed Dec. 5, 2012, 24 pages.
Final Office Action dated Mar. 3, 2015, issued in connection with U.S. Appl. No. 13/864,251, filed Apr. 17, 2013, 13 pages.
Final Office Action dated Mar. 4, 2015, issued in connection with U.S. Appl. No. 13/848,904, filed Mar. 22, 2013, 16 pages.
Final Office Action dated Mar. 5, 2015, issued in connection with U.S. Appl. No. 13/888,203, filed May 6, 2013, 13 pages.
Final Office Action dated Jan. 7, 2015, issued in connection with U.S. Appl. No. 13/848,932, filed Mar. 22, 2013, 14 pages.
Final Office Action dated Mar. 9, 2015, issued in connection with U.S. Appl. No. 14/516,867, filed Oct. 17, 2014, 14 pages.
Final Office Action dated Aug. 10, 2015, issued in connection with U.S. Appl. No. 14/290,493, filed May 29, 2014, 26 pages.
Final Office Action dated Feb. 10, 2014, issued in connection with U.S. Appl. No. 13/013,740, filed Jan. 25, 2011, 13 pages.
Final Office Action dated Aug. 11, 2015, issued in connection with U.S. Appl. No. 13/864,247, filed Apr. 17, 2013, 15 pages.
Final Office Action dated Feb. 11, 2015, issued in connection with U.S. Appl. No. 14/184,526, filed Feb. 19, 2014, 13 pages.
Final Office Action dated Feb. 11, 2015, issued in connection with U.S. Appl. No. 14/184,935, filed Feb. 20, 2014, 17 pages.
Final Office Action dated Feb. 12, 2015, issued in connection with U.S. Appl. No. 14/184,522, filed Feb. 19, 2014, 20 pages.
Final Office Action dated Apr. 13, 2017, issued in connection with U.S. Appl. No. 14/563,515, filed Dec. 8, 2014, 13 pages.
Final Office Action dated Dec. 13, 2016, issued in connection with U.S. Appl. No. 14/629,937, filed Feb. 24, 2015, 14 pages.
Final Office Action dated Oct. 13, 2011, issued in connection with U.S. Appl. No. 11/853,790, filed Sep. 11, 2007, 10 pages.
Final Office Action dated Oct. 13, 2011, issued in connection with U.S. Appl. No. 12/035,112, filed Feb. 21, 2008, 10 pages.
Final Office Action dated Jul. 15, 2015, issued in connection with U.S. Appl. No. 14/504,812, filed Oct. 2, 2014, 18 pages.
Final Office Action dated Jun. 15, 2015, issued in connection with U.S. Appl. No. 14/184,522, filed Feb. 19, 2014, 25 pages.
Final Office Action dated Dec. 17, 2014, issued in connection with U.S. Appl. No. 13/533,105, filed Jun. 26, 2012, 36 pages.
Final Office Action dated Oct. 19, 2016, issued in connection with U.S. Appl. No. 13/848,921, filed Mar. 22, 2013, 14 pages.
Final Office Action dated Jan. 21, 2010, issued in connection with U.S. Appl. No. 11/906,702, filed Oct. 2, 2007, 27 pages.
Final Office Action dated Oct. 22, 2014, issued in connection with U.S. Appl. No. 14/186,850, filed Feb. 21, 2014, 12 pages.
Final Office Action dated Oct. 23, 2014, issued in conection with U.S. Appl. No. 13/705,176, filed Dec. 5, 2012, 23 pages.
Final Office Action dated Feb. 24, 2016, issued in connection with U.S. Appl. No. 13/871,795, filed Apr. 26, 2013, 28 pages.
Final Office Action dated May 25, 2016, issued in connection with U.S. Appl. No. 14/290,493, filed May 29, 2014, 33 pages.
Final Office Action dated Apr. 28, 2015, issued in connection with U.S. Appl. No. 14/186,850, filed Feb. 21, 2014, 20 pages.
Final Office Action dated Jun. 29, 2015, issued in connection with U.S. Appl. No. 14/465,457, filed Aug. 21, 2014, 13 pages.
Final Office Action dated Nov. 30, 2015, issued in connection with U.S. Appl. No. 13/871,795, filed Apr. 26, 2013, 26 pages.
Final Office Action dated Apr. 6, 2017, issued in connection with U.S. Appl. No. 14/620,937, filed Feb. 12, 2015, 15 pages.
Final Office Action dated Dec. 7, 2017, issued in connection with U.S. Appl. No. 14/619,813, filed Feb. 11, 2015, 11 pages.
Fireball DVD and Music Manager DVDM-100 Installation and User's Guide, Copyright 2003, 185 pages.
Fireball MP-200 User's Manual, Copyright 2006, 93 pages.
Fireball Remote Control Guide WD006-1-1, Copyright 2003, 19 pages.
Fireball SE-D1 User's Manual, Copyright 2005, 90 pages.
First Action Interview Office Action Summary dated Apr. 15, 2015, issued in connection with U.S. Appl. No. 14/505,027, filed Oct. 2, 2014, 6 pages.
Fober et al., “Clock Skew Compensation over a High Latency Network,” Proceedings of the ICMC, 2002, pp. 548-552.
Fries et al. “The MP3 and Internet Audio Handbook: Your Guide to the Digital Music Revolution.” 2000, 320 pages.
Fulton et al., “The Network Audio System: Make Your Application Sing (as Well as Dance)!” The X Resource, 1994, 14 pages.
Gaston et al., “Methods for Sharing Stereo and Multichannel Recordings Among Planetariums,” Audio Engineering Society Convention Paper 7474, 2008, 15 pages.
General Event Notification Architecture Base: Client to Arbiter (Apr. 2000) (23 pages).
Hans et al., “Interacting with Audio Streams for Entertainment and Communication,” Proceedings of the Eleventh ACM International Conference on Multimedia, ACM, 2003, 7 pages.
Herre et al., “The Reference Model Architecture for MPEG Spatial Audio Coding,” Audio Engineering Society Convention Paper (Presented at the 118th Convention), May 28-31, 2005, 13 pages.
Non-Final Office Action dated Dec. 22, 2014, issued in connection with U.S. Appl. No. 13/458,558, filed Apr. 27, 2012, 11 pages.
Non-Final Office Action dated Sep. 22, 2016, issued in connection with U.S. Appl. No. 15/088,906, filed Apr. 1, 2016, 9 pages.
Non-Final Office Action dated Sep. 22, 2016, issued in connection with U.S. Appl. No. 15/155,149, filed May 16, 2016, 7 pages.
Non-Final Office Action dated Jun. 23, 2015, issued in connection with U.S. Appl. No. 13/705,176, filed Dec. 5, 2012, 30 pages.
Non-Final Office Action dated Mar. 23, 2015, issued in connection with U.S. Appl. No. 14/299,847, filed Jun. 9, 2014, 14 pages.
Non-Final Office Action dated Oct. 23, 2014, issued in connection with U.S. Appl. No. 13/848,904, filed Mar. 22, 2013, 11 pages.
Non-Final Office Action dated Oct. 23, 2014, issued in connection with U.S. Appl. No. 13/864,251, filed Apr. 17, 2013, 11 pages.
Non-Final Office Action dated Oct. 23, 2014, issued in connection with U.S. Appl. No. 13/888,203, filed May 6, 2013, 9 pages.
Non-Final Office Action dated Sep. 23, 2014, issued in connection with U.S. Appl. No. 13/630,565, filed Sep. 28, 2012, 7 pages.
Non-Final Office Action dated Feb. 24, 2017, issued in connection with U.S. Appl. No. 14/619,813, filed Feb. 11, 2015, 9 pages.
Non-Final Office Action dated May 24, 2016, issued in connection with U.S. Appl. No. 15/134,767, filed Apr. 21, 2016, 12 pages.
Non-final Office Action dated Oct. 24, 2014, issued in connection with U.S. Appl. No. 13/435,776, filed Mar. 30, 2012, 14 pages.
Non-Final Office Action dated Apr. 25, 2018, issued in connection with U.S. Appl. No. 15/130,919, filed Apr. 15, 2016, 13 pages.
Non-Final Office Action dated Feb. 26, 2015, issued in connection with U.S. Appl. No. 14/186,850, filed Feb. 21, 2014, 25 pages.
Non-Final Office Action dated Mar. 26, 2015, issued in connection with U.S. Appl. No. 14/184,528, filed Feb. 19, 2014, 18 pages.
Non-Final Office Action dated Jan. 27, 2015, issued in connection with U.S. Appl. No. 14/465,457, filed Aug. 21, 2014, 11 pages.
Non-Final Office Action dated Jun. 27, 2008, issued in connection with U.S. Appl. No. 10/861,653, filed Jun. 5, 2004, 19 pages.
Non-Final Office Action dated Mar. 27, 2015, issued in connection with U.S. Appl. No. 13/705,178, filed Dec. 5, 2012, 14 pages.
Non-Final Office Action dated Sep. 27, 2013, issued in connection with U.S. Appl. No. 13/013,740, filed Jan. 25, 2011, 12 pages.
Non-Final Office Action dated Sep. 27, 2016, issued in connection with U.S. Appl. No. 15/228,685, filed Aug. 4, 2016, 8 pages.
Non-Final Office Action dated Dec. 28, 2015, issued in connection with U.S. Appl. No. 14/290,493, filed May 29, 2014, 29 pages.
Non-Final Office Action dated Dec. 28, 2016, issued in connection with U.S. Appl. No. 15/343,000, filed Nov. 3, 2016, 11 pages.
Non-Final Office Action dated Jan. 29, 2016, issued in connection with U.S. Appl. No. 14/937,523, filed Nov. 10, 2015, 10 pages.
Non-Final Office Action dated Jun. 29, 2016, issued in connection with U.S. Appl. No. 14/629,937, filed Feb. 24, 2015, 12 pages.
Non-Final Office Action dated Apr. 30, 2012, issued in connection with U.S. Appl. No. 13/204,511, filed Aug. 5, 2011, 16 pages.
Non-Final Office Action dated Jan. 30, 2015, issued in connection with U.S. Appl. No. 14/290,493, filed May 29, 2014, 29 pages.
Non-Final Office Action dated Jan. 30, 2015, issued in connection with U.S. Appl. No. 14/504,812, filed Oct. 2, 2014, 13 pages.
Non-Final Office Action dated Nov. 30, 2016, issued in connection with U.S. Appl. No. 15/243,186, filed Aug. 22, 2016, 12 pages.
Non-Final Office Action dated Oct. 30, 2018, issued in connection with U.S. Appl. No. 16/128,443, filed Sep. 11, 2018, 21 pages.
Non-Final Office Action dated Sep. 30, 2016, issued in connection with U.S. Appl. No. 13/864,249, filed Apr. 17, 2013, 12 pages.
Non-Final Office Action dated Oct. 31, 2016, issued in connection with U.S. Appl. No. 14/806,070, filed Jul. 22, 2015, 11 pages.
North American MPEG-2 Information, “The MPEG-2 Transport Stream,” Retrieved from the Internet: URL: http://www.coolstf.com/mpeg/#ts, 2006, pp. 1-5.
Notice of Allowability dated Apr. 18, 2013, issued in connection with U.S. Appl. No. 11/853,790, filed Sep. 11, 2007, 4 pages.
Notice of Allowance date Jan. 31, 2013, issued in connection with U.S. Appl. No. 13/298,090, filed Nov. 16, 2011, 19 pages.
Notice of Allowance dated Dec. 1, 2016, issued in connection with U.S. Appl. No. 15/088,283, filed Apr. 1, 2016, 9 pages.
Notice of Allowance dated Jun. 1, 2017, issued in connection with U.S. Appl. No. 14/808,397, filed Jul. 24, 2015, 5 pages.
Notice of Allowance dated Dec. 2, 2016, issued in connection with U.S. Appl. No. 15/088,532, filed Apr. 1, 2016, 9 pages.
Notice of Allowance dated Dec. 2, 2016, issued in connection with U.S. Appl. No. 15/088,678, filed Apr. 1, 2016, 9 pages.
Notice of Allowance dated Dec. 2, 2016, issued in connection with U.S. Appl. No. 15/089,758, filed Apr. 4, 2016, 9 pages.
Notice of Allowance dated Dec. 2, 2016, issued in connection with U.S. Appl. No. 15/155,149, filed May 16, 2016, 9 pages.
Notice of Allowance dated Jul. 2, 2015, issued in connection with U.S. Appl. No. 13/848,904, filed Mar. 22, 2013, 17 pages.
Notice of Allowance dated Jul. 2, 2015, issued in connection with U.S. Appl. No. 13/888,203, filed May 6, 2013, 19 pages.
Notice of Allowance dated Jul. 2, 2015, issued in connection with U.S. Appl. No. 14/184,935, filed Feb. 20, 2014, 23 pages.
Notice of Allowance dated Jun. 2, 2014, issued in connection with U.S. Appl. No. 13/083,499, filed Apr. 8, 2011, 5 pages.
Notice of Allowance dated Sep. 3, 2015, issued in connection with U.S. Appl. No. 13/705,174, filed Dec. 5, 2012, 4 pages.
Notice of Allowance dated Aug. 4, 2015, issued in connection with U.S. Appl. No. 14/516,867, filed Oct. 17, 2014, 13 pages.
Notice of Allowance dated Dec. 5, 2014, issued in connection with U.S. Appl. No. 14/256,434, filed Apr. 18, 2014, 7 pages.
Notice of Allowance dated Oct. 5, 2012, issued in connection with U.S. Appl. No. 13/204,511, filed Aug. 5, 2011, 11 pages.
Notice of Allowance dated Mar. 6, 2014, issued in connection with U.S. Appl. No. 13/827,653, filed Mar. 14, 2013, 17 pages.
Notice of Allowance dated May 6, 2011, issued in connection with U.S. Appl. No. 11/801,468, filed May 9, 2007, 10 pages.
Home Networking with Universal Plug and Play, IEEE Communications Magazine, vol. 39 No. 12 (Dec. 2001) (D+M_0402025-40) (16 pages).
“Home Theater Control Systems,” Cinema Source, 2002, 19 pages.
Horwitz, Jeremy, “Logic3 i-Station25,” retrieved from the internet: http://www.ilounge.com/index.php/reviews/entry/logic3-i-station25/, last visited Dec. 17, 2013, 5 pages.
Huang C.M., et al., “A Synchronization Infrastructure for Multicast Multimedia at the Presentation Layer,” IEEE Transactions on Consumer Electronics, 1997, pp. 370-380, vol. 43, No. 3.
IBM Home Director Installation and Service Manual, Copyright1998, 124 pages.
IBM Home Director Owner's Manual, Copyright 1999, 67 pages.
ID3 tag version 2.4.0—Native Frames, Draft Specification, copyright 2000, 41 pages.
Implicit, LLC v. Sonos, Inc. (No. 14-1330-RGA), Defendant's Original Complaint (Mar. 3, 2017) (15 pages).
Integra Audio Network Receiver NAC 2.3 Instruction Manual, 68 pages.
Integra Audio Network Server NAS 2.3 Instruction Manual, pp. 1-32.
Integra Service Manual, Audio Network Receiver Model NAC-2.3, Dec. 2002, 44 pages.
Intel Designing a UPnP AV Media Renderer, v. 1.0 (“Intel AV Media Renderer”) (May 20, 2003) (SONDM000115117-62) (46 pages).
Intel Media Renderer Device Interface (“Intel Media Renderer”) (Sep. 6, 2002) (62 pages).
Intel SDK for UPnP Devices Programming Guide, Version 1.2.1, (Nov. 2002) (30 pages).
International Bureau, International Preliminary Report on Patentability dated Jan. 8, 2015, issued in connection with International Application No. PCT/US2013/046372, filed on Jun. 18, 2013, 6 pages.
International Bureau, International Preliminary Report on Patentability, dated Jan. 8, 2015, issued in connection with International Application No. PCT/US2013/046386, filed on Jun. 18, 2013, 8 pages.
International Bureau, International Preliminary Report on Patentability, dated Oct. 17, 2013, issued in connection with International Application No. PCT/IB2012/052071, filed on Apr. 26, 2012, 7 pages.
International Bureau, International Preliminary Report on Patentability dated Jan. 30, 2014, issued in connection with International Application No. PCT/US2012/045894, filed on Jul. 9, 2012, 6 pages.
International Searching Authority, International Search Report dated Aug. 1, 2008, in connection with International Application No. PCT/US2004/023102, 5 pages.
International Searching Authority, International Search Report dated Aug. 23, 2012, issued in connection with International Application No. PCT/IB2012/052071, filed on Apr. 26, 2012, 3 pages.
International Searching Authority, International Search Report dated Aug. 26, 2013, issued in connection with International Application No. PCT/US2013/046372, filed on Jun. 18, 2013, 3 pages.
International Searching Authority, International Search Report dated Dec. 26, 2012, issued in connection with International Application No. PCT/US2012/045894, filed on Jul. 9, 2012, 3 pages.
International Searching Authority, International Search Report dated Sep. 30, 2013, issued in connection with International Application No. PCT/US2013/046386, filed on Jun. 18, 2013, 3 pages.
International Searching Authority, Written Opinion dated Aug. 23, 2012, issued in connection with International Application No. PCT/IB2012/052071, filed on Apr. 26, 2012, 6 pages.
International Searching Authority, Written Opinion dated Aug. 26, 2013, issued in connection with International Application No. PCT/US2013/046372, filed on Jun. 18, 2013, 4 pages.
International Searching Authority, Written Opinion dated Dec. 26, 2012, issued in connection with International Application No. PCT/US2012/045894, filed on Jul. 9, 2012, 4 pages.
International Searching Authority, Written Opinion dated Sep. 30, 2013, issued in connection with International Application No. PCT/US2013/046386, filed on Jun. 18, 2013, 6 pages.
Ishibashi et al., “A Comparison of Media Synchronization Quality Among Reactive Control Schemes,” IEEE Infocom, 2001, pp. 77-84.
Ishibashi et al., “A Group Synchronization Mechanism for Live Media in Multicast Communications,” IEEE Global Telecommunications Conference, 1997, pp. 746-752, vol. 2.
Ishibashi et al., “A Group Synchronization Mechanism for Stored Media in Multicast Communications,” IEEE Information Revolution and Communications, 1997, pp. 692-700, vol. 2.
Issues with Mixed IEEE 802.b/802.11g Networks, AVAGO0058, Agere Systems, Feb. 2004, 5 pages.
Japanese Patent Office, Decision of Rejection dated Jul. 8, 2014, issued in connection with Japanese Patent Application No. 2012-178711, 3 pages.
Japanese Patent Office, Notice of Rejection, dated Feb. 3, 2015, issued in connection with Japanese Patent Application No. 2014-521648, 7 pages.
Japanese Patent Office, Notice of Rejection dated Sep. 15, 2015, issued in connection with Japanese Patent Application No. 2014-220704, 7 pages.
Japanese Patent Office, Office Action dated Nov. 1, 2016, issued in connection with Japanese Application No. 2015-238682, 7 pages.
Japanese Patent Office, Office Action dated Jan. 6, 2015, issued in connection with Japanese Patent Application No. 2014-503273, 5 pages.
Japanese Patent Office, Office Action dated May 24, 2016, issued in connection with Japanese Patent Application No. 2014-220704, 7 pages.
Japanese Patent Office, Office Action dated Mar. 29, 2016, issued in connection with Japanese Patent Application No. JP2015-520288, 12 pages.
Japanese Patent Office, Office Action Summary dated Feb. 2, 2016, issued in connection with Japanese Patent Application No. 2015-520286, 6 pages.
Japanese Patent Office, Office Action Summary dated Sep. 8, 2015, issued in connection with Japanese Patent Application No. 2014-503273, 4 pages.
Japanese Patent Office, Office Action Summary dated Nov. 19, 2013, issued in connection with Japanese Patent Application No. 2012-178711, 5 pages.
Jo et al., “Synchronized One-to-many Media Streaming with Adaptive Playout Control,” Proceedings of SPIE, 2002, pp. 71-82, vol. 4861.
Jones, Stephen, “Dell Digital Audio Receiver: Digital upgrade for your analog stereo,” Analog Stereo, Jun. 24, 2000 retrieved Jun. 18, 2014, 2 pages.
Kou et al., “RenderingControl:1 Service Template Verion 1.01,” Contributing Members of the UPnP Forum, Jun. 25, 2002, 63 pages.
Lake Processors: Lake® LM Series Digital Audio Processors Operation Manual, 2011, 71 pages.
Levergood et al., “AudioFile: A Network-Transparent System for Distributed Audio Applications,” Digital Equipment Corporation, 1993, 109 pages.
LG: RJP-201M Remote Jack Pack Installation and Setup Guide, 2010, 24 pages.
Lienhart et al., “On the Importance of Exact Synchronization for Distributed Audio Signal Processing,” Session L: Poster Session II—ICASSP'03 Papers, 2002, 1 page.
LinkSys by Cisco, Wireless Home Audio Controller, Wireless-N Touchscreen Remote DMRW1000 Datasheet, Copyright 2008, 2 pages.
LinkSys by Cisco, Wireless Home Audio Controller, Wireless-N Touchscreen Remote DMRW1000 User Guide, Copyright 2008, 64 pages.
“ZR-8630AV MultiZone Audio/Video Receiver, Installation and Operation Guide,” Niles Audio Corporation, 2003, 86 pages.
“884+ Automatic Matrix Mixer Control System,” Ivie Technologies, Inc., 2000, pp. 1-4.
Advanced Driver Tab User Interface WaveLan GUI Guide, AVAGO0009, Agere Systems, Feb. 2004, 4 pages.
Advisory Action dated Feb. 2, 2016, issued in connection with U.S. Appl. No. 13/848,921, filed Mar. 22, 2013, 8 pages.
Advisory Action dated Sep. 18, 2008, issued in connection with U.S. Appl. No. 10/816,217, filed Apr. 1, 2004, 8 pages.
Advisory Action dated Feb. 1, 2016, issued in connection with U.S. Appl. No. 13/864,247, filed Apr. 17, 2013, 6 pages.
Advisory Action dated Jun. 1, 2015, issued in connection with U.S. Appl. No. 14/516,867, filed Oct. 17, 2014, 11 pages.
Advisory Action dated Mar. 2, 2015, issued in connection with U.S. Appl. No. 13/848,932, filed Mar. 22, 2013, 3 pages.
Advisory Action dated Jan. 5, 2012, issued in connection with U.S. Appl. No. 12/035,112, filed Feb. 21, 2008, 3 pages.
Advisory Action dated Oct. 5, 2015, issued in connection with U.S. Appl. No. 13/458,558, filed Apr. 27, 2012, 4 pages.
Advisory Action dated Sep. 5, 2014, issued in connection with U.S. Appl. No. 13/907,666, filed May 31, 2013, 3 pages.
Advisory Action dated Oct. 6, 2016, issued in connection with U.S. Appl. No. 13/458,558, filed Apr. 27, 2012, 4 pages.
Advisory Action dated Jan. 8, 2015, issued in connection with U.S. Appl. No. 13/705,176, filed Dec. 5, 2012, 4 pages.
Advisory Action dated Jun. 9, 2016, issued in connection with U.S. Appl. No. 13/871,795, filed Apr. 25, 2013, 3 pages.
Advisory Action dated Feb. 10, 2016, issued in connection with U.S. Appl. No. 13/871,795, filed Apr. 26, 2013, 3 pages.
Advisory Action dated Nov. 12, 2014, issued in connection with U.S. Appl. No. 13/907,666, filed May 31, 2013, 6 pages.
Advisory Action dated Apr. 15, 2015, issued in connection with U.S. Appl. No. 14/184,526, filed Feb. 19, 2014, 9 pages.
Advisory Action dated Apr. 15, 2015, issued in connection with U.S. Appl. No. 14/184,935, filed Feb. 20, 2014, 9 pages.
Advisory Action dated Dec. 22, 2011, issued in connection with U.S. Appl. No. 11/853,790, filed Sep. 11, 2007, 2 pages.
Advisory Action dated Mar. 25, 2015, issued in connection with U.S. Appl. No. 13/533,105, filed Jun. 26, 2012, 5 pages.
Advisory Action dated Feb. 26, 2015, issued in connection with U.S. Appl. No. 14/184,528, filed Feb. 19, 2014, 3 pages.
Advisory Action dated Nov. 26, 2014, issued in connection with U.S. Appl. No. 14/186,850, filed Feb. 21, 2014, 9 pages.
Advisory Action dated Jul. 28, 2015, issued in connection with U.S. Appl. No. 14/184,522, filed Feb. 19, 2014, 7 pages.
Advisory Action dated Sep. 28, 2009, issued in connection with U.S. Appl. No. 10/816,217, filed Apr. 1, 2004, 4 pages.
Agere Systems' Voice-over-Wireless LAN (VoWLAN) Station Quality of Service, AVAGO0015, Agere Systems, Jan. 2005, 5 pages.
Akyildiz et al., “Multimedia Group Synchronization Protocols for Integrated Services Networks,” IEEE Journal on Selected Areas in Communications, 1996 pp. 162-173, vol. 14, No. 1.
Anonymous, “Information technology—Generic coding of moving pictures and associated audio information—Part 3: Audio,” ISO/IEC 13818-3, Apr. 1998, pp. 11.
Anonymous, “Transmission Control Protocol,” RFC: 793, USC/Information Sciences Institute, Sep. 1981, 91 pages.
Audio Authority: How to Install and Use the Model 1154 Signal Sensing Auto Selector, 2002, 4 pages.
Audio Authority: Model 1154B High Definition AV Auto Selector, 2008, 8 pages.
AudioSource: AMP 100 User Manual, 2003, 4 pages.
AudioTron Quick Start Guide, Version 1.0, Mar. 2001, 24 pages.
AudioTron Reference Manual, Version 3.0, May 2002, 70 pages.
AudioTron Setup Guide, Version 3.0, May 2002, 38 pages.
Automatic Profile Hunting Functional Description, AVAGO0013, Agere Systems, Feb. 2004, 2 pages.
“A/V Surround Receiver AVR-5800,” Denon Electronics, 2000, 2 pages.
“A/V System Controleer, Owner's Manual,” B&K Compontents, Ltd., 1998, 52 pages.
AVTransport:1 Service Template Version 1.01 for UPnP, Version 1.0 (Jun. 25, 2002) (66 pages).
AXIS Communication: AXIS P8221 Network I/O Audio Module, 2009, 41 pages.
Baldwin, Roberto. “How-To: Setup iTunes DJ on Your Max and iPhone”, available at http://www.maclife.com/article/howtos/howto_setup_itunes_dj_your mac_and_iphone, archived on Mar. 17, 2009, 4 pages.
Balfanz et al., “Network-in-a-Box: How to Set Up a Secure Wireless Network in Under a Minute,” 13th USENIX Security Symposium—Technical Paper, 2002, 23 pages.
Balfanz et al., “Talking to Strangers: Authentication in Ad-Hoc Wireless Networks,” Xerox Palo Alto Research Center, 2002, 13 pages.
Barham et al., “Wide Area Audio Synchronisation,” University of Cambridge Computer Laboratory, 1995, 5 pages.
Baudisch et al., “Flat Volume Control: Improving Usability by Hiding the Volume Control Hierarchy in the User Interface,” 2004, 8 pages.
Benslimane Abderrahim, “A Multimedia Synchronization Protocol for Multicast Groups,” Proceedings of the 26th Euromicro Conference, 2000, pp. 456-463, vol. 1.
Biersack et al., “Infra- and Inter-Stream Synchronization for Stored Multimedia Streams,” IEEE International Conference on Multimedia Computing and Systems, 1996, pp. 372-381.
Blakowski G. et al., “A Media Synchronization Survey: Reference Model, Specification, and Case Studies,” Jan. 1996, pp. 5-35, vol. 14, No. 1.
Bluetooth. “Specification of the Bluetooth System: The ad hoc SCATTERNET for affordable and highly functional wireless connectivity,” Core, Version 1.0 A, Jul. 26, 1999, 1068 pages.
Bluetooth. “Specification of the Bluetooth System: Wireless connections made easy,” Core, Version 1.0 B, Dec. 1, 1999, 1076 pages.
Bogen Communications, Inc., ProMatrix Digitally Matrixed Amplifier Model PM3180, Copyright1996, 2 pages.
Brassil et al., “Enhancing Internet Streaming Media with Cueing Protocols,” 2000, 9 pages.
Sonos, Inc. v. D&M Holdings Inc. el al., Defendant's Preliminary Identification of Prior Art References, provided Jul. 9, 2016, 5 pages.
Sonos, Inc. v. D&M Holdings Inc. el al., Exhibit A: Defendants' Second Amended Answer to Plaintiffs' Third Amended Complaint, provided Oct. 12, 2016, 43 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Reply Brief in Support of Defendants' Motion for Leave to Amend their Answel to Add the Defense of Inequitable Conduct, provided Sep. 9, 2016, 16 pages.
Notice of Incomplete Re-Exam Request mailed on May 25, 2017, issued in connection with U.S. Appl. No. 90/013,959, filed Apr. 1, 2016, 10 pages.
Notice of Intent to Issue Re-Examination Certificate mailed on Mar. 24, 2017, issued in connection with U.S. Appl. No. 90/013,859, filed Nov. 4, 2016, 10 pages.
Office Action in Ex Parte Reexamination mailed on Oct. 20, 2017, issued in connection with U.S. Appl. No. 90/013,959, filed Jun. 16, 2017, 50 pages.
Rothermel et al., “An Adaptive Stream Synchronization Protocol,” 5th International Workshop on Network and Operating System Support for Digital Audio and Video, Apr. 18-21, 1995, 12 pages.
Rothermel et al., “Clock Hierarchies—An Abstraction for Grouping and Controlling Media Streams,” University of Stuttgart Institute of Parallel and Distributed High-Performance Systems, Jan. 1996, 23 pages.
Rothermel et al., “Synchronization in Joint-Viewing Environments,” University of Stuttgart Institute of Parallel and Distributed High-Performance Systems, 1992, 13 pages.
Rothermel, Kurt, “State-of-the-Art and Future Research in Stream Synchronization,” University of Stuttgart, 3 pages.
“RVL-6 Modular Multi-Room Controller, Installation & Operation Guide,” Nile Audio Corporations, 1999, 46 pages.
Schmandt et al., “Impromptu: Managing Networked Audio Applications for Mobile Users,” 2004, 11 pages.
Schulzrinne et al., “RTP: A Transport Protocol for Real-Time Applications,” Network Working Group, RFC: 3550, Standards Track, Jul. 2003, 104 pages.
Schulzrinne H., et al., “RTP: A Transport Protocol for Real-Time Applications, RFC 3550,” Network Working Group, 2003, pp. 1-89.
Simple Network Time Protocol (SNTPI), RFC 1361 (Aug. 1992) (D+M_0397537-46) (10 pages).
Simple Network Time Protocol (SNTPII), RFC 1769 (Mar. 1995) (D+M_0397663-76) (14 pages).
Simple Service Discovery Protocol/1.0 Operating without an Arbiter (Oct. 28, 1999) (24 pages).
Sonos Controller for iPad Product Guide; copyright 2004-2013; 47 pages.
Sonos Digital Music System User Guide, Version: 050801, Aug. 2005, 114 pages.
Sonos, Inc. v D&M Holdings, D&M Supp Opposition Brief including Exhibits, Mar. 17, 2017, 23 pages.
Sonos, Inc. v. D&M Holdings, Expert Report of Jay P. Kesan including Appendices A-P, Feb. 20, 2017, 776 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Complaint for Patent Infringement, filed Oct. 21, 2014, 20 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Amended Invalidity Contentions, filed Sep. 14, 2016, 100 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Initial Invalidity Contentions, filed Apr. 15, 2016, 97 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Preliminary Identification of Indefinite Terms, provided Jul. 29, 2016, 8 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendants' 35 U.S.C. § 282 Notice filed Nov. 2, 2017, 31 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendants' Amended Answer, Defenses, and Counterclaims for Patent Infringement, filed Nov. 30, 2015, 47 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendants' Answer to Plaintiff's Second Amended Complaint, filed Apr. 30, 2015, 19 pages.
Sonos, Inc. v. D&M Holdings Inc. el al., Defendants' First Amended Answer to Plaintiffs' Third Amended Complaint, filed Sep. 7, 2016, 23 pages.
Sonos, Inc. v. D&M Holdings Inc. el al., Defendants' Reply in Support of Partial Motion for Judgment on the Pleadings, filed Jun. 10, 2016, 15 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Exhibit A: Defendants' First Amended Answer to Plaintiffs' Third Amended Complaint, provided Aug. 1, 2016, 26 pages.
Sonos, Inc. v. D&M Holdings Inc. el al., Exhibit A: Defendants' Second Amended Answer to Plaintiffs' Third Amended Complaint, filed Sep. 9, 2016, 43 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Exhibit A: Defendants'Second Amended Answer to Plaintiffs' Third Amended Complaint, provided Sep. 9, 2016, 88 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., First Amended Complaint for Patent Infringement, filed Dec. 17, 2014, 26 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Joint Claim Construction Chart, vol. 1 of 3 with Exhibits A-O, filed Aug. 17, 2016, 30 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Opening Brief in Support of Defendants' Partial Motion for Judgment on the Pleadings for Lack of Patent-Eligible Subject Matter, filed May 6, 2016, 27 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Plaintiff Sonos, Inc.'s Opening Claim Construction Brief, filed Sep. 9, 2016, 26 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Plaintiff Sonos, Inc.'s Response in Opposition to Defendants' Partial Motion for Judgment on the Pleadings, filed May 27, 2016, 24 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Second Amended Complaint for Patent Infringement, filed Feb. 27, 2015, 49 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Sonos's Motion to Strike Defendants' New Amended Answer Submitted with their Reply Brief, provided Sep. 15, 2016, 10 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Sonos's Opposition to Defendants' Motion for Leave to Amend their Answer to Add the Defense of Inequitable Conduct, provided Oct. 31, 2016, 26 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Third Amended Complaint for Patent Infringement, filed Jan. 29, 2016, 47 pages.
Sonos, Inc. v. D&M Holdings, Inc. (No. 14-1330-RGA), Defendants' Final Invalidity Contentions (Jan. 18, 2017) (106 pages).
Sonos, Inc. v. D&M Holdings (No. 14-1330-RGA), DI 226, Opinion Denying Inequitable Conduct Defenses, Feb. 6, 2017, updated, 5 pages.
Sonos, Inc. v. D&M Holdings (No. 14-1330-RGA), DI 242, US District Judge Andrews 101 Opinion, Mar. 13, 2017, 16 pages.
Sonos, Inc. v D&M Holdings, Sonos Supp Opening Markman Brief including Exhibits, Mar. 3, 2017, 17 pages.
Sonos, Inc. v. D&M Holdings, Sonos Supp Reply Markman Brief including Exhibits, Mar. 29, 2017, 36 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Declaration of Steven C. Visser, executed Sep. 9, 2016, 40 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Amended Invalidity Contentions Exhibit 1: Defendants' Invalidity Contentions for U.S. Pat. No. 7,571,014 filed Sep. 16, 2016, 270 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Amended Invalidity Contentions Exhibit 10: Defendants' Invalidity Contentions for U.S. Pat. No. 9,219,959 filed Sep. 27, 2016, 236 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Amended Invalidity Contentions Exhibit 11: Defendants' Invalidity Contentions for Design U.S. Pat. No. D. 559,197 filed Sep. 27, 2016, 52 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Amended Invalidity Contentions Exhibit 2: Defendants' Invalidity Contentions for U.S. Pat. No. 8,588,949 filed Sep. 27, 2016, 224 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Amended Invalidity Contentions Exhibit 3: Defendants' Invalidity Contentions for U.S. Pat. No. 8,843,224 filed Sep. 27, 2016, 147 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Amended Invalidity Contentions Exhibit 4: Defendants' Invalidity Contentions for U.S. Pat. No. 8,938,312 filed Sep. 27, 2016, 229 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Amended Invalidity Contentions Exhibit 5: Defendants' Invalidity Contentions for U.S. Pat. No. 8,938,637 filed Sep. 27, 2016, 213 pages.
Sonos, Inc. v. D&M Holdings Inc. et al., Defendant's Amended Invalidity Contentions Exhibit 6: Defendants' Invalidity Contentions for U.S. Pat. No. 9,042,556 filed Sep. 27, 2016, 162 pages.
AuviTran AVB32-ES User's Manual, 2005, 25 pages.
AuviTran AVKIT-ES for AD8HR User's Manual, 2005, 15 pages.
Chinese Patent Office, Second Office Action and Translation dated Jun. 27, 2019, issued in connection with Chinese Application No. 201610804134.8, 15 pages.
Chinese Patent Office, Translation of Office Action dated Jun. 27, 2019, issued in connection with Chinese Application No. 201610804134.8, 10 pages.
CobraNet Manager, Direct control over your audio network. www.peakaudiacom/CobraNet/FAQ.html, 2005 [retrieved online Jul. 12, 2019 at web.archive.org/web/20050403214230/http://www.peakaudio.com/CobraNet/FAQ] 13 pages.
European Patent Office, European Office Action dated Sep. 16, 2019, issued in connection with European Application No. 17198867.8, 6 pages.
Final Office Action dated Nov. 14, 2018, issued in connection with U.S. Appl. No. 15/130,919, filed Apr. 15, 2016, 12 pages.
Final Office Action dated Jan. 3, 2019, issued in connection with U.S. Appl. No. 15/405,931, filed Jan. 13, 2017, 16 pages.
Japanese Patent Office, Final Office Action dated Jun. 4, 2019, issued in connection with Japanese Patent Application No. 2017-211958, 8 pages.
Japanese Patent Office, Office Action dated Dec. 18, 2018, issued in connection with Japanese Application No. 2017-211958, 8 pages.
Japanese Patent Office, Translation of Final Office Action dated Jun. 4, 2019, issued in connection with Japanese Patent Application No. 2017-211958, 5 page.
Japanese Patent Office, Translation of Office Action dated Dec. 18, 2018, issued in connection with Japanese Application No. 2017-211958, 6 pages.
Machine Translation of JP2004-193868A Wireless Transmission and Reception System and Wireless Transmission and Reception Method, 12 pages.
Machine Translation of JP2007-2888405A Video Sound Output System, Video Sound Processing Method, and Program, 64 pages.
Non-Final Office Action dated Nov. 1, 2018, issued in connection with U.S. Appl. No. 16/129,758, filed Sep. 12, 2018, 23 pages.
Non-Final Office Action dated Jul. 17, 2019, issued in connection with U.S. Appl. No. 15/130,919, filed Apr. 15, 2016, 15 pages.
Non-Final Office Action dated Sep. 27, 2019, issued in connection with U.S. Appl. No. 15/405,931, filed Jan. 13, 2017, 13 pages.
Non-Final Office Action dated Aug. 28, 2019, issued in connection with U.S. Appl. No. 16/422,160, filed May 24, 2019, 14 pages.
Non-Final Office Action dated Jul. 5, 2019, issued in connection with U.S. Appl. No. 16/383, 561, filed Apr. 12, 2019, 12 pages.
Non-Final Office Action dated Jul. 5, 2019, issued in connection with U.S. Appl. No. 16/383,565, filed Apr. 12, 2019, 11 pages.
Notice of Allowance dated Sep. 9, 2019, issued in connection with U.S. Appl. No. 16/383,561, filed Apr. 12, 2019, 18 pages.
Notice of Allowance dated Jan. 15, 2019, issued in connection with U.S. Appl. No. 15/487,686, filed Apr. 14, 2017, 8 pages.
Notice of Allowance dated May 30, 2019, issued in connection with U.S. Appl. No. 16/129,758, filed Sep. 12, 2018, 7 pages.
Notice of Allowance dated Sep. 5, 2019, issued in connection with U.S. Appl. No. 16/383,565, filed Apr. 12, 2019, 14 pages.
Notice of Allowance dated Dec. 7, 2018, issued in connection with U.S. Appl. No. 15/228,812, filed Aug. 4, 2016, 7 pages.
Sonos, Inc. v. Implicit, LLC: Declaration of Roman Chertov in Support of the Inter Partes Review of U.S. Pat. No. 7,391,791 dated Mar. 9, 2018, 92 pages.
Sonos, Inc. v. Implicit, LLC: Declaration of Roman Chertov in Support of the Inter Partes Review of U.S. Pat. No. 8,942,252 dated Mar. 9, 2018, 81 pages.
Sonos, Inc. v. Lenbrook Industries Limited et al., Defendants' Answer to Plaintiff's Complaint—Exhibit A, filed Oct. 14, 2019, 3 pages.
Sonos, Inc. v. Lenbrook Industries Limited et al., Defendants' Answer to Plaintiff's Complaint—Exhibit C, filed Oct. 14, 2019, 16 pages.
Sonos, Inc. v. Lenbrook Industries Limited et al., Defendants' Answer to Plaintiff's Complaint—Exhibit D, filed Oct. 14, 2019, 36 pages.
Sofas, Inc. v. Lenbrook Industries Limited et al., Defendants' Answer to Plaintiff's Complaint—Exhibit E, filed Oct. 14, 2019, 21 pages.
Sonos, Inc. v. Lenbrook Industries Limited et al., Defendants' Answer to Plaintiff's Complaint, filed Oct. 14, 2019, 56 pages.
Notice of Allowance dated Nov. 4, 2019, issued in connection with U.S. Appl. No. 16/422160, filed May 24, 2019, 13 pages.
Sofas, Inc. v. Lenbrook Industries Limited et al., Defendants' First Amended Answer and Counterclaims to Plaintiff's Complaint, filed Nov. 14, 2019, 66 pages.
Related Publications (1)
Number Date Country
20190014413 A1 Jan 2019 US
Provisional Applications (1)
Number Date Country
60825407 Sep 2006 US
Continuations (3)
Number Date Country
Parent 14629937 Feb 2015 US
Child 16128443 US
Parent 14299847 Jun 2014 US
Child 14629937 US
Parent 13083499 Apr 2011 US
Child 14299847 US
Continuation in Parts (2)
Number Date Country
Parent 13013740 Jan 2011 US
Child 13083499 US
Parent 11853790 Sep 2007 US
Child 13013740 US