This application relates generally to playing audio, and more particularly to synchronizing audio playback from multiple outputs to an orientation of a device, or video playing on a device.
The rise of portable electronic devices has provided unprecedented access to information and entertainment. Many people use portable computing devices, such as smart phones, tablet computing devices, portable content players, and the like to store and play back both audio and audiovisual content. For example, it is common to digitally store and play music, movies, home recordings and the like.
Many modern portable electronic devices may be turned by a user to re-orient information displayed on a screen of the device. As one example, some people prefer to read documents in a portrait mode while others prefer to read documents shown in a landscape format. As yet another example, many users will turn an electronic device on its side while watching widescreen video to increase the effective display size of the video.
Many current electronic devices, even when re-oriented in this fashion, continue to output audio as if the device is in a default orientation. That is, left channel audio may be omitted from the same speaker(s) regardless of whether or not the device is turned or otherwise re-oriented; the same is true for right channel audio and other audio channels.
One embodiment described herein takes the form of a method for outputting audio from a plurality of speakers associated with an electronic device, including the operations of: deter mining an orientation of video displayed by the electronic device; using the determined orientation of video to determine a first set of speakers generally on a left side of the video being displayed by the electronic device; using the determined orientation of video to determine a second set of speakers generally on a right side of the video being displayed by the electronic device; routing left channel audio to the first set of speakers for output therefrom; and routing right channel audio to the second set of speakers for output therefrom.
Another embodiment takes the form of an apparatus for outputting audio, including: a processor; an audio processing router operably connected to the processor; a first speaker operably connected to the audio processing router; a second speaker operably connected to the audio processing router; a video output operably connected to the processor, the video output operative to display video; an orientation sensor operably connected to the audio processing router and operative to output an orientation of the apparatus; wherein the audio processing router is operative to employ at least one of the orientation of the apparatus and an orientation of the video displayed on the vide output to route audio to the first speaker and second speaker for output.
Still another embodiment takes the form of a method for outputting audio from an electronic device, including the operations of: determining a first orientation of the electronic device; based on the first orientation, routing a first audio channel to a first set of speakers; based on the first orientation, routing a second audio channel to a second set of speakers; determining that the electronic device is being re-oriented from the first orientation to a second orientation; based on the determination that the electronic device is being re-oriented, transitioning the first audio channel to a third set of speakers; and based on the determination that the electronic device is being re-oriented, transitioning the second audio channel to a fourth set of speakers; wherein the first set of speakers is different from the third set of speakers; the second set of speakers is different from the fourth set of speakers; and during the operation of transitioning the first set of audio, playing at least a portion of the first audio channel and the second audio channel from at least one of the first set of speakers and third set of speakers.
Generally, embodiments described herein may take the form of devices and methods for matching an audio output to an orientation of a device providing the audio output. Thus, for example, as a device is rotated, audio may be routed to device speakers in accordance with the video orientation. To elaborate, consider a portable device having two speakers, as shown in
When the device 100 is rotated 180 degrees, as shown in
It should be appreciated that certain embodiments may have more than two speakers, or may have two speakers positioned in different locations than those shown in
Example Portable Device
Turning now to
Audio may be inputted to the device through an audio input 170 or may be stored on the storage medium 150 as a digital file. Audio may be it or stored alone, as part of audiovisual content (e.g., movies, television shows, presentations and the like), or as part of a data file or structure (such as a video game or other digital file incorporating audio). The audio may be formatted for any number of channels and/or subchannels, such as 5.1 audio, 7.1 audio, stereo and the like. Similarly, the audio may be encoded or processed in any industry-standard fashion. Including any of the various processing techniques associated with DOLBY Laboratories, THX, and the like.
The processor 130 generally controls various operations, inputs and outputs of the electronic device. The processor 130 may receive user inputs from a variety of user interfaces, including buttons, touch-sensitive surfaces, keyboards, mice and the like, (For simplicity's sake, no user interfaces are shown in
The speakers 110, 120 output audio in accordance with an audio routing determined by the audio processing router 140 (discussed below). The speakers may output any audio provided to them by the audio processing router and/or the processor 130.
The storage medium 150 generally stores digital data, optionally including audio files. Sample digital audio files suitable for storage on the storage medium 150 include MPEG-3 and MPEG-4 audio, Advanced Audio Coding audio, Waveform Audio Format audio files, and the like. The storage medium 150 may also store other types of data, software, and the like. In some embodiments, the audio processing router 140 may be embodied as software and stored on the storage medium. The storage medium may be any type of digital storage suitable for use with the electronic device 100, including magnetic storage, flash storage such as flash memory, solid-state storage, optical storage and so on.
Generally, the electronic device 100 may use the orientation sensor 160 to determine an orientation or motion of the device; this sensed orientation and/or motion may be inputted to the audio processing router 140 in order to route or re-route audio to or between speakers. As one example, the orientation sensor 160 may detect a rotation of the device 100. The output of the orientation sensor may be inputted to the orientation sensor, which changes the routing of certain audio channels from a first speaker configuration to a second speaker configuration. The output of the orientation sensor may be referred to herein as “sensed motion” or “sensed orientation.”
It should be appreciated that the orientation sensor 160 may detect motion, orientation, absolute position and/or relative position. The orientation sensor may be an accelerometer, gyroscope, global positioning system sensor, infrared or other electromagnetic sensor, and the like. As one example, the orientation sensor may be a gyroscope and detect rotational motion of the electronic device 100. As another example the orientation sensor may be a proximity sensor and detect motion of the device relative to a user. In some embodiments, multiple sensors may be used or aggregated. The use of multiple sensors is contemplated and embraced by this disclosure, although only a single sensor is shown in
The audio processing router 140 is generally responsible for receiving an audio input and a sensed motion and determining an appropriate audio output that is relayed to the speakers 110, 120. Essentially, the audio processing router 140 connects a number of audio input channels to a number of speakers for audio output. “Input channels” or “audio channels,” as used herein, refers to the discrete audio tracks that may each be outputted from a unique speaker, presuming the electronic device 100 (and audio processing router 140) is configured to recognize and decode the audio channel format and has sufficient speakers to output each channel from a unique speaker. Thus, 5.1 audio generally has five channels: front left; center; front right; rear left; and rear right. The “5” in “5.1” is the number of audio channels, while the “0.1” represents the number of subwoofer outputs supported by this particular audio format. (As bass frequencies generally sound omnidirectional, many audio formats send all audio below a certain frequency to a common subwoofer or sub roofers.)
The audio processing router 140 initially may receive audio and determine the audio format, including the number of channels. As part of its input signal processing operations, the audio processing router may map the various channels to a default speaker configuration, thereby producing a default audio map. For example, presume an audio source is a 5.1 source, as discussed above. If the electronic device 100 has two speakers 110, 120 as shown in
As the electronic device 100 is rotated or re-oriented, the sensor 160 may detect these motions and produce a sensed motion or sensed orientation signal. This signal may indicate to the audio processing router 140 and/or processor 130 the current orientation of the electronic device, and thus the current position of the speakers 110, 120. Alternatively, the signal may indicate changes in orientation or a motion of the electronic device. If the signal corresponds to a change in orientation or a motion, the audio routing processor 140 or the processor 130 may use the signal to calculate a current orientation. The current orientation, or the signal indicating the current orientation, may be used to determine a current position of the speakers 110, 120. This current position, in turn, may be used to determine which speakers are considered left speakers, right speakers, center speakers and the like and thus which audio channels are mapped to which speakers.
It should be appreciated that this input signal processing performed by the audio processing router 140 alternatively may be done without reference to the orientation of the electronic device 100. In addition to input signal processing, the audio processing router 140 may perform output signal processing. When performing output signal processing, the audio processing router 140 may use the sensed motion or sensed orientation to re-route audio to speakers in an arrangement different from the default output map.
The audio input 170 may receive audio from a source outside the electronic device 100. The audio input 170 may, for example, accept a jack or plug that connects the electronic device 100 to an external audio source. Audio received through the audio input 170 is handed by the audio processing router 140 in a manner similar to audio retrieved from a storage device 150.
Example of Operation
In operation 410, the audio processing router 140 creates an initial audio map. The audio map generally matches the audio channels of the audio source to the speaker configuration of the device. Typically, although not necessarily, the audio processing router attempts to ensure that left and right channel audio outputs (whether front or back) are sent to speakers on the left and right sides of the device, respectively, given the device's current orientation. Thus, front and rear left channel audio may be mixed and sent to the left speaker(s) while the front and rear right channel audio may be mixed and sent to the right speaker(s). In alternative embodiments, the audio processing router may create or retrieve a default audio map based on the number of input audio channels and the number of speakers in the device 100 and assume a default or baseline orientation, regardless of the actual orientation of the device.
Center channel audio may be distributed across multiple speakers or sent to a single speaker, as necessary. As one example, if there is no approximately centered speaker for the electronic device 100 in its current orientation, center channel audio may be sent to one or more speakers on both the left and right sides on the device. If there are more speakers on one side than the other, gain may be applied to the center channel to compensate for the disparity in speakers. As yet another option, the center channel may be suppressed entirely if no centered speaker exists.
Likewise, the audio processing router 140 may use gain or equalization to account for differences in the number of speakers on the left and right sides of the electronic device 100. Thus, if one side has more speakers than the other, equalization techniques may normalize the volume of the audio emanating from the left-side and right-side speaker(s). It should be noted that “left-side” and “right-side” speakers may refer not only to speakers located at or adjacent the left or right sides of the electronic device, but also speakers that are placed to the left or right side of a centerline of the device. Again, it should be appreciated that these terms are relative to a device's current orientation.
A sensed motion and/or sensed orientation may be used to determine the orientation of the speakers. The sensed motion/orientation provided by the sensor may inform the audio routing processor of the device's current orientation, or of motion that may be used, with a prior known orientation, to determine a current orientation. The current speaker configuration (e.g., which speakers 110 are located on a left or right side or left or right of a centerline of the device 100) may be determined from the current device orientation.
Once the audio map is created, the embodiment may determine in operation 415 if the device orientation is locked. Many portable devices permit a user to lock an orientation, so that images displayed on the device rotate as the device rotates. This orientation lock may likewise be useful to prevent audio outputted by the device 100 from moving from speaker to speaker to account for rotation of the device.
If the device orientation is locked, then the method 400 proceeds to operation 425. Otherwise, operation 420 is accessed. In operation 420, the embodiment may determine if the audio map corresponds to an orientation of any video being played on the device 100. For example, the audio processing router 140 or processor 130 may make this determination in some embodiments. A dedicated processor or other hardware element may also make such a determination. Typically, as with creating an audio map, an output from an orientation and/or location sensor may be used in this determination. The sensed orientation/motion may either permit the embodiment to determine the present orientation based on a prior, known orientation and the sensed changes, or may directly include positional data, it should be noted that the orientation of the video may be different than the orientation of the device itself. As one example, a user may employ software settings to indicate that widescreen-formatted video should always be displayed in landscape mode, regardless of the orientation of the device. As another example, a user may lock the orientation of video on the device, such that it does not reorient as the device 100 is rotated.
In some embodiments, it may be useful to determine if the audio map matches an orientation of video being played on the device 100 in addition to, or instead of, determining if the audio map matches a device orientation. The video may be oriented differently from the device either through user preference, device settings (including software settings), or some other reason. A difference between video orientation and audio orientation (as determined through the audio map) may lead to a dissonance in user perception as well as audio and/or video miscues. It should be appreciated that operations 420 and 425 may both be present in some embodiments, although other embodiments may omit one or the other.
In the event that the audio map matches the video orientation in operation 420, operation 430 is executed as described below. Otherwise, operation 425 is accessed, in operation 435, the embodiment determines if the current audio map matches the device orientation. That is, the embodiment determines if the assumptions regarding speaker 110 location that are used to create the audio map are correct, given the current orientation of the device 100. Again, this operation may be bypassed or may not be present in certain embodiments, while in other embodiments it may replace operation 420.
If the audio map does match the device 100 orientation, then operation 430 is executed. Operation 430 will be described in more detail below. If the audio map and device orientation do not match in operation 425, then the embodiment proceeds to operation 435. In operation 435, the embodiment creates a new audio map using the presumed locations and orientations of the speakers, given either or both of the video orientation and device 100 orientation. The process for creating a new audio map is similar to that described previously.
Following operation 435, the embodiment executes operation 440 and transitions the audio between the old and new audio maps. The “new” audio map is that created in operation 435, while the “old” audio map is the one that existed prior to the new audio map's creation, in order to avoid abrupt changes in audio presentation (e.g., changing the speaker 110 from which a certain audio channel emanates), the audio processing router 140 or processor 130 may gradually shift audio outputs between the two maps. The embodiment may convolve the audio channels from the first map to the second map, as one example. As another example, the embodiment may linearly transition audio between the two audio maps. As yet another example, if rotation was detected in operation 430, the embodiment may determine or receive a rate of rotation and attempt to generally match the change between audio maps to the rate of rotation (again, convolution may be used to perform this function).
Thus, one or more audio channels may appear to fade out from a first speaker and fade in from a second speaker during the audio map transition. Accordingly, it is conceivable that a single speaker may be outputting both audio from the old audio map and audio from the new audio map simultaneously. In many cases, the old and new audio outputs may be at different levels to create the effect that the old audio map transitions to the new audio map. The old audio channel output may be negatively gained (attenuated) while the new audio channel output is positively gained across some time period to create this effect. Gain, equalization, filtering, time delays and other signal processing may be employed during this operation. Likewise, the time period for transition between first and second orientations may be used to determine the transition, or rate of transition, from an old audio map to a new audio map. In various embodiments, the period of transition may be estimated from the rate of rotation or other reorientation, may be based on past rotation or other reorientation, or may be a fixed, default value. Continuing this concept, transition between audio maps may happen on the fly for smaller angles; as an example, a 10 degree rotation of the electronic device may result in the electronic device reorienting audio between speakers to match this 10 degree rotation substantially as the rotation occurs.
In some embodiments, the transition between audio maps (e.g., the reorientation of the audio output) may occur only after a reorientation threshold has been passed. For example, remapping of audio channels to outputs may occur only once the device has rotated at least 90 degrees. In certain embodiment, the device may not remap audio until the threshold has been met and the device and stops rotating for a period of time. Transitioning audio from a first output to a second output may take place over a set period of time (such as one that is aesthetically pleasing to an average listener), in temporal sync (or near-sync) to the rotation of the device, or substantially instantaneously.
After operation 435, end state 440 is entered. It should be appreciated that the end state 440 is used for convenience only. In actuality, an embodiment may continuously check for re-orientation of a device 100 or video playing on a device and adjust audio outputs accordingly. Thus, a portion or all of this flowchart may be repeated.
Operation 430 will now be discussed. As previously mentioned, the embodiment may execute operation 430 upon a positive determination from either operations 420 or 425. In operation 430, the orientation sensor 160 determines if the device 100 is being rotated or otherwise reoriented. If not, end state 445 is executed. If so, operation 435 is executed as described above.
It should be appreciated that any or all of the foregoing operations may be omitted in certain embodiments. Likewise, operations may be shifted in order. For example, operations 420, 425 and 430 may all be rearranged with respect to one another. Thus,
As shown generally in at least
The number of speakers 110 present in an electronic device 100 typically influences the audio map created by the audio processing router 140 or processor 130. First, the numbers of speakers generally indicates how many left and/or right speakers exist and thus which audio channels may be mapped to which speakers. To elaborate, consider the electronic device 500 in the orientation shown in
For example, the audio processing router may downmix both the left front and left rear channels of a 5 channel audio source and send them to the first speaker 510. The right front and right rear channels may be downmixed and sent to the second speaker 520 in a similar fashion. Center audio may be mapped to the third speaker 530, as it is approximately at the vertical centerline of the device 500.
When the device is rotated 90 degrees, as shown in
It should be appreciated that alternative audio maps may be created, depending on a variety of factors such as user preference, programming of the audio processing router 140, importance or frequency of audio on a given channel and the like. As one example, the center channel may be played through all three speakers 510, 520, 530 when the device 500 is oriented as in
As another example, the audio processing router 140 may downmix the left front and left rear channels for presentation on the third speaker 530 in the configuration of
Gain may be applied to audio routed to a particular set of speakers. In certain situations, gain is applied in order to equalize audio of the left and right channels (front, rear or both, as the case may be). As one example, consider the orientation of the device 500 in
Gain may be used to equalize or normalize audio, or a user's perception of audio, in the event an electronic device 100 is laterally moved toward or away from a user. The device 100 may include a motion sensor sensitive to lateral movement such as a GPS sensor, accelerometer and the like. In some embodiments, a camera integrated into the device 100 may be used; the camera may capture images periodically and compare one to the other. The device 100, through the processor, may recognize a user, for example by extracting the user from the image using known image processing techniques, if the user's position or size changes from one captured image to another, the device may infer that the user has moved in a particular position. This information may be used to adjust the audio being outputted. In yet another embodiment, a presence detector (such as an infrared presence detector or the like) may be used for similar purposes.
For example, if the user (or a portion of the user's body, such as his head) appears smaller, the user has likely moved away from the device and the volume or gain may be increased. If the user appears larger, the user may have moved closer and volume/gain may be decreased, if the user shifts position in an image, he may have moved to one side or the device may have been moved with respect to him. Again, gain may be applied to the audio channels to compensate for this motion. As one example, speakers further away from the user may have a higher gain than speakers near a user; likewise, gain may be increased more quickly for speakers further away than those closer when the relative position of the user changes.
Time delays may also be introduced into one or more audio channels. Time delays may be useful for syncing up audio outputted by a first set of the device's 100 speakers 110 nearer a user and audio outputted by a second set of speakers. The audio emanating from the first set of speakers may be slightly time delayed in order to create a uniform sound with the audio emanating from the second set of speakers, for example. The device 100 may determine what audio to time delay by determining which speakers may be nearer a user based on the devices orientation, as described above, or by determining a distance of various speakers from a user, also as described above.
The foregoing description has broad application. For example, while examples disclosed herein may focus on utilizing a smart phone or mobile computing device, it should be appreciated that the concepts disclosed herein may equally apply to other devices that output audio. As one example, an embodiment may determine an orientation of video outputted by a projector or on a television screen, and route audio according to the principles set forth herein to a variety of speakers in order to match the video orientation. As another example, certain embodiments may determine an orientation of displayed video on an electronic device and match audio outputs to corresponding speakers, as described above. However, if the device determines that a video orientation is locked (e.g., the orientation of the video does not rotate as the device rotates), then the device may ignore video orientation and use the device's orientation to create and employ an audio map.
Similarly, although the audio routing method may be discussed with respect to certain operations and orders of operations, it should be appreciated that the techniques disclosed herein may be employed with certain operations omitted, other operations added or the order of operations changed. Accordingly, the discussion of any embodiment is meant only to be an example and is not intended to suggest that the scope of the disclosure, including the claims, is limited to these examples.
This application is a continuation of co-pending U.S. application Ser. No. 13/302,673 filed on Nov. 22, 2011.
Number | Name | Date | Kind |
---|---|---|---|
1893291 | Kwartin | Jan 1933 | A |
4068103 | King et al. | Jan 1978 | A |
4081631 | Feder | Mar 1978 | A |
4089576 | Barchet | May 1978 | A |
4245642 | Skubitz et al. | Jan 1981 | A |
4466441 | Skubitz et al. | Aug 1984 | A |
4658425 | Julstrom | Apr 1987 | A |
4684899 | Carpentier | Aug 1987 | A |
5060206 | DeMetz | Oct 1991 | A |
5106318 | Endo et al. | Apr 1992 | A |
5121426 | Baumhauer, Jr. et al. | Jun 1992 | A |
5293002 | Grenet et al. | Mar 1994 | A |
5335011 | Addeo et al. | Aug 1994 | A |
5406038 | Reiff et al. | Apr 1995 | A |
5570324 | Geil | Oct 1996 | A |
5604329 | Kressner et al. | Feb 1997 | A |
5619583 | Page et al. | Apr 1997 | A |
5649020 | McClurg et al. | Jul 1997 | A |
5691697 | Carvalho et al. | Nov 1997 | A |
5733153 | Takahashi et al. | Mar 1998 | A |
5879598 | McGrane | Mar 1999 | A |
6036554 | Koeda et al. | Mar 2000 | A |
6069961 | Nakazawa | May 2000 | A |
6073033 | Campo | Jun 2000 | A |
6129582 | Wilhite et al. | Oct 2000 | A |
6138040 | Nicholls et al. | Oct 2000 | A |
6151401 | Annaratone | Nov 2000 | A |
6154551 | Frenkel | Nov 2000 | A |
6192253 | Charlier et al. | Feb 2001 | B1 |
6246761 | Cuddy | Jun 2001 | B1 |
6278787 | Azima | Aug 2001 | B1 |
6317237 | Nakao et al. | Nov 2001 | B1 |
6324294 | Azima et al. | Nov 2001 | B1 |
6332029 | Azima et al. | Dec 2001 | B1 |
6342831 | Azima | Jan 2002 | B1 |
6469732 | Chang et al. | Oct 2002 | B1 |
6618487 | Azima | Sep 2003 | B1 |
6757397 | Buecher et al. | Jun 2004 | B1 |
6813218 | Antonelli et al. | Nov 2004 | B1 |
6829018 | Lin et al. | Dec 2004 | B2 |
6882335 | Saarinen | Apr 2005 | B2 |
6914854 | Heberley | Jul 2005 | B1 |
6934394 | Anderson | Aug 2005 | B1 |
6980485 | McCaskill | Dec 2005 | B2 |
7003099 | Zhang et al. | Feb 2006 | B1 |
7054450 | McIntosh et al. | May 2006 | B2 |
7082322 | Harano | Jul 2006 | B2 |
7130705 | Amir et al. | Oct 2006 | B2 |
7154526 | Foote et al. | Dec 2006 | B2 |
7158647 | Azima et al. | Jan 2007 | B2 |
7190798 | Yasuhara | Mar 2007 | B2 |
7194186 | Strub et al. | Mar 2007 | B1 |
7263373 | Mattisson | Aug 2007 | B2 |
7266189 | Day | Sep 2007 | B1 |
7346315 | Zurek et al. | Mar 2008 | B2 |
7378963 | Begault et al. | May 2008 | B1 |
7527523 | Yohn et al. | May 2009 | B2 |
7536029 | Choi et al. | May 2009 | B2 |
7570772 | Sorensen et al. | Aug 2009 | B2 |
7679923 | Inagaki et al. | Mar 2010 | B2 |
7848529 | Zhang et al. | Dec 2010 | B2 |
7867001 | Ambo et al. | Jan 2011 | B2 |
7878869 | Murano et al. | Feb 2011 | B2 |
7912242 | Hikichi | Mar 2011 | B2 |
7966785 | Zadesky et al. | Jun 2011 | B2 |
8030914 | Alameh et al. | Oct 2011 | B2 |
8031853 | Bathurst et al. | Oct 2011 | B2 |
8055003 | Mittleman et al. | Nov 2011 | B2 |
8116505 | Kawasaki-Hedges et al. | Feb 2012 | B2 |
8116506 | Kuroda et al. | Feb 2012 | B2 |
8135115 | Hogg et al. | Mar 2012 | B1 |
8184180 | Beaucoup | May 2012 | B2 |
8226446 | Kondo et al. | Jul 2012 | B2 |
8300845 | Zurek et al. | Oct 2012 | B2 |
8401210 | Freeman | Mar 2013 | B2 |
8447054 | Bharatan et al. | May 2013 | B2 |
8452019 | Fomin et al. | May 2013 | B1 |
8488817 | Mittleman et al. | Jul 2013 | B2 |
8574004 | Tarchinski et al. | Nov 2013 | B1 |
8620162 | Mittleman | Dec 2013 | B2 |
8965014 | Castor-Perry | Feb 2015 | B2 |
20010011993 | Saarinen | Aug 2001 | A1 |
20010017924 | Azima et al. | Aug 2001 | A1 |
20010026625 | Azima et al. | Oct 2001 | A1 |
20020012442 | Azima et al. | Jan 2002 | A1 |
20020037089 | Usuki et al. | Mar 2002 | A1 |
20020044668 | Azima et al. | Apr 2002 | A1 |
20020150219 | Jorgenson et al. | Oct 2002 | A1 |
20030048911 | Furst et al. | Mar 2003 | A1 |
20030053643 | Bank et al. | Mar 2003 | A1 |
20030161493 | Hosler | Aug 2003 | A1 |
20030171936 | Sall et al. | Sep 2003 | A1 |
20030236663 | Dimitrova et al. | Dec 2003 | A1 |
20040013252 | Craner et al. | Jan 2004 | A1 |
20040156527 | Stiles et al. | Aug 2004 | A1 |
20040203520 | Schirtzinger et al. | Oct 2004 | A1 |
20040263636 | Cutler et al. | Dec 2004 | A1 |
20050129267 | Azima | Jun 2005 | A1 |
20050147273 | Azima et al. | Jul 2005 | A1 |
20050152565 | Jouppi et al. | Jul 2005 | A1 |
20050182627 | Tanaka et al. | Aug 2005 | A1 |
20050209848 | Ishii | Sep 2005 | A1 |
20050226455 | Aubauer et al. | Oct 2005 | A1 |
20050238188 | Wilcox | Oct 2005 | A1 |
20050271216 | Lashkari | Dec 2005 | A1 |
20060005156 | Korpipaa | Jan 2006 | A1 |
20060023898 | Katz | Feb 2006 | A1 |
20060045294 | Smyth | Mar 2006 | A1 |
20060072248 | Watanabe et al. | Apr 2006 | A1 |
20060206560 | Kanada | Sep 2006 | A1 |
20060239471 | Mao et al. | Oct 2006 | A1 |
20060256983 | Kenoyer et al. | Nov 2006 | A1 |
20060279548 | Geaghan | Dec 2006 | A1 |
20070011196 | Ball et al. | Jan 2007 | A1 |
20070025555 | Gonai | Feb 2007 | A1 |
20070188901 | Heckerman et al. | Aug 2007 | A1 |
20070291961 | Shin | Dec 2007 | A1 |
20080063211 | Kusunoki | Mar 2008 | A1 |
20080130923 | Freeman | Jun 2008 | A1 |
20080175408 | Mukund et al. | Jul 2008 | A1 |
20080204379 | Perez-Noguera | Aug 2008 | A1 |
20080292112 | Valenzuela et al. | Nov 2008 | A1 |
20080310663 | Shirasaka et al. | Dec 2008 | A1 |
20090018828 | Nakadai et al. | Jan 2009 | A1 |
20090048824 | Amada | Feb 2009 | A1 |
20090060222 | Jeong | Mar 2009 | A1 |
20090070102 | Maegawa | Mar 2009 | A1 |
20090094029 | Koch et al. | Apr 2009 | A1 |
20090247237 | Mittleman et al. | Oct 2009 | A1 |
20090274315 | Carnes et al. | Nov 2009 | A1 |
20090304198 | Herre et al. | Dec 2009 | A1 |
20090316943 | Munoz et al. | Dec 2009 | A1 |
20100062627 | Ambo et al. | Mar 2010 | A1 |
20100066751 | Ryu et al. | Mar 2010 | A1 |
20100080084 | Chen et al. | Apr 2010 | A1 |
20100103776 | Chan | Apr 2010 | A1 |
20100110232 | Zhang | May 2010 | A1 |
20110002487 | Panther | Jan 2011 | A1 |
20110033064 | Johnson et al. | Feb 2011 | A1 |
20110038489 | Visser et al. | Feb 2011 | A1 |
20110087491 | Wittenstein et al. | Apr 2011 | A1 |
20110150247 | Oliveras | Jun 2011 | A1 |
20110161074 | Pance et al. | Jun 2011 | A1 |
20110164141 | Tico | Jul 2011 | A1 |
20110193933 | Ryu et al. | Aug 2011 | A1 |
20110243369 | Wang | Oct 2011 | A1 |
20110274303 | Filson et al. | Nov 2011 | A1 |
20110316768 | McRae | Dec 2011 | A1 |
20120082317 | Pance et al. | Apr 2012 | A1 |
20120177237 | Shukla et al. | Jul 2012 | A1 |
20120230497 | Dressler | Sep 2012 | A1 |
20120243698 | Elko et al. | Sep 2012 | A1 |
20120250928 | Pance et al. | Oct 2012 | A1 |
20120263019 | Armstrong-Munter | Oct 2012 | A1 |
20120306823 | Pance et al. | Dec 2012 | A1 |
20120330660 | Jaiswal | Dec 2012 | A1 |
20130017738 | Asakuma et al. | Jan 2013 | A1 |
20130028443 | Pance et al. | Jan 2013 | A1 |
20130028446 | Krzyzanowski | Jan 2013 | A1 |
20130051601 | Hill et al. | Feb 2013 | A1 |
20130108054 | Groh | May 2013 | A1 |
20130129122 | Johnson | May 2013 | A1 |
20130142355 | Isaac | Jun 2013 | A1 |
20130142356 | Isaac | Jun 2013 | A1 |
20130164999 | Ge et al. | Jun 2013 | A1 |
20130259281 | Filson et al. | Oct 2013 | A1 |
20130280965 | Kojyo | Oct 2013 | A1 |
Number | Date | Country |
---|---|---|
2094032 | Aug 2009 | EP |
2310559 | Aug 1997 | GB |
2342802 | Apr 2000 | GB |
62-189898 | Aug 1987 | JP |
2102905 | Apr 1990 | JP |
2003-032776 | Jan 2003 | JP |
2004153018 | May 2004 | JP |
2006297828 | Nov 2006 | JP |
2007-081928 | Mar 2007 | JP |
WO 0193554 | Dec 2001 | WO |
WO03049494 | Jun 2003 | WO |
WO04025938 | Mar 2004 | WO |
WO 2007045908 | Apr 2007 | WO |
WO 2007083894 | Jul 2007 | WO |
WO08153639 | Dec 2008 | WO |
WO2009017280 | Feb 2009 | WO |
WO2011057346 | May 2011 | WO |
WO2011061483 | May 2011 | WO |
Entry |
---|
European Extended Search Report, EP 12178106.6, dated Jul. 11, 2012, 8 pages. |
PCT International Search Report and Written Opinion, PCT/US2012/057909, dated Feb. 19, 2013, 14 pages. |
PCT International Preliminary Report on Patentability, PCT/US2011/052589, dated Apr. 11, 2013, 9 pages. |
Baechtle et al., “Adjustable Audio Indicator,” IBM, 2 pages, Jul. 1, 1984. |
Pingali et al., “Audio-Visual Tracking for Natural Interactivity,” Bell Laboratories, Lucent Technologies, pp. 373-382, Oct. 1999. |
“Snap fit theory”, Feb. 23, 2005, DSM, p. 2. |
PCT International Search Report and Written Opinion, PCT/US2011/052589, dated Feb. 23, 2012, 13 pages. |
PCT International Search Report and Written Opinion, PCT/US2012/0045967 dated Feb. 6, 2014,10 pages. |
Number | Date | Country | |
---|---|---|---|
20150023533 A1 | Jan 2015 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13302673 | Nov 2011 | US |
Child | 14507582 | US |