The invention relates to audio and, more particularly, to multi-channel (two or more) loudspeaker reproduction, indoor navigation, and near-field-communication (NFC).
An electronic device typically comprises a variety of components and/or features that enable users to interact with the electronic device. Some considerations when providing these features in a portable electronic device may include, for example, compactness, suitability for mass manufacturing, durability, and ease of use. Increase of computing power of portable devices is turning them into versatile portable computers, which can be used for multiple different purposes. Therefore versatile components and/or features are needed in order to take full advantage of capabilities of mobile devices.
Some electronic devices may be used with a multi-channel audio file which a listener seeks to play back. Richness when playing back the multi-channel audio file is enhanced by having the loudspeakers also properly placed, but the audio file is of course not tied to any particular set of loudspeakers. Additionally, in some instances the physical location of portable wireless speakers can be arbitrary. This can prevent the listener from experiencing an aimed spatial audio experience. Regardless of the listener's familiarity with specifics of audio technology, an aimed spatial experience is what people have come to expect from a 5:1 or even 7:1 arrangement for multi-channel audio related for example to watching movies. Hardwired speakers are typically spatially situated purposefully to achieve a proper surround sound. A similar spatial pre-arrangement of wireless loudspeakers with assigned audio channels tends to lose effectiveness over time when individual wireless loudspeakers are relocated away from the position designated for the surround-sound channel provided to it.
Additionally, whether the loudspeakers are wired or wireless those previous audio systems that rely on pre-arranged spatial positioning of the speakers had the centralized host device that is handling the audio file (for example, a conventional stereo amplifier or a host/master mobile phone) output different ones of the audio channels to different speakers or different speaker-hosting devices.
Various aspects of examples of the invention are set out in the claims.
In accordance with one aspect of the invention, a method is disclosed. Near field communication is detected between at least two devices. A location of at least one of the at least two devices is determined based on the detected near field communication. An audio channel of a multi-channel audio file is assigned based on the determined location of the at least one of the at least two devices.
In accordance with another aspect of the invention, an apparatus is disclosed. The apparatus includes at least one processor and at least one memory including computer program code. The at least one memory and the computer program code configured to, with the at least one processor, cause the apparatus to perform at least the following, detect near field communication, determine a location of a device based on the detected near field communication, and assign a channel of a multi-channel audio file based on the determined location of the device.
In accordance with another aspect of the invention, a computer program product including a non-transitory computer-readable medium bearing computer program code embodied therein for use with a computer is disclosed. The computer program code includes code for detecting near field communication. Code for determining a location of a device based on the detected near field communication. Code for assigning a channel of a multi-channel audio file based on the determined location of the device.
For a more complete understanding of example embodiments of the present invention, reference is now made to the following descriptions taken in connection with the accompanying drawings in which:
Example embodiments of the invention and its potential advantages are understood by referring to
The exemplary and non-limiting embodiments detailed below present a way for discovering the physical positions of different loudspeakers relative to one another, and then selecting an arrangement of those loudspeakers that is appropriate for playing back a multi-channel audio file. The arrangement has multiple distinct speakers, each outputting different channels of the unitary multi-channel audio file. For a better context of the flexibility of these teachings the examples below consider a variety of different-type of audio devices; some may be mobile terminals such as smart phones, some may be only stand-alone wireless speakers which may or may not have the capability of ‘knowing’ the relative position of other audio devices in the arrangement, and some may be MP3-only devices with some limited radio capability, to name a few non-limiting examples. So long as some other audio device has the capacity to discover neighboring audio devices, the discovering audio device can discover any other audio devices which themselves lack such discovering capability, learn the relative positions of all the various neighbor audio devices according to the teachings below, and then form an audio arrangement appropriate for the sound file to be played back. In the below examples any of the above types of host devices for a loudspeaker are within the scope of the term ‘audio device’, which refers to the overall host device rather than any individual loudspeaker unit. In some implementations each audio device may have wireless connectivity with the other audio devices, as well as the capability for sound reproduction/play back and possibly also sound capture/recording.
Any given audio device is not limited to hosting only one loudspeaker. In different implementations of these teachings any such audio device can host one loudspeaker which outputs only one of the audio multi-channels, or it may host two (or possibly more) loudspeakers which can output the same audio multi-channel (such as for example a mobile handset having two speakers which when implementing these teachings are considered too close to one another to output different audio multi-channels), or one audio device may host two (or possibly more) loudspeakers which each output different audio multi-channels (such as for example two speakers of a single mobile handset outputting different left- and right-surround audio channels). Other implementations may employ combinations of the above. In more basic implementations where individual audio devices are not distinguished in the discovery phase as having one or multiple loudspeakers, each host audio device may be assumed to have only one loudspeaker and the same audio channel allocated to the device is played out over all the loudspeakers hosted in that individual audio device.
Referring to
The arrangement shown in
It should be noted that while various exemplary embodiments of the invention have been described in connection with the arrangement 10 as comprising five audio devices 12, 14, 16, 18, 20, one skilled in the art will appreciate that the various exemplary embodiments of the invention are not necessarily so limited and that any suitable number of audio devices (and/or speakers) may be provided. For example, in some embodiments of the invention, the arrangement 10 may comprise two audio devices where a first device is used to play back/output the front channels L, R, and a second device is used to play back/output the rear channels Ls (left surround), Rs (right surround). In some other embodiments of the invention, the arrangement 10 may comprise three audio devices arranged such that a first device plays back front L and R audio channels, a second device plays back rear audio channel Ls and video-L channel, and a third device plays back rear audio channel Rs and video-R channel. In yet another embodiment of the invention, the arrangement 10 may comprise four audio devices, wherein a first device plays back front L audio channel and left video-L channel, a second device plays back front audio channel R and right video-R channel, a third device plays back rear audio channel Ls, and a fourth device plays back rear audio channel Rs. Additionally, in other embodiments of the invention, the arrangement 10 may comprise more than five audio devices.
The arrangements mentioned above are exemplary of the set of audio devices which is discovered and selected for multi-channel playback according to these teachings. These are not intended to be comprehensive but rather serve as various examples of the possibilities which result from the various audio device arrangements. According to some embodiments of the invention, knowing this arrangement allows the audio system and audio devices to know what is the role of particular speakers in the whole system (for example, so one device can know that it is the right front channel speaker in the system of devices and another device can know that it is the left front channel speaker in the system of devices, etc.).
By determining locations of the audio devices, a mesh of speakers can be formed. Each audio device is a “node” and the distance between two nodes is a “path”. Eventually, the path between each node is known and hence the arrangement of speakers can be found. The arrangement might be static or in some cases as with mobile terminals it may be dynamic, and so to account for the latter case in some implementations the audio device discovery is periodically or continuously updated.
In some embodiments such as where each audio device has the capability for direct radio communications with each other audio device (for example, they are each a mobile terminal handset), synchronous operation can be enabled by a single (master) mobile terminal allocating the audio channels to the different other audio devices/mobile terminals via radio-frequency signaling (for example, via Bluetooth/personal area network including Bluetooth Smart, wireless local area network WLAN/WiFi, ANT+, device-to-device D2D communications, or any other radio access technology which is available among the audio devices), and the different audio devices/mobile terminals then synchronously play out their respectively assigned audio channel for a much richer audio environment. Or in other embodiments each audio device has the identical multi-channel audio file and only plays out its respectively assigned or allocated audio channels synchronously with the other audio devices.
Synchronous play back or recording can be achieved when one device, termed herein as the ‘master’ device, provides a synchronization signal for that playback, or alternatively deciding what (third-party) signal will serve as the synchronization signal. For example, the master device may choose that a beacon broadcast by some nearby WiFi network will be the group-play synchronization signal. The master device will in this case send to the ‘slave’ audio devices some indication of what is to be the synchronization signal the various audio devices should all use for the group play back. Whether master or slave device is grounded in synchronization; it may be that the extent of control that the master device exercises over all of the other ‘slave’ audio devices is limited only to controlling timing of the audio file playback, which in the above examples is accomplished via selection of the synchronization signal.
In general, devices for smart environments benefit from having their physical location known, but even if low-cost Bluetooth based solutions for indoor navigation emerge in the market soon, and high-precision outdoor location is already available with differential GPS, it is impractical to implement them in every device, and also, it makes little engineering or economical sense to fit devices that are seldom moved with indoor navigation capabilities. According to various exemplary embodiments of the invention, one common use case where determining the location of devices with respect to the user's location is in multi-channel sound reproduction with loudspeakers, where at least the propagation delay, and possibly also sound level and/or equalization settings are adjusted to correspond to the distance between the user and the sound sources, and the channel selection generally conforms to the physical arrangement of the loudspeakers in the listening space (for example, left and right or front and back may not be reversed).
According to various exemplary embodiments of the invention, the location of each peripheral device is determined with an appropriate navigation system when the device is brought into a close proximity interaction (or near field communication) with a portable device. In some embodiments, the close proximity interaction may be a contact (or devices placed in close proximity to each other) determined with near field communications (NFC) or any other suitable short distance communication method. In some other exemplary embodiments of the invention, the close proximity interaction may be a contact that can be detected by tapping the object with the terminal device, so accelerometer or the microphone triggers the detection of the location of objects without near-field communication capabilities (room boundaries, conventional loudspeakers, for example) can be determined. The location information is stored then in the portable device, peripheral device, or both, and transmitted to other devices (such as, other loudspeakers, amplifier, for example). Also the listening location can be determined by the portable device, which can used for playback or which acts as a connection hub for the loudspeaker system. According to some embodiments of the invention, the basic settings (such as, delay and level, for example) are determined from the location information, and if the order of channels assigned to each loudspeaker does not correspond to the physical layout of the loudspeakers the channel assignments are changed accordingly. Additionally, more detailed direction information can be used to adjust the rendering of spatial audio (channel mixing, or object-based rendering). It should be noted that, according to various exemplary embodiments, the near field communication or contact can be determined by a near field communication (NFC) method, or one or more sensors. Additionally, the one or more sensors could be an accelerometer or a microphone, for example.
It should be noted that the term ‘close proximity interaction’ mentioned above and throughout the specification refers to any type of contact between the devices or placement between the devices such that they are in close proximity to each other. For example, close proximity interactions include tapping devices together, contacting devices together, as well as near field communications (NFC) or any other suitable short distance communication method between the devices.
Referring now also to
The method 100 provides a detection flowchart, or a loudspeaker setup, wherein the user assigns the channel and object information.
Referring now also to
The method 200 provides a detection flowchart, or a loudspeaker setup, wherein the system automatically assigns channel information to each loudspeaker based on their location.
According to some embodiments of the invention, the application may display loudspeaker setup information on the master device. For example,
According to some exemplary embodiments of the invention, the implementing application may find a ‘best-fit’for play back from among those discovered devices (or from among whichever plurality of devices remain available for inclusion in the arrangement). For example,
In general, there is an idealized spatial arrangement for 5:1 surround sound. Some exemplary embodiments of the invention, the implementing device selects which audio devices best fit the idealized spatial arrangement and selects those as members of the arrangement. The implementing device can of course select more devices than there are channels, for example if there were two devices found near the position of device 18, for example, the implementing device can select them both and allocate the same right surround channel to them both. If for example the file to be played back is 5:1 surround sound but the implementing device finds only three devices, the spatial effect to be presented will be 3:1 surround sound because 5:1 is not possible given the discovered devices. For the more specific embodiment where the arrangement is selected for a best fit to an idealized spatial arrangement for achieving the intended spatial audio effect, in this example the best fit may then be 3:1 surround sound so the best fit for the case of play back does not have to match the multi-channel profile of the file that is to be played back.
For the case of multi-channel recording, the implementing device selects a type or profile of the multi-channel file as the spatial audio effect it would like the recording to present, such as for example stereo or 5:1 surround sound. For example, the implementing device may choose the spatial audio effect it would like to achieve based on what is the spatial arrangement of the devices it discovers. The implementing device may find there are several possible arrangements, and in the more particular ‘best fit’ embodiment choose the ‘best fit’ as the one which is deemed to record the richest audio environment. If there are only 4 devices found but their spatial arrangement is such that the best fit is 3:1 surround sound (L, C and R channels), the master device may then choose 3:1 and allocate channels accordingly.
Additionally, in some exemplary embodiments, the method may further include determining, based on the detected near field communication, a location of at least one of: one of the at least two devices; one or more other devices using one of the at least two devices; and the other of the at least two devices. For example, according to some embodiments, the multi-channel audio file is assigned based on the detected loudspeakers in example embodiments but the actual portable device (for example, the mobile phone) may be just used for detecting and assigning audio channels to other detected devices/loudspeakers. The portable device (mobile phone) may not be used for audio playback in alternative embodiments. According to some embodiments, if a first device is a portable device (such as a mobile phone, for example) and then a second device is a loudspeaker, then the location the first device is not generally determined (apart from the determination of the listening position) as the location of the portable device continuously changes, such as when the user taps a first loudspeaker and then a second loudspeaker and then third so on, for example. In some alternative embodiments, the first device location may be important if the phone is used for sound reproduction. In these alternative embodiments, then when the method comprises determining, based on the detected near field communication, a location of one of the at least two devices, this may correspond to a loudspeaker location detected, for example. When the method comprises determining, based on the detected near field communication, a location of one or more other devices using one of the at least two devices, this may correspond to one or more other loudspeaker locations detected, for example. When the method comprises determining, based on the detected near field communication, a location of the other of the at least two devices, this may correspond to portable device location detected, for example.
Various exemplary embodiments of the invention provide for discovering the physical positions of different loudspeakers relative to one another, and then selecting an arrangement of those loudspeakers that is appropriate for playing back a multi-channel audio file, the arrangement has multiple distinct speakers, each outputting different channels of the unitary multi-channel audio file. However it should be noted that the speakers are not required to output only in the audible range. For example in some embodiments, the speakers could include metadata in the inaudible range (such as, above 20 kHz, for example), wherein there is provided the use of sub-bands to transmit the metadata alongside the audio. Examples for the metadata could include, for example, general data transfer, speaker identification, sync signaling, distance estimation/confirmation, and distance change detection. Additionally, as the signaling is done in the inaudible audio range, it can happen simultaneously with the audio playback.
Referring now also to
Referring now also to
Referring now also to
Technical effects of any one or more of the exemplary embodiments provide a location based loudspeaker system configuration providing various advantages when compared to conventional configurations. Many of the conventional configurations use adjustment of multi-channel audio systems based on acoustical measurement of the transfer function between the loudspeakers and a microphone (system) in the listening location. In the conventional configurations this measurement is time-consuming, and some embedded systems suffer from reliability problems. Also, the basic acoustical measurements often rely on only one microphone, and thus are unable to determine the directions of the loudspeakers with respect to the listening position.
Without in any way limiting the scope, interpretation, or application of the claims appearing below, a technical effect of one or more of the example embodiments disclosed herein is that in multi-channel audio reproduction the information about system geometry is more precise and reliable than with sound-based system measurement. Although various exemplary embodiments may require some user interaction, it is still fast as compared to the conventional methods. Room geometry information can be used as initial data for more advanced room correction or measurement systems. However, the information does not take into account the acoustics of the loudspeakers and the listening room, so further corrections may be used, but the location based information is useful as a baseline even if additional acoustical measurements are used.
Another technical effect of one or more of the example embodiments disclosed herein is to provide for a location based loudspeaker system configuration to determine the speaker locations in a room or a listening area. Another technical effect of one or more of the example embodiments disclosed herein is related to multi-channel audio playback configuration wherein the playback geometry comprising the location of playback devices and the listening position are determined based on near field communication so as to reproduce optimized audio channels based on the known locations.
Another technical effect of one or more of the example embodiments disclosed herein is that the nodes (speakers) need not be active in order determine the arrangement/constellation, for example, any surface of the device/speaker can be located. Another technical effect of one or more of the example embodiments disclosed herein is that the channel order may be locked by assigning one speaker channel, for example, in the phone UI, while tapping the speaker location, wherein all the other speaker channels can be assigned optimally based on this one channel information. Furthermore, by tapping the listening position, the speaker channel assignment optimization can be further improved. Another technical effect of one or more of the example embodiments disclosed herein is that audio signaling to determine speaker locations is not needed/required. Another technical effect of one or more of the example embodiments disclosed herein is that the device/speaker does not need any extra features (such as a speaker having active multiple microphones and required DSP capabilities [for beamforming]), and any “dummy” object can be located.
Another technical effect of one or more of the example embodiments disclosed herein is the speaker arrangement/constellation estimation does not require active elements in the speaker, and/or additional software in the phone/device (such as various types of detection software). Various exemplary embodiments of the invention use the phone's, already existing, location information and assign that to different physical (or even virtual) objects, which allows locating any object in the environment. The accuracy of the system is determined by the accuracy of the phone's (indoor) positioning accuracy.
The above teachings may be implemented as a stored software application and hardware that allows the several distinct mobile devices to be configured to make a synchronized stereo/multichannel recording or play back together, in which each participating device contributes one or more channels of the recording or of the play back. In a similar fashion, a 3D video recording can be made using cameras of the various devices in the arrangement, with a stereo base that is much larger than the maximum dimensions of any one of the individual devices. Any two participating devices that are spaced sufficiently far apart could be selected for the arrangement of devices that will record the three dimensional video.
According to some embodiments of the invention, there may be one application running on the master device only, which controls the other slave devices to play out or record the respective channel that the master device assigns. Or in other embodiments some or all of the participating devices are each running their own application which aids in determining speaker and/or listener locations in the room.
In some embodiments of the invention, such as for the case of play back the slave devices can get the whole multi-channel file, or only their respective channel(s), from the master device. For the case of recording each can learn their channel assignments from the master device, and then after the individual channels are recorded they can send their respectively recorded channels to the master device for combining into a multi-channel file, or all the participating devices can upload their respective channel recordings to a web server which does the combining and makes the multi-channel file available for download.
The various participating devices do not need to be of the same type. If the arrangement of devices are not all of the same model it is inevitable that there will be frequency response and level differences between them, but these may be corrected automatically by the software application; for recording by the devices these corrections can be done during mixing of the final multi-channel recording, and for play back these can be done even dynamically using the microphone of mobile-terminal type devices to listen to the acoustic environment during play back and dynamically adjust amplitude or synthesis of their respective channel play back because any individual device knowing the arrangement and distances can estimate how the sound environment should sound be at its own microphone.
The master device and the other participating devices may for example be implemented as user mobile terminals or more generally referred to as user equipments UEs.
At least the master audio device 510 includes a controller, such as a computer or a data processor (DP) 510A, a computer-readable memory (MEM) 510B that tangibly stores a program of computer-readable and executable instructions (PROG) 510C such as the software application detailed in the various embodiments above, and in embodiments where the links 515A, 515B are radio links also a suitable radio frequency (RF) transmitter 510D and receiver 510E for bidirectional wireless communications over those RF wireless links 515A, 515B via one or more antennas 510F (two shown). The master audio device 510 may also have a Bluetooth, WLAN or other such limited-area network module whose antenna may be inbuilt into the module, which in
The illustrated slave audio devices 520, 530 each also includes a controller/DP 520A/530A, a computer-readable memory (MEM) 520B/530B tangibly storing a program computer-readable and executable instructions (PROG) 520C/530C (a software application), and a suitable radio frequency (RF) transmitter 520D/530D and receiver 520E/530E for bidirectional wireless communications over the respective wireless links 515A/515B via one or more antennas 520F/530F. The slave audio devices 520, 530 may also have a Bluetooth, WLAN or other such limited-area network module and one or more microphones 520H/530H and possibly also a camera 520J/530J, all powered by a portable power source such as a battery.
At least one of the PROGs in at least the master device 510 but possibly also in one or more of the slave devices 520, 530 is assumed to include program instructions that, when executed by the associated DP, enable the device to operate in accordance with the exemplary embodiments of this invention, as detailed above. That is, the exemplary embodiments of this invention may be implemented at least in part by computer software executable by the DP of the master and/or slave devices 510, 520, 530; or by hardware, or by a combination of software and hardware (and firmware).
In general, the various embodiments of the audio devices 510, 520, 530 can include, but are not limited to: cellular telephones; personal digital assistants (PDAs) having wireless communication and at least audio recording and/or play back capabilities; portable computers (including laptops and tablets) having wireless communication and at least audio recording and/or play back capabilities; image capture and sound capture/play back devices such as digital video cameras having wireless communication capabilities and a speaker and/or microphone; music capture, storage and playback appliances having wireless communication capabilities; Internet appliances having at least audio recording and/or play back capability; audio adapters, headsets, and other portable units or terminals that incorporate combinations of such functions.
The computer readable MEM in the audio devices 510, 520, 530 may be of any type suitable to the local technical environment and may be implemented using any suitable data storage technology, such as semiconductor based memory devices, flash memory, magnetic memory devices and systems, optical memory devices and systems, fixed memory and removable memory. The DPs may be of any type suitable to the local technical environment, and may include one or more of general purpose computers, special purpose computers, microprocessors, digital signal processors (DSPs) and processors based on a multicore processor architecture, as non-limiting examples.
In general, the various exemplary embodiments may be implemented in hardware or special purpose circuits, software, logic or any combination thereof. For example, some aspects may be implemented in hardware, while other aspects may be implemented in embodied firmware or software which may be executed by a controller, microprocessor or other computing device, although the invention is not limited thereto. While various aspects of the exemplary embodiments of this invention may be illustrated and described as block diagrams, flow charts, or using some other pictorial representation, it is well understood that these blocks, apparatus, systems, techniques or methods described herein may be implemented in, as non-limiting examples, hardware, embodied software and/or firmware, special purpose circuits or logic, general purpose hardware or controller or other computing devices, or some combination thereof, where general purpose elements may be made special purpose by embodied executable software.
It should thus be appreciated that at least some aspects of the exemplary embodiments of the inventions may be practiced in various components such as integrated circuit chips and modules, and that the exemplary embodiments of this invention may be realized in an apparatus that is embodied as an integrated circuit. The integrated circuit, or circuits, may comprise circuitry (as well as possibly firmware) for embodying at least one or more of a data processor or data processors, a digital signal processor or processors, and circuitry described herein by example.
It should be understood that components of the invention can be operationally coupled or connected and that any number or combination of intervening elements can exist (including no intervening elements). The connections can be direct or indirect and additionally there can merely be a functional relationship between components.
As used in this application, the term ‘circuitry’ refers to all of the following: (a) hardware-only circuit implementations (such as implementations in only analog and/or digital circuitry) and (b) to combinations of circuits and software (and/or firmware), such as (as applicable): (i) to a combination of processor(s) or (ii) to portions of processor(s)/software (including digital signal processor(s)), software, and memory(ies) that work together to cause an apparatus, such as a mobile phone or server, to perform various functions) and (c) to circuits, such as a microprocessor(s) or a portion of a microprocessor(s), that require software or firmware for operation, even if the software or firmware is not physically present.
This definition of ‘circuitry’ applies to all uses of this term in this application, including in any claims. As a further example, as used in this application, the term “circuitry” would also cover an implementation of merely a processor (or multiple processors) or portion of a processor and its (or their) accompanying software and/or firmware. The term “circuitry” would also cover, for example and if applicable to the particular claim element, a baseband integrated circuit or applications processor integrated circuit for a mobile phone or a similar integrated circuit in server, a cellular network device, or other network device.
Below are provided further descriptions of various non-limiting, exemplary embodiments. The below-described exemplary embodiments may be practiced in conjunction with one or more other aspects or exemplary embodiments. That is, the exemplary embodiments of the invention, such as those described immediately below, may be implemented, practiced or utilized in any combination (e.g., any combination that is suitable, practicable and/or feasible) and are not limited only to those combinations described herein and/or included in the appended claims.
In one exemplary embodiment, a method comprising detecting close proximity interactions; determining locations of a plurality of devices based on the detected close proximity interactions; and assigning a channel of a multi-channel audio file based on the determined locations of the plurality of devices.
A method as above, wherein the close proximity interactions comprise short distance communication methods.
A method as above, wherein the close proximity interactions comprise near field communications.
A method as above, wherein the close proximity interactions comprise tapping or contacting devices.
A method as above, further comprising reproducing optimized audio channels based on the determined locations of the plurality of devices.
A method as above, wherein further comprising determining a listening location relative to the determined locations of the plurality of devices.
A method as above, further comprising determining distances between the listening location and the plurality of devices.
A method as above, wherein the assigning of the channel further comprises assigning a center sound channel, a right front sound channel, a left front sound channel, a right surround sound channel, or a left surround sound channel.
A method as above, wherein each one of the plurality of devices comprises an audio device having at least one speaker.
In another exemplary embodiment, a method comprising detecting near field communication between at least two devices; determining a location of at least one of the at least two devices based on the detected near field communication; and assigning an audio channel of a multi-channel audio file based on the determined location of the at least one of the at least two devices.
A method as above, further comprising determining, based on the detected near field communication, a location of at least one of: one of the at least two devices; one or more other devices using one of the at least two devices; and the other of the at least two devices.
A method as above, further comprising assigning audio channels of the multi-channel audio file based on the determined location of at least one of the at least two devices and/or the one or more other devices.
A method as above, wherein the near field communication comprises tapping or contacting devices.
A method as above, further comprising reproducing optimized audio channels based on the determined location of the at least one of the at least two devices.
A method as above, further comprising determining a listening location relative to the determined location of the at least one of the at least two devices.
A method as above, further comprising determining a distance between the listening location and the at least one of the at least two devices.
A method as above, wherein the assigning of the channels further comprises assigning a center sound channel, a right front sound channel, a left front sound channel, a right surround sound channel, or a left surround sound channel.
A method as above, wherein each one of the at least two devices comprises an audio device having at least one speaker.
A method as above, wherein the determined location is stored and transmitted.
A method as above, wherein each one of the at least two devices comprises an audio system having at least one speaker.
In another exemplary embodiment, an apparatus comprising: at least one processor; and at least one memory including computer program code, the at least one memory and the computer program code configured to, with the at least one processor, cause the apparatus to perform at least the following: detect near field communication; determine a location of a device based on the detected near field communication; and assign a channel of a multi-channel audio file based on the determined location of the device.
An apparatus as above, wherein the at least one memory and the computer program code are further configured to, with the at least one processor, cause the apparatus to determine a listening location relative to the determined location of the device.
An apparatus as above, wherein the apparatus comprises a mobile device, and wherein the listening location is determined by the mobile device.
An apparatus as above, wherein the at least one memory and the computer program code are further configured to, with the at least one processor, cause the apparatus to determine a distance between the listening location and the device.
An apparatus as above, wherein the device comprises at least two devices, and wherein the at least one memory and the computer program code are further configured to, with the at least one processor, cause the apparatus to the determine, based on the detected near field communication, a location of at least one of: one of the at least two devices; one or more other devices using one of the at least tow devices; and the other of the at least two devices.
An apparatus as above, wherein the at least one memory and the computer program code are further configured to, with the at least one processor, cause the apparatus to assign channels of the multi-channel audio file based on the determined location of at least one of the at least two devices and/or the one or more other devices.
An apparatus as above, wherein the near field communication comprises tapping or contacting devices.
An apparatus as above, wherein each one of the devices comprises an audio system having at least one speaker.
An apparatus as above, wherein the near field communication comprises one of: a near field communication (NFC) method, or one or more sensors.
In another exemplary embodiment, a computer program product comprising a non-transitory computer-readable medium bearing computer program code embodied therein for use with a computer, the computer program code comprising: code for detecting near field communication; code for determining a location of a device based on the detected near field communication; and code for assigning a channel of a multi-channel audio file based on the determined location of the device.
A computer program product as above, further comprising code for determining locations of at least two devices based on the detected near field communication.
A computer program product as above, wherein the near field communication comprises tapping or contacting devices.
If desired, the different functions discussed herein may be performed in a different order and/or concurrently with each other. Furthermore, if desired, one or more of the above-described functions may be optional or may be combined.
Although various aspects of the invention are set out in the independent claims, other aspects of the invention comprise other combinations of features from the described embodiments and/or the dependent claims with the features of the independent claims, and not solely the combinations explicitly set out in the claims.
It should be understood that the foregoing description is only illustrative of the invention. Various alternatives and modifications can be devised by those skilled in the art without departing from the invention. Accordingly, the invention is intended to embrace all such alternatives, modifications and variances which fall within the scope of the appended claims.
Number | Name | Date | Kind |
---|---|---|---|
7630501 | Blank et al. | Dec 2009 | B2 |
8199941 | Hudson et al. | Jun 2012 | B2 |
8279709 | Choisel et al. | Oct 2012 | B2 |
8316154 | Yoneda | Nov 2012 | B2 |
9277321 | Toivanen | Mar 2016 | B2 |
20020122003 | Patwari et al. | Sep 2002 | A1 |
20030119523 | Bulthuis | Jun 2003 | A1 |
20040034655 | Tecu | Feb 2004 | A1 |
20040071294 | Halgas, Jr. et al. | Apr 2004 | A1 |
20040209654 | Cheung et al. | Oct 2004 | A1 |
20050190928 | Noto | Sep 2005 | A1 |
20070116306 | Riedel et al. | May 2007 | A1 |
20080019576 | Senftner | Jan 2008 | A1 |
20080177822 | Yoneda | Jul 2008 | A1 |
20080207123 | Andersen | Aug 2008 | A1 |
20080240460 | Sugii | Oct 2008 | A1 |
20090310790 | Sinton et al. | Dec 2009 | A1 |
20100105325 | Halla | Apr 2010 | A1 |
20100119072 | Ojanpera | May 2010 | A1 |
20100135118 | Van Leest et al. | Jun 2010 | A1 |
20100260348 | Bhow | Oct 2010 | A1 |
20110191823 | Huibers | Aug 2011 | A1 |
20120128160 | Kim et al. | May 2012 | A1 |
20130044894 | Samsudin et al. | Feb 2013 | A1 |
20130130714 | Huibers | May 2013 | A1 |
20130226324 | Hannuksela et al. | Aug 2013 | A1 |
20130315405 | Kanishima | Nov 2013 | A1 |
20140050454 | Slotte | Feb 2014 | A1 |
20140133672 | Lakkundi | May 2014 | A1 |
20140169569 | Toivanen | Jun 2014 | A1 |
20140362995 | Backman et al. | Dec 2014 | A1 |
Number | Date | Country |
---|---|---|
1894439 | Aug 2010 | EP |
WO-2006131894 | Dec 2006 | WO |
WO-2011144795 | Dec 2011 | WO |
Entry |
---|
Koskela, et al.; Evolution Towards Smart Home Environments Empirical Evaluation of Three User Interfaces ; Jul. 3-4, 2004; Institute of Software Systems, Tampere University of Technology, P.O. Box 553, 3101, Tampere, Finland; http://dl.acm.org/citation.cfm?id=1012660 (2 pages). |
Raykar, et al.; IEEE Transactions on Speech and Audio Processing; Position Calibration of Microphones and Loudspeakers in Distributed Computing Platforms; Aug. 27, 2003; http://www.umiacs.umd.edu/labs/labs/cvl/pirl/vikas/publications/IEEETSAP—v12—onecolumn.pdf (29 pages). |
Toivanen, et al.; U.S. Appl. No. 13/716,660, filed Dec. 17, 2012; Device Discovery and Constellation Selection. |
“In Stereo”, http://instereo.com.au/, (2011), (2 pages). |
Summit Semiconductor, LLC, “SpeakerFinder Technology” Automated Home Theater Speaker Configuration and Setup, (Feb. 9, 2011), (14 pages). |
http://store.sony.com/webapp/wcs/stores/servlet/StoreCatalogDisplay?langId=-1&storeId-10151&catal . . . ,(2012), (4 pages). |
Number | Date | Country | |
---|---|---|---|
20140362995 A1 | Dec 2014 | US |