Various embodiments of the present application relate to the wireless exchange of data between devices in live events. More specifically aspects of the present disclosure relate to improving the auditory experience and enhance the user-engagement of the audience before, during and after live events.
Live events include among others performances such as music, theater, dance, opera, etc. as well as other types of events such as sports, political gatherings, festivals, religious ceremonies, TV shows, games etc. The global financial impact of such events is massive and event organizers are interested in maximizing their financial revenues by creating a great user-experience for the event audience. The term audience here refers to not only those who are physically present in live events but also everyone who experiences live events via any medium, for example via broadcasting, recording, virtual reality reproduction, etc. Live events can be experienced either in real time or anytime after the actual time of the event. In all said cases, a very important aspect of the overall live events' user-experience is the auditory experience of the audience. Therefore, there's a need for new methods and systems that improve the auditory experience of live events.
In an indoor or outdoor live event, no matter how small or large, the main Public Address (PA) system is typically setup and tuned in an empty venue, e.g. without an audience present. Typically dedicated engineers take care to ensure homogeneous coverage of all audience positions in terms of sound pressure, loudness, frequency response or any other parameter. Such setup and tuning ensures high-quality auditory experience for the audience. However, this setup and tuning of the PA system is time-consuming and requires expensive equipment and highly-skilled professionals. Therefore in many live events, careful setup and tuning of the PA system is not performed and as a result the auditory performance can be bad or mediocre. Furthermore, even in cases where a careful setup and tuning of the PA system is performed, there's no way to achieve a perfect result since: (a) the behavior of the PA system will change over time according to environmental conditions (temperature, humidity, etc.) and (b) the appearance of the audience alters significantly the acoustic characteristics, mainly in indoor venues. In addition, the success of the setup and tuning of a PA system is limited by another fact: it's extremely difficult to perform measurements in all audience positions especially in larger venues. Therefore, only indicative measurements or coarse simulations are typically performed, resulting in a sub-optimal result for several venue positions. Therefore there's a need for methods and systems that perform continuous measurements in several venue positions at the time of live events.
Although live events are sometimes equipped with adequate professional equipment for reinforcing, recording and broadcasting, there are often limitations on the equipment quantity and quality, especially when the production budget is low. In addition even for expensive productions, there can be always limitations on the equipment placement. For example, a live sound engineer of a concert cannot place microphones in between the concert crowd. On the other hand, modern audience members carry with them portable devices including but not limited to smartphones, tablets, video cameras and portable recorders. These devices typically have sensors such as microphones and cameras, as well as significant processing power and they can transmit data wirelessly. Therefore, there is a need to harness the computing power and/or exploit the sensors of such devices in order to enhance among others the quality and quantity of the live event reinforcement, recording and broadcasting. Another factor that enhances the user-experience of live events is the user-engagement at the time of the event or later on. During each live event, the event audience can be engaged by actively participating in it. By giving said option to the live event audience, the event organizers can create immersive experiences for the users, increase the user-satisfaction and as a result transform the event audience from one-time users to loyal fans. Since live event audience already carries with them their portable devices, it might also make sense to allow them to use said devices in order to interact with or participate in the event. Therefore there is a need for new methods and systems that give the event audience the option to participate actively in live events by using their portable devices.
Aspects of the invention relate to a method for wireless data exchange between devices in live events.
Aspects of the invention relate to a method for exploring data from multiple devices in order to get information on the acoustic paths of venues.
Aspects of the invention relate to a method for exploring data captured from microphones of devices of live event's audience.
For a more complete understanding of the invention, reference is made to the following description and accompanying drawings, in which:
Hereinafter, embodiments of the present invention will be described in detail in accordance with the references to the accompanying drawings. It is understood that other embodiments may be utilized and structural changes may be made without departing from the scope of the present application.
The exemplary systems and methods of this invention will sometimes be described in relation to audio systems. However, to avoid unnecessarily obscuring the present invention, the following description omits well-known structures and devices that may be shown in block diagram form or otherwise summarized.
For purposes of explanation, numerous details are set forth in order to provide a thorough understanding of the present invention. It should be appreciated however that the present invention may be practiced in a variety of ways beyond the specific details set forth herein. The terms determine, calculate and compute, and variations thereof, as used herein are used interchangeably and include any type of methodology, process, mathematical operation or technique.
where F{} is the forward Fourier transform, F−1{} is the inverse Fourier transform and h(t) is the IR of the acoustic path. Alternatively the IR can be extracted in the time domain or with any suitable technique. There are several methods to measure the IR of an acoustic path using various types of excitation signals x(t) as for example described in [Stan, Guy-Bart and Embrechts, Jean-Jacques and Archambeau, Dominique, “Comparison of Different Impulse Response Measurement Techniques”, J. Audio Eng. Soc., 50(4), p. 249-262]. All or any of these can be applied to this invention. The acquired signal and/or the IR can be used in order to perform meaningful acoustic analysis such as fractional-octave analysis, sound-level measurements, power spectra, frequency response measurements, transient analysis, etc. The analysis of the captured signal can be used in order to tune or calibrate any stage or aspect of the sound system, mainly by changing settings and/or adding/removing components to the system. The tuning of the system can be done either manually or automatically. In some embodiments, the tuning of the system might increase the sound quality either subjectively (for the audience during and after the event) or objectively (for recordings, broadcasting, etc).
In a particular embodiment, acoustic measurements are performed in a venue. In a typical venue, there can be none, one or more stages where the event primarily takes place 401, none, one or more loudspeakers that reinforce the sound 402, 403, 404 and none, one or more consoles/mixers 408 with multiple inputs 409, 410, 411 and outputs 412, 413, 414. The inputs may be wireless 409, 410 or wired 411. In order to improve the acoustic characteristics, the acoustic path from the sound system to each position of interest can be improved. In some embodiments, one or more microphones are placed inside the venue 405, 406, 407 and connected wirelessly 405, 406 or via a wired connection 407 to the inputs of the console mixer. Alternatively the microphones can be connected to the inputs of any device that can perform acoustic measurements or acquire sound. In a particular example, a sound signal can be routed from the console to the loudspeakers and captured via the microphones. The captured signal can be processed in order to extract meaningful information for the acoustic path. Ideally, every location of interest must be measured and an infinite number of microphones must be placed inside the venue. Since this is practically impossible, in prior art a limited number of measurements are usually performed using one or more microphones and typically the acoustic measurements take place before the event. However, the acoustic conditions during the event change significantly, due to different environmental conditions and the presence of the audience. Therefore, the practical value of such acoustic measurements is limited.
The data from the sound-capturing devices can be used to extract the impulse or frequency response of the acoustic path in each position. In addition the captured data can be used to extract parameters like: spectrum magnitude and phase, coherence, correlation, delay, spectrogram or any other time, frequency or time-frequency representation, stereo power balance, signal envelopes and transients, sound power or sound pressure level, loudness, peak or RMS level values, Reverberation Time (RT), Early Decay Time (EDT), Clarity (C), Definition or Deutlichkeit (D), Center of Gravity (TS), Interaural Cross Correlation (IACC), Lateral Fraction (LF/LFC), Direct to Reverberation Ratio, Speech Transmission Index (STI), Room Acoustics Speech Transmission Index (RASTI), Speech Transmission Index for Public Address Systems (STIPA), Articulation Loss of Consonants (%ALCons), Signal to Noise Ratio (SNR), Segmental Signal to Noise Ratio, Weighted Spectral Slope (WSS), Perceptual Evaluation of Speech Quality (PESQ), Perceptual Evaluation of Audio Quality (PEAQ), Log-Likelihood Ratio (LLR), Itakura-Saito Distance, Cepstrum Distance, Signal to Distortion Index, Signal to Interference Index or any other quantity that gives information on the acoustic paths or the emitted signals. Any such quantity can be presented/transmitted for the full audible frequency range or for any subset of the audible frequencies. Such information can be used in order to calibrate or tune the sound system and/or alter the captured signals. In another embodiment, sound-capturing devices carried by the audience might use the captured data and their own processing power in order to calculate any quantity that gives information on the acoustic paths or the signals. The calculated quantities can be transmitted with or without the captured sound signals to the console/mixer 508, any storage device 511 or any other appropriate device 514. The transmitted quantities can be used in order to manually or automatically change settings at any stage of the sound system or change the sound system topology. In some embodiments, the captured data can be sometimes transmitted together with location data and used in order to produce acoustic maps, i.e. graphic representations of the distributions of a certain quantities in a given region of the venue. In another embodiment, location data of each sound-capturing device can be determined by any appropriate technique at the console/mixer. These acoustic maps can be used to calibrate the sound system even during the live event, in a way that an improved auditory experience will be ensured for all audience positions.
In the present embodiment, detailed acoustic maps can be available to a sound engineer in real-time during the event so that she/he can continuously improve the auditory experience of the audience. In the present embodiment, instead of creating the acoustic maps via simulations or sparse measurements, accurate acoustic maps are created via real-time data acquisition using the sensors of the sound-capturing devices of the audience. Note that since the acoustic maps of the present embodiment can be dynamically updated in real-time any change of the acoustic conditions can be taken into account. For example, sometimes due to equipment malfunctions, sound engineers may replace the sound gear (e.g. microphones, guitar or bass cabins and amplifiers, etc) at the time of the live event. In the present embodiment the sound system can be automatically or manually re-tuned to compensate for any change in the acoustic conditions. In other embodiments, loudspeakers (typically monitor speakers for the musicians) and microphones located on stage of the live event can be used in order to produce acoustic maps with meaningful acoustic data for the stage area. Since sound engineering techniques rely heavily on the manipulation of the spectral content, sometimes there might not be a need for data transmission of the whole audible frequency range. In another embodiment, sound data limited in frequency bands can be provided from the sound-capturing devices in a way that a potential problem might be identified in a specific spectral region. By limiting the frequency band of interest, the amount of transmitted data can be efficiently reduced. Generally, any subset of the captured signal can be transmitted from the sound capturing devices. In all cases, when a sound engineer has access to detailed acoustic maps she/he can use typical engineering tools and techniques to tune the sound system including but not limited to hardware or software equalizers, dynamic range compressors, change of the microphone and/or source positions, etc.
In some embodiments, special signals including but not limited to sine sweeps, MLS noise, etc can be reproduced from the loudspeakers and captured from the sound-capturing devices in order to better estimate the acoustic paths. In other embodiments, said special signals can be presented alone or “hidden” in the music of the main event. For example, if such signals are not audible to the audience (because, e.g. they are masked by other sounds) they do not have a negative effect on the auditory experience while providing valuable information to better estimate acoustic paths. In other embodiments, the frequency content of these signals can be in the non-audible range.
In another embodiment, the use of a bidirectional communication channel (audio and video) between the sound-capturing devices and the mixing/console can enable the sound engineer to route audio or video to the device's speakers in order to create effects during the concert. For example, a sound effect where the main PA system is muted and thousands of speakers of the crowd's devices are activated can be created. In another embodiment, real-time video from and to the main stage can be transmitted using the crowd's devices. Such user-experience enhancements can be combined with other applications including but not limited to in-concert competitions, crowd balloting for the next songs, multimedia contests, sales of tickets for future concerts, in-app sales of music, etc.
In another embodiment, data from sound-capturing devices can be explored in order to compliment the main microphones when mixing or processing the live concert and resulting for example in multichannel audio reproduction, new sound effects, specific directivity patterns, better speech intelligibility and sound clarity, spatial allocation of sounds or sound sources, etc. A signal decomposition step might be also used in order to produce more meaningful input signals as proposed in the U.S. patent application Ser. No. 14/265,560.
In another embodiment, sound-capturing devices of audience that participate in the event through broadcasting can exchange data wirelessly with the event console/mixer. Therefore, sound or video data from remote audience members can be available to the sound engineer.
In some embodiments, the network of the sound-capturing devices can be an ad-hoc network. In other embodiments, the network of the sound-capturing devices can be a centralized network. A server acting as a router or access point may manage the network. The server can be located in the mixing/console of the live event or in any other appropriate location.
In particular embodiments, the sound-capturing devices may transmit data wirelessly. For this, any wireless data transmission may be used included but not limited to Bluetooth, communication protocols described in IEEE 802.11 (including any IEEE 802.11 revisions), communication protocols described in IEEE 802.1 (including any IEEE 802.1 revisions), cellular technology (such as GSM, CDMA, UMTS, EV-DO, WiMAX, or LTE), Zigbee, or any communications technologies used for the Internet of Things (IoT).
In particular embodiments, time data can be transmitted from the sound-capturing devices. Said time-data can be autonomous or linked with sound data, sometimes resulting in time-stamped sound data. In another embodiment, location data determining the exact or relative location of each sound-capturing device can be transmitted. In some embodiments, the location of a sound-capturing device relative to a second sound-capturing device or a plurality of sound-capturing devices can be determined or the location can be pre-determined beforehand. In another embodiment, the receiving device (for example the mixer/console) or any other device can determine the location of each sound-capturing device. This can be done via any standard location-tracking technique including but not limited to triangulation, trilateration, multilateration, WiFi beaconing, magnetic beaconing, etc. In another embodiment, the data can be transmitted continuously, periodically, as requested by the receiver, or in response to any another trigger. In another embodiment, data from other sensors can be transmitted from the sound-capturing devices including but not limited to video cameras, still cameras, Global Positioning System (GPS) receivers, infra red sensors, optical sensors, biosensors, Radio Frequency Identification (RFID) systems, wireless sensors, pressure sensors, temperature sensors, magnetometers, accelerometers, gyroscopes, and/or compasses.
While the above-described flowcharts have been discussed in relation to a particular sequence of events, it should be appreciated that changes to this sequence can occur without materially effecting the operation of the invention. Additionally, the exemplary techniques illustrated herein are not limited to the specifically illustrated embodiments but can also be utilized and combined with the other exemplary embodiments and each described feature is individually and separately claimable.
Additionally, the systems, methods and protocols of this invention can be implemented on a special purpose computer, a programmed micro-processor or microcontroller and peripheral integrated circuit element(s), an ASIC or other integrated circuit, a digital signal processor, a hard-wired electronic or logic circuit such as discrete element circuit, a programmable logic device such as PLD, PLA, FPGA, PAL, a modem, a transmitter/receiver, any comparable means, or the like. In general, any device capable of implementing a state machine that is in turn capable of implementing the methodology illustrated herein can be used to implement the various communication methods, protocols and techniques according to this invention.
Furthermore, the disclosed methods may be readily implemented in software using object or object-oriented software development environments that provide portable source code that can be used on a variety of computer or workstation platforms. Alternatively the disclosed methods may be readily implemented in software on an embedded processor, a micro-processor or a digital signal processor. The implementation may utilize either fixed-point or floating point operations or both. In the case of fixed point operations, approximations may be used for certain mathematical operations such as logarithms, exponentials, etc. Alternatively, the disclosed system may be implemented partially or fully in hardware using standard logic circuits or VLSI design. Whether software or hardware is used to implement the systems in accordance with this invention is dependent on the speed and/or efficiency requirements of the system, the particular function, and the particular software or hardware systems or microprocessor or microcomputer systems being utilized. The systems and methods illustrated herein can be readily implemented in hardware and/or software using any known or later developed systems or structures, devices and/or software by those of ordinary skill in the applicable art from the functional description provided herein and with a general basic knowledge of the audio processing arts.
Moreover, the disclosed methods may be readily implemented in software that can be stored on a storage medium, executed on programmed general-purpose computer with the cooperation of a controller and memory, a special purpose computer, a microprocessor, or the like. In these instances, the systems and methods of this invention can be implemented as program embedded on personal computer such as an applet, JAVA.RTM. or CGI script, as a resource residing on a server or computer workstation, as a routine embedded in a dedicated system or system component, or the like. The system can also be implemented by physically incorporating the system and/or method into a software and/or hardware system, such as the hardware and software systems of an electronic device.
It is therefore apparent that there has been provided, in accordance with the present invention, systems and methods for wireless exchange of data between devices in live events. While this invention has been described in conjunction with a number of embodiments, it is evident that many alternatives, modifications and variations would be or are apparent to those of ordinary skill in the applicable arts. Accordingly, it is intended to embrace all such alternatives, modifications, equivalents and variations that are within the spirit and scope of this invention.
This application claims the benefit of and priority under 35 U.S.C. §119(e) to U.S. Patent Application No. 61/952,636 filed Mar. 13, 2014, entitled “Ad-Hoc Wireless Exchange of Data Between Devices with Microphones and Speakers”. In addition, this application is related to U.S. patent application Ser. No. 14/265,560, filed Apr. 30, 2014, entitled “Methods and Systems for Processing and Mixing Signals Using Signal Decomposition,” each of which are incorporated herein by reference in their entirety.
Number | Date | Country | |
---|---|---|---|
61952636 | Mar 2014 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 14645713 | Mar 2015 | US |
Child | 15218884 | US |