This application is based upon and claims the benefit of priority of Japanese Patent Application No. 2019-071116 filed on Apr. 3, 2019, the contents of which are incorporated herein by reference in its entirety.
An embodiment of this invention relates to a sound signal processor that performs various processing on a sound signal.
JP-A-2014-103456 discloses an audio system that localizes a sound source in a position specified by the user through a mobile terminal such as a smartphone. The mobile terminal detects information about a posture of the own terminal and transmits it to the audio system together with position information about a placement position of the sound source desired by the user. The audio system localizes the sound source based on the received information, and generates audio signals to be supplied to respective speakers. According to such technical contents, the placement position of the sound source can be moved in real time by changing the posture of the mobile terminal.
However, it is complicated and difficult for the user to manually control the placement position of the sound source.
An object of this invention is to provide a sound signal processor capable of automatically controlling processing on a sound signal of a sound source.
A sound signal processor according to an aspect of the present invention includes a memory storing instructions and a processor configured to implement the stored instructions to execute a plurality of tasks, the tasks including a sound signal input task configured to obtain a sound signal, a beat detection task configured to detect a beat in the sound signal, and a processing task configured to perform an effect processing on the sound signal in accordance with a timing of the detected beat.
According to the above-described aspect, the sound signal processor is capable of automatically performing control since effect processing is performed on the sound signal in accordance with the timing of the beat contained in the sound signal.
In the present embodiment, as an example, the listening environment is a rectangular parallelepiped room R. The speakers SP1 to SP8 are placed in the room R. The speaker SP1 and the speaker SP2 are front speakers which are placed in both corners on one side of the floor of the room R. The speaker SP3 and the speaker SP4 are rear speakers which are placed in both corners on the other side of the floor of the room R. The speaker SP5 is a center speaker which is placed between the speaker SP1 and the speaker SP2. The speaker SP6 and the speaker SP7 are ceiling speakers which are placed on the ceiling of the room R. The speaker SP8 is a subwoofer which is placed near the speaker SP5. The speakers SP1 to SP8 are each connected to the sound signal processor 1.
The CPU 17 reads an operation program (firmware) stored in the flash memory 18 to the RAM 19, and integrally controls the sound signal processor 1.
The sound signal input portion 11 is, for example, an HDMI (trademark) interface, or a communication interface such as a network interface. In the present embodiment, the sound signal input portion 11 receives sound signals corresponding to a plurality of sound sources, and outputs them to the signal processing portion 13. Further, the sound signal input portion 11 outputs the sound signals to the CPU 17. Here, sound source information contained in the sound signals, for example, position information of respective sound sources and information such as the level information are also outputted to the CPU 17.
The signal processing portion 13 is configured by, for example, a DSP. In the present embodiment, the signal processing portion 13 performs signal processing such as delay, reverb or equalizer on the sound signal corresponding to each of sound sources according to the setting and an instruction of the CPU 17. After the signal processing, the sound signal corresponding to each of the sound sources is inputted to the localization processing portion 14.
The localization processing portion 14 is configured by, for example, a DSP. In the present embodiment, the localization processing portion 14 performs localization processing to localize a sound image according to an instruction of the CPU 17. The localization processing portion 14 distributes sound signals corresponding to each sound source to the speakers SP1 to SP8 with predetermined gains so that the sound images are localized in positions corresponding to the position information of respective sound sources specified by the CPU 17. The localization processing portion 14 inputs the sound signals corresponding to the speakers SP1 to SP8 to the D/A converter 15.
The D/A converter 15 converts the sound signals corresponding to the speakers SP1 to SP8 into analog signals. The amplifier 16 amplifies the analog sound signals corresponding to the speakers SP1 to SP8, and inputs them to the speakers SP1 to SP8.
In the above-described embodiment, the sound signal input portion 11 obtains sound signals corresponding to a plurality of sound sources, and outputs them directly to the signal processing portion 13. However, in another embodiment, a decoder (not shown) may be further provided between the sound signal input portion 11 and the signal processing portion 13. The decoder is configured by, for example, a DSP. In such a structure, when the sound signal input portion 11 obtains contents data, the decoder decodes the contents data, and extracts a sound signal from the contents data. When the contents data is data conforming to the object base method, the decoder further extracts sound source information from the contents data. According to the object base method, a plurality of sound sources (objects) contained in contents are stored as independent sound signals. The decoder inputs the sound signals corresponding to the sound sources to the signal processing portion 13 and the CPU 17. The sound source information contains information such as the position information and the levels of the sound sources. The decoder inputs the position information and the level information of the sound sources to the CPU 17.
The localization processing portion 14 performs effect processing related to a two-or-more-dimensional space on the sound signals, that is, processing to change the positions of the sound sources on a two-dimensional plane or in a three-dimensional space according to an instruction of the CPU 17. Moreover, the signal processing portion 13 performs signal processing such as delay, reverb or equalizer according to an instruction of the CPU 17. Accordingly, a DSP including the signal processing portion 13 and the localization processing portion 14, and the CPU 17 may be treated as one processing portion. The signal processing portion 13, the localization processing portion 14 and the decoder may be implemented in one DSP by means of software, or may be implemented by individual DSPs by means of hardware. In this embodiment, the signal processing portion 13 and the localization processing portion 14 perform effect processing (sound source position change and signal processing) for each of the sound sources, on the sound signals corresponding to a plurality of sound sources.
In this embodiment, the beat detection portion 171 obtains sound signals from the sound signal input portion 11 (S11). After obtaining the sound signals corresponding to a plurality of sound sources, the beat detection portion 171 detects beats from the sound signals (S12). The beat detection portion 171 may perform beat detection on the sound signal corresponding to a specific sound source or may perform beat detection on all the sound signals. The beat detection portion 171, for example, calculates the amplitude average value of the sound signal per unit time, and compares the calculated amplitude average value with the amplitude values of the sound signals. A beat is detected when the amplitude value of a sound signal is higher than the amplitude average value by not less than a certain degree (for example, not less than +6 dB). However, a threshold value of beat detection is not limited to +6 dB. Moreover, the beat detection method is not limited to the above-described method.
When beat detection is finished, the beat detection portion 171 notifies the signal processing portion 13 of the result of the beat detection (S13). To be specific, the beat detection portion 171 notifies the signal processing portion 13 of the positions of the detected beats, that is, the timing where the beats are detected within the sound signals. Then, in accordance with timings of the detected beats, the signal processing portion 13 performs signal processing, for example, processing to adjust the depth of the reverb and the delay on the sound signals. That is, the signal processing portion 13 changes the depth of the reverb and the length of the delay for each timing of beat detection. In this embodiment, the signal processing portion 13 performs signal processing for each sound source, on the sound signals corresponding to a plurality of sound sources.
As an example of the signal processing, the signal processing portion 13 adjusts the volume of the sound signal in accordance with the timing of the detected beat. For example, the signal processing portion 13 increases the gain of the sound signal at the timing where the beat is detected, and decreases the gain of the sound signal at a timing other than the timing of the beat. That is, the signal processing portion 13 increases the level (volume) of a part of the sound signal where the beat is detected, and decreases the level of a part of the sound signal other than the part of the sound signal where the beat is detected.
As another example of the signal processing, the signal processing portion 13 replaces a sound signal of the sound source with a sound signal of another sound source which is different in kind from the sound source in accordance with the timing of the detected beat. To implement this processing, the CPU 17 further includes a sound signal generation portion (not shown). The sound signal generation portion previously generates the sound signal of the another sound source and sends it to the signal processing portion 13. Then, the signal processing portion 13 replaces an existing sound signal with the previously prepared sound signal of the another sound source in accordance with the result of the beat detection. In such processing, the sound signal processor 1 can create a new piece of music.
According to the above-described processing, when the speakers SP1 to SP8 output sounds based on the sound signals, various expressions with musicality can be performed.
The position control portion 173 changes the position information of the sound sources in accordance with timings of the detected beats based on the result of the beat detection (S14′). As an example of the change of the position information of the sound sources, the position control portion 173 randomly moves the position of each of the sound sources. However, the change of the position information of the sound sources is not limited to the random one. As a second example, the position control portion 173 virtually rotates the position of each of the sound sources about a predetermined axis. As a third example, the position control portion 173 virtually moves the position of each of the sound sources upward or downward every beat detection. After changing the position information of the sound sources, the position control portion 173 outputs the changed position information to the localization processing portion 14 (S15′).
The localization processing portion 14 performs localization processing to localize a sound image, based on the changed position information. That is, the localization processing portion 14 distributes the sound signal of each of the sound sources to the speakers SP1 to SP8 with a predetermined gain so that the sound image is localized in the position corresponding to the changed position information of each of the sound sources from the CPU 17 in accordance with the timings of the detected beats.
According to the above-described change of the sound image localization position of the sound source, when the speakers SP1 to SP8 output sounds based on the sound signals, various new expressions with musicality can be performed.
In the flowchart shown in
In the above-described embodiment, the beat detection portion 171 performs a beat detection on a sound signal of a predetermined range (for example, one piece of music). However, in another embodiment, the beat detection portion 171 may detect beats in real time on sequentially inputted sound signals. The sound signal processor 1 can detect beats from sound signals and instantly perform effect processing in accordance with timings of the detected beats.
In the embodiment of the present invention, the sound signal processor 1 may output the result of the beat detection in real time or collectively to an external control device 30 or operation device through the interface 20 as illustrated in
The signal processing portion 13 does not perform signal processing in accordance with the timings of the detected beats on the low-frequency components of the sound signals. The signal processing portion 13 outputs the low-frequency components of the sound signals to the localization processing portion 14 without conducting a beat-based signal processing. The localization processing portion 14 distributes the low-frequency components of the sound signals corresponding to the respective sound sources only to the speaker SP8. That is, the low-frequency components of the sound signals are outputted to the subwoofer. In such a structure, the low-frequency components of the sound signals are outputted from a stable position through the subwoofer.
In the above-described embodiment, the sound signals and contents data obtained by the sound signal input portion 11 conform to the object base method. However, the sound signals and contents data that the present invention can handle are not limited thereto. In another embodiment, the sound signals and contents data obtained by the sound signal input portion 11 may conform to a channel base method.
The signal processing portion 13 calculates, for example, the level of the sound signal of each of channels and the cross-correlation between the channels. The signal processing portion 13 estimates the position of the sound source based on the level of the sound signal of each of the channels and the cross-correlation between the channels. For example, in a case where the correlation value between the L channel and the SL channel is high and the level of the L channel and the level of the SL channel are high (exceed a predetermined threshold value), the signal processing portion 13 estimates that a sound source is present between the L channel and the SL channel. The signal processing portion 13 estimates the position of the sound source based on the level of the L channel and the level of the SL channel. For example, when the ratio between the level of the L channel and the level of the SL channel is 1:1, the signal processing portion 13 estimates that the position of the sound source is just at the middle point between the L channel and the SL channel. The larger the number of channels is, the more accurately the position of the sound source can be estimated. By calculating the correlation value between a multiplicity of channels, the signal processing portion 13 can substantially uniquely identify the position of the sound source.
In a case where the inputted sound signals and contents data conform to the channel base method, the beat detection portion 171 having obtained sound signals from the sound signal input portion 11 and the decoder detects beats on at least one of the sound signals of a plurality of channels. The beat detection portion 171 outputs the result of the beat detection to the signal processing portion 13 in real time or collectively. The signal processing portion 13 performs signal processing such as delay, reverb or equalizer on the sound signals in accordance with the timings of the detected beats.
Further, the position control portion 173 changes the position information of the sound sources in accordance with the timings of the detected beats based on the result of the beat detection. The position control portion 173 outputs the changed position information to the localization processing portion 14. Then, the localization processing portion 14 performs localization processing to localize a sound image, based on the changed position information.
In the above-described embodiment, the sound signal processor 1 continuously performs effect processing on the sound signals conforming to the channel base method. However, the present invention is not limited thereto. Signal processing such as delay, reverb or equalizer or sound source position change may be separately performed on the sound signals conforming to the channel base method.
The descriptions of the present embodiment are illustrative in all respects and not restrictive. The scope of the present invention is shown not by the above-described embodiments but by the scope of the claims. Further, it is intended that all changes within the meaning and the scope equivalent to the scope of the claims are embraced by the scope of the present invention.
Number | Date | Country | Kind |
---|---|---|---|
JP2019-071116 | Apr 2019 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
5614687 | Yamada | Mar 1997 | A |
20030174845 | Hagiwara | Sep 2003 | A1 |
20140033902 | Sasai | Feb 2014 | A1 |
20150264502 | Aoki | Sep 2015 | A1 |
20160125867 | Jarvinen | May 2016 | A1 |
20170263230 | Takehisa | Sep 2017 | A1 |
20200043453 | Lang | Feb 2020 | A1 |
Number | Date | Country |
---|---|---|
1347668 | Sep 2003 | EP |
2014103456 | Jun 2014 | JP |
Entry |
---|
Extended European Search Report issued in European Appln. No. 20167831.5 dated Jul. 10, 2020. |
Office Action issued in Chinese Appln. No. 202010185419.4 dated Mar. 5, 2021. English machine translation provided. |
Number | Date | Country | |
---|---|---|---|
20200322746 A1 | Oct 2020 | US |