Priority is claimed on Japanese Patent Application No. 2011-075457, filed on Mar. 30, 2011, the contents of which are incorporated herein by reference in its entirety.
Field of the Invention
The present invention relates to a signal-processing device, an imaging apparatus, and a signal-processing program.
Description of Related Art
In the related art, in order to remove noise mixed in a voice signal, a method is known in which a time domain signal is converted into a frequency domain signal frame by frame, a noise is estimated using a non-voice component signal, and the noise is reduced by subtracting the estimated noise from the frequency domain signal (refer to Japanese Unexamined Patent Application No. 2005-195955).
However, the method disclosed in Japanese Unexamined Patent Application No. 2005-195955 is to reduce the noise simply by subtracting the estimated noise from the frequency domain signal and therefore has a problem in that the noise cannot always be adequately reduced.
According to an aspect of the present invention, it is desirable to provide a signal-processing device, an imaging apparatus, and a signal-processing program which can adequately reduce noise.
According to an aspect of the present invention, there is provided a signal-processing device including: a determination section that compares a frequency spectrum and a floor spectrum of an input audio signal to each other for each frequency bin and determines whether the input audio signal should be subjected to noise reduction processing or not for each of the frequency bins; and a noise reduction-processing section that subtracts a noise frequency spectrum from the frequency spectrum of the input audio signal for each of the frequency bins on the basis of the result determined by the determination section for each of the frequency bins.
In addition, according to another aspect of the present invention, there is provided an imaging apparatus including the signal-processing device according to the above-described aspect.
In addition, according to still another aspect of the present invention, there is provided a signal-processing program causing a computer as a signal-processing device to execute: a determination process of comparing a frequency spectrum and a floor spectrum of an input audio signal to each other for each frequency bin and determining whether the input audio signal should be subjected to noise reduction processing or not for each of the frequency bins; and a noise reduction process of subtracting a noise frequency spectrum from the frequency spectrum of the input audio signal for each of the frequency bins on the basis of the result determined in the determination process for each of the frequency bins.
According to the aspects of the present invention, an advantage of adequately reducing noise can be exhibited.
Hereinafter, an embodiment of the present invention will be described with reference to the drawings.
As illustrated in
The imaging section 110 includes an optical system 400, an imaging element 119, and an A/D conversion section 120, is controlled by the CPU 190 in accordance with set imaging conditions (for example, an aperture value and an exposure value) to form an optical image on the imaging element 119 using the optical system 400, and generates image data based on the optical image which is converted into a digital signal by the A/D conversion section 120.
The optical system 400 includes a zoom lens 114, a lens for reducing vibration (hereinafter, referred to as a VR (Vibration Reduction) lens) 113, a lens for adjusting a focal point (hereinafter, referred to as an AF (Auto Focus) lens) 112, a zoom encoder 115, a lens-driving section 116, an AF encoder 117, and a vibration reduction section 118.
The optical system 400 guides the optical image which has been passed through the zoom lens 114, the VR lens 113, and the AF lens 112 to a light-receiving surface of the imaging element 119.
The lens-driving section 116 controls the position of the AF lens 112 or the zoom lens 114 on the basis of a drive control signal input from the CPU 190, which will be described below.
The vibration reduction section 118 controls the position of the VR lens 113 on the basis of the drive control signal input from the CPU 190, which will be described below. The vibration reduction section 118 may detect the position of the VR lens 113.
The zoom encoder 115 detects a zoom position indicating the position of the zoom lens 114 and outputs the detected zoom position to the CPU 190.
The AF encoder 117 detects a focus position indicating the position of the AF lens 112 and outputs the detected zoom position and focus position to the CPU 190.
In addition, the above-described optical system 400 may be integrally attached to the imaging apparatus 100 or may be detachably attached to the imaging apparatus 100.
The imaging element 119 converts, for example, the optical image formed on the light-receiving surface into an electric signal to output to the A/D conversion section 120.
In addition, the imaging element 119 stores image data, which is obtained when a photography instruction is received through the manipulation section 180, in a storage medium 200 through the A/D conversion section 120 or the image-processing section 140 as photography image data of a photographed still image.
On the other hand, for example, in a case where a photography instruction is not received through the manipulation section 180, the imaging element 119 outputs image data, which is continuously obtained, to the CPU 190 and the display section 150 as a through image data via the A/D conversion section 120 or the image-processing section 140.
The A/D conversion section 120 A/D-converts the electric signal which is converted by the imaging element 119 and outputs image data as the converted digital signal.
The manipulation section 180 includes, for example, a power supply switch, a shutter button, and other manipulation keys, receives a manipulation input by a user manipulating the manipulation section, and outputs the manipulation input to the CPU 190.
The image-processing section 140 performs image processing for the image data stored in the buffer memory 130 or the storage medium 200 with reference to image processing conditions stored in the storage section 160.
The display section 150 is a liquid crystal display, for example, and displays image data obtained by the imaging section 110, a manipulation screen, and the like.
The storage section 160 stores determination conditions which are referred to when a scene is determined by the CPU 190, imaging conditions, and the like. The storage section 160 includes a floor spectrum storage section 161, a noise storage section 162, and a mode information storage section 163. The floor spectrum storage section 161 stores a floor spectrum, which will be described below. The noise storage section 162 stores noise, which will be described below.
The mode information storage section 163 stores mode information which is information regarding which mode is selected between a quality-emphasized mode (first mode) which emphasizes the quality of an audio signal input by the manipulation of the user through the manipulation section 180 and a noise reduction-emphasized mode (second mode) which emphasizes reducing noise from the input audio signal.
The quality-emphasized mode described herein represents a mode in which a target sound such as a voice is output as is almost without any change, although the noise thereof is reduced, for example. In addition, the noise reduction-emphasized mode described herein represents a mode in which the noise is reduced as much as possible.
The microphone 230 collects a sound and outputs an audio signal corresponding to the collected sound. The audio signal is an analog signal.
The A/D conversion section 240 converts the audio signal, which is the analog signal input from the microphone 230, into a digital audio signal.
The signal-processing section 250 performs audio signal processing such as noise reduction on the audio signal which is converted into the digital signal by the A/D conversion section 240 and stores the audio signal subjected to the audio signal processing in the storage medium 200. In addition, the signal-processing section 250 performs the audio signal processing such as noise reduction in accordance with the mode information stored in the mode information storage section 163 of the storage section 160. The details of the signal-processing section 250 will be described below.
In addition, the audio signal subjected to audio signal processing by the signal-processing section 250 may be stored in the storage medium 200 to be time-associated with the image data imaged by the imaging element 119 or may be stored therein as a moving image containing the audio signal.
The buffer memory section 130 temporarily stores the image data imaged by the imaging section 110, the audio signal converted by the signal-processing section 250, and the like.
The communication section 170 is connected to the detachable storage medium 200 such as a card memory and stores, reads, or deletes information in or from the storage medium 200.
The storage medium 200 is a storage section detachably connected to the imaging apparatus 100, and stores, for example, the image data generated (photographed) by the imaging section 110 and the audio signal subjected to the audio signal processing by the signal-processing section 250.
The CPU 190 controls the entire imaging apparatus 100, for example, generates the drive control signal which controls the positions of the zoom lens 114 and the AF lens 112 on the basis of the zoom position input from the zoom encoder 115, the focus position input from the AF encoder 117, and the manipulation input which is input from the manipulation section 180. The CPU 190 controls the positions of the zoom lens 114 and the AF lens 112 through the lens-driving section 116 on the basis of the drive control signal.
In addition, the CPU 190 includes a timing detection section 191. The timing detection section 191 detects timing when an operation section included in the imaging apparatus 100 operates.
The operation section described herein represents, for example, the zoom lens 114, the VR lens 113, the AF lens 112, or the manipulation section 180 which is described above, and is a component which generates a sound (or having a possibility of generating a sound) by operating or being operated, in the imaging apparatus 100.
In addition, the operation section has a configuration in which the microphone 230 collects (or has a possibility of collecting) the sound which is generated by the component in the imaging apparatus 100 operating or being operated.
The timing detection section 191 may detect the timing when the operation section operates, on the basis of a control signal which operates the operation section. This control signal is a control signal which causes the operation section to operate the operate section or a control signal which operates the operation section.
For example, in order to drive the zoom lens 114, the VR lens 113, or the AF lens 112, the timing detection section 191 may detect the timing when the operation section operates, on the basis of the drive control signal which is input to the lens-driving section 116 or the vibration reduction section 118 or on the basis of the drive control signal generated by the CPU 190.
In addition, when the CPU 190 generates the drive control signal, the timing detection section 191 may detect the timing when the operation section operates, on the basis of processing or a command which is executed on the CPU 190.
In addition, the timing detection section 191 may detect the timing when the operation section operates, on the basis of a signal which is input from the manipulation section 180 and indicates that the zoom lens 114 or the AF lens 112 is to be driven.
In addition, the timing detection section 191 may detect the timing when the operation section operates, on the basis of a signal indicating that the operation section is operated.
For example, the timing detection section 191 may detect the timing when the operation section operates by detecting that the zoom lens 114 or the AF lens 112 is driven on the basis of the output from the zoom encoder 115 or the AF encoder 117.
In addition, the timing detection section 191 may detect the timing when the operation section operates by detecting that the VR lens 113 is driven on the basis of the output from the vibration reduction section 118.
In addition, the timing detection section 191 may detect the timing when the operation section operates by detecting that the manipulation section 180 is manipulated on the basis of the input from the manipulation section 180.
In addition, the timing detection section 191 detects the timing when the operation section included in the imaging apparatus 100 operates, and outputs the signal indicating the detected timing to the signal-processing section 250 (refer to
The bus 300 is connected to the imaging section 110, the CPU 190, the manipulation section 180, the image-processing section 140, the display section 150, the storage section 160, the buffer memory section 130, the communication section 170, and the signal-processing section 250, and transmits data output from the respective sections and the like.
<Specific Configuration of Signal-Processing Section 250>
Next, the details of the signal-processing section 250 illustrated in
Here, a case will be described in which the signal which is input from the timing detection section 191 and indicates the timing and the audio signal which is converted into the digital signal by the A/D conversion section 240 are input to the signal-processing section 250 illustrated in
In
In the example illustrated in
In addition, in the example illustrated in
In response to such an operation of the operation section, after the midway section of the frame No. 45, noise is superimposed on the waveform (d) of the audio signal input from the A/D conversion section 240. Here, when focusing on the relationship between each frame and a noise occurrence zone, it can be seen that noise is collected on frames subsequent to frame No. 45 (46, 47, 48, and . . . ) on the basis of the fact that the detected signal rises midway through the frame No. 45. In addition, before the frame No. 44 (43, 42, 41, and . . . ), noise is not collected at all. After the frame No. 46 (46, 47, 48, and . . . ), noise is collected in the entire frame zone.
In the present embodiment, the following configuration has been described: the signal-processing section 250 divides the audio signal, which is converted into the digital signal by the A/D conversion section 240, into frames, performs Fourier transform on the audio signal of each of the divided frames, and generates a frequency spectrum of the audio signal in each of the frames; the signal-processing section 250 performs noise reduction processing on the frequency spectrum of the audio signal for each of the frames, as will be described below with reference to
The floor spectrum estimation section 251 estimates a floor spectrum from the audio signal, which is converted into the digital signal by the A/D conversion section 240, on the basis of the timing when the operation section operates which is detected by the timing detection section 191. The floor spectrum represents a frequency spectrum of an audio signal in a frame immediately before the timing when the operation section operates or represents a frequency spectrum of an audio signal in a period where the operation section does not operate. In addition, the floor spectrum estimation section 251 stores the estimated floor spectrum in the floor spectrum storage section 161.
For example, the floor spectrum estimation section 251 estimates as the floor spectrum the frequency spectrum of the audio signal in the frame immediately before the timing when the operation section operates, on the basis of the timing when the operation section operates which is detected by the timing detection section 191. In
In the following description, the frequency spectrum (=S44) of the audio signal in the frame No. 44 will be referred to as the floor spectrum FS. In addition, in the following description, the intensity values of the respective frequency bins (the respective frequency domains) in the floor spectrum FS will be respectively referred to as F1, F2, F3, F4, and F5 in order from low frequency to high frequency (refer to (a) of
The noise estimation section 252 estimates noise from the audio signal which is converted into the digital signal by the A/D conversion section 240, on the basis of the timing when the operation section operates which is detected by the timing detection section 191. In addition, the noise estimation section 252 stores the estimated noise in the noise storage section 162.
For example, the noise estimation section 252 estimates as a noise frequency spectrum (noise spectrum) the difference between the frequency spectrum of the audio signal in the frame immediately after the timing when the operation section operates (and in the frame where the operation section operates across the entire frame) and the frequency spectrum of the audio signal in the frame immediately before the timing when the operation section operates (and in the frame where the operation section does not operate across the entire frame), on the basis of the timing when the operation section operates which is detected by the timing detection section 191.
In
In the following description, the frequency spectrum of the audio signal in the frame No. 46 will be referred to as the frequency spectrum S46 (refer to (b) of
The noise estimation section 252 estimates the frequency spectrum calculated by the subtraction as the noise frequency spectrum ((d) of
Hereinafter, the noise frequency spectrum estimated by the noise estimation section 252 will be referred to as a noise NS. In addition, the intensity values of the respective frequency bins in the noise NS will be respectively referred to as N1, N2, N3, N4, and N5 in order from low frequency to high frequency (refer to (d) of
The noise frequency spectrum thus obtained is subtracted from the frequency spectrum in the frame containing the noise (for example, frame No. 46, 47, 48, and . . . ). By converting the subtracted result into a time domain, the noise in the frame containing the noise is reduced (eliminated).
That is, the signal-processing section 250 performs spectral subtraction processing on the audio signal on the basis of the noise frequency spectrum, thereby reducing the noise of the audio signal. First, the spectral subtraction processing is a method of reducing the noise of the audio signal by converting the audio signal into the frequency domain by Fourier transform and the noise is reduced in the frequency domain, followed by inverse Fourier transform.
In addition, the signal-processing section 250 may perform Fast Fourier Transform (FFT) or Inverse Fast Fourier Transform (IFFT) as the Fourier transform or the inverse Fourier transform.
Referring to
<Quality-Emphasized Mode>
First, the respective configurations of the signal-processing section 250 in the quality-emphasized mode will be described with reference to
The determination section 253 compares the frequency spectrum and the floor spectrum to each other of the input audio signal for each of the spectrum bins and determines whether the input audio signal should be subjected to the noise reduction processing or not for each of the frequency bins. “The frequency spectrum of the input audio signal” described herein represents a frequency spectrum in which the audio signal converted into the digital signal by the A/D conversion section 240 is divided into the frames by the signal-processing section 250 and the audio signal in each of the frames is further Fourier-transformed into the frequency spectrum.
For example, the determination section 253 compares the frequency spectrum (frequency spectrum in the frame No. 46; refer to (b) of
Here, with respect to a frequency bin where the frequency spectrum of the input audio signal (frequency spectrum in the frame No. 46; refer to (b) of
On the other hand, with respect to a frequency bin where the frequency spectrum of the input audio signal (frequency spectrum in the frame No. 46; refer to (b) of
In the frequency bin Nos. 1 to 4 illustrated in (a) and (b) of
Therefore, the determination section 253 determines that the input audio signal in the frequency bin Nos. 1 to 4 should be subjected to the noise reduction processing (refer to four Symbols O indicated from the low frequency side (left side) in (d) of
<Noise Reduction-Processing Section 254>
In the quality-emphasized mode, the noise reduction-processing section 254 subtracts the noise frequency spectrum from the frequency spectrum of the input audio signal for each of the frequency bins, on the basis of the result determined by the determination section 253 for each of the frequency bins.
For example, in the quality-emphasized mode, with respect to a frequency bin where the determination section 253 determines that the input audio signal should be subjected to the noise reduction processing, the noise reduction-processing section 254 subtracts the noise frequency spectrum from the frequency spectrum of the input audio signal.
In addition, in the quality-emphasized mode, with respect to a frequency bin where the determination section 253 determines that the input audio signal should not be subjected to the noise reduction processing, the noise reduction-processing section 254 outputs the frequency spectrum of the input audio signal as is.
Based on the result determined by the determination section 253 (refer to (d) of
Accordingly, the noise reduction-processing section 254 calculates a frequency spectrum SA with the intensity values of A1 (=B1-N1), A2 (=B2−N2), A3 (=B3−N3), A4 (=B4−N4), and A5 (=B5) in order from the frequency bin Nos. 1 to 5 (refer to (c) of
In the quality-emphasized mode, the substitution section 255 selects a candidate frequency bin for substitution among the frequency bins of the frequency spectrum subtracted by the noise reduction-processing section 254, on the basis of the result determined by the determination section 253 for each of the frequency bins. Next, the substitution section 255 compares the frequency spectrum subtracted by the noise reduction-processing section 254 for each of the frequency bins and the floor spectrum to each other for each of the frequency bins in the selected frequency bin. Then, with respect to a frequency bin where the floor spectrum has an intensity value larger than that of the frequency spectrum subtracted by the noise reduction-processing section 254, the substitution section 255 substitutes the frequency spectrum subtracted by the noise reduction-processing section 254 with the floor spectrum.
For example, in the quality-emphasized mode, the substitution section 255 selects the frequency bin Nos. 1 to 4 as candidate frequency bins for substitution among the frequency bins of the frequency spectrum SA (refer to (c) of
Next, the substitution section 255 compares the frequency spectrum SA (refer to (c) of
Then, with respect to a frequency bin where the floor spectrum FS has an intensity value larger than that of the frequency spectrum SA subtracted by the noise reduction-processing section 254, the substitution section 255 substitutes the frequency spectrum SA subtracted by the noise reduction-processing section 254 with the floor spectrum FS. In this case, the substitution section 255 substitutes the frequency spectrum SA with the floor spectrum FS in the frequency bin Nos. 2 and 4. Accordingly, the substitution section 255 calculates a frequency spectrum SC with the intensity values of A1, F2, A3, F4, and B5 in order from the frequency bin Nos. 1 to 5 (refer to (f) of
Thereafter, the signal-processing section 250 performs inverse Fourier transform on the frequency spectrum SC illustrated in (f) of
As described above with reference to
<Noise Reduction-Emphasized Mode>
Next, the respective configurations of the signal-processing section 250 in the noise reduction-emphasized mode will be described with reference to
In the noise reduction-emphasized mode, the noise reduction-processing section 254 subtracts the noise frequency spectrum from the frequency spectrum of the input audio signal for each of the frequency bins.
For example, in the noise reduction-emphasized mode, the noise reduction-processing section 254 subtracts the noise frequency spectrum NS (refer to (b) of
The frequency spectrum SA illustrated in (c) of
In the example illustrated in (a) and (b) of
Therefore, in the frequency spectrum SA calculated by the noise reduction-processing section 254, the intensity values of A1, A2, A3, and A4 in the frequency bin Nos. 1 to 4 are positive (plus) values and the intensity value A5 in the frequency bin No. 5 is a negative (minus) value.
Here, in the noise reduction-emphasized mode, when the result of subtracting the noise frequency spectrum from the frequency spectrum of the input audio signal for each of the frequency bins is a negative value, the noise reduction-processing section 254 changes the result to 0.
For example, in the example illustrated in (c) of
Next, in the noise reduction-emphasized mode, the substitution section 255 compares the frequency spectrum SN (refer to (d) of
Then, with respect to a frequency bin where the floor spectrum FS (refer to (e) of
In (f) of
Therefore, the substitution section 255 substitutes the intensity values only in the frequency bin Nos. 1, 2, and 4 among the frequency bins of the frequency spectrum SA′ (refer to (d) of
Thereafter, similar to the case of the frequency spectrum SC illustrated in (f) of
As described above with reference to
As described above with reference to
In addition, in either case of the quality-emphasized mode or the noise reduction-emphasized mode, the substitution section 255 of the signal-processing section 250 according to the present embodiment substitutes the frequency spectrum subtracted by the noise reduction-processing section 254 for each of the frequency bins with the floor spectrum for each of the frequency bins, on the basis of the result of comparing the frequency spectrum subtracted by the noise reduction-processing section 254 for each of the frequency bins and the floor spectrum to each other for each of the frequency bins (refer to (e) and (f) of
In addition, when the noise is subtracted from the audio signal, there is a possibility of generating musical noise. On the other hand, as described above, the substitution section 255 of the signal-processing section 250 subtracts the noise from the audio signal and then performs so-called flooring processing on the basis of the result of comparing with the floor spectrum. Accordingly, the substitution section 255 of the signal-processing section 250 can reduce the possibility of generating musical noise.
In addition, the substitution section 255 of the signal-processing section 250 does not simply perform the flooring processing but performs the flooring processing according to the quality-emphasized mode and the noise reduction-emphasized mode (refer to (e) and (f) of
In addition, the noise reduction-processing section 254 does not simply subtract the noise frequency spectrum from the frequency spectrum of the input audio signal for each of the frequency bins but subtracts the noise frequency spectrum from the frequency spectrum of the input audio signal for each of the frequency bin on the basis of the result determined by the determination section 253 for each of the frequency bins. Accordingly, the noise reduction-processing section 254 can adequately reduce the noise from the input audio signal.
<Regarding Processes after Frame No. 47 in
In the above description with reference to
For example, in the case of the audio signal in the frame No. 47 and the quality-emphasized mode, the signal-processing section 250 changes the frequency spectrum S46 in the frame No. 46 illustrated in (b) of
In addition, for example, in the case of the audio signal in the frame No. 47 and the noise reduction-emphasized mode, the signal-processing section 250 changes the frequency spectrum S46 in the frame No. 46 illustrated in (a) of
In this way, similar to the case of the frame No. 46, the signal-processing section 250 can perform the noise reduction processing on the audio signals in the frame Nos. 47, 48, and . . . which are the audio signals after the frame No. 46 in either case of the quality-emphasized mode or the noise reduction-emphasized mode.
<Regarding Estimation of Floor Spectrum>
In the above description with reference to
For example, the floor spectrum estimation section 251 respectively converts the audio signals in plural frames before the timing when the operation section operates into the frequency spectra, on the basis of the timing when the operation section operates which is detected by the timing detection section 191. Furthermore, the floor spectrum estimation section 251 may estimate the average frequency spectrum, which is obtained by averaging the plural frequency spectra for each of the frequency bins, as the floor spectrum.
In addition, when the plural frequency spectra are averaged for each of the frequency bins, the floor spectrum estimation section 251 may weight the plural frequency spectra to calculate the average. The weighted value may be lowered as the frequency spectrum becomes distant from a frame (start frame) of an audio signal as a target of the flooring processing.
In addition, when the floor spectrum is estimated, the floor spectrum estimation section 251 desirably estimates the floor spectrum at least on the basis of the frames after the timing when the operation section has operated immediately before. This is because the frequency spectrum of the audio signal in the frame where the operation section does not operate is desirable as the floor spectrum. In addition, this is also because the frame of the audio signal generating the floor spectrum is less appropriate for the floor spectrum with respect to the audio signal as it becomes temporally distant from the audio signal as the target to be subjected to the flooring processing.
In addition, the floor spectrum storage section 161 may store the floor spectrum in advance. For example, the floor spectrum storage section 161 may store the floor spectrum in advance to be associated with environment information indicating the surrounding sound circumstances during photographing or photography mode information indicating a photography mode, according to the situation. The signal-processing section 250 may read out the floor spectrum which is associated with the environment information or photography mode information selected by a user from the floor spectrum storage section 161, and may perform the noise reduction processing described above with reference to
<Regarding Estimation of Noise>
In addition, in the above description with reference to
Instead of the floor spectrum FS which is the frequency spectrum of the audio signal in the frame No. 44, the noise estimation section 252 can estimate the floor spectrum FS by an arbitrary method in which the above-described floor spectrum estimation section 251 estimates the floor spectrum FS.
In addition, instead of the frequency spectrum S46 of the audio signal in the frame No. 46, the noise estimation section 252 may use the frequency spectrum which is obtained by averaging the frequency spectra of the audio signals in the plural frames for each of the frequency bins at the timing when the operation section operates on the basis of the timing when the operation section operates which is detected by the timing detection section 191. For example, instead of the frequency spectrum S46 of the audio signal in the frame No. 46, the noise estimation section 252 may use the frequency spectrum which is obtained by averaging the frequency spectra of the audio signals in the plural frames, such as the frame Nos. 46, 47, and 48, for each of the frequency bins.
In addition, when the plural frequency spectra are averaged for each of the frequency bins, the noise estimation section 252 may weight the frequency spectra to calculate the average. The weighted value may be lowered as the frequency spectrum becomes distant from a frame (start frame) of an audio signal as a target of the flooring processing. In addition, similar to the case of the floor spectrum, the noise frequency spectrum may be stored in the noise storage section 162 in advance.
<Regarding Overlap of Frames in
In addition, in the above description with reference to
In addition, the signal-processing section 250 may convert the audio signal of each of the frames into the frequency spectrum after multiplying the audio signal of each of the frames by a window function such as Hamming window.
In addition, in the above description with reference to
However, the present invention is not limited thereto. The signal-processing section 250 may control the position such that the audio signal is divided into the frames according to the signal (a) which is input from the timing detection section 191 and indicates the timing, that is, the signal which indicates the timing when the operation section operates. For example, the signal-processing section 250 may generate the frames with respect to the audio signal such that the frame boundaries of the audio signal matches the position (refer to Symbol O of
The signal-processing section 250 may perform the above-described noise reduction processing on the basis of the period before the operation section operates and the period in which the operation section operates, according to the signal indicating the timing when the operation section operates.
In the above description, a case where the signal-processing section 250 performs the signal processing on the audio signal collected by the microphone 230 is described. However, the above-described processing of the signal-processing section 250 according to the present embodiment is not applied only to the audio signal collected in this way in real time.
For example, the signal-processing section 250 according to the present embodiment can also perform the above-described processing on an audio signal recorded in advance, that is, perform the above-described processing even in a case where a storage section such as the storage medium 200 stores the timing when the operation section of a device which records this audio signal operates, to be associated with the audio signal.
In the above description, the noise superimposed on the audio signal is mainly the sound generated by driving the optical system 400. However, the noise is not limited thereto. For example, the same shall be applied to a sound generated by pressing a button or the like of the manipulation section 180. In this case, a signal generated by pressing the button or the like of the manipulation section 180 is input to the timing detection section 191 of the CPU 190. Accordingly, similar to the case of driving the optical system 400, the timing detection section 191 can detect the timing when the manipulation section 180 or the like operates.
In addition, in the above description, the imaging apparatus 100 includes the signal-processing section 250. However, the signal-processing section 250 may be included in a sound recorder, a mobile phone, or a communication terminal.
The signal-processing section 250 in
Instead, the signal-processing section 250 or the respective components of the signal-processing section 250 may include a memory and a CPU (Central Processing Unit) and realize the functions thereof by loading a program for realizing the functions on the memory and executing the program.
In addition, the signal-processing section 250 in
In addition, “the computer system” includes a homepage-providing environment (or a homepage display environment) when using the World Wide Web system.
In addition, “the computer-readable recording medium” refers to storage devices including flexible discs, magneto-optical discs, portable media such as ROM and CD-ROM, and hard discs built into the computer systems. Furthermore, “the computer-readable recording medium” includes: media dynamically holding the program in a short period of time, for example, a communication line of a case where the program is transmitted through a network such as the Internet or a communication line such as a telephone line; and media holding the program for a given time, for example, a volatile memory built into a computer system as a server or client in the above case where the program is transmitted through the communication line. In addition, the above-described program may partially realize the above-described functions. Furthermore, the above-described functions may be realized in combination with a program stored in advance in a computer system.
Hereinbefore, the embodiment of the present invention has been described with reference to the drawings. However, the specific configurations are not limited to the embodiment and include designs and the like within a range not departing from the scope of the present invention.
While preferred embodiments of the invention have been described and illustrated above, it should be understood that these are exemplary of the invention and are not to be considered as limiting. Additions, omissions, substitutions, and other modifications can be made without departing from the spirit or scope of the present invention. Accordingly, the invention is not to be considered as being limited by the foregoing description, and is only limited by the scope of the appended claims.
Number | Date | Country | Kind |
---|---|---|---|
2011-075457 | Mar 2011 | JP | national |
Number | Name | Date | Kind |
---|---|---|---|
6339758 | Kanazawa et al. | Jan 2002 | B1 |
7698133 | Ichikawa | Apr 2010 | B2 |
20060293882 | Giesbrecht | Dec 2006 | A1 |
20090177466 | Rui et al. | Jul 2009 | A1 |
20100296665 | Ishikawa et al. | Nov 2010 | A1 |
20120163622 | Karthik et al. | Jun 2012 | A1 |
Number | Date | Country |
---|---|---|
A-2000-047699 | Feb 2000 | JP |
2004-341339 | Dec 2004 | JP |
A-2004-341339 | Dec 2004 | JP |
A-2005-195955 | Jul 2005 | JP |
A-2005-203981 | Jul 2005 | JP |
A-2006-163231 | Jun 2006 | JP |
2006279185 | Oct 2006 | JP |
A-2010-271411 | Dec 2010 | JP |
A-2011-097335 | May 2011 | JP |
Entry |
---|
Japanese Office Action issued in Application No. 2011-075457; Dated Jan. 29, 2013 (With Translation). |
Number | Date | Country | |
---|---|---|---|
20120250885 A1 | Oct 2012 | US |