This disclosure relates to reducing octave errors during pitch determination for noisy audio signals, such as with voice enhancement of noisy audio signals.
One aspect of the disclosure relates to a system configured to perform voice enhancement on noisy audio signals, in accordance with one or more implementations. Because pitch determines harmonic spacing, any integer divider of pitch can explain a harmonic signal. Any multiple of the pitch can explain a large fraction of a signal. This may create an ambiguity in the pitch estimation producing “octave errors.” As such, the system may be configured to reduce octave errors during pitch determination for such noisy audio signals. Octave errors may be reduced during pitch determination for noisy audio signals. Pitch may be tracked over time by determining amplitudes at harmonics for individual time windows of an input signal. Octave errors may be reduced in individual time windows by fitting amplitudes of corresponding harmonics across successive time windows to identify spurious harmonics caused by octave error. A given harmonics in a given time window may be associated with a fitting function that fits amplitudes of harmonics corresponding to the given harmonic in time windows proximate to the given time window. The given harmonic may be identified as either being associated with the same pitch as adjacent harmonics in the given time window or being spurious based on parameters of the fitting function.
The communications platform may be configured to execute computer program modules. The computer program modules may include one or more of an input module, a pitch tracking module, an octave error reduction module, one or more extraction modules, a reconstruction module, an output module, and/or other modules.
The input module may be configured to receive an input signal from a source. The input signal may include human speech (or some other wanted signal) and noise. The waveforms associated with the speech and noise may be superimposed in input signal.
The pitch tracking module may be configured to track pitch over time. This may include determining amplitudes at harmonics for individual time windows of the input signal. Tracked pitch in the first time window may be associated with a number of harmonics including a first harmonic and a second harmonic. The first harmonic may have a first amplitude and the second harmonic may have a second amplitude. The first harmonic and the second harmonic may be adjacent but either associated with the same pitch or different pitches resulting from an octave error. An octave error in the pitch may determine whether harmonics correspond to the actual signal or are spurious.
Generally speaking, the extraction module(s) may be configured to extract harmonic information from the input signal. The extraction module(s) may include one or more of a transform module, a formant model module, and/or other modules.
The transform module may be configured to perform a transform on individual time windows of the input signal to obtain corresponding sound models of the input signal in the individual time windows. A given sound model may be a mathematical representation of harmonics in a given time window of the input signal.
The octave error reduction module may be configured to reduce octave errors in individual time windows. Reducing octave errors may include fitting amplitudes of corresponding harmonics across successive time windows to identify spurious harmonics caused by octave error. Harmonics in the first time window, including the first harmonic and the second harmonic, may be fitted using the corresponding sound model provided by the transform module. The fit may be performed at a plurality of times within the first time window. A determination may be made as to the probabilities of whether the first harmonic and/or the second harmonic are a part of the actual signal or are spurious. The determination may be made based on the quality of the fit of the sound model to the harmonics. The determination may be made based on the pattern and alternation of the harmonics. According to some implementations, pitch probabilities estimated across larger time periods may be computed by compounding the probabilities of the individual pitches in each individual time within the first time window. Continuity of pitch may be used as a prior assumption on the computation of the pitch probabilities.
The formant model module may be configured to model harmonic amplitudes based on a formant model. Generally speaking, a formant may be described as the spectral resonance peaks of the sound spectrum of the voice. One formant model—the source-filter model—postulates that vocalization in humans occurs via an initial periodic signal produced by the glottis (i.e., the source), which is then modulated by resonances in the vocal and nasal cavities (i.e., the filter).
The reconstruction module may be configured to reconstruct the speech component of the input signal with the noise component of the input signal being suppressed. The reconstruction may be performed once each of the parameters of the formant model has been determined. The reconstruction may be performed by interpolating all the time-dependent parameters and then resynthesizing the waveform of the speech component of the input signal.
The output module may be configured to transmit an output signal to a destination. The output signal may include the reconstructed speech component of the input signal.
These and other features, and characteristics of the present technology, as well as the methods of operation and functions of the related elements of structure and the combination of parts and economies of manufacture, will become more apparent upon consideration of the following description and the appended claims with reference to the accompanying drawings, all of which form a part of this specification, wherein like reference numerals designate corresponding parts in the various figures. It is to be expressly understood, however, that the drawings are for the purpose of illustration and description only and are not intended as a definition of the limits of the invention. As used in the specification and in the claims, the singular form of “a”, “an”, and “the” include plural referents unless the context clearly dictates otherwise.
Octave errors may be reduced during pitch determination for noisy audio signals. Pitch may be tracked over time by determining amplitudes at harmonics for individual time windows of an input signal. Octave errors may be reduced in individual time windows by fitting amplitudes of corresponding harmonics across successive time windows to identify spurious harmonics caused by octave error. A given harmonics in a given time window may be associated with a fitting function that fits amplitudes of harmonics corresponding to the given harmonic in time windows proximate to the given time window. The given harmonic may be identified as either being associated with the same pitch as adjacent harmonics in the given time window or being spurious based on parameters of the fitting function.
The communications platform 102 may be configured to execute computer program modules. The computer program modules may include one or more of an input module 104, a preprocessing module 106, one or more extraction modules 112, a reconstruction module 114, an output module 116, and/or other modules.
The input module 104 may be configured to receive an input signal 118 from a source 120. The input signal 118 may include human speech (or some other wanted signal) and noise. The waveforms associated with the speech and noise may be superimposed in input signal 118. The input signal 118 may include a single channel (i.e., mono), two channels (i.e., stereo), and/or multiple channels. The input signal 118 may be digitized.
Speech is the vocal form of human communication. Speech is based upon the syntactic combination of lexicals and names that are drawn from very large vocabularies (usually in the range of about 10,000 different words). Each spoken word is created out of the phonetic combination of a limited set of vowel and consonant speech sound units. Normal speech is produced with pulmonary pressure provided by the lungs which creates phonation in the glottis in the larynx that is then modified by the vocal tract into different vowels and consonants. Various differences among vocabularies, syntax that structures individual vocabularies, sets of speech sound units associated with individual vocabularies, and/or other differences create the existence of many thousands of different types of mutually unintelligible human languages.
The noise included in input signal 118 may include any sound information other than a primary speaker's voice. The noise included in input signal 118 may include structured noise and/or unstructured noise. A classic example of structured noise may be a background scene where there are multiple voices, such as a café or a car environment. Unstructured noise may be described as noise with a broad spectral density distribution. Examples of unstructured noise may include white noise, pink noise, and/or other unstructured noise. White noise is a random signal with a flat power spectral density. Pink noise is a signal with a power spectral density that is inversely proportional to the frequency.
An audio signal, such as input signal 118, may be visualized by way of a spectrogram. A spectrogram is a time-varying spectral representation that shows how the spectral density of a signal varies with time. Spectrograms may be referred to as spectral waterfalls, sonograms, voiceprints, and/or voicegrams. Spectrograms may be used to identify phonetic sounds.
Referring again to
The preprocessing module 106 may be configured to segment input signal 118 into discrete successive time windows. According to some implementations, a given time window may have a duration in the range of 30-60 milliseconds. In some implementations, a given time window may have a duration that is shorter than 30 milliseconds or longer than 60 milliseconds. The individual time windows of segmented input signal 118 may have equal durations. In some implementations, the duration of individual time windows of segmented input signal 118 may be different. For example, the duration of a given time window of segmented input signal 118 may be based on the amount and/or complexity of audio information contained in the given time window such that the duration increases responsive to a lack of audio information or a presence of stable audio information (e.g., a constant tone).
The pitch tracking module 108 may be configured to track pitch over time. This may include determining amplitudes at harmonics for individual time windows of the input signal. Tracked pitch in a given time window being associated with a first harmonic having a first amplitude, a second harmonic having a second amplitude, and/or other harmonics having corresponding amplitudes. By way of non-limiting illustration,
The octave error reduction module 110 may be configured to reduce octave errors in individual time windows. The octave error reduction module 110 is described further in conjunction with extraction module(s) 112.
Generally speaking, extraction module(s) 112 may be configured to extract harmonic information from input signal 118. The extraction module(s) 112 may include one or more of a transform module 112A, a formant model module 112B, and/or other modules.
The transform module 112A may be configured to obtain a sound model over individual time windows of input signal 118. In some implementations, transform module 112A may be configured to obtain a linear fit in time of a sound model over individual time windows of input signal 118. A sound model may be described as a mathematical representation of harmonics in an audio signal. A harmonic may be described as a component frequency of the audio signal that is an integer multiple of the fundamental frequency (i.e., the lowest frequency of a periodic waveform or pseudo-periodic waveform). That is, if the fundamental frequency is f, then harmonics have frequencies 2f, 3f, 4f, etc. The harmonics of a given sound model may include a first harmonic and/or a second harmonic depending on whether the first harmonic and/or the second harmonic are identified as either being associated with the same pitch or being spurious based on parameters of the first fitting function and the second fitting function, as discussed in connection with octave error reduction module 110.
The transform module 112A may be configured to model input signal 118 as a superposition of harmonics that all share a common pitch and chirp. Such a model may be expressed as:
where φ is the base pitch and χ is the fractional chirp rate
where c is the actual chirp), both assumed to be constant. Pitch is defined as the rate of change of phase over time. Chirp is defined as the rate of change of pitch (i.e., the second time derivative of phase). The model of input signal 118 may be assumed as a superposition of Nh harmonics with a linearly varying fundamental frequency. Ah is a complex coefficient weighting all the different harmonics. Being complex, Ah carries information about both the amplitude and about the initial phase for each harmonic.
The model of input signal 118 as a function of Ah may be linear, according to some implementations. In such implementations, linear regression may be used to fit the model, such as follows:
The best value for Ā may be solved via standard linear regression in discrete time, as follows:
Ā=M(φ,χ)\s, EQN. 3
where the symbol \ represents matrix left division (e.g., linear regression).
Due to input signal 118 being real, the fitted coefficients may be doubled with their complex conjugates as:
The optimal values of φ,χ may not be determinable via linear regression. A nonlinear optimization step may be performed to determine the optimal values of φ,χ. Such a nonlinear optimization may include using the residual sum of squares as the optimization metric:
where the minimization is performed on φ,χ at the value of Ā given by the linear regression for each value of the parameters being optimized.
The transform module 112A may be configured to impose continuity to different fits over time. That is, both continuity in the pitch estimation and continuity in the coefficients estimation may be imposed to extend the model set forth in EQN. 1. If the pitch becomes a continuous function of time (i.e., φ=φ(t)), then the chirp may be not needed because the fractional chirp may be determined by the derivative of φ(t) as
According to some implementations, the model set forth by EQN. 1 may be extended to accommodate a more general time dependent pitch as follows:
where Φ(t)=2π∫0tφ(τ)dτ is integral phase.
According to model set forth in EQN. 6, the harmonic amplitudes Ah(t) are time dependent. The harmonic amplitudes may be assumed to be piecewise linear in time such that linear regression may be invoked to obtain Ah(t) for a given integral phase Φ(t):
where
and ΔAhi, are time-dependent harmonic coefficients. The time-dependent harmonic coefficients ΔAhi, represent the variation on the complex amplitudes at times ti.
EQN. 7 may be substituted into EQN. 6 to obtain a linear function of the time-dependent harmonic coefficients ΔAhi. The time-dependent harmonic coefficients ΔAhi may be solved using standard linear regression for a given integral phase Φ(t). Actual amplitudes may be reconstructed by
The linear regression may be determined efficiently due to the fact that the correlation matrix of the model associated with EQN. 6 and EQN. 7 has a block Toeplitz structure, in accordance with some implementations.
A given integral phase Φ(t) may be optimized via nonlinear regression. Such a nonlinear regression may be performed using a metric similar to EQN. 5. In order to reduce the degrees of freedom, Φ(t) may be approximated with a number of time points across which to interpolate by Φ(t)=interp(Φ1=Φ(t1), Φ2=Φ(t2), . . . , ΦN
The different Φi may be optimized one at a time with multiple iterations across them. Because each Φi affects the integral phase only around ti, the optimization may be performed locally, according to some implementations.
The octave error reduction module 110 may be configured to reduce octave errors in individual time windows. According to some implementations, reducing octave errors in individual time windows may include fitting amplitudes of corresponding harmonics across successive time windows to identify spurious harmonics caused by octave error. Referring again to plot 300 in
Referring now to formant model module 112B in
where A(t) is a global amplitude scale common to all the harmonics, but time dependent. G characterizes the source as a function of glottal parameters g(t). Glottal parameters g(t) may be a vector of time dependent parameters. In some implementations, G may be the Fourier transform of the glottal pulse. F describes a resonance (e.g., a formant). The various cavities in a vocal tract may generate a number of resonances F that act in series. Individual formants may be characterized by a complex parameter fr(t). R represents a parameter-independent filter that accounts for the air impedance.
In some implementations, the individual formant resonances may be approximated as single pole transfer functions:
where f(t)=jp(t)+d(t) is a complex function, p(t) is the resonance peak p(t), and d(t) is a dumping coefficient. The fitting of one or more of these functions may be discretized in time in a number of parameters pi,di corresponding to fitting times ti.
According to some implementations, R may be assumed to be R(t)=1−jω(t), which corresponds to a high pass filter.
The Fourier transform of the glottal pulse G may remain fairly constant over time. In some implementations, G=g(t) g E(g(t))t. The frequency profile of G may be approximated in a nonparametric fashion by interpolating across the harmonics frequencies at different times.
Given the model for the harmonic amplitudes set forth in EQN. 9, the model parameters may be regressed using the sum of squares rule as:
The regression in EQN. 11 may be performed in a nonlinear fashion assuming that the various time dependent functions can be interpolated from a number of discrete points in time. Because the regression in EQN. 11 depends on the estimated pitch, and in turn the estimated pitch depends on the harmonic amplitudes (see, e.g., EQN. 8), it may be possible to iterate between EQN. 11 and EQN. 8 to refine the fit.
In some implementations, the fit of the model parameters may be performed on harmonic amplitudes only, disregarding the phases during the fit. This may make the parameter fitting less sensitive to the phase variation of the real signal and/or the model, and may stabilize the fit. According to one implementation, for example:
In accordance with some implementations, the formant estimation may occur according to:
EQN. 10 may be extended to include the pitch in one single minimization as:
The minimization may occur on a discretized version of the time-dependent parameter, assuming interpolation among the different time samples of each of them.
The final residual of the fit on the HAM(Ah(t)) for both EQN. 10 and EQN. 11 may be assumed to be the glottal pulse. The glottal pulse may be subject to smoothing (or assumed constant) by taking an average:
The reconstruction module 114 may be configured to reconstruct the speech component of input signal 118 with the noise component of input signal 118 being suppressed. The reconstruction may be performed once each of the parameters of the formant model has been determined. The reconstruction may be performed by interpolating all the time-dependent parameters and then resynthesizing the waveform of the speech component of input signal 118 according to:
The output module 116 may be configured to transmit an output signal 122 to a destination 124. The output signal 122 may include the reconstructed speech component of input signal 118, as determined by EQN. 13. The destination 124 may include a speaker (i.e., an electric-to-acoustic transducer), a remote device, and/or other destination for output signal 122. By way of non-limiting illustration, where communications platform 102 is a mobile communications device, a speaker integrated in the mobile communications device may provide output signal 122 by converting output signal 122 to sound to be heard by a user. As another illustration, output signal 122 may be provided from communications platform 102 to a remote device. The remote device may have its own speaker that converts output signal 122 to sound to be heard by a user of the remote device.
In some implementations, one or more components of system 100 may be operatively linked via one or more electronic communication links. For example, such electronic communication links may be established, at least in part, via a network such as the Internet, a telecommunications network, and/or other networks. It will be appreciated that this is not intended to be limiting, and that the scope of this disclosure includes implementations in which one or more components of system 100 may be operatively linked via some other communication media.
The communications platform 102 may include electronic storage 126, one or more processors 128, and/or other components. The communications platform 102 may include communication lines, or ports to enable the exchange of information with a network and/or other platforms. Illustration of communications platform 102 in
The electronic storage 126 may comprise electronic storage media that electronically stores information. The electronic storage media of electronic storage 126 may include one or both of system storage that is provided integrally (i.e., substantially non-removable) with communications platform 102 and/or removable storage that is removably connectable to communications platform 102 via, for example, a port (e.g., a USB port, a firewire port, etc.) or a drive (e.g., a disk drive, etc.). The electronic storage 126 may include one or more of optically readable storage media (e.g., optical disks, etc.), magnetically readable storage media (e.g., magnetic tape, magnetic hard drive, floppy drive, etc.), electrical charge-based storage media (e.g., EEPROM, RAM, etc.), solid-state storage media (e.g., flash drive, etc.), and/or other electronically readable storage media. The electronic storage 126 may include one or more virtual storage resources (e.g., cloud storage, a virtual private network, and/or other virtual storage resources). The electronic storage 126 may store software algorithms, information determined by processor(s) 128, information received from a remote device, information received from source 120, information to be transmitted to destination 124, and/or other information that enables communications platform 102 to function as described herein.
The processor(s) 128 may be configured to provide information processing capabilities in communications platform 102. As such, processor(s) 128 may include one or more of a digital processor, an analog processor, a digital circuit designed to process information, an analog circuit designed to process information, a state machine, and/or other mechanisms for electronically processing information. Although processor(s) 128 is shown in
It should be appreciated that although modules 104, 106, 108, 110, 112A, 112B, 114, and 116 are illustrated in
In some embodiments, method 400 may be implemented in one or more processing devices (e.g., a digital processor, an analog processor, a digital circuit designed to process information, an analog circuit designed to process information, a state machine, and/or other mechanisms for electronically processing information). The one or more processing devices may include one or more devices executing some or all of the operations of method 400 in response to instructions stored electronically on an electronic storage medium. The one or more processing devices may include one or more devices configured through hardware, firmware, and/or software to be specifically designed for execution of one or more of the operations of method 400.
At an operation 402, an input signal may be segmented into discrete successive time windows. The input signal may convey audio comprising a speech component superimposed on a noise component. The time windows may include a first time window. Operation 402 may be performed by one or more processors configured to execute a preprocessing module that is the same as or similar to preprocessing module 106, in accordance with one or more implementations.
At an operation 404, pitch may be tracked over time by determining amplitudes at harmonics for individual time windows of the input signal. Tracked pitch in the first time window may be associated with a first harmonic having a first amplitude and a second harmonic having a second amplitude. The first harmonic and the second harmonic may be adjacent but either associated with the same pitch or different pitches resulting from an octave error. Operation 404 may be performed by one or more processors configured to execute a pitch tracking module that is the same as or similar to pitch tracking module 108, in accordance with one or more implementations.
At an operation 406, octave errors may be reduced in individual time windows by fitting amplitudes of corresponding harmonics across successive time windows to identify spurious harmonics caused by octave error. Operation 406 may be performed by one or more processors configured to execute an octave error reduction module that is the same as or similar to octave error reduction module 110, in accordance with one or more implementations.
Although the present technology has been described in detail for the purpose of illustration based on what is currently considered to be the most practical and preferred implementations, it is to be understood that such detail is solely for that purpose and that the technology is not limited to the disclosed implementations, but, on the contrary, is intended to cover modifications and equivalent arrangements that are within the spirit and scope of the appended claims. For example, it is to be understood that the present technology contemplates that, to the extent possible, one or more features of any implementation can be combined with one or more features of any other implementation.
Number | Name | Date | Kind |
---|---|---|---|
5774837 | Yeldener et al. | Jun 1998 | A |
5815580 | Craven et al. | Sep 1998 | A |
5978824 | Ikeda | Nov 1999 | A |
6195632 | Pearson | Feb 2001 | B1 |
6594585 | Gersztenkorn | Jul 2003 | B1 |
7085721 | Kawahara | Aug 2006 | B1 |
7117149 | Zakarauskas | Oct 2006 | B1 |
7249015 | Jiang et al. | Jul 2007 | B2 |
7389230 | Nelken | Jun 2008 | B1 |
7664640 | Webber | Feb 2010 | B2 |
7668711 | Chong et al. | Feb 2010 | B2 |
8015002 | Li et al. | Sep 2011 | B2 |
8380331 | Smaragdis | Feb 2013 | B1 |
20030177002 | Chen | Sep 2003 | A1 |
20040066940 | Amir | Apr 2004 | A1 |
20040111266 | Coorman et al. | Jun 2004 | A1 |
20040128130 | Rose et al. | Jul 2004 | A1 |
20040158462 | Rutledge et al. | Aug 2004 | A1 |
20040167777 | Hetherington et al. | Aug 2004 | A1 |
20040176949 | Wenndt et al. | Sep 2004 | A1 |
20040220475 | Szabo et al. | Nov 2004 | A1 |
20050114128 | Hetherington et al. | May 2005 | A1 |
20050149321 | Kabi et al. | Jul 2005 | A1 |
20060053003 | Suzuki et al. | Mar 2006 | A1 |
20060100866 | Alewine et al. | May 2006 | A1 |
20060100868 | Hetherington et al. | May 2006 | A1 |
20060130637 | Crebouw | Jun 2006 | A1 |
20060136203 | Ichikawa | Jun 2006 | A1 |
20070010997 | Kim | Jan 2007 | A1 |
20080033585 | Zopf | Feb 2008 | A1 |
20080052068 | Aguilar et al. | Feb 2008 | A1 |
20080082323 | Bai et al. | Apr 2008 | A1 |
20080234959 | Joublin | Sep 2008 | A1 |
20080262836 | Goto | Oct 2008 | A1 |
20080312913 | Goto | Dec 2008 | A1 |
20090012638 | Lou | Jan 2009 | A1 |
20090016434 | Amonou et al. | Jan 2009 | A1 |
20090076822 | Sanjaume | Mar 2009 | A1 |
20100131086 | Itoyama | May 2010 | A1 |
20100174534 | Vos | Jul 2010 | A1 |
20100211384 | Qi et al. | Aug 2010 | A1 |
20100260353 | Ozawa | Oct 2010 | A1 |
20100299144 | Barzelay et al. | Nov 2010 | A1 |
20100332222 | Bai et al. | Dec 2010 | A1 |
20110016077 | Vasilache et al. | Jan 2011 | A1 |
20110060564 | Hoge | Mar 2011 | A1 |
20110286618 | Vandali et al. | Nov 2011 | A1 |
20120010881 | Avendano | Jan 2012 | A1 |
20120072209 | Krishnan | Mar 2012 | A1 |
20120191450 | Pinson | Jul 2012 | A1 |
20120243694 | Bradley et al. | Sep 2012 | A1 |
20120243705 | Bradley et al. | Sep 2012 | A1 |
20120243707 | Bradley et al. | Sep 2012 | A1 |
20130046533 | Nyquist et al. | Feb 2013 | A1 |
20130158923 | Stanton et al. | Jun 2013 | A1 |
20130165788 | Osumi et al. | Jun 2013 | A1 |
20130255473 | Abe et al. | Oct 2013 | A1 |
Number | Date | Country |
---|---|---|
WO 2012129255 | Sep 2012 | WO |
WO 2012134991 | Oct 2012 | WO |
WO 2012134993 | Oct 2012 | WO |
Entry |
---|
S. Saha and S. M. Kay, “Maximum likelihood parameter estimation of superimposed chirps using Monte Carlo importance sampling,” in IEEE Transactions on Signal Processing, vol. 50, No. 2, pp. 224-230, Feb. 2002. |
Y. Pantazis, O. Rosec and Y. Stylianou, “Chirp rate estimation of speech based on a time-varying quasi-harmonic model,” 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, Taipei, 2009, pp. 3985-3988. |
Kumar et al., “Speaker Recognition Using GMM”, International Journal of Engineering Science and Technology, vol. 2, No. 6, 2010, retrieved from the Internet: http://www.ijest.info/docs/IJEST10-02-06-112.pdf, pp. 2428-2436. |
Kamath et al, “Independent Component Analysis for Audio Classification”, IEEE 11th Digital Signal Processing Workshop & IEEE Signal Processing Education Workshop, 2004, retrieved from the Internet: http://2002.114.89.42/resource/pdf/1412.pdf, pp. 352-355. |
Vargas-Rubio et al., “An Improved Spectrogram Using the Multiangle Centered Discrete Fractional Fourier Transform”, Proceedings of International Conference on Acoustics, Speech, and Signal Processing, Philadelphia, 2005, retrieved from the internet: <URL: http://www.ece.unm.edu/faculty/beanthan/PUB/ICASSP-05-JUAN.pdf>, 4 pages. |
U.S. Appl. No. 13/961,811 Office Action dated Apr. 20, 2015 citing prior art, 9 pages. |
U.S. Appl. No. 13/961,811, Aug. 7, 2013, 30 pages. |
Saha, S.; Kay, S.M., “Maximum likelihood parameter estimation of superimposed chirps using Monte Carlo importance sampling,” in Signal Processing, IEEE Transactions on , vol. 50, No. 2, pp. 224-230, Feb. 2002. |
Vargas-Rubio, J.G.; Santhanam, B., An improved spectrogram using the multiangle centered discrete fractional Fourier transform,: in Acoustics, Speech, and Signal Processing, 2005. Proceedings. (ICASSP '05). IEEE International Conference on , vol. 4, No., pp. iv/505-iv/508 vol. 4, Mar. 18-23, 2005. |
Pantazis, Y.; Rosec, O.; Stylianou, Y., “Chirp rate estimation of speech based on a time-varying quasi-Harmonic-model,” in Acoustics, Speech and Signal Processing, 2009. ICASSP 2009. IEEE International Conference on , vol., No., pp. 3985-3988, Apr. 19-24, 2009. |
Luis Weruaga, Marian Kepesi, The fan-chirp transform for non-stationary harmonic signals, Signal Processing, vol. 87, Issue 6, Jun. 2007, pp. 1504-1522, ISSN 0165-1684, http://dx.doi.org/10.1016/j.sigpro.2007.01.006. (http://www.sciencedirect.com/science/article/pii/S0165168407000114). |