One embodiment is directed generally to a device, and more particularly, to a device that produces haptic effects.
Haptics is a tactile and force feedback technology that takes advantage of a user's sense of touch by applying haptic feedback effects (i.e., “haptic effects”), such as forces, vibrations, and motions, to the user. Devices, such as mobile devices, touchscreen devices, and personal computers, can be configured to generate haptic effects. In general, calls to embedded hardware capable of generating haptic effects (such as actuators) can be programmed within an operating system (“OS”) of the device. These calls specify which haptic effect to play. For example, when a user interacts with the device using, for example, a button, touchscreen, lever, joystick, wheel, or some other control, the OS of the device can send a play command through control circuitry to the embedded hardware. The embedded hardware then produces the appropriate haptic effect.
Such devices can also be configured to play audio data, such as a digital audio signal. For example, such devices can include applications configured to play video data, such as a movie or video game, that contains an audio portion, or audio data, such as a song. Similar to haptics, calls to additional embedded hardware capable of generating audio effects (such as speakers) can be programmed within the OS of the device. Thus, the OS of the device can send a play command through control circuitry to the additional embedded hardware, where the additional embedded hardware then produces the appropriate audio effect.
One embodiment is a system that converts an audio signal into one or more haptic effects. The system receives a data frame of the audio signal. The system further generates a maximum value of the data frame. The system further transforms the data frame. The system further generates at least one sine carrier waveform based on a maximum value of the transformed data frame. The system further mixes the transformed data frame with the at least one sine carrier waveform to generate a modulated data frame. The system further generates a haptic signal based on the maximum value, and the modulated data frame. The system further sends the haptic signal to an actuator to generate the one or more haptic effects.
Further embodiments, details, advantages, and modifications will become apparent from the following detailed description of the preferred embodiments, which is to be taken in conjunction with the accompanying drawings.
One embodiment is a haptic conversion system that can intercept one or more frames of audio data, such as a digital audio signal, transform the one or more frames, convert the one or more transformed frames into a haptic signal, and play the created haptic signal through an actuator to produce one or more haptic effects. The haptic signal can include a waveform, where a waveform is a set of one or more signal values in a pulse-coded modulation (“PCM”) format. The haptic signal can be applied to a type of actuator configured to receive a waveform, where the actuator can utilize the waveform to generate the one or more haptic effects.
A computer-readable medium may be any available medium that can be accessed by processor 22 and may include both a volatile and nonvolatile medium, a removable and non-removable medium, a communication medium, and a storage medium. A communication medium may include computer readable instructions, data structures, program modules or other data in a modulated data signal such as a carrier wave or other transport mechanism, and may include any other form of an information delivery medium known in the art. A storage medium may include RAM, flash memory, ROM, erasable programmable read-only memory (“EPROM”), electrically erasable programmable read-only memory (“EEPROM”), registers, hard disk, a removable disk, a compact disk read-only memory (“CD-ROM”), or any other form of a storage medium known in the art.
In one embodiment, memory 14 stores software modules that provide functionality when executed by processor 22. The modules include an operating system 15 that provides operating system functionality for system 10, as well as the rest of a mobile device in one embodiment. The modules further include a haptic conversion module 16 that converts an audio signal into a haptic signal that is used to produce one or more haptic effects, as disclosed in more detail below. In certain embodiments, haptic conversion module 16 can comprise a plurality of modules that each provide specific individual functionality for converting an audio signal into a haptic signal that is used to produce one or more haptic effects. System 10 will typically include one or more additional application modules 18 to include additional functionality, such as MOTIV® Haptic Development Platform by Immersion Corporation.
System 10, in embodiments that transmit and/or receive data from remote sources, further includes a communication device 20, such as a network interface card, to provide mobile wireless network communication, such as infrared, radio, Wi-Fi, cellular network, or other next-generation wireless-data network communication. In other embodiments, communication device 20 provides a wired network connection, such as an Ethernet connection or a modem.
Processor 22 is further coupled via bus 12 to a display 24, such as a Liquid Crystal Display (“LCD”), for displaying a graphical representation or user interface to a user. The display 24 may be a touch-sensitive input device, such as a touch screen, configured to send and receive signals from processor 22, and may be a multi-touch touch screen.
System 10 further includes one or more actuators 26. Processor 22 may transmit a haptic signal associated with a haptic effect to actuator 26, which in turn outputs haptic effects. Actuator 26 may be, for example, an electric motor, an electro-magnetic actuator, a voice coil, a shape memory alloy, an electro-active polymer, a solenoid, an eccentric rotating mass motor (“ERM”), a linear resonant actuator (“LRA”), a piezoelectric actuator, a high bandwidth actuator, an electroactive polymer (“EAP”) actuator, an electrostatic friction display, or an ultrasonic vibration generator.
In some embodiments, system 10 further includes one or more speakers 28. Processor 22 may transmit an audio signal to speaker 28, which in turn outputs audio effects. Speaker 28 may be, for example, a dynamic loudspeaker, an electrodynamic loudspeaker, a piezoelectric loudspeaker, a magnetostrictive loudspeaker, an electrostatic loudspeaker, a ribbon and planar magnetic loudspeaker, a bending wave loudspeaker, a flat panel loudspeaker, a heil air motion transducer, a plasma arc speaker, and a digital loudspeaker.
The haptic conversion system also includes a haptic conversion module 230, according to the embodiment. In certain embodiments, haptic conversion module 230 is identical to haptic conversion module 16 of
According to the embodiment, reverb module 231 is configured to intercept the one or more audio buffers that audio track module 210 streams to speaker 220. Reverb module 231 is further configured to send each audio data frame of each audio buffer to boombox module 232. Boombox module 232 is configured to analyze each audio data frame and calculate a maximum value for each audio data frame. The calculating of the maximum value for each audio data frame, performed by boombox module 232, is further described in greater detail in relation to
Boombox module 232 is further configured to send each audio data frame of each audio buffer to band-pass filter 233. Band-pass filter 233 is configured to band-pass filter each audio data frame of each audio buffer. By band-pass filtering each audio data frame, band-pass filter 233 can filter on one or more frequency bands of frequency signals of any given range in a given audio buffer (e.g., a single frequency band of frequency signals in the 0-500 Hz range). Thus, all signals within this frequency range can be extracted to create a new audio buffer of only these signals. The effect yielded from this filtering can be a “bass-boost” or “subwoofer” type of signal. Band-pass filter 233 is further configured to return each band-pass filtered audio data frame to boombox module 232. In alternate embodiments, band-pass filter 233 is not a filter, but a transformation module (such as a digital signal processor (“DSP”) process, a state machine, or other type of programmatic logic). In these alternate embodiments, the transformation module can transform the received audio data frame into a new audio data frame, where the transformation is not necessary a filtering transformation.
Boombox module 232 is further configured, for each band-pass filtered audio data frame, to generate a sine-wave periodic carrier frequency (also identified as a “sine carrier waveform”) based on the maximum value of the band-pass filtered audio data frame. The generation of the sine-wave periodic carrier frequency is further described in greater detail in relation to
The haptic conversion system also includes a haptic effect player module 240 and an actuator 250. Haptic effect player module 240 is a module that is embedded within a device (such as a mobile device), and that is configured to play one or more haptic effects at an actuator by sending one or more haptic signals to the actuator. In one embodiment, haptic effect player module 240 is a TouchSense Player® module by Immersion Corporation. Actuator 250 is an actuator configured to receive one or more haptic signals, and configured to output one or more haptic effects. In certain embodiments, actuator 250 is an actuator configured to receive a waveform, where the waveform is used to control a magnitude and precision of a waveform effect when played through the actuator. Actuator 250 can be, for example, a piezoelectric actuator, a high bandwidth actuator, or an EAP actuator.
According to the embodiment, reverb module 231 is configured to determine whether each maximum value of each audio data frame is greater than a specific threshold. When the maximum value of the audio data frame is greater than the specific threshold, reverb module 231 is configured to call an application program interface (“API”) of haptic effect player module 240, and send haptic effect player module 240 the corresponding modulated audio data frame as a parameter of the API. When the maximum value of the audio data frame is not greater than the specific threshold, reverb module 231 is configured to ignore the audio data frame, and not send the corresponding modulated audio data frame to haptic effect player module 240. In certain embodiments, the API of haptic effect player module 240 is a “ImmVibeAppendWaveformEffect” API of a TouchSense Player® module by Immersion Corporation. In other embodiments the API could be “ImmVibePlayMagSweepEffect” or “ImmVibePlayPeriodicEffect.” The API of haptic effect player module 240 sends the modulated audio data frame to actuator 250, where actuator 250 is configured to play an appropriate waveform based on the modulated audio data frame. According to the embodiment, the modulated audio data frame serves as a waveform that is played by actuator 250, where the waveform includes one or more frequency bands that are emphasized in addition to a periodic haptic signal, and where the modulated audio data frame can control a precision of the waveform played by actuator 250. By playing a waveform, actuator 250 produces a haptic effect.
According to the embodiment, the haptic conversion system includes an operating system 300, a reverb module 301, a boombox module 302, and an infinite impulse response (“IIR”) filter 303. Operating system 300 is an operating system of a device (such as a mobile device). In the illustrated embodiment, operating system 300 is an Android® operating system for a mobile device. Reverb module 301, boombox module 302, and IIR filter 303 are sub-modules of a haptic conversion module (such as haptic conversion module 16 of
According to the embodiment, at 305, an application configured to create an audio signal is created. The application is any application that can be executed within operating system 300. In the illustrated embodiment, the application is an Android® application. The flow proceeds to 310. At 310, a “SoundPool” object is created by the application, where a SoundPool object is an instance of a SoundPool API (which is an Android® API), and where a SoundPool object manages and plays audio resources for the application. The SoundPool object can include a collection of audio samples that can be loaded into a memory of a device from a resource stored within the device, or from a file stored within a file system of the device. The SoundPool object can decode the collection of audio samples into a raw 16-bit PCM mono or stereo buffer. The flow proceeds to 315. At 315, other audio objects are created by the application, where the other audio objects are instances of other audio APIs. The other audio objects can also decode audio samples into a 16-bit PCM audio buffer. The flow proceeds to 320. At 320, an AudioTrack object is created by either the SoundPool object or one of the other audio objects. An AudioTrack object is configured to receive one or more PCM audio buffers and stream the one or more PCM audio buffers.
The flow proceeds to 325. At 325, a reverb object is created. A reverb object is an instance of reverb module 301, where the reverb object is configured to process one or more audio data frames of a PCM audio buffer sent by the AudioTrack object, as previously described in relation to
The flow proceeds to 340. At 340, the reverb object sends an audio data frame of a PCM audio buffer, that is received from an AudioTrack object, to the boombox object. The flow proceeds to 345. At 345, the boombox object samples data contained within the audio data frame. Such sampling is further described in greater detail in relation to
The flow proceeds to 355. At 355, the reverb object performs a thresholding task. More specifically, the reverb object determines whether the maximum value of the audio data frame, received from the boombox object, is greater than a specified threshold. The maximum value returned is first linearly mapped to a force value, which in one embodiment can vary between 0 and 10,000. For example, the threshold may be specified as 1,000. In this example, if the value of the force is greater than 1,000, then the audio data frame is written to an audio buffer, to be subsequently sent to an actuator. If the value of the force is less than or equal to 1,000, then the audio data frame is “dropped” (i.e., is not written to the audio buffer), and is not subsequently sent to an actuator. Thus, according to the embodiment, not every audio data frame is necessarily played at the actuator. The flow proceeds to 360. At 360, the reverb object reads the audio data frame from the audio buffer, assuming that the audio data frame was not “dropped” at 355.
The flow proceeds to 365. At 365, the reverb object sends the audio data frame, that was stored within the audio data buffer, to the boombox object. The flow proceeds to 370. At 370, the boombox object resamples the audio data frame, for example, at 8 KHz mono audio. The boombox object subsequently sends the resampled audio data frame to the IIR filter 303. The flow proceeds to 375. At 375, IIR filter 303 low-pass filters the received audio data frame, for example, at 500 Hz. In the example, by low-pass filtering the audio data frame, IIR filter 303 filters on a single frequency band of low-frequency signals within the audio data frame (e.g., 0-500 Hz). IIR filter 303 subsequently returns the low-pass filtered audio data frame to the boombox object. In alternate embodiments, IIR filter 303 is not a filter, but a transformation module (such as a DSP process, a state machine, or other type of programmatic logic). In these alternate embodiments, the transformation module can transform the received audio data frame into a new audio data frame, where the transformation is not necessary a filtering transformation.
The flow proceeds to 380. At 380, the boombox object boosts the audio data signal contained within the filtered audio data frame. Boosting the audio data signal can involve multiplying all the PCM values of the audio data signal by a constant. The flow proceeds to 385. At 385, the boombox object creates a sine carrier waveform (also identified as a “sine-wave periodic carrier frequency” or “sine wave”) at a resonant frequency of an actuator, or any other frequency desired, and mixes the sine carrier waveform with the audio data signal contained within the filtered audio data frame. More specifically, the boombox object scans the filtered audio data frame for its maximum value (i.e., amplitude), identified as A. Then, the boombox object utilizes a sine function (such as a C++ programming language sin( ) function) to retrieve a calculated sine value of the filtered audio data frame, using the formula, val=sin(2*pi*frequency*phase), where frequency is a frequency of the filtered audio data frame, and phase is a phase of the filtered audio data frame. Subsequently, the boombox object multiplies the calculated sine value (i.e., val) by the maximum value, or amplitude, of the filtered audio frame (i.e., A). This calculation creates a full sine carrier waveform. The boombox object then calculates a mixed value, where the mixed value represents a mix of the filtered audio data frame and the full sine carrier waveform. More specifically, the boombox object calculates the mixed value according to the formula, final=(mix*bpSample)+((1−mix)*(A*val)), where final is a mixed value representing a mix of the filtered audio data frame and the full sine carrier waveform, bpSample is the filtered audio data frame, mix is a decimal value between 0 and 1, A is the maximum value, or amplitude, of the filtered audio frame, and val is the calculated sine value of the full sine carrier waveform. Thus, according to the embodiment, the mix of the filtered audio data frame and the full sine carrier waveform is normalized by ensuring the mixed value representing a mix of the filtered audio data frame and the full sine carrier waveform (i.e., final) never exceeds 1. If, instead, the filtered audio data frame (i.e., bpSample) is merely added to a product of the maximum value, or amplitude, of the filtered audio frame (i.e., A) and the calculated sine value of the full sine carrier waveform (i.e., val), then the result yields an overpowering, undesired, and noisy haptic signal. This can be avoided by normalizing the mix of the filtered audio data frame and the full sine carrier waveform, as described above. The boombox object then returns the mix of the filtered audio data frame and the full sine carrier waveform to the reverb object. In other embodiments, multiple simultaneous carrier signals (at different frequencies) may be used to create a high bandwidth output signal. In these embodiments, summation of individual carriers is performed using an extension of the algorithm presented here.
The flow proceeds to 390. At 390, the reverb object plays a waveform effect at an actuator using the mix of the filtered audio data frame and the full sine carrier waveform. By playing the waveform effect at the actuator, the reverb object causes the actuator to produce a haptic effect.
As previously described, a haptic conversion module (such as haptic conversion module 16 of
According to certain embodiments, an envelope of the audio signal is first extracted. An envelope can be extracted using all frequencies of an original audio signal or a filtered version of the original audio signal. However, the envelope itself does not have the same frequency content as the original audio signal.
In one embodiment, an audio data frame is taken from the audio signal. As an example, an audio data frame may have a frame length of 1 ms. As another example, the audio data frame may have a frame length of 10 ms. A 1 ms audio data frame captures the envelope of frequencies above 500 Hz but lower frequencies are not captured, and “leak” through. A 10 ms frame captures the envelope of frequencies above 50 Hz, etc. In one embodiment, an absolute value of each sample in the frame is computed. In an alternate embodiment, a square of each sample in the frame is computed, rather than an absolute value. Such processing constitutes the sampling of an audio data frame performed by a haptic conversion module (such as haptic conversion module 16 of
A maximum value “V1” of the absolute sample values (or the square sample values in an alternate embodiment) is computed. The maximum value, “V1” can then be converted using a linear conversion function, as described below in relation to
Thus, according to the embodiment, an envelope of an audio signal is mapped to an amplitude of a haptic signal, where the amplitude of the haptic signal is filtered. In an alternate embodiment, the envelope of the audio signal can first be filtered, and then the filtered envelope of the audio signal can be mapped to an amplitude of a haptic signal.
Magnitude filtering, performed by a haptic conversion module (such as haptic conversion module 16 of
Each frame of certain ms duration is processed in the frequency domain. Thus, in certain embodiments, a fast fourier transformation (FFT) of the frame can be used to extract the frequency content. In other embodiments, band-pass filters can be used to extract the frequency content.
In one embodiment, the frame information can be divided in bass or low frequency (for example, lower than 200 Hz), middle frequency (for example, between 240 Hz and 4000 Khz), and high frequency (for example, higher than 4400 KHz), where the middle frequency content and high frequency content is extracted using the content of the signal.
According to the embodiment, when the signal starts to be processed, the first frame contains more middle frequency content (i.e., graph 730), given the maximum value obtained which is bigger than a frequency value in the low frequency band (i.e., graph 720) and high frequency band (i.e., graph 740).
According to the embodiment, a haptic effect can be played when the content frequency of the frame is different from the frequency content of the previous frame. As illustrated in
In alternate embodiments, variations of this technique can include: playing a pulse at every frequency change with magnitude as follows: maximum magnitude of the unprocessed frame; maximum magnitude corresponding to the frequency band; maximum magnitude allowed in TouchSense software only for the first frame.
In certain alternate embodiments, if the following frames have the same frequency content then: the magnitude of the unprocessed frame can be played; or the magnitude of the processed frame can be played; or a constant vibration, perhaps small, can be played as long as the frame has the same frequency content; or the magnitude can be varied based on power content in the frequency band.
In certain alternate embodiments, a different haptic effect can be played when a frame with a different frequency content is found.
Furthermore, human perception of sounds is not linear and depends on frequency of the sound. More particularly, humans are more sensitive to low frequencies than to high frequencies. Gender and age also affect the perception. Thus, in certain embodiments, there could be a compensation factor that boosts certain frequencies. The boosting factor can be based on human perception of sound or even preference of users where some frequency bands are enhanced manually. The software can enhance, in general, the high frequency content where it is known that users have a lower sensitivity to sound. This will allow the haptic effect for a specific band to be of greater strength, and thus, enhancing the sound that, while perceived and noticed by a user's auditive system, might not have enough power.
In certain embodiments, a user of the haptic conversion system can customize the sound-to-haptic effect conversion algorithm. More specifically, the user can choose whether a digital audio signal is filtered using a low-pass filter, a high-pass filter, or a band-pass filter. Alternatively, or in addition to customizing a filter type, the user can specify filter parameters rather than choosing between pre-defined parameters. According to these embodiments, a haptic conversion system can display one or more user-defined parameters within a graphical user interface. The user can then either choose from one or more user-defined parameters, or specify one or more user-defined parameters, using the graphical user interface. For example, a user can choose a type of filter utilized in the sound-to-haptic effect conversion algorithm from among a low-pass filter, a band-pass filter, and a high-pass filter. As another example, the user can specify one or more cut-off frequencies, where instead of a value between 250 and 500 Hz, the user can specify any values. As yet another example, the user can specify a gain that is used in the sound-to-haptic effect conversion algorithm.
In certain embodiments, rather than utilizing pre-defined filter parameters, the haptic conversion system can calculate filter parameters at runtime based on user-defined parameters. More specifically, a user can select a filter type. For example, a user can select a Butterworth design, a Bessel design, or a Chebyshev design, where for each design, a user can select a low-pass filter, a high-pass filter, a band-pass filter, or a band-stop filter. If a user selects a Chebyshev design, a user can also specify a ripple. A user can also specify a filter order, and a sample rate. A user can also specify one corner frequency if the user selected a low-pass filter or a high-pass filter, and two corner frequencies if the user selected a band-pass filter or a band-stop filter. A user can also select an optional additional zero value, and an optional lower limit of a magnitude scale. Furthermore, a user can also select whether the filter is designed by a bilinear transform method, or a matched z-transform method. The haptic conversion system can then calculate one or more filter parameters at runtime based on the parameters specified by the user.
At 920, a maximum value of the data frame is generated. In certain embodiments, the maximum value is generated by sampling the data frame to generate one or more sample values of the data frame, wherein the maximum value is a maximum value of the one or more sample values. In one embodiment, the one or more sample values are absolute values. In an alternate embodiment, the one or more sample values are square values. In certain embodiments, the maximum value is converted from an original maximum value using a linear conversion function. The flow proceeds to 930.
At 930, the data frame is transformed. In certain embodiments, the data frame is band-pass filtered using a band-pass filter. In these embodiments, a single frequency band of frequency signals within the data frame is filtered on. In alternate embodiments, the data frame is low-passed filtered using an IIR filter. In these embodiments, a single frequency band of frequency signals within the data frame can be filtered on. For example, a band of 0-500 Hz can be filtered on. In certain embodiments, before the data frame is transformed, the data frame is resampled. In certain embodiments, after the data frame is transformed, the digital audio signal contained with the data frame is boosted. The flow proceeds to 940.
At 940, at least one sine carrier waveform is generated based on a maximum value of the transformed data frame. In certain embodiments, the transformed data frame is scanned, and the maximum value of the transformed data frame is determined. Next, a sine function is used to retrieve a calculated sine value of the transformed data frame, based on a frequency of the transformed data frame, and a phase of the transformed data frame. Subsequently, the calculated sine value is multiplied by the maximum value of the transformed data frame, in order to create the at least one sine carrier waveform.
At 950, the transformed data frame is mixed with the at least one sine carrier waveform to generate a modulated data frame. In certain embodiments, the transformed data frame is mixed with the at least one sine carrier waveform according to a formula where the mix of the transformed data frame and the at least one sine carrier waveform is normalized.
At 960, a haptic signal is generated based on the maximum value and the modulated data frame. In certain embodiments, the maximum value determines whether the haptic signal is generated, and the modulated data frame defines a precision of the haptic signal. In certain embodiments where the maximum value determines whether the haptic signal is generated, the haptic signal is generated only when the maximum value is greater than a specified threshold. The flow proceeds to 970.
At 970, the haptic signal is sent through an actuator to generate the one or more haptic effects. In certain embodiments, the haptic signal is a waveform effect. Also, in certain embodiments, the generating and the sending the haptic signal further includes passing the modulated data frame as a parameter to an application programming interface of a haptic effect player module.
Thus, according to an embodiment, a haptic conversion system intercepts audio data frames that are part of an audio signal, converts the audio data frames into a haptic signal, where a maximum value of each audio data frame defines a sine-wave periodic carrier frequency that is mixed with each filtered audio data frame to define a waveform of the haptic signal, and plays the converted haptic signal through an actuator to generate haptic effects. The modulation of each audio data frame performed by the haptic conversion system showcases a large amount of precision and realism felt within the actuator itself, and results in more flexibility in the haptic signal that is played at the actuator.
The features, structures, or characteristics of the invention described throughout this specification may be combined in any suitable manner in one or more embodiments. For example, the usage of “one embodiment,” “some embodiments,” “certain embodiment,” “certain embodiments,” or other similar language, throughout this specification refers to the fact that a particular feature, structure, or characteristic described in connection with the embodiment may be included in at least one embodiment of the present invention. Thus, appearances of the phrases “one embodiment,” “some embodiments,” “a certain embodiment,” “certain embodiments,” or other similar language, throughout this specification do not necessarily all refer to the same group of embodiments, and the described features, structures, or characteristics may be combined in any suitable manner in one or more embodiments.
One having ordinary skill in the art will readily understand that the invention as discussed above may be practiced with steps in a different order, and/or with elements in configurations which are different than those which are disclosed. Therefore, although the invention has been described based upon these preferred embodiments, it would be apparent to those of skill in the art that certain modifications, variations, and alternative constructions would be apparent, while remaining within the spirit and scope of the invention. In order to determine the metes and bounds of the invention, therefore, reference should be made to the appended claims.
This application is a continuation of U.S. patent application Ser. No. 15/433,488, filed on Feb. 15, 2017, which is a continuation of U.S. patent application Ser. No. 14/709,672, filed May 12, 2015 and issued on Mar. 28, 2017 as U.S. Pat. No. 9,606,627, which is a continuation of U.S. patent application Ser. No. 14/175,163, filed on Feb. 7, 2014, now U.S. Pat. No. 9,064,387, issued Jun. 23, 2015, which is a continuation of U.S. patent application Ser. No. 13/366,010, filed on Feb. 3, 2012, now U.S. Pat. No. 8,717,152, issued May 6, 2014, which claims priority of U.S. Provisional Patent Application Ser. No. 61/441,792, filed on Feb. 11, 2011. The disclosures of the above applications are all hereby incorporated by reference in their entirety.
Number | Name | Date | Kind |
---|---|---|---|
5675709 | Chiba | Oct 1997 | A |
5684722 | Thorner et al. | Nov 1997 | A |
5842163 | Weintraub | Nov 1998 | A |
6183367 | Kaji et al. | Feb 2001 | B1 |
6285351 | Chang et al. | Sep 2001 | B1 |
6422941 | Thorner et al. | Jul 2002 | B1 |
7091948 | Chang et al. | Aug 2006 | B2 |
7623114 | Rank | Nov 2009 | B2 |
7979146 | Ullrich et al. | Jul 2011 | B2 |
8000825 | Ullrich et al. | Aug 2011 | B2 |
8378964 | Ullrich et al. | Feb 2013 | B2 |
8688251 | Ullrich et al. | Apr 2014 | B2 |
8717152 | Bhatia et al. | May 2014 | B2 |
8754757 | Ullrich et al. | Jun 2014 | B1 |
8754758 | Ullrich et al. | Jun 2014 | B1 |
8860563 | Cruz-Hernandez et al. | Oct 2014 | B2 |
9064387 | Bhatia et al. | Jun 2015 | B2 |
9092059 | Bhatia | Jul 2015 | B2 |
9349378 | Bharitkar et al. | May 2016 | B2 |
9354704 | Lacroix et al. | May 2016 | B2 |
9606627 | Bhatia et al. | Mar 2017 | B2 |
10055950 | Bhatia et al. | Aug 2018 | B2 |
10152296 | Trestain | Dec 2018 | B2 |
20020082754 | Robichaux et al. | Jun 2002 | A1 |
20030067440 | Rank | Apr 2003 | A1 |
20050134561 | Tierling et al. | Jun 2005 | A1 |
20070079138 | Chou | Apr 2007 | A1 |
20070097073 | Takashima et al. | May 2007 | A1 |
20070146137 | Claesson | Jun 2007 | A1 |
20070236449 | Lacroix et al. | Oct 2007 | A1 |
20070242040 | Ullrich et al. | Oct 2007 | A1 |
20080084384 | Gregorio et al. | Apr 2008 | A1 |
20090002315 | Chu | Jan 2009 | A1 |
20090007758 | Schlosser et al. | Jan 2009 | A1 |
20090009481 | Yatsu et al. | Jan 2009 | A1 |
20090231276 | Ullrich et al. | Sep 2009 | A1 |
20100066512 | Rank | Mar 2010 | A1 |
20100141408 | Doy et al. | Jun 2010 | A1 |
20100156818 | Burrough et al. | Jun 2010 | A1 |
20100231539 | Cruz-Hernandez et al. | Sep 2010 | A1 |
20100287311 | Cruz-Hernandez et al. | Nov 2010 | A1 |
20100316228 | Baran et al. | Dec 2010 | A1 |
20110102160 | Heubel et al. | May 2011 | A1 |
20110102161 | Heubel et al. | May 2011 | A1 |
20110115709 | Cruz-Hernandez | May 2011 | A1 |
20110128132 | Ullrich et al. | Jun 2011 | A1 |
20110169908 | Lee et al. | Jul 2011 | A1 |
20110202155 | Ullrich et al. | Aug 2011 | A1 |
20110215913 | Ullrich et al. | Sep 2011 | A1 |
20120026114 | Lee et al. | Feb 2012 | A1 |
20120206246 | Cruz-Hernandez et al. | Aug 2012 | A1 |
20120206247 | Bhatia et al. | Aug 2012 | A1 |
20120306631 | Hughes | Dec 2012 | A1 |
20130088339 | Lim et al. | Apr 2013 | A1 |
20130131851 | Ullrich et al. | May 2013 | A1 |
20130207917 | Cruz-Hernandez et al. | Aug 2013 | A1 |
20130265286 | Da Costa et al. | Oct 2013 | A1 |
20140064516 | Cruz-Hernandez et al. | May 2014 | A1 |
20140118125 | Bhatia | May 2014 | A1 |
20150070261 | Saboune et al. | Mar 2015 | A1 |
20150070144 | Weddle et al. | May 2015 | A1 |
20150070260 | Saboune et al. | May 2015 | A1 |
20150154966 | Bharitkar et al. | Jun 2015 | A1 |
20150355713 | Lacroix et al. | Dec 2015 | A1 |
20160063850 | Yang et al. | Mar 2016 | A1 |
20160370862 | Colgate et al. | Dec 2016 | A1 |
Number | Date | Country |
---|---|---|
1599925 | Mar 2005 | CN |
0144774 | Jun 1985 | EP |
2487557 | Aug 2012 | EP |
H10-98344 | Apr 1998 | JP |
H10-506508 | Jun 1998 | JP |
H11-196497 | Jul 1999 | JP |
H11-296126 | Oct 1999 | JP |
2003-099177 | Apr 2003 | JP |
2005-506613 | Mar 2005 | JP |
2008-283305 | Nov 2008 | JP |
2011-141890 | Jul 2011 | JP |
10-2004-0062559 | Jul 2004 | KR |
9609617 | Mar 1996 | WO |
03032289 | Apr 2003 | WO |
2010054014 | May 2010 | WO |
2010085575 | Jul 2010 | WO |
2010104953 | Sep 2010 | WO |
2011059703 | May 2011 | WO |
Entry |
---|
Notice of Preliminary Rejection issued by KIPO in KR Application No. 10-2018-0082337, dated Aug. 27, 2018. |
Office Action dated Feb. 15, 2018 in Japanese Patent Application No. 2017-007414. |
Communication pursuant to Article 94(3) EPC dated Jan. 31, 2018 in European Patent Application No. 12 154 829.1. |
Decision to Grant dated Nov. 20, 2017 in Chinese Patent Application No. 201210131186.5. |
Notice of Reasons for Rejection issued in JP Application No. 2017-007414, dated Sep. 5, 2017. |
Notice of Preliminary Rejection issued by KIPO in KR Application No. 10-2012-0013995, dated Aug. 29, 2017. |
Final Office Action dated Feb. 17, 2016 in U.S. Appl. No. 13/365,984. |
Notice of Allowance dated Jun. 9, 2016 in U.S. Appl. No. 13/365,984. |
Non-Final Office Action dated Jun. 19, 2015 in U.S. Appl. No. 13/365,984. |
Non-Final Office Action dated Jul. 7, 2014 in U.S. Appl. No. 13/365,984. |
Final Office Action dated Dec. 18, 2014 in U.S. Appl. No. 13/365,984. |
International Search Report issued in PCT/US2010/026829, dated May 4, 2010. |
Tony Fisher, “Butterworth/Bessel/Chebyshev Filters,” http://www-users.cs.york.ac.uk/˜fisher/mkfilter/trad.html (last visited Jan. 29, 2012). |
Notice of Preliminary Rejection from Korean Intellectual Property Office issued in Korean Application No. 10-2012-0013978, dated Oct. 16, 2018. |
Office Action from CN Appl. 201810109500.7 dated Dec. 5, 2018. |
Communication under Rule 71(3) EPC, issued in European Application No. 12154829.1, dated Dec. 20, 2018. |
Notice of Reasons for Rejection issued in Japanese Patent Application No. 2018-115848, dated May 7, 2019. |
Number | Date | Country | |
---|---|---|---|
20190005782 A1 | Jan 2019 | US |
Number | Date | Country | |
---|---|---|---|
61441792 | Feb 2011 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 15433488 | Feb 2017 | US |
Child | 16102464 | US | |
Parent | 14709672 | May 2015 | US |
Child | 15433488 | US | |
Parent | 14175163 | Feb 2014 | US |
Child | 14709672 | US | |
Parent | 13366010 | Feb 2012 | US |
Child | 14175163 | US |