Various technologies have been used to detect a touch input on a display area. The most popular technologies today include capacitive and resistive touch detection technology. Using resistive touch technology, often a glass panel is coated with multiple conductive layers that register touches when physical pressure is applied to the layers to force the layers to make physical contact. Using capacitive touch technology, often a glass panel is coated with material that can hold an electrical charge sensitive to a human finger. By detecting the change in the electrical charge due to a touch, a touch location can be detected. However, with resistive and capacitive touch detection technologies, the glass screen is required to be coated with a material that reduces the clarity of the glass screen. Additionally, because the entire glass screen is required to be coated with a material, manufacturing and component costs can become prohibitively expensive as larger screens are desired.
Another type of touch detection technology includes bending wave technology. One example includes the Elo Touch Systems Acoustic Pulse Recognition, commonly called APR, manufactured by Elo Touch Systems of 301 Constitution Drive, Menlo Park, Calif. 94025. The APR system includes transducers attached to the edges of a touchscreen glass that pick up the sound emitted on the glass due to a touch. However, the surface glass may pick up other external sounds and vibrations that reduce the accuracy and effectiveness of the APR system to efficiently detect a touch input. Another example includes the Surface Acoustic Wave-based technology, commonly called SAW, such as the Elo IntelliTouch Plus™ of Elo Touch Systems. The SAW technology sends ultrasonic waves in a guided pattern using reflectors on the touch screen to detect a touch. However, sending the ultrasonic waves in the guided pattern increases costs and may be difficult to achieve. Detecting additional types of inputs, such as multi-touch inputs, may not be possible or may be difficult using SAW or APR technology. Therefore there exists a need for a better way to detect an input on a surface.
Various embodiments of the invention are disclosed in the following detailed description and the accompanying drawings.
The invention can be implemented in numerous ways, including as a process; an apparatus; a system; a composition of matter; a computer program product embodied on a computer readable storage medium; and/or a processor, such as a processor configured to execute instructions stored on and/or provided by a memory coupled to the processor. In this specification, these implementations, or any other form that the invention may take, may be referred to as techniques. In general, the order of the steps of disclosed processes may be altered within the scope of the invention. Unless stated otherwise, a component such as a processor or a memory described as being configured to perform a task may be implemented as a general component that is temporarily configured to perform the task at a given time or a specific component that is manufactured to perform the task. As used herein, the term ‘processor’ refers to one or more devices, circuits, and/or processing cores configured to process data, such as computer program instructions.
A detailed description of one or more embodiments of the invention is provided below along with accompanying figures that illustrate the principles of the invention. The invention is described in connection with such embodiments, but the invention is not limited to any embodiment. The scope of the invention is limited only by the claims and the invention encompasses numerous alternatives, modifications and equivalents. Numerous specific details are set forth in the following description in order to provide a thorough understanding of the invention. These details are provided for the purpose of example and the invention may be practiced according to the claims without some or all of these specific details. For the purpose of clarity, technical material that is known in the technical fields related to the invention has not been described in detail so that the invention is not unnecessarily obscured.
Detecting a location of a touch input is disclosed. For example, a user touch input on a glass surface of a display screen is detected. In some embodiments, a plurality of transmitters coupled to a propagating medium (e.g., glass) emits signals that are distinguishable from other signals emitted from other transmitters. For example, a signal such as an acoustic or ultrasonic signal is propagated freely through a propagating medium with a touch input surface from each transmitter coupled to the propagating medium. In some embodiments, the signals emitted by the transmitters are distinguishable from each other by varying a phase of the signals (e.g., code division multiplexing, code division multiple access (CDMA), spread spectrum multiple access (SSMA)), a frequency range of the signals (e.g., frequency division multiplexing, frequency division multiple access (FDMA)) or a timing of the signals (e.g., time division multiplexing, time division multiple access (TDMA)).
At least one receiver is coupled to the propagating medium, and the receiver is configured to receive the signals from the transmitters to detect the location of the touch input on a surface of the propagating medium as indicated by the effect of the touch input on each of the distinguishable signals. For example, when the surface of the propagating medium is touched, the emitted signals propagating through the propagating medium are disturbed (e.g., the touch causes an interference with the propagated signals). In some embodiments, by processing the received signals and comparing it against corresponding expected signals without the disturbance, a location on the surface associated with the touch input is at least in part determined. For example, a relative time difference between when the disturbance was detected in the received signals is used to determine the touch input location on the surface. In various embodiments, the touch input includes a physical contact to a surface using a human finger, pen, pointer, stylus, and/or any other body parts or objects that can be used to contact or disturb the surface. In some embodiments, the touch input includes an input gesture and/or a multi-touch input.
In some embodiments, the received signal is used to determine one or more of the following associated with a touch input: a gesture, a coordinate position, a time, a time frame, a direction, a velocity, a force magnitude, a proximity magnitude, a pressure, a size, and other measurable or derived parameters. In some embodiments, by detecting disturbances of a freely propagated signal, touch input detection technology can be applied to larger surface regions with less or no additional cost due to a larger surface region as compared to certain previous touch detection technologies. Additionally, the optical transparency of a touch screen may not have to be affected as compared to resistive and capacitive touch technologies. Merely by way of example, the touch detection described herein can be applied to a variety of objects such as a kiosk, an ATM, a computing device, an entertainment device, a digital signage apparatus, a cell phone, a tablet computer, a point of sale terminal, a food and restaurant apparatus, a gaming device, a casino game and application, a piece of furniture, a vehicle, an industrial application, a financial application, a medical device, an appliance, and any other objects or devices having surfaces.
Examples of transmitters 104, 106, 108, and 110 include piezoelectric transducers, electromagnetic transducers, transmitters, sensors, and/or any other transmitters and transducers capable of propagating a signal through medium 102. Examples of sensors 112, 114, 116, and 118 include piezoelectric transducers, electromagnetic transducers, laser vibrometer transmitters, and/or any other sensors and transducers capable of detecting a signal on medium 102. In some embodiments, the transmitters and sensors shown in
Touch detector 120 is connected to the transmitters and sensors shown in
A signal detected from a sensor such as sensor 112 of
At 304, signal transmitters and sensors are calibrated. In some embodiments, calibrating the transmitter includes calibrating a characteristic of a signal driver and/or transmitter (e.g., strength). In some embodiments, calibrating the sensor includes calibrating a characteristic of a sensor (e.g., sensitivity). In some embodiments, the calibration of 304 is performed to optimize the coverage and improve signal-to-noise transmission/detection of a signal (e.g., acoustic or ultrasonic) to be propagated through a medium and/or a disturbance to be detected. For example, one or more components of the system of
At 306, surface disturbance detection is calibrated. In some embodiments, a test signal is propagated through a medium such as medium 102 of
At 308, a validation of a touch detection system is performed. For example, the system of
In some embodiments, sending the signal includes determining the signal to be transmitted by a transmitter such that the signal is distinguishable from other signal(s) transmitted by other transmitters. In some embodiments, sending the signal includes determining a phase of the signal to be transmitted (e.g., utilize code division multiplexing/CDMA). For example, an offset within a pseudorandom binary sequence to be transmitted is determined. In this example, each transmitter (e.g., transmitters 104, 106, 108, and 110 of
In some embodiments, sending the signal includes determining a frequency of the signal to be transmitted (e.g., utilize frequency division multiplexing/FDMA). For example, a frequency range to be utilized for the signal is determined. In this example, each transmitter (e.g., transmitters 104, 106, 108, and 110 of
In some embodiments, sending the signal includes determining a timing of the signal to be transmitted (e.g., utilize time division multiplexing/TDMA). For example, a time when the signal should be transmitted is determined. In this example, each transmitter (e.g., transmitters 104, 106, 108, and 110 of
At 404, the active signal that has been disturbed by a disturbance of the surface region is received. The disturbance may be associated with a user touch indication. In some embodiments, the disturbance causes the active signal that is propagating through a medium to be attenuated and/or delayed. In some embodiments, the disturbance in a selected portion of the active signal corresponds to a location on the surface that has been indicated (e.g., touched) by a user.
At 406, the received signal is processed to at least in part determine a location associated with the disturbance. In some embodiments, determining the location includes extracting a desired signal from the received signal at least in part by removing or reducing undesired components of the received signal such as disturbances caused by extraneous noise and vibrations not useful in detecting a touch input. In some embodiments, components of the received signal associated with different signals of different transmitters are separated. For example, different signals originating from different transmitters are isolated from other signals of other transmitters for individual processing. In some embodiments, determining the location includes comparing at least a portion of the received signal (e.g., signal component from a single transmitter) to a reference signal (e.g., reference signal corresponding to the transmitter signal) that has not been affected by the disturbance. The result of the comparison may be used with a result of other comparisons performed using the reference signal and other signal(s) received at a plurality of sensors. The location, in some embodiments, is a location (e.g., a location coordinate) on the surface region where a user has provided a touch input. In addition to determining the location, one or more of the following information associated with the disturbance may be determined at 406: a gesture, simultaneous user indications (e.g., multi-touch input), a time, a status, a direction, a velocity, a force magnitude, a proximity magnitude, a pressure, a size, and other measurable or derived information. In some embodiments, the location is not determined at 406 if a location cannot be determined using the received signal and/or the disturbance is determined to be not associated with a user input. Information determined at 406 may be provided and/or outputted.
Although
At 502, a received signal is conditioned. In some embodiments, the received signal is a signal including a pseudorandom binary sequence that has been freely propagated through a medium with a surface that can be used to receive a user input. For example, the received signal is the signal that has been received at 404 of
At 504, an analog to digital signal conversion is performed on the signal that has been conditioned at 502. In various embodiments, any number of standard analog to digital signal converters may be used. The resulting digital signal is used to perform a first correlation at 506. In some embodiments, components of the received signal associated with different signals/transmitters are separated. For example, different signals originating from different transmitters are isolated from other signals of other transmitters. In some embodiments, performing the first correlation includes correlating at least a portion of the converted signal (e.g., signal component from a single transmitter) with a reference signal (e.g., corresponding reference signal of the signal transmitter signal). Performing the correlation includes cross-correlating or determining a convolution (e.g., interferometry) of the converted signal with a reference signal to measure the similarity of the two signals, as a time-lag is applied to one of the signals. By performing the correlation, the location of a portion of the converted signal that most corresponds to the reference signal can be located. For example, a result of the correlation can be plotted as a graph of time within the received and converted signal (e.g., time-lag between the signals) vs. a measure of similarity. The associated time value of the largest value of the measure of similarity corresponds to the location where the two signals most correspond. By comparing this measured time value against a reference time value (e.g., at 306 of
At 508, a second correlation is performed based on a result of the first correlation. Performing the second correlation includes correlating (e.g., cross-correlation or convolution similar to step 506) at least a portion of the converted signal in 504 with a second reference signal. The second reference signal is a more complex/detailed (e.g., more computationally intensive) reference signal as compared to the first reference signal used in 506. In some embodiments, the second correlation is performed in 508 because using the second reference signal in 506 may be too computationally intensive for the time interval required to be correlated in 506. Performing the second correlation based on the result of the first correlation includes using one or more time values determined as a result of the first correlation. For example, using a result of the first correlation, a range of likely time values (e.g., time-lag) that most correlate between the received signal and the first reference signal is determined and the second correlation is performed using the second reference signal only across the determined range of time values to fine tune and determine the time value that most corresponds to where the second reference signal (and, by association, also the first reference signal) matched the received signal. In various embodiments, the first and second correlations have been used to determine a portion within the received signal that corresponds to a disturbance caused by a touch input at a location on a surface of a propagating medium. In other embodiments, the second correlation is optional. For example, only a single correlation step is performed.
At 510, a result of the second correlation is used to at least in part determine a location associated with a disturbance. In some embodiments, determining the location includes comparing a determined time value where the signals of the second correlation are most correlated and comparing the determined time value with a reference time value (e.g., determined at 306 of
The propagated signal may contain a predetermined signal and the predetermined signal is received at the various sensors. Each of the received signals is correlated with a reference signal to determine the results received at 602. In some embodiments, the received results are associated with a same signal content (e.g., same binary sequence) that has been freely propagated on a medium at the same time. In some embodiments, the received results are associated with different signal contents that have been disturbed by the same disturbance. In some embodiments, the received signal at a receiver/sensor includes components of a plurality of distinguishable signals transmitted by different transmitters and the received signal is separated into different received signals each corresponding to a transmitted signal of a signal transmitter for individual analysis to at least in part determine the results received at 602.
At 604, time differences associated with the plurality of results are used to determine a location associated with the disturbance. In some embodiments, each of the time differences is associated with a time when signals used in the correlation are most correlated. In some embodiments, the time differences are associated with a determined time delay/offset or phase difference caused on the received signal due to the disturbance. This time delay may be calculated by comparing a time value determined using a correlation with a reference time value that is associated with a scenario where a touch input has not been specified. The result of the comparison may be used to calculate a location of the disturbance relative to the locations of transmitters and/or sensors that received the plurality of signals. By using the location of the transmitters and/or sensors relative to a surface of a medium that has propagated the received signal, a location on the surface where the disturbance originated may be determined.
Although the foregoing embodiments have been described in some detail for purposes of clarity of understanding, the invention is not limited to the details provided. There are many alternative ways of implementing the invention. The disclosed embodiments are illustrative and not restrictive.
This application is a continuation in part of co-pending U.S. patent application Ser. No. 13/451,288 entitled METHOD AND APPARATUS FOR ACTIVE ULTRASONIC TOUCH DEVICES filed Apr. 19, 2012, which is incorporated herein by reference for all purposes, which claims priority to U.S. Provisional Application No. 61/479,331, entitled METHOD AND APPARATUS FOR ACTIVE ULTRASONIC TOUCH DEVICES filed Apr. 26, 2011 which is incorporated herein by reference for all purposes.
Number | Name | Date | Kind |
---|---|---|---|
4488000 | Glenn | Dec 1984 | A |
5334805 | Knowles et al. | Aug 1994 | A |
5451723 | Huang et al. | Sep 1995 | A |
5563849 | Hall et al. | Oct 1996 | A |
5573077 | Knowles | Nov 1996 | A |
5637839 | Yamaguchi et al. | Jun 1997 | A |
5708460 | Young et al. | Jan 1998 | A |
5739479 | Davis-Cannon et al. | Apr 1998 | A |
5784054 | Armstrong et al. | Jul 1998 | A |
5883457 | Rinde et al. | Mar 1999 | A |
5912659 | Rutledge et al. | Jun 1999 | A |
6091406 | Kambara et al. | Jul 2000 | A |
6232960 | Goldman | May 2001 | B1 |
6254105 | Rinde et al. | Jul 2001 | B1 |
6473075 | Gomes et al. | Oct 2002 | B1 |
6492979 | Kent et al. | Dec 2002 | B1 |
6535147 | Masters et al. | Mar 2003 | B1 |
6567077 | Inoue et al. | May 2003 | B2 |
6630929 | Adler et al. | Oct 2003 | B1 |
6633280 | Matsumoto et al. | Oct 2003 | B1 |
6636201 | Gomes et al. | Oct 2003 | B1 |
6788296 | Ikeda et al. | Sep 2004 | B2 |
6798403 | Kitada et al. | Sep 2004 | B2 |
6856259 | Sharp | Feb 2005 | B1 |
6891527 | Chapman et al. | May 2005 | B1 |
6948371 | Tanaka et al. | Sep 2005 | B2 |
7000474 | Kent | Feb 2006 | B2 |
7006081 | Kent et al. | Feb 2006 | B2 |
7116315 | Sharp et al. | Oct 2006 | B2 |
7119800 | Kent et al. | Oct 2006 | B2 |
7187369 | Kanbara et al. | Mar 2007 | B1 |
7193617 | Kanbara et al. | Mar 2007 | B1 |
7204148 | Tanaka et al. | Apr 2007 | B2 |
7218248 | Kong et al. | May 2007 | B2 |
7274358 | Kent | Sep 2007 | B2 |
RE39881 | Flowers | Oct 2007 | E |
7315336 | North et al. | Jan 2008 | B2 |
7345677 | Ing et al. | Mar 2008 | B2 |
7411581 | Hardie-Bick | Aug 2008 | B2 |
7456825 | Kent et al. | Nov 2008 | B2 |
7511711 | Ing et al. | Mar 2009 | B2 |
7545365 | Kent et al. | Jun 2009 | B2 |
7554246 | Maruyama et al. | Jun 2009 | B2 |
7583255 | Ing | Sep 2009 | B2 |
7649807 | Ing | Jan 2010 | B2 |
7683894 | Kent | Mar 2010 | B2 |
7880721 | Suzuki | Feb 2011 | B2 |
7920133 | Tsumura et al. | Apr 2011 | B2 |
8085124 | Ing | Dec 2011 | B2 |
8228121 | Benhamouda et al. | Jul 2012 | B2 |
8237676 | Duheille et al. | Aug 2012 | B2 |
8319752 | Hardie-Bick | Nov 2012 | B2 |
8325159 | Kent et al. | Dec 2012 | B2 |
8358277 | Mosby et al. | Jan 2013 | B2 |
8378974 | Aroyan et al. | Feb 2013 | B2 |
8392486 | Ing | Mar 2013 | B2 |
8427423 | Tsumura | Apr 2013 | B2 |
8436806 | Almalki et al. | May 2013 | B2 |
8436808 | Chapman et al. | May 2013 | B2 |
8493332 | D'Souza | Jul 2013 | B2 |
8576202 | Tanaka et al. | Nov 2013 | B2 |
8619063 | Chaine et al. | Dec 2013 | B2 |
8638318 | Gao et al. | Jan 2014 | B2 |
8648815 | Kent et al. | Feb 2014 | B2 |
8659579 | Nadjar et al. | Feb 2014 | B2 |
8670290 | Aklil et al. | Mar 2014 | B2 |
8681128 | Scharff et al. | Mar 2014 | B2 |
8692809 | D'Souza | Apr 2014 | B2 |
8692810 | Ing | Apr 2014 | B2 |
8692812 | Hecht | Apr 2014 | B2 |
8730213 | D'Souza et al. | May 2014 | B2 |
8749517 | Aklil | Jun 2014 | B2 |
8823685 | Scharff et al. | Sep 2014 | B2 |
8854339 | Kent et al. | Oct 2014 | B2 |
8890852 | Aroyan et al. | Nov 2014 | B2 |
8896429 | Chaine | Nov 2014 | B2 |
8896564 | Scharff et al. | Nov 2014 | B2 |
8917249 | Buuck et al. | Dec 2014 | B1 |
8941624 | Kent et al. | Jan 2015 | B2 |
8946973 | Pelletier | Feb 2015 | B2 |
8994696 | Berget et al. | Mar 2015 | B2 |
9030436 | Ikeda | May 2015 | B2 |
9046959 | Schevin et al. | Jun 2015 | B2 |
9046966 | D'Souza | Jun 2015 | B2 |
9058071 | Esteve | Jun 2015 | B2 |
9099971 | Lynn et al. | Aug 2015 | B2 |
20010050677 | Tosaya | Dec 2001 | A1 |
20020185981 | Dietz et al. | Dec 2002 | A1 |
20030197691 | Fujiwara | Oct 2003 | A1 |
20030206162 | Roberts | Nov 2003 | A1 |
20040133366 | Sullivan | Jul 2004 | A1 |
20040160421 | Sullivan | Aug 2004 | A1 |
20040183788 | Kurashima et al. | Sep 2004 | A1 |
20040239649 | Ludtke | Dec 2004 | A1 |
20040246239 | Knowles et al. | Dec 2004 | A1 |
20060071912 | Hill et al. | Apr 2006 | A1 |
20060114233 | Radivojevic et al. | Jun 2006 | A1 |
20060139340 | Geaghan | Jun 2006 | A1 |
20060166681 | Lohbihler | Jul 2006 | A1 |
20060262104 | Sullivan et al. | Nov 2006 | A1 |
20070109274 | Reynolds | May 2007 | A1 |
20070165009 | Sakurai et al. | Jul 2007 | A1 |
20070171212 | Sakurai et al. | Jul 2007 | A1 |
20070211022 | Boillot | Sep 2007 | A1 |
20070214462 | Boillot | Sep 2007 | A1 |
20070229479 | Choo et al. | Oct 2007 | A1 |
20070279398 | Tsumura et al. | Dec 2007 | A1 |
20080018618 | Hill et al. | Jan 2008 | A1 |
20080030479 | Lowles et al. | Feb 2008 | A1 |
20080081671 | Wang et al. | Apr 2008 | A1 |
20080105470 | Van De Ven et al. | May 2008 | A1 |
20080111788 | Rosenberg et al. | May 2008 | A1 |
20080174565 | Chang et al. | Jul 2008 | A1 |
20080198145 | Knowles et al. | Aug 2008 | A1 |
20080231612 | Hill et al. | Sep 2008 | A1 |
20080284755 | Hardie-Bick | Nov 2008 | A1 |
20090103853 | Daniel | Apr 2009 | A1 |
20090146533 | Leskinen et al. | Jun 2009 | A1 |
20090160728 | Emrick et al. | Jun 2009 | A1 |
20090167704 | Terlizzi et al. | Jul 2009 | A1 |
20090237372 | Kim et al. | Sep 2009 | A1 |
20090271004 | Zecchin et al. | Oct 2009 | A1 |
20090273583 | Norhammar | Nov 2009 | A1 |
20090315848 | Ku et al. | Dec 2009 | A1 |
20100026667 | Bernstein | Feb 2010 | A1 |
20100044121 | Simon et al. | Feb 2010 | A1 |
20100045635 | Soo | Feb 2010 | A1 |
20100079264 | Hoellwarth | Apr 2010 | A1 |
20100117993 | Kent | May 2010 | A1 |
20100156818 | Burrough et al. | Jun 2010 | A1 |
20100165215 | Shim | Jul 2010 | A1 |
20100185989 | Shiplacoff et al. | Jul 2010 | A1 |
20100188356 | Vu et al. | Jul 2010 | A1 |
20100245265 | Sato et al. | Sep 2010 | A1 |
20100269040 | Lee | Oct 2010 | A1 |
20100277431 | Klinghult | Nov 2010 | A1 |
20100315373 | Steinhauser et al. | Dec 2010 | A1 |
20100321312 | Han et al. | Dec 2010 | A1 |
20100321325 | Springer et al. | Dec 2010 | A1 |
20100321337 | Liao et al. | Dec 2010 | A1 |
20110012717 | Pance et al. | Jan 2011 | A1 |
20110018695 | Bells et al. | Jan 2011 | A1 |
20110042152 | Wu | Feb 2011 | A1 |
20110057903 | Yamano et al. | Mar 2011 | A1 |
20110063228 | St. Pierre | Mar 2011 | A1 |
20110080350 | Almalki et al. | Apr 2011 | A1 |
20110084912 | Almalki | Apr 2011 | A1 |
20110156967 | Oh et al. | Jun 2011 | A1 |
20110167391 | Momeyer et al. | Jul 2011 | A1 |
20110175813 | Sarwar et al. | Jul 2011 | A1 |
20110199342 | Vartanian | Aug 2011 | A1 |
20110260990 | Ali et al. | Oct 2011 | A1 |
20110298670 | Jung et al. | Dec 2011 | A1 |
20110316790 | Ollila et al. | Dec 2011 | A1 |
20120001875 | Li et al. | Jan 2012 | A1 |
20120026114 | Lee et al. | Feb 2012 | A1 |
20120050230 | Harris | Mar 2012 | A1 |
20120062564 | Miyashita et al. | Mar 2012 | A1 |
20120068939 | Pemberton-Pigott | Mar 2012 | A1 |
20120068970 | Pemberton-Pigott | Mar 2012 | A1 |
20120081337 | Camp et al. | Apr 2012 | A1 |
20120088548 | Yun et al. | Apr 2012 | A1 |
20120120031 | Thuillier | May 2012 | A1 |
20120126962 | Ujii et al. | May 2012 | A1 |
20120127088 | Pance et al. | May 2012 | A1 |
20120188889 | Sambhwani et al. | Jul 2012 | A1 |
20120194466 | Posamentier | Aug 2012 | A1 |
20120200517 | Nikolovski | Aug 2012 | A1 |
20120229407 | Harris et al. | Sep 2012 | A1 |
20120232834 | Roche et al. | Sep 2012 | A1 |
20120272089 | Hatfield et al. | Oct 2012 | A1 |
20130059532 | Mahanfar et al. | Mar 2013 | A1 |
20130249831 | Harris | Sep 2013 | A1 |
20140185834 | Fromel et al. | Jul 2014 | A1 |
Number | Date | Country |
---|---|---|
101373415 | Feb 2009 | CN |
101669088 | Mar 2010 | CN |
2948787 | Feb 2011 | FR |
1020040017272 | Feb 2004 | KR |
1020050092179 | Feb 2004 | KR |
1020070005580 | Jan 2007 | KR |
1020080005990 | Jan 2008 | KR |
2006115947 | Nov 2006 | WO |
2006115947 | Nov 2006 | WO |
2011010037 | Jan 2011 | WO |
2011024434 | Mar 2011 | WO |
2011048433 | Apr 2011 | WO |
2011051722 | May 2011 | WO |
2012010912 | Jan 2012 | WO |
Entry |
---|
Liu et al., “Acoustic Wave Approach for Multi-Touch Tactile Sensing”, Micro-NanoMechatronics and Human Science, 2009. MHS 2009. International Symposium, Nov. 9-11, 2009. |
Number | Date | Country | |
---|---|---|---|
20140078112 A1 | Mar 2014 | US |
Number | Date | Country | |
---|---|---|---|
61479331 | Apr 2011 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13451288 | Apr 2012 | US |
Child | 14033316 | US |