Simultaneous sensing arrangement

Information

  • Patent Grant
  • 10871850
  • Patent Number
    10,871,850
  • Date Filed
    Thursday, December 15, 2016
    8 years ago
  • Date Issued
    Tuesday, December 22, 2020
    4 years ago
Abstract
Multi-touch touch-sensing devices and methods are described herein. The touch sensing devices can include multiple sense points, each located at a crossing of a drive line and a sense line. In some embodiments, multiple drive lines may be simultaneously or nearly simultaneously stimulated with drive signals having unique characteristics, such as phase or frequency. A sense signal can occur on each sense line that can be related to the drive signals by an amount of touch present at sense points corresponding to the stimulated drive lines and the sense line. By using processing techniques based on the unique drive signals, an amount of touch corresponding to each sense point can be extracted from the sense signal. The touch sensing methods and devices can be incorporated into interfaces for a variety of electronic devices such as a desktop, tablet, notebook, and handheld computers, personal digital assistants, media players, and mobile telephones.
Description
BACKGROUND

Recently, interest has developed in touch and/or proximity-based input systems for electronic devices and computer systems that are capable of recognizing multiple touch and/or hover events simultaneously. Many of these systems, for example those based on mutual capacitance or certain optical sensing arrangements, involve applying periodic stimulus waveforms to a plurality of sense points and detecting sense waveforms that can be related to the periodic stimulus waveform by the amount of touch and/or proximity present at the sense point. In some embodiments, these systems apply periodic stimulus waveforms to drive lines that are coupled to sense lines at the sense points. Typically, stimulus waveforms have been applied to these drive lines one at a time. Because devices typically include a plurality of these drive lines, each drive line has been driven sequentially.


SUMMARY

According to one embodiment of the invention, a method of deriving touch information from a touch sensitive surface is provided. The touch sensitive device can include a plurality of sensing points. Each sensing point can be located at or near a crossing of a drive line and a sense line. For example, the method can include simultaneously (or substantially simultaneously) stimulating a plurality of the drive lines with one or more unique drive signals. For example, the signals may have predetermined phase and/or frequency relationships. The method can further include sensing a sense signal on at least one of the sense lines. The sense signal can relate to the drive signals by touch or proximity of one or more objects to one or more sensing points located at or near the crossing of the plurality of drive lines and the at least one sense line. The method can also include, for example, deriving touch information from the sense signal. Touch may be derived from the sense signal by deriving a plurality of values from the sense signal, e.g., by integrating the sense signal over one or more time periods and deriving touch information from a mathematical combination of the plurality of values.


In another embodiment, the invention can relate to a multi-touch sensing device. The touch sensing device can include, for example, a touch sensitive surface with a plurality of sensing points located at a crossings of drive lines and sense lines. The touch sensing device can also include drive circuitry configured to simultaneously apply unique drive signals to a plurality of the drive lines. For example, the signals may have predetermined phase and/or frequency relationships. The touch sensing device can also include sense circuitry that is configured to detect a sense signal in at least one sense line and derive touch information from this sense signal for one or more of the sense points. Such a touch sensing device may be based, for example, on self or mutual capacitance.


In yet another embodiment, the invention can relate to an electronic device or computer system incorporating a touch sensing arrangement or implementing a touch sensing method, as discussed above. The electronic device can take a variety of forms, including, for example, a desktop computer, a tablet computer, a notebook computer, a handheld computer, a personal digital assistant, a media player, or a mobile telephone. Other form factors are also possible.


In still another embodiment, the present invention can relate to a method of deriving touch information from a touch sensitive surface. The method can include performing a coarse scan of a region of the touch sensitive surface to determine whether a touch is present within the first region. If a touch is present, fine scanning of the region may be performed to determine more exact data about the touch or touches that are present within the region. If a touch is not present, fine scanning any be omitted, and a coarse scan of another region may begin. By eliminating unnecessary fine scans, time and power savings can result.





BRIEF DESCRIPTION OF THE DRAWINGS

The aforementioned and other aspects of the invention may best be understood by reference to the following description taken in conjunction with the accompanying drawings in which:



FIG. 1 illustrates a multi-touch sensing device used as an input device to a computer system in accordance with an embodiment of the present invention.



FIG. 2 illustrates a plurality of contact patch areas corresponding to an object in proximity to a plurality of sense points of a multi-touch surface in accordance with an embodiment of the present invention.



FIG. 3 illustrates a simplified schematic diagram of a mutual capacitance sensing circuit that may be used in an embodiment of the present invention.



FIG. 4 illustrates a process for operating a multi-touch sensing device in accordance with an embodiment of the present invention.



FIG. 5 illustrates a multi-touch sensing device in accordance with an embodiment of the present invention.



FIG. 6 illustrates a process for performing multi-line stimulation in accordance with an embodiment of the present invention.



FIG. 7 illustrates a single-line stimulation arrangement according to the prior art.



FIG. 8 illustrates a two-line simultaneous stimulation arrangement in accordance with an embodiment of the present invention.



FIG. 9 illustrates a four-line simultaneous stimulation arrangement in accordance with an embodiment of the present invention.



FIG. 10 illustrates a variety of electronic device and computer system form factors that may be used in accordance with an embodiment of the present invention.



FIG. 11 illustrates a frequency-based four-line simultaneous stimulation arrangement in accordance with an embodiment of the present invention.



FIGS. 12A-E illustrate generally various alternatives for stimulus windows in accordance with an embodiment of the present invention.





DETAILED DESCRIPTION

Recognizing multiple simultaneous or near-simultaneous touch events may be accomplished with a multi-touch sensing arrangement as illustrated in FIG. 1. Multi-touch sensing arrangement 100 can detect and monitor multiple touch attributes (including, for example, identification, position, velocity, size, shape, and magnitude) across touch sensitive surface 101, at the same time, nearly the same time, at different times, or over a period of time. Touch sensitive surface 101 can provide a plurality of sensor points, coordinates, or nodes 102 that function substantially independently of one another and that represent different points on a touch sensitive surface. Sensing points 102 may be positioned in a grid or a pixel array, with each sensing point capable of generating a signal at the same time. Sensing points 102 may be considered as mapping touch sensitive surface 101 into a coordinate system, for example, a Cartesian or polar coordinate system.


A touch sensitive surface may, for example, be in the form of a tablet or a touch screen. To produce a touch screen, the capacitance sensing points and other associated electrical structures can be formed with a substantially transparent conductive medium, such as indium tin oxide (ITO). The number and configuration of sensing points 102 may be varied. The number of sensing points 102 generally depends on the desired resolution and sensitivity. In touch-screen applications, the number of sensing points 102 may also depend on the desired transparency of the touch screen.


Using a multi-touch sensing arrangement, like that described in greater detail below, signals generated at nodes 102 of multi-touch sensor 101 may be used to produce an image of the touches at a particular point in time. For example, each object (e.g., finger, stylus, etc.) in contact with or in proximity to touch sensitive surface 101 can produce contact patch area 201, as illustrated in FIG. 2. Each of contact patch area 201 may cover several nodes 102. Covered nodes 202 may detect the object, while remaining nodes 102 do not. As a result, a pixilated image of the touch surface plane (which may be referred to as a touch image, a multi-touch image, or a proximity image) can be formed. The signals for each contact patch area 201 may be grouped together. Each contact patch area 201 may include high and low points based on the amount of touch at each point. The shape of contact patch area 201, as well as the high and low points within the image, may be used to differentiate contact patch areas 201 that are in close proximity to one another. Furthermore, the current image can be compared to previous images to determine how the objects may be moving over time, and what corresponding action should be performed in a host device as a result thereof.


Many different sensing technologies can be used in conjunction with these sensing arrangements, including resistive, capacitive, optical, etc. In capacitance-based sensing arrangements, as an object approaches touch-sensitive surface 101, a small capacitance forms between the object and sensing points 102 in proximity to the object. By detecting changes in capacitance at each of the sensing points 102 caused by this small capacitance, and by noting the position of the sensing points, a sensing circuit 103 can detect and monitor multiple touches. The capacitive sensing nodes may be based on self capacitance or mutual capacitance.


In self capacitance systems, the “self” capacitance of a sensing point is measured relative to some reference, e.g., ground. Sensing points 102 may be spatially separated electrodes. These electrodes are coupled to driving circuitry 104 and sensing circuitry 103 by conductive traces 105a (drive lines) and 105b (sense lines). In some self-capacitance embodiments, a single conductive trace to each electrode may be used as both a drive and sense line.


In mutual capacitance systems, the “mutual” capacitance between a first electrode and a second electrode can be measured. In mutual capacitance sensing arrangements, the sensing points may be formed by the crossings of patterned conductors forming spatially separated lines. For example, driving lines 105a may be formed on a first layer and sensing lines 105b may be formed on a second layer 105b such that the drive and sense lines cross or “intersect” one another at sensing points 102. The different layers may be different substrates, different sides of the same substrate, or the same side of a substrate with some dielectric separation. Because the drive and sense lines are separated, there is a capacitive coupling node at each “intersection.”


The manner in which the drive and sense lines are arranged may vary. For example, in a Cartesian coordinate system (as illustrated), the drive lines may be formed as horizontal rows, while the sense lines may be formed as vertical columns (or vice versa), thus forming a plurality of nodes that may be considered as having distinct x and y coordinates. Alternatively, in a polar coordinate system, the sense lines may be a plurality of concentric circles with the drive lines being radially extending lines (or vice versa), thus forming a plurality of nodes that may be considered as having distinct r and angle coordinates. In either case, drive lines 105a may be connected to drive circuit 104, and sensing lines 105b may be connected to sensing circuit 103.


During operation, a drive signal (e.g., a periodic voltage) is applied to each drive line 105a. When driven, the charge impressed on drive line 105a can capacitively couple to the intersecting sense lines 105b through nodes 102. This can cause a detectable, measurable current and/or voltage in sense lines 105b. The relationship between the drive signal and the signal appearing on sense lines 105b is a function of the capacitance coupling the drive and sense lines, which, as noted above, may be affected by an object in proximity to node 102. Capacitance sensing circuit (or circuits) 103 may sense sensing lines 105b and may determine the capacitance at each node as described in greater detail below.


As discussed above, conventional drive lines 105a were driven one at a time, while the other drive lines were grounded. This process was repeated for each drive line 105a until all the drive lines had been driven, and a touch image (based on capacitance) was built from the sensed results. Once all the lines 105a had been driven, the sequence would repeat to build a series of touch images. However, in some embodiments of the present invention, multiple drive lines may be driven simultaneously or nearly simultaneously, as described, for example, below. As used herein, “simultaneously” encompasses precisely simultaneous as well as nearly simultaneous events. For example, simultaneous events may begin at about the same time, end at about the same time, and/or take place over at least partially overlapping time periods.



FIG. 3 illustrates a simplified schematic diagram of mutual capacitance circuit 300 corresponding to the arrangement described above. Mutual capacitance circuit 300 may include drive line 105a and sense line 105b, which are spatially separated thereby forming capacitive coupling node 102. Drive line 105a may be electrically (i.e., conductively) coupled to drive circuit 104 represented by voltage source 301. Sense line 105b may be electrically coupled to capacitive sensing circuit 103. Both drive line 105aand sense line 105b may, in some cases, include some parasitic capacitance 302.


As noted above, in the absence of a conductive object proximate the intersection of drive line 105a and sense line 105b, the capacitive coupling at node 102 stays fairly constant. However, if an electrically conductive object (for example, a user's finger, stylus, etc.) comes in proximity to node 102, the capacitive coupling (i.e., the capacitance of the local system) changes. The change in capacitive coupling changes the current (and/or voltage) carried by sense line 105b. Capacitance sensing circuit 103 may note the capacitance change and the position of node 102 and report this information in some form to processor 106 (FIG. 1).


With reference to FIG. 1, sensing circuit 103 may acquire data from touch surface 101 and supply the acquired data to processor 106. In some embodiments, sensing circuit 103 may be configured to send raw data (e.g., an array of capacitance values corresponding to each sense point 102) to processor 106. In other embodiments, sensing circuit 103 may be configured to process the raw data itself and deliver processed touch data to processor 106. In either case, the processor may then use the data it receives to control operation of computer system 107 and/or one or more applications running thereon. Various implementations along these lines are described in the applications referenced above, and include a variety of computer systems having touch pads and touch screens.


In some embodiments, sensing circuit 103 may include one or more microcontrollers, each of which may monitor one or more sensing points 102. The microcontrollers may be application specific integrated circuits (ASICs), that work with firmware to monitor the signals from touch sensitive surface 101, process the monitored signals, and report this information to processor 106. The microcontrollers may also be digital signal processors (DSPs). In some embodiments, sensing circuit 103 may include one or more sensor ICs that measure the capacitance in each sensing line 105b and report measured values to processor 106 or to a host controller (not shown) in computer system 107. Any number of sensor ICs may be used. For example, a sensor IC may be used for all lines, or multiple sensor ICs may be used for a single line or group of lines.



FIG. 4 illustrates at a high level process 400 for operating a multi-touch sensing arrangement, like that described above. The process may begin at block 401 where plurality of sensing points 102 are driven. Following block 401, the process flow can proceed to block 402, where the outputs from sensing points 102 are read. For example, a capacitance value for each sensing point 102 can be obtained. Following block 402, the process can proceed to block 403 where an image or other form of data (signal or signals) of the touch at one moment in time can be produced and thereafter analyzed to determine where objects touching or in proximity to the touch sensor may be located. Following block 403, the process can proceed to block 404, where the current image or signal may be compared to one or more past images or signals to determine a change in one or more of the shape, size, location, direction, speed, acceleration, pressure, etc. for each object. This information can be subsequently used (in step 405) to perform an action in computer system 107, ranging from moving a pointer or cursor to complex gesture-based interactions.


As noted above, enhanced operation of multi-touch sensing arrangements may be achieved by driving multiple rows simultaneously. An example multi-touch sensing device 500 with which multi-row stimulation may be employed is illustrated in FIG. 5 and has reference numerals generally corresponding to sensing arrangement 101 illustrated in FIG. 1. In the given example, touch sensitive surface 501 has sixteen drive rows 505a, although any number of drive rows could be used. The drive rows may be divided, for example, into four groups, e.g., Group 1, Group 2, Group 3, and Group 4, each including four drive rows 505a. Other numbers of groups and numbers of rows per group may also be employed.


Scanning of multi-touch sensor arrays is described in various references, including U.S. patent application Ser. No. 11/381,313, which is hereby incorporated by reference. The process may be briefly summarized by reference to FIG. 7. In general, a periodic waveform (e.g., a series of square wave pulses) is applied sequentially to drive rows 505a. For example, a first group of pulses 701 may be applied to Row A, followed by a second group of pulses 702 applied to row B, followed by a third group of pulses 703 applied to Row C, followed by a fourth group of pulses 704 applied to Row D. These periodic waveforms are capacitively coupled into sense lines 505b, generally resulting in sensed waveform 705 on each sense line. Sensed waveform 705 can be input integrated over a predetermined time period to measure the capacitive coupling between drive lines 505a and sense lines 505b at each node 502. Other processing, such as filtering, demodulation, etc., may also occur.


In sensed waveform 7051 the time period from to t0 t1 corresponds to the stimulus of Row A. Integrating the sensed waveform over this time period results in a numerical value X1, which can correspond to the capacitance of a node at the intersection of Row A and the column being sensed. Similarly, the time period from t1 to t2 corresponds to Row B, the time period from t2 to t3 corresponds to Row C, and the time period from t3 to t4 corresponds to Row D. Integrating the sensed waveform over each of these time periods can give numerical values X2, X3, and X4 corresponding to the capacitance of nodes at the intersection of the Rows B, C, and D and the column being sensed.



FIG. 8 illustrates a multi-line stimulation arrangement in which two lines may be stimulated simultaneously. Specifically, Row A and Row B may be simultaneously (or nearly simultaneously) stimulated with waveforms 801 and 802. As can be seen, the phase of waveform 801 may be adjusted, e.g., after the fourth pulse. As a result, the remaining pulses of waveforms 801 and 802 may be 180° out of phase. Similarly, Row C and Row D may be simultaneously (or nearly simultaneously) stimulated with waveforms 806 and 807. Again, the phase of waveform 806 may be adjusted, e.g., after the fourth pulse. As a result, the remaining pulses of waveforms 806 and 807 may be 180° out of phase. The phase relationships between the waveforms may be easily understood with reference to the + and − signs above.


Unlike the example described above with reference to FIG. 7, the time periods t0 to t1, t1 to t2, t2 to t3, and t3 to t4 may no longer uniquely correspond to Rows A-D. Likewise, the measured values X1, X2, X3, and X4 resulting from integrating the sensed waveform (not shown) over these time periods no longer uniquely correspond to the capacitance of a node at the intersection of a particular row and the column being sensed. In the example of FIG. 8, time periods t0 to t1 and t1 to t2, along with their corresponding measured values X1 and X2, correspond together to both Row A and Row B. Similarly, time periods t2 to t3 and t3 to t4, along with their corresponding measured values X3 and X4, correspond together to both Row C and Row D. However, because of the phase difference between the stimulus waveforms 801 and 802, the effects of stimulating Row A only and Row B only can be isolated from each other. Similarly, because of the phase difference between the stimulus waveforms 806 and 807, the effects of stimulating Row C only and Row D only can be isolated from each other.


Specifically, because stimulus waveforms 801 and 802 are in phase over time period t0 to t1 and out of phase over time period t1 to t2 the integration result X1 plus the integration result X2 (i.e., X1+X2) can yield a value corresponding to the capacitance of a node at the intersection of Row A and the line being sensed, i.e., only the effect of the stimulus of Row A. Similarly, the integration result X1 minus the integration result X2 (i.e., X1−X2) can yield a value corresponding to the capacitance of a node at the intersection of Row B and the line being sensed, i.e., only the effect of the stimulus of Row B.


The same applies to Rows C and D. Because stimulus waveforms 806 and 807 are in phase over time period t2 to t3 and out of phase over time period t3 to t4, the integration result X3 plus the integration result X4 (i.e., X3+X4) can yield a value corresponding to the capacitance of a node at the intersection of Row C and the line being sensed, i.e., only the effect of the stimulus of Row C. Similarly, the integration result X3 minus the demodulation result X4 (i.e., X3−X4) can yield a value corresponding to the capacitance of a node at the intersection of Row D and the line being sensed, i.e., only the effect of the stimulus of Row D.



FIG. 9 illustrates an exemplary multi-line stimulation arrangement in which four lines are stimulated simultaneously. During a first period from t0 to t1, the periodic waveforms 901-904 applied to Rows A-D are in phase. At time t1, e.g., after the fourth pulse, the phase of waveforms 901 and 902 can be adjusted so that waveforms 901 and 902 are 180° out of phase with respect to waveforms 903 and 904. Similarly at time t2, e.g., after the next four pulses, the phases of waveforms 901 and 903 are adjusted. This can result in waveforms 901 and 904 being 180° out of phase with respect to waveforms 902 and 903. Finally, at time t3, e.g., after four more pulses, the phase of waveforms 901 and 902 can again be adjusted. This can result in waveforms 901 and 903 being 180° out of phase with respect to waveforms 902 and 904. The phase relationships between the various waveforms can be understood with reference to the + and − signs in the figure.


As in the preceding example, the phase relationships between waveforms 901-904 allow the effects of the stimulus on each individual row to be isolated as mathematical combinations of the measured integration results. Specifically, the effects of the stimulus of Row A can be determined by the expression X1+X2+X3+X4. The effects of the stimulus of Row B can be determined by summing X1+X2−X3−X4. The effects of the stimulus of Row C can be determined by X1−X2−X3+X4. The effects of the stimulus of row D can be determined by X1−X2+X3−X4.


The process of operation 600 of the multi-touch sensing arrangement 500 with multi-row stimulation may be further understood with reference to the flow chart of FIG. 6. First, the DC content for a group of rows can be obtained 601. In the example of FIG. 8, the DC content of the group of rows including Rows A and B can be obtained by integrating the sensed waveform over time period t0 to t1. This DC content can indicate (somewhat coarsely) whether there is any touch present within a given group, e.g., Rows A and B, corresponding to a particular region of touch surface 501. If at decision block 602, it is determined that there is no touch within a given group/region the next group (e.g., Rows C and D) can be similarly scanned. This may be implemented, for example, by implementing counter 603. If at decision block 602, it is determined that there is touch within a given group/region as indicated by the DC content of the group, a fine scan of the group is performed in block 604. The results of the fine scan may be combined to extract the signals corresponding to each row as described above. Once all groups have been scanned, the process repeats.


Turning back to FIGS. 5 and 6, each group of rows may be scanned according to the principles described in the foregoing paragraphs. A multi-touch sensing arrangement may comprise any number of rows and any number of groups. In some embodiments, a multi-touch sensing arrangement can employ multi-line stimulation on a single group, i.e., all lines of the device may be stimulated simultaneously or nearly simultaneously. Additionally, although described in terms of rows and columns, the drive and sense lines may be arranged in any geometric arrangement.


Multi-line stimulation as described above may provide a number of advantages. For example, when multiple rows are stimulated simultaneously, the stimulus voltage can be reduced. Specifically, the additive effect of multiple-row stimulus can result in the same sensed waveform amplitude for a lower “per row” stimulus voltage. For example, a single-line scanning arrangement using an 18Vpp (volts peak-to-peak) stimulus voltage could use a 9Vpp stimulus voltage with two lines being simultaneously stimulated or with a 4.5Vpp stimulus voltage with four lines being simultaneously stimulated, etc., to obtain similar sensed waveform amplitude.


Reducing the stimulus voltage can allow drive signals to be supplied directly from a driver chip without requiring a high voltage booster. The reduced voltage can also help to avoid fringe field and transistor breakdown issues. The reduced voltage can also result in reduced power consumption. Because power scales as square of voltage, cutting voltage by a factor of four (for four row simultaneous stimulation) cuts the power per row by a factor of 16. However, because there are four rows being driven, the actual power savings may only be a factor of 4. However, additional power may also be saved by not doing a fine-scan when there is no touch detected during a DC scan of the region, as described above.


A variation on the multi-row stimulation techniques described above may be referred to as differential multi-row stimulation. Differential multi-row stimulation may be understood with reference to the table below, which shows the polarities of the stimulating waveforms for the multi-row stimulation example of FIG. 9 above (Multi-Row Stimulation) compared to the polarities of the stimulating waveforms for a differential multi-row stimulation example (Differential Multi-Row Stimulation). Differential multi-row stimulation may generally correspond to the multi-row stimulation examples discussed above, except that the polarities of the stimulating waveforms may be rearranged as indicated below.












Stimulus Voltage Phase Comparison



















Row



Row
X1
X2
X3
X4
Sum











Multi−Row Stimulation














A
+ 
+
+
+
+4  



B
+ 
+


0



C
+ 


+
0



D
+ 

+

0



Col.
+4
0
0
0




Sum












Differential Multi−Row Stimulation














A
− 
+

+
0



B
+ 
+


0



C
+ 

+

0



D
− 

+
+
0



Col.
  0
0
0
0




Sum










As can be seen from the table, in the multi-row stimulation example, the net polarity applied across row A can have a DC component of four times the amplitude of the stimulus waveform. Similarly, the first time period (during which the value X1 may be measured) also can have a net DC component of four times the amplitude of the stimulus waveform. In the differential multi-row stimulation example, the polarities may be rearranged such that no row nor time period has a DC component. The absence of a DC component can result in a number of advantages, including allowing the charge sensing circuitry to operate with zero offset, automatic baseline removal, inherent centroids computation, and increased signal to noise ratio. In some embodiments, it may be desired to stimulate all rows of the sensor simultaneously (i.e., have only one group of rows) as differences in DC offset from one group to another may be lost because of an absence of DC content in the stimulus.


Other variations of the multi-row stimulation concept include phase or frequency-based multi-row stimulation. In the foregoing examples, different rows can be stimulated with waveforms having polarity (phase) differences so that effects of a given row may be isolated in the aggregate sense waveform. Another way to allow this type of isolation, illustrated in FIG. 11, is to stimulate each row of a group with a stimulus waveform having a different frequency. One or more demodulation circuits can then separate these frequencies in the sense waveform so that the contributions of each stimulated line may be isolated. In many embodiments demodulator circuits may already be present for noise filtering.


Examples of other possibilities for stimulus waveforms according to the principles described herein may be understood with reference to FIG. 12. Timing signal 1201 can define a stimulus time window. Each line can have a corresponding timing signal. During the stimulus time window, e.g., when signal 1201 is high, a stimulation waveform can be applied to the corresponding line (or lines). This stimulation waveform may take a variety of forms, including a square wave, a sine wave, an exponentially decaying sine wave, a pulsed sine wave, etc.


Simultaneous stimulation, as used herein, means that at least one stimulus is applied to at least two lines during the same time period (e.g., window, which can include one or more pulses of any shape and in any combination). In other words, simultaneous stimulation involves at least two lines having stimulus windows that at least partially overlap in time. For example, in FIG. 12B, a stimulus window for Row A, defined by timing signal 1201a can begin at the same time, extend for the same duration, and end at the same time as a stimulus window for Row B, defined by timing signal 1201b. Alternatively, as illustrated in FIG. 12C, the stimulus windows for Row A (defined by timing signal 1201a) and Row B (defined by timing signal 1201b) may begin and end at different times, but have at least some overlapping portion. Another alternative, illustrated in FIG. 12D, is for stimulus windows for Row A (defined by timing signal 1201a) and Row B (defined by timing signal 1201b) may begin at the same time, but end at different times. Still another alternative, illustrated in FIG. 12E, is for stimulus windows for Row A (defined by timing signal 1201a) and Row B (defined by timing signal 1201b) to begin at different times but end at the same time. These various arrangements can also be extended to a number of rows greater than two, with complete flexibility so long as there is some stimulation overlap between at least some rows.


The principles described herein may be used to devise input devices for a variety of electronic devices and computer systems. These electronic devices and computer system may be any of a variety of types illustrated in FIG. 10, including desktop computers 1001, notebook computers 1002, tablet computers 1003, handheld computers 1004, personal digital assistants 1005, media players 1006, mobile telephones 1007, and the like. Additionally, the electronic devices and computer systems may be combinations of these types, for example, a device that is a combination of a personal digital assistant, media player, and mobile telephone.


Other alternations, permutations, and combinations of the aforementioned embodiments are also possible. For example, multiple touch and proximity systems may be designed based on infrared/optical sensing arrangements that rely on periodic waveform stimulus and reflections from hand parts or other touch objects to detect touch and/or hover events. The principles herein, though described with reference to capacitive systems, are equally applicable to any systems in which touch or proximity sensing depends on information extracted from periodic stimulus waveforms. It is therefore intended that the following claims be interpreted as including all alterations, permutations, combinations and equivalents of the foregoing.

Claims
  • 1. A touch sensing device comprising: drive circuitry configured to: apply a first drive signal to a first drive line of a plurality of drive lines during a first stimulus window; andapply a second drive signal to a second drive line of the plurality of drive lines during a second stimulus window, the second stimulus window at least partially overlapping the first stimulus window during an overlapping stimulus window, wherein the first and second drive signals have a predetermined frequency relationship and are in phase during a first portion of the overlapping stimulus window and out of phase during a second portion of the overlapping stimulus window; andsense circuitry configured to: detect a sense signal from at least one sense line, the sense signal being related to the first and second drive signals by touch or proximity of one or more objects to one or more sensing points associated with the at least one sense line, a sensing point being associated with at least one of the plurality of drive lines and at least one of a plurality of sense lines; andderive touch information for the one or more sensing points from the sense signal.
  • 2. The touch sensing device of claim 1, wherein the sense circuitry comprises at least one microcontroller.
  • 3. The touch sensing device of claim 2, wherein the at least one microcontroller is an application specific integrated circuit (ASIC).
  • 4. The touch sensing device of claim 2, wherein the at least one microcontroller is a digital signal processor (DSP).
  • 5. The touch sensing device of claim 1, wherein the sense circuitry derives touch information from the sense signal by deriving a plurality of values from the sense signal and deriving touch information from a mathematical combination of the plurality of values.
  • 6. The touch sensing device of claim 5, wherein the sense circuitry derives a plurality of values from the sense signal by integrating the sense signal over time.
  • 7. The touch sensing device of claim 5, wherein the plurality of values correspond to a capacitance measurement at a plurality of sensing points.
  • 8. The touch sensing device of claim 1, the drive circuitry further configured to: apply a third drive signal to a third drive line of the plurality of drive lines during a third stimulus window; andapply a fourth drive signal to a fourth drive line of the plurality of drive lines during a fourth stimulus window, the first, second, third and fourth stimulus windows at least partially overlapping during the overlapping stimulus window;wherein the first, second, third and fourth drive signals have the predetermined frequency relationship and two of the first, second, third and fourth drive signals are out of phase with another two of the first, second, third and fourth drive signal during a third portion of the overlapping stimulus window.
  • 9. The touch sensing device of claim 1, wherein a phase relationship between the first and second drive lines during the overlapping stimulus window is selected to eliminate a DC component of the sense signal.
  • 10. The touch sensing device of claim 1, wherein the plurality of drive lines and the plurality of sense lines are part of a touch screen.
  • 11. A method of stimulating a touch sensitive surface, the touch sensitive surface comprising a plurality of sensing points, a sensing point being associated with at least one of a plurality of drive lines and at least one of a plurality of sense lines, the method comprising: stimulating a first drive line of the plurality of drive lines with a first drive signal during a first stimulus window; andstimulating a second drive line of the plurality of drive lines with a second drive signal during a second stimulus window, the second stimulus window at least partially overlapping the first stimulus window during an overlapping stimulus window;wherein the first and second drive signals have a predetermined frequency relationship and are in phase during a first portion of the overlapping stimulus window and out of phase during a second portion of the overlapping stimulus window.
  • 12. The method of claim 11, the method further comprising: sensing a sense signal on at least one sense line, wherein the sense signal is related to the first and second drive signals by touch or proximity of one or more objects to one or more sensing points associated with at least one of the first and second drive lines and the at least one sense line; andderiving touch information from the sense signal;wherein deriving touch information from the sense signal comprises: deriving a plurality of values from the sense signal; andderiving touch information from a mathematical combination of the plurality of values.
  • 13. The method of claim 12, wherein deriving touch information from the sense signal comprises integrating the sense signal over time.
  • 14. The method of claim 12, wherein the plurality of values from the sense signal include a capacitance measurement.
  • 15. The method of claim 11, further comprising: stimulating a third drive line of the plurality of drive lines with a third drive signal during a third stimulus window; andstimulating a fourth drive line of the plurality of drive lines with a fourth drive signal during a fourth stimulus window, the first, second, third and fourth stimulus windows at least partially overlapping during the overlapping stimulus window;wherein the first, second, third and fourth drive signals have the predetermined frequency relationship and two of the first, second, third and fourth drive signals are out of phase with another two of the first, second, third and fourth drive signals during a third portion of the overlapping stimulus window.
  • 16. The method of claim 15, wherein the first, second, third and fourth drive signals are in phase during a fourth portion of the overlapping stimulus window.
  • 17. The method of claim 11, wherein a predetermined phase relationship of the first and second drive signals is selected to eliminate a DC component of a sense signal.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application is a continuation of U.S. patent application Ser. No. 14/482,979, filed Sep. 10, 2014 and published on Dec. 25, 2014 as U.S. Publication No. 2014-0375612, which is a continuation of U.S. patent application Ser. No. 14/019,264, filed Sep. 5, 2013, now U.S. Pat. No. 8,928,617, issued Jan. 6, 2015, which is a continuation of U.S. patent application Ser. No. 12/874,184, filed Sep. 1, 2010, now U.S. Pat. No. 8,552,998, issued Oct. 8, 2013, which is a division of U.S. patent application Ser. No. 11/619,433, filed Jan. 3, 2007, now U.S. Pat. No. 7,812,827, issued Oct. 12, 2010, the entire disclosures of which are also incorporated herein by reference. The present application is also related to the following U.S. Patents and Patent Applications, each of which is hereby incorporated by reference in its entirety: U.S. patent application Ser. No. 09/236,513, titled “Method and Apparatus for Integrating Manual Input,” filed Jan. 25, 1999, now U.S. Pat. No. 6,323,846, issued Nov. 27, 2001;U.S. patent application Ser. No. 10/840,862, titled “Multipoint Touchscreen,” filed May 6, 2004, now U.S. Pat. No. 7,663,607, issued Feb. 16, 2010; andU.S. patent application Ser. No. 11/381,313, titled “Multipoint Touch Surface Controller,” filed May 2, 2006, now U.S. Pat. No. 8,279,180, issued Oct. 2, 2012.

US Referenced Citations (284)
Number Name Date Kind
3342935 Leifer et al. Sep 1967 A
3732369 Cotter May 1973 A
3767858 Rodgers Oct 1973 A
3818133 Cotter Jun 1974 A
3875472 Schermerhorn Apr 1975 A
3886539 Gould, Jr. May 1975 A
4071691 Pepper, Jr. Jan 1978 A
4080515 Anderson Mar 1978 A
4103252 Bobick Jul 1978 A
4129747 Pepper, Jr. Dec 1978 A
4293734 Pepper, Jr. Oct 1981 A
4444998 House Apr 1984 A
4550221 Mabusth Oct 1985 A
4560830 Perl Dec 1985 A
4680429 Murdock et al. Jul 1987 A
4698460 Krein et al. Oct 1987 A
4698461 Meadows et al. Oct 1987 A
4733222 Evans Mar 1988 A
4853498 Meadows et al. Aug 1989 A
4916308 Meadows Apr 1990 A
4922061 Meadows et al. May 1990 A
5105186 May Apr 1992 A
5218174 Gray et al. Jun 1993 A
5270711 Knapp Dec 1993 A
5305017 Gerpheide Apr 1994 A
5355149 Casebolt Oct 1994 A
5357266 Tagawa Oct 1994 A
5475711 Betts et al. Dec 1995 A
5483261 Yasutake Jan 1996 A
5488204 Mead et al. Jan 1996 A
5495077 Miller et al. Feb 1996 A
5526294 Ono et al. Jun 1996 A
5543590 Gillespie et al. Aug 1996 A
5565658 Gerpheide et al. Oct 1996 A
5606346 Kai Feb 1997 A
5621425 Hoshino et al. Apr 1997 A
5634207 Yamaji et al. May 1997 A
5650597 Redmayne Jul 1997 A
5650801 Higashi Jul 1997 A
5686705 Conroy et al. Nov 1997 A
5691512 Obi Nov 1997 A
5766463 Janik et al. Jun 1998 A
5790106 Hirano et al. Aug 1998 A
5790107 Kasser et al. Aug 1998 A
5816225 Koch et al. Oct 1998 A
5825352 Bisset et al. Oct 1998 A
5831600 Inoue et al. Nov 1998 A
5835079 Shieh Nov 1998 A
5838308 Knapp et al. Nov 1998 A
5861583 Schediwy et al. Jan 1999 A
5880411 Gillespie et al. Mar 1999 A
5920309 Bisset et al. Jul 1999 A
5930309 Knutson et al. Jul 1999 A
5945980 Moissev et al. Aug 1999 A
5996082 Cortopassi Nov 1999 A
6025726 Gershenfeld Feb 2000 A
6043810 Kim et al. Mar 2000 A
6057903 Colgan et al. May 2000 A
6075520 Inoue et al. Jun 2000 A
6177918 Colgan et al. Jan 2001 B1
6188391 Seely et al. Feb 2001 B1
6204897 Colgan et al. Mar 2001 B1
6222528 Gerpheide Apr 2001 B1
6229856 Diab et al. May 2001 B1
6246729 Richardson Jun 2001 B1
6310610 Beaton et al. Oct 2001 B1
6323846 Westerman et al. Nov 2001 B1
6327011 Kim Dec 2001 B2
6380931 Gillepsie et al. Apr 2002 B1
6424094 Feldman Jul 2002 B1
6429857 Masters et al. Aug 2002 B1
6452514 Philipp Sep 2002 B1
6483498 Colgan et al. Nov 2002 B1
6492979 Kent et al. Dec 2002 B1
6501529 Kurihara et al. Dec 2002 B1
6559658 Brandt May 2003 B1
6583676 Krah et al. Jun 2003 B2
6621484 Yee Sep 2003 B1
6658245 Li et al. Dec 2003 B2
6667740 Ely et al. Dec 2003 B2
6680448 Kawashima et al. Jan 2004 B2
6690387 Zimmerman et al. Feb 2004 B2
6730863 Gerpheide et al. May 2004 B1
6816750 Klass Nov 2004 B1
7015894 Morohoshi Mar 2006 B2
7030782 Ely et al. Apr 2006 B2
7031886 Hargreaves Apr 2006 B1
7042444 Cok May 2006 B2
7050046 Park et al. May 2006 B1
7129714 Baxter Oct 2006 B2
7129935 Mackey Oct 2006 B2
7129939 Toyozawa et al. Oct 2006 B2
7133032 Cok Nov 2006 B2
7184064 Zimmerman et al. Feb 2007 B2
7202856 Cok Apr 2007 B2
7230608 Cok Jun 2007 B2
7230609 Chao et al. Jun 2007 B2
7248625 Chien Jul 2007 B2
7277087 Hill et al. Oct 2007 B2
7280167 Choi et al. Oct 2007 B2
7289142 Silverbrook Oct 2007 B2
7339499 Khlat Mar 2008 B2
7362313 Geaghan et al. Apr 2008 B2
7372455 Perski et al. May 2008 B2
7379054 Lee May 2008 B2
7436393 Hong et al. Oct 2008 B2
7643011 O'Connor et al. Jan 2010 B2
7649524 Haim et al. Jan 2010 B2
7663607 Hotelling et al. Feb 2010 B2
7734261 Bury Jun 2010 B2
7812827 Hotelling Oct 2010 B2
7859522 Takahashi et al. Dec 2010 B2
7876311 Krah et al. Jan 2011 B2
7986193 Krah Jul 2011 B2
8026904 Westerman Sep 2011 B2
8120591 Krah et al. Feb 2012 B2
8144125 Peng Mar 2012 B2
8144126 Wright Mar 2012 B2
8175549 Faust et al. May 2012 B2
8232970 Krah et al. Jul 2012 B2
8237667 Krah Aug 2012 B2
8479122 Hotelling et al. Jul 2013 B2
8493331 Krah et al. Jul 2013 B2
8508244 Seguine Aug 2013 B2
8514185 Hotelling Aug 2013 B2
8552998 Hotelling Oct 2013 B2
8592697 Hotelling et al. Nov 2013 B2
8593423 Hotelling et al. Nov 2013 B2
8659556 Wilson Feb 2014 B2
8659568 Krah et al. Feb 2014 B2
8754867 Krah et al. Jun 2014 B2
8791920 Krah Jul 2014 B2
8928617 Hotelling Jan 2015 B2
8976124 Wright Mar 2015 B1
8988390 Krah et al. Mar 2015 B1
9036650 Wilson et al. May 2015 B2
9069408 Hotelling et al. Jun 2015 B2
9086750 Krah Jul 2015 B2
9092086 Krah et al. Jul 2015 B2
9348451 Wilson et al. May 2016 B2
9430087 Krah et al. Aug 2016 B2
9483141 Hotelling et al. Nov 2016 B2
9552115 Hotelling Jan 2017 B2
9606663 Yousefpor et al. Mar 2017 B2
9715306 Hotelling et al. Jul 2017 B2
9990084 Krah et al. Jun 2018 B2
10042396 Gupta et al. Aug 2018 B1
10042472 Hotelling et al. Aug 2018 B2
10042476 Wilson et al. Aug 2018 B2
10139890 Wilson et al. Nov 2018 B2
20020015024 Westerman et al. Feb 2002 A1
20020067348 Masters et al. Jun 2002 A1
20020067845 Griffis Jun 2002 A1
20020084992 Agnew Jul 2002 A1
20020136269 Kurabe et al. Sep 2002 A1
20020140689 Huang et al. Oct 2002 A1
20020167488 Hinckley et al. Nov 2002 A1
20020196066 Krah et al. Dec 2002 A1
20030025676 Cappendijk Feb 2003 A1
20030048261 Yamamoto et al. Mar 2003 A1
20030063073 Geaghan et al. Apr 2003 A1
20030067447 Geaghan et al. Apr 2003 A1
20030132922 Philipp Jul 2003 A1
20030197691 Fujiwara et al. Oct 2003 A1
20040056845 Harkcom et al. Mar 2004 A1
20040081339 Benkley, III Apr 2004 A1
20040109097 Mai Jun 2004 A1
20040141096 Mai Jul 2004 A1
20040151235 Olson et al. Aug 2004 A1
20040183787 Geaghan et al. Sep 2004 A1
20040183833 Chua Sep 2004 A1
20040189587 Jung et al. Sep 2004 A1
20040227743 Brown Nov 2004 A1
20050052427 Wu et al. Mar 2005 A1
20050052582 Mai Mar 2005 A1
20050094038 Choi et al. May 2005 A1
20050104867 Westerman et al. May 2005 A1
20050146512 Hill et al. Jul 2005 A1
20050146513 Hill et al. Jul 2005 A1
20050151727 Kwong Jul 2005 A1
20050231487 Ming Oct 2005 A1
20050243023 Reddy et al. Nov 2005 A1
20060007165 Yang et al. Jan 2006 A1
20060022955 Kennedy Feb 2006 A1
20060022959 Geaghan Feb 2006 A1
20060026521 Hotelling et al. Feb 2006 A1
20060097991 Hotelling et al. May 2006 A1
20060114247 Brown Jun 2006 A1
20060114650 Wang et al. Jun 2006 A1
20060132462 Geaghan Jun 2006 A1
20060145365 Halls et al. Jul 2006 A1
20060146033 Chen et al. Jul 2006 A1
20060146034 Chen et al. Jul 2006 A1
20060161870 Hotelling et al. Jul 2006 A1
20060197750 Kerr et al. Sep 2006 A1
20060197753 Hotelling Sep 2006 A1
20060244733 Geaghan Nov 2006 A1
20060244736 Tseng Nov 2006 A1
20060279548 Geaghan Dec 2006 A1
20060284856 Soss Dec 2006 A1
20070018969 Chen et al. Jan 2007 A1
20070062739 Philipp et al. Mar 2007 A1
20070075977 Chen et al. Apr 2007 A1
20070109274 Reynolds May 2007 A1
20070176905 Shih et al. Aug 2007 A1
20070216657 Konicek Sep 2007 A1
20070229468 Peng et al. Oct 2007 A1
20070257890 Hotelling et al. Nov 2007 A1
20070262967 Rho Nov 2007 A1
20070268272 Perski et al. Nov 2007 A1
20070273663 Park et al. Nov 2007 A1
20070274411 Lee et al. Nov 2007 A1
20080006453 Hotelling Jan 2008 A1
20080012835 Rimon et al. Jan 2008 A1
20080018618 Hill et al. Jan 2008 A1
20080042964 Sako et al. Feb 2008 A1
20080048989 Yoon et al. Feb 2008 A1
20080048994 Lee et al. Feb 2008 A1
20080055221 Yabuta et al. Mar 2008 A1
20080055268 Yoo et al. Mar 2008 A1
20080062147 Hotelling et al. Mar 2008 A1
20080067528 Choi et al. Mar 2008 A1
20080074401 Chung et al. Mar 2008 A1
20080079697 Lee et al. Apr 2008 A1
20080088594 Liu et al. Apr 2008 A1
20080129898 Moon Jun 2008 A1
20080136980 Rho et al. Jun 2008 A1
20080143683 Hotelling Jun 2008 A1
20080150901 Lowles et al. Jun 2008 A1
20080156546 Hauck Jul 2008 A1
20080157867 Krah Jul 2008 A1
20080157882 Krah Jul 2008 A1
20080157893 Krah Jul 2008 A1
20080158167 Hotelling et al. Jul 2008 A1
20080158169 O'Connor et al. Jul 2008 A1
20080158172 Hotelling et al. Jul 2008 A1
20080158175 Hotelling et al. Jul 2008 A1
20080158180 Krah et al. Jul 2008 A1
20080158184 Land et al. Jul 2008 A1
20080162996 Krah et al. Jul 2008 A1
20080162997 Vu et al. Jul 2008 A1
20080165203 Pantfoerder Jul 2008 A1
20080278143 Cox et al. Nov 2008 A1
20080309625 Krah et al. Dec 2008 A1
20080309628 Krah et al. Dec 2008 A1
20090009483 Hotelling et al. Jan 2009 A1
20090189867 Krah et al. Jul 2009 A1
20090278479 Platner et al. Nov 2009 A1
20090283340 Liu et al. Nov 2009 A1
20090314621 Hotelling Dec 2009 A1
20090315840 Park et al. Dec 2009 A1
20100059295 Hotelling et al. Mar 2010 A1
20100060589 Wilson Mar 2010 A1
20100060590 Wilson et al. Mar 2010 A1
20100060591 Krah Mar 2010 A1
20100060593 Krah Mar 2010 A1
20100060608 Yousefpor Mar 2010 A1
20100214232 Chan et al. Aug 2010 A1
20100328265 Hotelling et al. Dec 2010 A1
20110025634 Krah et al. Feb 2011 A1
20110042152 Wu Feb 2011 A1
20110063993 Wilson et al. Mar 2011 A1
20110084857 Marino et al. Apr 2011 A1
20120019467 Hotelling et al. Jan 2012 A1
20120044194 Peng et al. Feb 2012 A1
20120280932 Krah et al. Nov 2012 A1
20120299880 Krah Nov 2012 A1
20130271410 Westerman et al. Oct 2013 A1
20140022203 Karpin et al. Jan 2014 A1
20140043293 Hotelling et al. Feb 2014 A1
20140092063 Krah Apr 2014 A1
20140168143 Hotelling et al. Jun 2014 A1
20140240287 Krah et al. Aug 2014 A1
20140306913 Krah Oct 2014 A1
20140375612 Hotelling et al. Dec 2014 A1
20150234535 Hotelling et al. Aug 2015 A1
20150261285 Wilson et al. Sep 2015 A1
20150301681 Krah et al. Oct 2015 A1
20160266718 Wilson et al. Sep 2016 A1
20160364078 Krah et al. Dec 2016 A1
20170010744 Hotelling et al. Jan 2017 A1
20170322669 Hotelling et al. Nov 2017 A1
20180275820 Krah et al. Sep 2018 A1
20180348957 Wilson et al. Dec 2018 A1
Foreign Referenced Citations (28)
Number Date Country
1175315 Mar 1998 CN
1254902 May 2000 CN
1773442 May 2006 CN
1914585 Feb 2007 CN
0 818 751 Jan 1998 EP
1 387 242 Feb 2004 EP
1 387 242 Feb 2004 EP
2 453 341 May 2012 EP
1 440 130 Jun 1976 GB
2 451 973 Feb 2009 GB
2 451 973 Feb 2009 GB
2000-163031 Jun 2000 JP
2002-342033 Nov 2002 JP
1998-0010726 Apr 1998 KR
WO-9618179 Jun 1996 WO
WO-9802964 Jan 1998 WO
WO-9807127 Feb 1998 WO
WO-2004099964 Nov 2004 WO
WO-2004099964 Nov 2004 WO
WO-2008010917 Jan 2008 WO
WO-2008085416 Jul 2008 WO
WO-2008085457 Jul 2008 WO
WO-2008085457 Jul 2008 WO
WO-2008085719 Jul 2008 WO
2008157252 Dec 2008 WO
WO-2008157245 Dec 2008 WO
WO-2010030706 Mar 2010 WO
WO-2010030709 Mar 2010 WO
Non-Patent Literature Citations (123)
Entry
Notice of Allowance dated Dec. 14, 2016, for U.S. Appl. No. 12/208,329, filed Sep. 10, 2008, seven pages.
Notice of Allowance dated Apr. 6, 2017, for U.S. Appl. No. 15/270,950, filed Sep. 20, 2016, seven pages.
Final Office Action dated Jun. 1, 2017, for U.S. Appl. No. 15/158,461, filed May 18, 2016, 17 pages.
Chinese Search Report dated Nov. 24, 2011, for CN Application No. ZL2008201335089, with English Translation, nine pages.
European Search Report dated Apr. 17, 2012, for EP Patent Application No. 11188985.3, six pages.
European Search Report dated Oct. 4, 2016, for EP Patent Application No. 16178444.2, four pages.
Fakatselis, J. (Aug. 1996). “Processing Gain for Direct Sequence Spread Spectrum Communication Systems and PRISM®,” Application Note AN9633, Intersil, four pages.
Final Office Action dated Feb. 1, 2011, for U.S. Appl. No. 11/818,345, filed Jun. 13, 2007, 20 pages.
Final Office Action dated Sep. 1, 2011, for U.S. Appl. No. 12/874,184, filed Sep. 1, 2010, 19 pages.
Final Office Action dated Aug. 17, 2012, for U.S. Appl. No. 13/250,984, filed Sep. 30, 2011, 14 pages.
Final Office Action dated Aug. 28, 2012, for U.S. Appl. No. 12/283,423, filed Sep. 10, 2008, nine pages.
Final Office Action dated Oct. 12, 2012, for U.S. Appl. No. 12/208,315, filed Sep. 10, 2008, 21 pages.
Final Office Action dated Dec. 5, 2013, for U.S. Appl. No. 12/208,315, filed Sep. 10, 2008, 22 pages.
Final Office Action dated Dec. 30, 2013, for U.S. Appl. No. 12/208,329, filed Sep. 10, 2008, 12 pages.
Final Office Action dated Oct. 7, 2014, for U.S. Appl. No. 12/208,315, filed Sep. 10, 2008, 28 pages.
Final Office Action dated Feb. 26, 2015, for U.S. Appl. No. 12/208,329, filed Sep. 10, 2008, 13 pages.
Final Office Action dated Dec. 24, 2015, for U.S. Appl. No. 12/208,315, filed Sep. 10, 2008, 28 pages.
Final Office Action dated Feb. 4, 2016, for U.S. Appl. No. 14/482,979, filed Sep. 10, 2014, 12 pages.
International Search Report dated Apr. 6, 2009, for PCT Application No. PCT/US2007/088750, filed Dec. 21, 2007, six pages.
International Search Report dated Oct. 30, 2009, for PCT Application No. PCT/US2008/066743, filed Jun. 12, 2008, six pages.
International Search Report dated Dec. 21, 2009, for PCT Application No. PCT/US2009/056413, filed Sep. 9, 2009, three pages.
International Search Report dated Dec. 22, 2009, for PCT Application No. PCT/US2009/056410, filed Sep. 9, 2009, three pages.
Kanda, E. et al. (2008). “55.2: Integrated Active Matrix Capacitive Sensors for Touch Panel LTPS-TFT LCDs,” SID 08 Digest, pp. 834-837.
Lee, S.K. et al. (Apr. 1985). “A Multi-Touch Three Dimensional Touch-Sensitive Tablet,” Proceedings of CHI: ACM Conference on Human Factors in Computing Systems, pp. 21-25.
Non-Final Office Action dated Nov. 4, 2009, for U.S. Appl. No. 11/619,433, filed Jan. 3, 2007, 33 pages.
Non-Final Office Action dated May 12, 2010, for U.S. Appl. No. 11/818,345, filed Jun. 13, 2007, 27 pages.
Non-Final Office Action dated Jun. 8, 2010, for U.S. Appl. No. 11/650,046, filed Jan. 3, 2007, 11 pages.
Non-Final Office Action dated Nov. 22, 2010, for U.S. Appl. No. 11/650,046, filed Jan. 3, 2007, 10 pages.
Non-Final Office Action dated Apr. 27, 2011, for U.S. Appl. No. 12/874,184, filed Sep. 1, 2010, 18 pages.
Non-Final Office Action dated May 20, 2011, for U.S. Appl. No. 12/283,435, filed Sep. 10, 2008, six pages.
Non-Final Office Action dated Sep. 8, 2011, for U.S. Appl. No. 12/208,334, filed Sep. 10, 2008, 16 pages.
Non-Final Office Action dated Nov. 7, 2011, for U.S. Appl. No. 13/250,984, filed Sep. 30, 2011, nine pages.
Non-Final Office Action dated Nov. 17, 2011, for U.S. Appl. No. 12/283,423, filed Sep. 10, 2008, nine pages.
Non-Final Office Action dated Nov. 25, 2011, for U.S. Appl. No. 12/208,329, filed Sep. 10, 2008, 31 pages.
Non-Final Office Action dated Dec. 19, 2011, for U.S. Appl. No. 12/208,315, filed Sep. 10, 2008, 21 pages.
Non-Final Office Action dated Apr. 13, 2012, for U.S. Appl. No. 13/250,984, filed Sep. 30, 2011, 14 pages.
Non-Final Office Action dated May 2, 2012, for U.S. Appl. No. 12/874,184, filed Sep. 1, 2010, 20 pages.
Non-Final Office Action dated Sep. 28, 2012, for U.S. Appl. No. 12/208,329, filed Sep. 10, 2008, 33 pages.
Non-Final Office Action dated Oct. 5, 2012, for U.S. Appl. No. 13/568,027, filed Aug. 6, 2012, six pages.
Non-Final Office Action dated Oct. 25, 2012, for U.S. Appl. No. 12/874,184, filed Sep. 1, 2010, 23 pages.
Non-Final Office Action dated Dec. 7, 2012, for U.S. Appl. No. 13/250,984, filed Sep. 30, 2011, 10 pages.
Non-Final Office Action dated Apr. 23, 2013, for U.S. Appl. No. 13/568,027, filed Aug. 6, 2012, 8 pages.
Non-Final Office Action dated Apr. 25, 2013, for U.S. Appl. No. 13/553,421, filed Jul. 19, 2012, 10 pages.
Non-Final Office Action dated Apr. 30, 2013, for U.S. Appl. No. 12/208,315, filed Sep. 10, 2008, 21 pages.
Non-Final Office Action dated Aug. 23, 2013, for U.S. Appl. No. 13/916,357, filed Jun. 12, 2013, 17 pages.
Non-Final Office Action dated Aug. 29, 2013, for U.S. Appl. No. 12/208,329, filed Sep. 10, 2008, 13 pages.
Non-Final Office Action dated Mar. 14, 2014, for U.S. Appl. No. 12/208,315, filed Sep. 10, 2008, 25 pages.
Non-Final Office Action dated May 13, 2014, for U.S. Appl. No. 14/056,841, filed Oct. 17, 2013, seven pages.
Non-Final Office Action dated Jun. 6, 2014, for U.S. Appl. No. 12/208,329, filed Sep. 10, 2008, 12 pages.
Non-Final Office Action dated Sep. 29, 2014, for U.S. Appl. No. 14/315,162, filed Jun. 25, 2014, seven pages.
Non-Final Office Action dated May 14, 2015, for U.S. Appl. No. 12/208,315, filed Sep. 10, 2008, 25 pages.
Non-Final Office Action dated Jul. 29, 2015, for U.S. Appl. No. 14/482,979, filed Sep. 10, 2014, 19 pages.
Non-Final Office Action dated Dec. 14, 2015, for U.S. Appl. No. 12/208,329, filed Sep. 10, 2008, 14 pages.
Non-Final Office Action dated Aug. 8, 2016, for U.S. Appl. No. 15/158,461, filed May 18, 2016, 15 pages.
Non-Final Office Action dated Oct. 28, 2016, for U.S. Appl. No. 15/270,950, filed Sep. 20, 2016, six pages.
Notice of Allowance dated Jun. 7, 2010, for U.S. Appl. No. 11/619,433, filed Jan. 3, 2007, 10 pages.
Notice of Allowance dated Mar. 28, 2012, for U.S. Appl. No. 11/650,046, filed Jan. 3, 2007, seven pages.
Notice of Allowance dated Apr. 5, 2012, for U.S. Appl. No. 12/283,435, filed Sep. 10, 2008, seven pages.
Notice of Allowance dated Mar. 21, 2013, for U.S. Appl. No. 12/208,329, filed Sep. 10, 2008, 14 pages.
Notice of Allowance dated Apr. 8, 2013, for U.S. Appl. No. 11/818,345, filed Jun. 13, 2007, 14 pages.
Notice of Allowance dated May 23, 2013, for U.S. Appl. No. 13/250,984, filed Sep. 30, 2011, 10 pages.
Notice of Allowance dated Jun. 4, 2013, for U.S. Appl. No. 12/874,184, filed Sep. 1, 2010, 12 pages.
Notice of Allowance dated Aug. 23, 2013, for U.S. Appl. No. 13/250,984, filed Sep. 30, 2011, 10 pages.
Notice of Allowance dated Aug. 29, 2013, for U.S. Appl. No. 12/283,423, filed Sep. 10, 2008, eight pages.
Notice of Allowance dated Sep. 27, 2013, for U.S. Appl. No. 13/553,421, filed Jul. 19, 2012, 8 pages.
Notice of Allowance dated Nov. 25, 2013, for U.S. Appl. No. 13/568,027, filed Aug. 6, 2012, nine pages.
Notice of Allowance dated Feb. 6, 2014, for U.S. Appl. No. 13/916,357, filed Jun. 12, 2013, 25 pages.
Notice of Allowance dated Mar. 11, 2014, for U.S. Appl. No. 13/568,027, filed Aug. 6, 2012, eight pages.
Notice of Allowance dated Jun. 13, 2014, for U.S. Appl. No. 14/019,264, filed Sep. 5, 2013, 13 pages.
Notice of Allowance dated Dec. 3, 2014, for U.S. Appl. No. 13/935,333, filed Jul. 3, 2013, nine pages.
Notice of Allowance dated Jan. 15, 2015, for U.S. Appl. No. 14/056,841, filed Oct. 17, 2013, ten pages.
Notice of Allowance (corrected) dated Mar. 19, 2015, for U.S. Appl. No. 14/056,841, filed Oct. 17, 2013, seven pages. (Supplemental).
Notice of Allowance dated Mar. 25, 2015, for U.S. Appl. No. 14/270,147, filed May 5, 2014, nine pages.
Notice of Allowance dated May 18, 2015, for U.S. Appl. No. 14/315,162, filed Jun. 25, 2014, eight pages.
Notice of Allowance dated Mar. 14, 2016, for U.S. Appl. No. 12/208,315, filed Sep. 10, 2008, 12 pages.
Notice of Allowance dated May 11, 2016, for U.S. Appl. No. 14/791,145, filed Jul. 2, 2015, 17 pages.
Notice of Allowance dated Jul. 15, 2016, for U.S. Appl. No. 14/704,885, filed May 5, 2015, seven pages.
Notice of Allowance dated Aug. 29, 2016, for U.S. Appl. No. 14/482,979, filed Sep. 10, 2014, 15 pages.
Notice of Allowance dated Sep. 14, 2016, for U.S. Appl. No. 12/208,329, filed Sep. 10, 2008, seven pages.
Rubine, D.H. (Dec. 1991). “The Automatic Recognition of Gestures,” CMU-CS-91-202, Submitted in Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy in Computer Science at Carnegie Mellon University, 285 pages.
Rubine, D.H. (May 1992). “Combining Gestures and Direct Manipulation,” CHI '92, pp. 659-660.
Search Report dated Oct. 15, 2008, for GB Patent Application No. 0808783.5, filed May 15, 2008, two pages.
Search Report dated Apr. 24, 2009, for NL Application No. 2001666, English translation, 12 pages.
Search Report dated May 25, 2015, for ROC (Taiwan) Patent Application No. 102100040, with English translation, two pages.
Taiwanese Search Report dated Aug. 16, 2012, for TW Patent Application No. 097100216, two pages.
Westerman, W. (Spring 1999). “Hand Tracking, Finger Identification, and Chordic Manipulation on a Multi-Touch Surface,” A Dissertation Submitted to the Faculty of the University of Delaware in Partial Fulfillment of the Requirements for the Degree of Doctor of Philosophy in Electrical Engineering, 364 pages.
Wikipedia, Online-Encyclopedia. (Jun. 6, 2007). “Phase (waves),” retrieved on Jan. 18, 2016, located at: http://en.wikipedia.org/w/index.php?title=phase_(wave)$oldid=136332745, two pages.
Wikipedia, Online-Encyclopedia. (Jun. 11, 2007). “Signal generator,” retrieved on Jan. 13, 2016, located at: http://en.wikipedia.org/w/index.php?title=signal_generator$oldid=137433567, two pages.
Notice of Allowance dated May 22, 2018, for U.S. Appl. No. 15/158,461, filed May 18, 2016, nine pages.
Notice of Allowance dated May 31, 2018, for U.S. Appl. No. 15/658,314, filed Jul. 24, 2017, seven pages.
Notice of Allowance dated Apr. 10, 2018, for U.S. Appl. No. 15/250,736, filed Aug. 29, 2016, nine pages.
Advisory Action received for U.S. Appl. No. 12/208,315, dated Feb. 26, 2014, 3 pages.
Advisory Action received for U.S. Appl. No. 12/874,184, dated Feb. 10, 2012, 3 pages.
Advisory Action received for U.S. Appl. No. 14/482,979, dated Aug. 1, 2016, 3 pages.
Advisory Action received for U.S. Appl. No. 14/715,351, dated Jul. 5, 2018, 5 pages.
Advisory Action received for U.S. Appl. No. 15/158,461, dated Oct. 31, 2017, 2 pages.
Applicant Initiated Interview Summary received for U.S. Appl. No. 14/482,979, dated Jul. 21, 2016, 3 pages.
Corrected Notice of Allowance received for U.S. Appl. No. 14/056,841, dated Mar. 6, 2015, 7 pages.
Final Office Action received for U.S. Appl. No. 11/650,046, dated May 3, 2011, 10 pages.
Final Office Action received for U.S. Appl. No. 12/557,814, dated Nov. 6, 2013, 21 pages.
Final Office Action received for U.S. Appl. No. 14/715,351, dated Feb. 13, 2018, 29 pages.
Final Office Action received for U.S. Appl. No. 16/056,180, dated Jul. 11, 2019, 13 pages.
First Action Interview Office Action received for U.S. Appl. No. 14/715,351, dated Aug. 10, 2017, 7 pages.
International Search Report received for PCT Patent Application No. PCT/US2007/026177, dated Jun. 11, 2008, 3 pages.
International Search Report received for PCT Patent Application No. PCT/US2008/066759, dated Oct. 6, 2008, 2 pages.
Non-Final Office Action received for U.S. Appl. No. 11/818,454, dated May 10, 2010, 17 pages.
Non-Final Office Action received for U.S. Appl. No. 12/283,423, dated Jan. 18, 2013, 10 pages.
Non-Final Office Action received for U.S. Appl. No. 12/557,814, dated Apr. 2, 2013, 25 pages.
Non-Final Office Action received for U.S. Appl. No. 12/904,012, dated Mar. 7, 2011, 26 pages.
Non-Final Office Action received for U.S. Appl. No. 15/997,541, dated Nov. 6, 2019, 8 pages.
Non-Final Office Action received for U.S. Appl. No. 16/056,180, dated Oct. 18, 2018, 15 pages.
Notice of Allowance received for U.S. Appl. No. 11/818,454, dated Sep. 29, 2010, 7 pages.
Notice of Allowance received for U.S. Appl. No. 12/557,814, dated Jan. 21, 2015, 5 pages.
Notice of Allowance received for U.S. Appl. No. 12/904,012, dated Nov. 16, 2011, 9 pages.
Notice of Allowance received for U.S. Appl. No. 12/904,012, dated Sep. 22, 2011, 10 pages.
Notice of Allowance received for U.S. Appl. No. 14/715,351, dated Jul. 27, 2018, 5 pages.
Notice of Allowance received for U.S. Appl. No. 15/997,541, dated Apr. 15, 2020, 8 pages.
Pre-Interview First Office Action received for U.S. Appl. No. 14/715,351, dated Mar. 29, 2017, 6 pages.
Restriction Requirement received for U.S. Appl. No. 11/619,433, dated Apr. 12, 2010, 5 pages.
Restriction Requirement received for U.S. Appl. No. 12/208,329, dated Jul. 15, 2011, 5 pages.
Restriction Requirement received for U.S. Appl. No. 12/283,435, dated Dec. 21, 2011, 6 pages.
Restriction Requirement received for U.S. Appl. No. 14/056,841, dated Jan. 16, 2014, 5 pages.
Search Report received for GB Patent Application No. 0808783.5, dated Jun. 25, 2012, 1 page.
Related Publications (1)
Number Date Country
20170097728 A1 Apr 2017 US
Divisions (1)
Number Date Country
Parent 11619433 Jan 2007 US
Child 12874184 US
Continuations (3)
Number Date Country
Parent 14482979 Sep 2014 US
Child 15380747 US
Parent 14019264 Sep 2013 US
Child 14482979 US
Parent 12874184 Sep 2010 US
Child 14019264 US