The present invention relates to interactive displays for electronic devices, and in particular to an interactive display that provide tactile feedback to a user when the user applies pressure to the interactive display.
An electronic device according to an exemplary embodiment of the present invention comprises: a housing comprising a first surface and a second surface; an interactive display having a keyboard-enabled mode and a keyboard-disabled mode, the interactive display comprising: a first image display device disposed at the first surface that displays image data; and a physical keypad disposed at the second surface that provides tactile feedback to a user only when the interactive display is in the keyboard-enabled mode, the physical keypad being substantially smooth when the interactive display is in the keyboard-disabled mode.
In at least one embodiment, the first image display device displays an imaged keypad comprising imaged keys when the interactive display is in the keyboard-enabled mode.
In at least one embodiment, the electronic device further comprises a second image display device disposed at the second surface.
In at least one embodiment, the physical keypad disposed at the second surface comprises one or more physical keys that control activation of the imaged keys within the imaged keypad.
In at least one embodiment, the physical keypad disposed at the second surface comprises one or more physical keys that perform a different function from the one or more imaged keys of the imaged keypad.
In at least one embodiment, the electronic device further comprises a sensor that detects user interaction with the physical keypad disposed on the second surface; and an image generator that displays within the first image display device an imaged representation of the user interaction with the physical keypad disposed on the second surface.
In at least one embodiment, the sensor comprises one or more of the following sensor types: a motion sensor, a thermal sensor and a pressure sensor.
In at least one embodiment, the imaged representation of the user interaction comprises an imaged depiction of the user's finger or thumb interacting with the imaged keypad.
In at least one embodiment, the first and second surfaces are opposite from one another.
In at least one embodiment, the first and second surfaces are angled relative to one another.
In at least one embodiment, the first and second surfaces are adjacent to one another.
In at least one embodiment, the physical keypad comprises piezoelectric material.
In at least one embodiment, the piezoelectric material is quartz.
In at least one embodiment, the physical keypad comprises magnetostrictive material.
In at least one embodiment, the image display device is selected from one of the following types of image display devices: liquid crystal displays, digital light processor displays, plasma displays and light emitting diode displays.
In at least one embodiment, the electronic device is selected from one of the following types of electronic devices: cell phones, personal digital assistants, gaming devices, e-books, automatic teller machines and data input devices.
An electronic device according to an exemplary embodiment of the present invention comprises: a housing comprising a first surface and a second surface; and an interactive display having an interactive mode and a non-interactive mode, the interactive display comprising: an image display device disposed at the first surface of the housing that displays a user-interactive imaged keypad in at least a portion of the image display device when the interactive display is in the interactive mode and that displays other image data in the at least a portion of the image display device when the interactive display is in the non-interactive mode; a substantially transparent physical keypad disposed at the second surface of the housing that provides tactile feedback to a user indicating location of keys within the imaged keypad; a sensor that detects user interaction with the physical keypad; and an image generator that displays an imaged representation of the user interaction with the physical keypad within the imaged keypad.
In at least one embodiment, the sensor comprises one or more of the following sensor types: a motion sensor, a thermal sensor and a pressure sensor.
In at least one embodiment, the imaged representation of the user interaction comprises an imaged depiction of the user's finger or thumb interacting with the imaged keypad.
An electronic device according to an exemplary embodiment of the present invention comprises: an interactive display having an interactive mode and a non-interactive mode, the interactive display comprising: an image display device that displays a user-interactive imaged keypad in at least a portion of the image display device when the interactive display is in the interactive mode and that displays other image data in the at least a portion of the image display device when the interactive display is in the non-interactive mode; and a substantially transparent physical keypad that provides tactile feedback to a user indicating location of one or more imaged keys within the imaged keypad in response to one or more actuators disposed below the image display device.
An electronic device according to an exemplary embodiment of the present invention comprises: a housing comprising a first surface and a second surface; and an interactive display having an interactive mode and a non-interactive mode, the interactive display comprising: an image display device disposed at the first surface of the housing that displays a user-interactive imaged keypad in at least a portion of the image display device when the interactive display is in the interactive mode and that displays other image data in the at least a portion of the image display device when the interactive display is in the non-interactive mode; and a substantially transparent physical keypad disposed at the second surface of the housing that provides tactile feedback to a user indicating location of keys within the imaged keypad.
In at least one embodiment, the first and second surfaces are opposite from one another.
In at least one embodiment, the first and second surfaces are angled relative to one another.
In at least one embodiment, the first and second surfaces are adjacent to one another.
In at least one embodiment, the image display device is structured so as to form a seal for the electronic device.
In at least one embodiment, the physical keypad comprises one or more physical keys corresponding to the imaged keys.
In at least one embodiment, the image display device and the physical keypad are integral to one another.
In at least one embodiment, the one or more physical keys are made of substantially transparent material so that the imaged keys may be viewed within the one or more physical keys.
In at least one embodiment, the imaged keypad comprises one or more sub-keypads, each of the one or more sub-keypads being selectively activated.
In at least one embodiment, one or more portions of the physical keypad are selectively activated to correspond to the one or more sub-keypads.
In at least one embodiment, the physical keypad provides tactile feedback only when the interactive display is in the interactive mode.
In at least one embodiment, the physical keypad comprises piezoelectric material.
In at least one embodiment, the piezoelectric material is quartz.
In at least one embodiment, the physical keypad is made of a magnetostrictive material.
In at least one embodiment, the image display device is selected from one of the following types of image display devices: liquid crystal displays, digital light processor displays, plasma displays and light emitting diode displays.
In at least one embodiment, the image display device is a touchscreen image display device.
In at least one embodiment, the physical keypad comprises physical keys that extend through corresponding openings in the touchscreen image display device.
In at least one embodiment, the one or more actuators move at least one of the physical keypad and the touchscreen display device relative to one another so that the one or more physical keys protrude through the openings in the touchscreen display device when the interactive display is in the interactive mode.
In at least one embodiment, one or more of the physical keys each comprises one or more magnetic elements.
In at least one embodiment, the touchscreen display device comprises one or more charged electrical circuit elements so that movement of the one or more physical keys comprising the one or more magnetic elements relative to the charged electrical circuits elements generate electricity.
In at least one embodiment, the electronic device further comprises a protective layer disposed between the physical keypad and the at least one magnetic actuator to prevent damage to the magnetic actuator resulting from contact with the physical keys.
In at least one embodiment, the electronic device is selected from one of the following types of electronic devices: cell phones, personal digital assistants, automatic teller machines and data input devices.
The above and related objects, features and advantages of the present invention will be more fully understood by reference to the following, detailed description of the preferred, albeit illustrative, embodiment of the present invention when taken in conjunction with the accompanying figures, wherein:
The present invention is directed to an electronic device including an interactive display having an interactive mode in which a user-interactive keypad is displayed in at least a portion of the interactive display and a non-interactive mode in which other image data is displayed in the portion of the interactive display. The interactive display includes a substantially transparent keypad portion that provides tactile feedback to allow the user to locate individual keys within the keypad portion. The present invention is applicable to any electronic device having a touchscreen display, including, for example, personal digital assistants (PDAs), cell phones, automated teller machines (ATMs), computers (including laptop and desktop computers), gaming devices, television monitors, video conferencing equipment, e-books (e.g., Amazon Kindle™ and Barnes & Noble Nook™) remote control devices and any general data input device.
In
In
The display device 130 may be any suitable display device, such as, for example, a liquid crystal display (LCD), a digital light processing (DLP) display, a plasma display or a light-emitting diode (LED) display, to name a few. As is known in the art, the display device 130 may include programmable elements that emit and/or block light to generate images. In the present embodiment, the display device 130 may display an imaged keyboard when the interactive display 100 is in the interactive mode.
The physical keypad 140 is a generally flat sheet or plate. The physical keys 145 are formed on the top surface of the physical keypad 140, and extend upwards towards the touchscreen panel 150. The physical keypad 140 is preferably made of a transparent material, such as, for example, plastic or glass. Any number of physical keys 145 may be formed on the physical keypad 140. In an exemplary embodiment, the number and shape of the physical keys 145 are made to correspond to the number and shape of the imaged keys in the imaged keyboard displayed by the display device 130.
The touchscreen panel 150 may be a transparent panel that generates a programming signal when pressure is applied to one or more areas on the touchscreen panel 150. Various programming signals generated by the touchscreen panel 150 may be sent to the display device 130, resulting in formation or manipulation of images in the display device 130. For example, a user may apply pressure to the touchscreen panel 150 to activate the imaged keyboard and place the interactive display 100 in the interactive mode. Any suitable touchscreen technology may be used for the touchscreen panel 150, such as, for example, resistive touchscreens, surface acoustic wave touchscreens, capacitive touchscreens, infrared touchscreens, strain gauge touchscreens, optical imaging, dispersive signal technology, acoustic pulse recognition, total internal reflection, and diffused laser imaging, to name a few.
The touchscreen panel 150 may include a number of openings 152 that correspond with the physical keys 145 of the physical keypad 140. The physical keys 145 extend through the openings 152. When the interactive display 100 is in the non-interactive mode, the top surfaces of the physical keys 145 may be co-planar with the top surface of the touchscreen panel 150. When the interactive display 100 is in the interactive mode, the top surfaces of the physical keys 145 may be raised or lowered relative to the top surface of the touchscreen panel 150, so that the touchscreen panel 150 includes a number of protrusions or indentations that correspond to the imaged keys in the imaged keypad. Thus, when the interactive display 100 is in the interactive mode, a user is able to feel the location of the various imaged keys based on the tactile feedback provided by the protrusions or indentations in the touchscreen panel 150.
In an exemplary embodiment of the present invention, first and second actuators 160, 162, located on either side of the frame element 110, may be used to manipulate the physical keypad 140 and thereby raise and lower the physical keys 145 relative to the touchscreen panel 150. The first and second actuators 160, 162 may be any suitable actuators, such as, for example, mechanical actuators, such as springs, microelectromechanical devices (MEMS), piezoelectric actuators and magnetostrictive actuators, to name a few. It should be appreciated that the number of actuators is not limited to two, and any number of actuators located at any suitable position relative to the physical keypad 140 may be used to raise and lower the physical keypad 140. Alternatively, the actuators 160, 162 may be used to raise and lower the touchscreen panel 150 rather than the physical keypad 140.
In various exemplary embodiments of the present invention, the physical keys in the physical keypad may all be connected to the physical keypad so as to form a unitary structure, with no independent movement of the physical keys. Alternatively, each physical key may be independently moveable relative to the physical keypad and the other physical keys so that each physical key may have the ability to be physically “pressed” by a user. In this regard, each physical key may include, for example, a spring mechanism or may itself be made of a resilient material that is able to flex under pressure from a user's finger/thumb. The physical keypad may also use conventional keypad technology, such as, for example, dome-switch keypad technology, so that manipulation of each physical key results in a corresponding function.
The physical keypad 240 in the present embodiment may include physical keys 245 defined by a material that changes shape under the influence of an electric or magnetic field. For example, the physical keys 245 may be made of a piezoelectric material, such as, for example, quartz, or a magnetostrictive material, such as, for example, ferromagnetic thin films. In the embodiment shown in
It should be appreciated that the present invention is not limited to the above described embodiments. For example, in other exemplary embodiments, each physical key may be made of two or more pieces of transparent material that are made to deflect relative to one another when the interactive display is switched between the interactive and non-interactive modes so as to form appropriate protrusions or indentations in the interactive display that demark the physical keys. In such an embodiment, an additional transparent layer may be disposed over the physical keypad, so that when the two or more pieces of material that form the physical keys are deflected, smooth bumps or indentations are formed in the transparent layer.
Further, as shown in
In other exemplary embodiments of the present invention, the interactive display may include separate regions, where one or more regions are interactive and one or more other regions are not interactive. Further, the interactive display may be composed of more than one display, where one or more of the displays are interactive and one or more other displays are not interactive.
The physical keypad 740 in the present embodiment may include physical keys 245. The physical keys 745 are preferably transparent and may be integrally formed with the remaining portions of the physical keypad 740 by a molding operation. In the exemplary embodiment shown in
As shown in
In the present embodiment, an actuator element 860 is disposed below the display device 830. The actuator element 860 may be any type of suitable actuator, such as, for example, piezoelectric actuators or magnetostrictive actuators. It should be appreciated that the number of actuators is not limited to one, and any number of actuators located at any suitable position relative to the physical keypad 840 may be used to raise and lower the physical keypad 845. Alternatively, the actuator element 860 may be used to raise and lower the touchscreen panel 850 rather than the physical keypad 840. The structure of the present embodiment allows the display device 830 to be sealed with the frame element 810 so as to protect the actuator element 860 and other internal components from being damaged.
The protective layer 930 is disposed between the sub-keypads 940 and the actuator element 960 so as to prevent damage to the actuator element 960 that may otherwise result from contact with the sub-keypads 940. The protective layer 930 may be made of any suitable protective material that does not interfere with the function of the actuator element 960, such as, for example, felt, cotton, plastic, insulators, cushioning material, etc. The protective material can even be air in one embodiment.
In the present embodiment, each key 1045 of the physical keypad 1040 forms part of the touchscreen display device 1030. In this regard, a touchscreen display element 1048 may be disposed within each key 1045, so that images, such as corresponding imaged keys, may be displayed on each key 1045 as part of the overall image displayed by the touchscreen display device 1030.
As shown in
In an alternative exemplary embodiment, the imaged keypad 2024 at the front of the electronic device 2000 and the physical keypad 2022 at the back of the electronic device 2000 together form a combination keypad made up of imaged keys 2026 and physical keys 2028, where the imaged keys 2026 perform functions different from those of the physical keys 2028. For example, the physical keys 2028 may include modifier keys, such as “control” and “shift” keys, lock keys, such as “num lock” and “caps lock”, navigation keys, such as arrow keys, “page up/page down” and “home/end”, and editing keys, such as “return”, “enter”, “backspace”, “insert”, “delete”, “tab” and space bar, while the image keys 2026 may include alphanumeric keys. In other embodiments, this configuration may be reversed, so that the image keys 2026 include modifier, navigation, lock, editing and navigation keys, while the physical keys 2028 include alphanumeric keys. Also, in other embodiments, the imaged keys 2026 and physical keys 2028 may include any suitable arrangement of modifier, navigation, lock, editing and alphanumeric keys. The combination of imaged keys 2026 and physical keys 2028 having varying functions and disposed on opposite surfaces of the electronic device 2000 allows for the use of both thumbs and fingers to manipulate the keys. For example, the fingers on the underside of the electronic device 2000 may be used to input letters and numbers, while the thumbs on the top of the electronic device 2000 may be used to control spacing, paragraph changes, etc. The imaged keys 2026 and physical keys 2028 may be rearranged based on user preference, so that the user is able to choose which keys are available on different surfaces of the electronic device 2000. For example, the user may choose to arrange certain keys that are typically activated by the user's thumb on one surface of the electronic device 2000 and certain keys that are typically activated by the user's fingers on another surface.
The physical keypad 2022 and imaged keypad 2024 may operate the same as in other exemplary embodiments of the present invention described herein, except that the physical keys 2028 provide tactile feedback to the user on a surface of the electronic device 2000 that is different from a surface on which the imaged keypad 2024 is displayed. For example, the physical keys 2028 may be automatically manipulated by actuators or through the use of material for the physical keys 2028 that changes shape under the influence of an electric or magnetic field. In this regard,
The display device 2130 may be any suitable display device, such as, for example, a liquid crystal display (LCD), a digital light processing (DLP) display, a plasma display or a light-emitting diode (LED) display, to name a few. As is known in the art, the display device 2130 may include programmable elements that emit and/or block light to generate images. In the present embodiment, the display device 2130 may display an imaged keyboard when the interactive display 2100 is in the keyboard-enabled mode.
The physical keypad 2140 is a generally flat sheet or plate. The physical keys 2145 are formed on the top surface of the physical keypad 2140, and extend downwards towards the bottom wall 2112 of the frame element 2110. The physical keypad 2140 may be made of a transparent material, such as, for example, plastic or glass. Any number of physical keys 2145 may be formed on the physical keypad 2140. In an exemplary embodiment, the number and shape of the physical keys 2145 are made to correspond to the number and shape of the imaged keys in the imaged keyboard displayed by the display device 2130.
The touchscreen panel 2150 may be a transparent panel that generates a programming signal when pressure is applied to one or more areas on the touchscreen panel 2150. Various programming signals generated by the touchscreen panel 2150 may be sent to the display device 2130, resulting in formation or manipulation of images in the display device 2130. For example, a user may apply pressure to the touchscreen panel 2150 to activate the imaged keyboard and place the interactive display 2100 in the interactive mode. Any suitable touchscreen technology may be used for the touchscreen panel 2150, such as, for example, resistive touchscreens, surface acoustic wave touchscreens, capacitive touchscreens, infrared touchscreens, strain gauge touchscreens, optical imaging, dispersive signal technology, acoustic pulse recognition, total internal reflection, and diffused laser imaging, to name a few.
The bottom wall 2112 of the frame element 2110 may include a number of openings 2152 that correspond with the physical keys 2145 of the physical keypad 2140. The physical keys 2145 extend through the openings 2152. When the interactive display 2100 is in the keyboard-disabled mode, the top surfaces of the physical keys 2145 may be co-planar with the bottom surface of the bottom wall 2112. When the interactive display 2100 is in the keyboard-enabled mode, the top surfaces of the physical keys 2145 may be raised or lowered relative to the top surface of the bottom wall 2112, so that the bottom wall 2112 includes a number of protrusions or indentations that correspond to the imaged keys in the imaged keypad. Thus, when the interactive display 2100 is in the keyboard-enabled mode, a user may be able to feel the location of the various imaged keys based on the tactile feedback provided by the protrusions or indentations in the bottom wall 2112.
In an exemplary embodiment of the present invention, first and second actuators 2160, 2162, located on either side of the frame element 2110, may be used to manipulate the physical keypad 2140 and thereby raise and lower the physical keys 2145 relative to the bottom wall 2112. The first and second actuators 2160, 2162 may be any suitable actuators, such as, for example, mechanical actuators, such as springs, microelectromechanical devices (MEMS), piezoelectric actuators and magnetostrictive actuators, to name a few. It should be appreciated that the number of actuators is not limited to two, and any number of actuators located at any suitable position relative to the physical keypad 2140 may be used to raise and lower the physical keypad 2140.
The physical keypad 3022 and imaged keypad 3024 may operate the same as in the previous exemplary embodiment. That is, the physical keys 3028 may provide tactile feedback to the user on a surface of the electronic device 3000 that is different from a surface on which the imaged keypad 3024 is displayed. For example, the physical keys 3028 may be automatically manipulated by actuators or through the use of material for the physical keys 3028 that changes shape under the influence of an electric or magnetic field. In this regard,
The display device 3130 may be any suitable display device, such as, for example, a liquid crystal display (LCD), a digital light processing (DLP) display, a plasma display or a light-emitting diode (LED) display, to name a few. As is known in the art, the display device 3130 may include programmable elements that emit and/or block light to generate images. In the present embodiment, the display device 3130 may display an imaged keyboard when the interactive display 2100 is in the keyboard-enabled mode.
The physical keypad 3140 is a generally flat sheet or plate. The physical keys 3145 are formed on the top surface of the physical keypad 3140, and extend downwards towards the bottom wall 3112 of the frame element 3110. The physical keypad 3140 may be made of a transparent material, such as, for example, plastic or glass. Any number of physical keys 3145 may be formed on the physical keypad 3140. In an exemplary embodiment, the number and shape of the physical keys 3145 are made to correspond to the number and shape of the imaged keys in the imaged keyboard displayed by the display device 3130.
The touchscreen panel 3150 may be a transparent panel that generates a programming signal when pressure is applied to one or more areas on the touchscreen panel 3150. Various programming signals generated by the touchscreen panel 3150 may be sent to the display device 3130, resulting in formation or manipulation of images in the display device 3130. For example, a user may apply pressure to the touchscreen panel 3150 to activate the imaged keyboard and place the interactive display 3100 in the keyboard-enabled mode. Any suitable touchscreen technology may be used for the touchscreen panel 3150, such as, for example, resistive touchscreens, surface acoustic wave touchscreens, capacitive touchscreens, infrared touchscreens, strain gauge touchscreens, optical imaging, dispersive signal technology, acoustic pulse recognition, total internal reflection, and diffused laser imaging, to name a few.
The bottom wall 3112 of the frame element 3110 may include a number of openings 3152 that correspond with the physical keys 3145 of the physical keypad 3140. The physical keys 3145 extend through the openings 3152. When the interactive display 3100 is in the keyboard-disabled mode, the top surfaces of the physical keys 3145 may be co-planar with the bottom surface of the bottom wall 3112. When the interactive display 3100 is in the keyboard-enabled mode, the top surfaces of the physical keys 3145 may be raised or lowered relative to the bottom surface of the bottom wall 3112, so that the bottom wall 3112 includes a number of protrusions or indentations that correspond to the imaged keys in the imaged keypad. Thus, when the interactive display 3100 is in the keyboard-enabled mode, a user is able to feel the location of the various imaged keys based on the tactile feedback provided by the protrusions or indentations in the bottom wall 3112.
In an exemplary embodiment of the present invention, first and second actuators 3160, 3162, located on either side of the frame element 3110, may be used to manipulate the physical keypad 3140 and thereby raise and lower the physical keys 3145 relative to the bottom wall 3112. The first and second actuators 3160, 3162 may be any suitable actuators, such as, for example, mechanical actuators, such as springs, microelectromechanical devices (MEMS), piezoelectric actuators and magnetostrictive actuators, to name a few. It should be appreciated that the number of actuators is not limited to two, and any number of actuators located at any suitable position relative to the physical keypad 3140 may be used to raise and lower the physical keypad 3140.
The interactive display 3100 may also include a sensor 3170 that detects user interaction with the physical keypad 3140. The sensor 3170 may be any known or later-discovered sensor that is able to detect the presence and/or motion of the digits of a user's hand relative to the physical keypad 3140. The sensor 3170 may be, for example, a motion sensor, a thermal sensor or a pressure sensor, or any combination of a variety of sensors. The sensor 3170 may generate electrical signals based on the detected user interaction, and the generated signals may be sent to an image generator. The image generator may use the generated signals to display in the imaged keypad 3023 an imaged representation of the user interaction. Thus, even though the user's thumbs or fingers may be out of view behind the electronic device 300, the user will still have the ability to determine the location of his thumbs or fingers relative to the physical keys 3145 by viewing the imaged representation in the imaged keypad 3023. In an exemplary embodiment, the electronic device 3000 may include a switch to allow the user to selectively activate the imaging of user interaction within the imaged keypad 3023. In exemplary embodiments of the invention, the sensor 3170 may include a number of individual sensors, where each sensor corresponds to a particular physical key 3145 within the imaged keypad 3140.
As shown in
The interactive display 4100 may also include a physical keypad 4140, a second surface touchscreen panel 4160 disposed over the physical keypad 4140, and a second surface display device 4180 disposed over the touchscreen panel 4160. The physical keypad 4140 in the present embodiment may include physical keys 4145 defined by a material that changes shape under the influence of an electric or magnetic field. For example, the physical keys 4145 may be made of a piezoelectric material, such as, for example, quartz, or a magnetostrictive material, such as, for example, ferromagnetic thin films. In the embodiment shown in
As in the previous embodiment, the interactive display 4100 may also include a sensor 4170 that detects user interaction with the physical keypad 4140. The sensor 4170 may generate electrical signals based on the detected user interaction, and the generated signals may be sent to an image generator. The image generator may use the generated signals to display in the imaged keypad an imaged representation of the user interaction.
Now that the preferred embodiments of the present invention have been shown and described in detail, various modifications and improvements thereon will become readily apparent to those skilled in the art. Accordingly, the spirit and scope of the present invention is to be construed broadly and limited only by the appended claims and not by the foregoing specification.
This application is a continuation of U.S. patent application Ser. No. 12/849,530, filed Aug. 3, 2010, entitled Interactive Display with Tactile Feedback, which is a non-provisional of U.S. Provisional Patent Application 61/353,509, filed Jun. 10, 2010, and a continuation-in-part of U.S. patent application Ser. No. 12/759,410, entitled Energy Efficient Interactive Display With Energy Regenerative Keyboard, filed Apr. 13, 2010, which is a continuation-in-part of U.S. patent application Ser. No. 12/339,721, entitled Apparatus and Method for Interactive Display With Tactile Feedback, filed Dec. 19, 2008, which is a continuation-in-part of U.S. patent application Ser. No. 12/142,624, entitled Apparatus and Method for Interactive Display With Tactile Feedback, filed Jun. 19, 2008, the contents of which are incorporated herein by reference in their entirety.
| Number | Name | Date | Kind |
|---|---|---|---|
| 3333160 | Gorski | Jul 1967 | A |
| 3541541 | Englebart | Nov 1970 | A |
| 3641568 | Brescia et al. | Feb 1972 | A |
| 3662105 | Hurst et al. | May 1972 | A |
| 3798370 | Hurst | Mar 1974 | A |
| 3995126 | Larson | Nov 1976 | A |
| 4127752 | Lowthorp | Nov 1978 | A |
| 4194097 | Bradam | Mar 1980 | A |
| 4246452 | Chandler | Jan 1981 | A |
| 4508942 | Inaba | Apr 1985 | A |
| 4550221 | Mabusth | Oct 1985 | A |
| 4654715 | Feldmann et al. | Mar 1987 | A |
| 4672364 | Lucas | Jun 1987 | A |
| 4672558 | Beckes et al. | Jun 1987 | A |
| 4692809 | Beining et al. | Sep 1987 | A |
| 4695827 | Beining et al. | Sep 1987 | A |
| 4707570 | Ide et al. | Nov 1987 | A |
| 4733222 | Evans | Mar 1988 | A |
| 4734685 | Watanabe | Mar 1988 | A |
| 4746770 | McAvinney | May 1988 | A |
| 4771276 | Parks | Sep 1988 | A |
| 4788384 | Bruere-Dawson et al. | Nov 1988 | A |
| 4806846 | Kerber | Feb 1989 | A |
| 4898555 | Sampson | Feb 1990 | A |
| 4940346 | Liljenquist | Jul 1990 | A |
| 4968877 | McAvinney et al. | Nov 1990 | A |
| 4990900 | Kikuchi | Feb 1991 | A |
| 5003919 | Noirjean | Mar 1991 | A |
| 5017030 | Crews | May 1991 | A |
| 5178477 | Gambaro | Jan 1993 | A |
| 5189403 | Franz et al. | Feb 1993 | A |
| 5194862 | Edwards | Mar 1993 | A |
| 5224861 | Glass et al. | Jul 1993 | A |
| 5241308 | Young | Aug 1993 | A |
| 5252951 | Tannenbaum et al. | Oct 1993 | A |
| 5281966 | Walsh | Jan 1994 | A |
| 5305017 | Gerpheide | Apr 1994 | A |
| 5311175 | Waldman | May 1994 | A |
| 5345543 | Capps et al. | Sep 1994 | A |
| 5376948 | Roberts | Dec 1994 | A |
| 5379057 | Clough et al. | Jan 1995 | A |
| 5398310 | Tchao et al. | Mar 1995 | A |
| 5412189 | Cragun | May 1995 | A |
| 5442742 | Greyson et al. | Aug 1995 | A |
| 5463388 | Boie et al. | Oct 1995 | A |
| 5463696 | Beernink et al. | Oct 1995 | A |
| 5483261 | Yasutake | Jan 1996 | A |
| 5488204 | Mead et al. | Jan 1996 | A |
| 5495077 | Miller et al. | Feb 1996 | A |
| 5513309 | Meier et al. | Apr 1996 | A |
| 5523775 | Capps | Jun 1996 | A |
| 5530455 | Gillick et al. | Jun 1996 | A |
| 5543590 | Gillespie et al. | Aug 1996 | A |
| 5543591 | Gillespie et al. | Aug 1996 | A |
| 5563632 | Roberts | Oct 1996 | A |
| 5563996 | Tchao | Oct 1996 | A |
| 5565658 | Gerpheide et al. | Oct 1996 | A |
| 5579036 | Yates, IV | Nov 1996 | A |
| 5581681 | Tchao et al. | Dec 1996 | A |
| 5583946 | Gourdol | Dec 1996 | A |
| 5590219 | Gourdol | Dec 1996 | A |
| 5592566 | Pagallo et al. | Jan 1997 | A |
| 5594810 | Gourdol | Jan 1997 | A |
| 5596694 | Capps | Jan 1997 | A |
| 5612719 | Beernink et al. | Mar 1997 | A |
| 5631805 | Bonsall | May 1997 | A |
| 5633955 | Bozinovic et al. | May 1997 | A |
| 5634102 | Capps | May 1997 | A |
| 5635958 | Murai | Jun 1997 | A |
| 5636101 | Bonsall et al. | Jun 1997 | A |
| 5642108 | Gopher et al. | Jun 1997 | A |
| 5644657 | Capps et al. | Jul 1997 | A |
| 5666113 | Logan | Sep 1997 | A |
| 5666502 | Capps | Sep 1997 | A |
| 5666552 | Greyson et al. | Sep 1997 | A |
| 5675361 | Santilli | Oct 1997 | A |
| 5675362 | Clough et al. | Oct 1997 | A |
| 5677710 | Thompson-Rohrlich | Oct 1997 | A |
| 5689253 | Hargreaves et al. | Nov 1997 | A |
| 5710844 | Capps et al. | Jan 1998 | A |
| 5729250 | Bishop et al. | Mar 1998 | A |
| 5730185 | Wilkins et al. | Mar 1998 | A |
| 5736976 | Cheung | Apr 1998 | A |
| 5741990 | Davies | Apr 1998 | A |
| 5745116 | Pisutha-Arnond | Apr 1998 | A |
| 5745716 | Tchao et al. | Apr 1998 | A |
| 5746818 | Yatake | May 1998 | A |
| 5748269 | Harris et al. | May 1998 | A |
| 5764222 | Shieh | Jun 1998 | A |
| 5767457 | Gerpheide et al. | Jun 1998 | A |
| 5767842 | Korth | Jun 1998 | A |
| 5790104 | Shieh | Aug 1998 | A |
| 5790107 | Kasser et al. | Aug 1998 | A |
| 5802516 | Shwarts et al. | Sep 1998 | A |
| 5808567 | McCloud | Sep 1998 | A |
| 5809267 | Moran et al. | Sep 1998 | A |
| 5821690 | Martens et al. | Oct 1998 | A |
| 5821930 | Hansen | Oct 1998 | A |
| 5823782 | Marcus et al. | Oct 1998 | A |
| 5825351 | Tam | Oct 1998 | A |
| 5825352 | Bisset et al. | Oct 1998 | A |
| 5835080 | Beeteson et al. | Nov 1998 | A |
| 5838138 | Henty | Nov 1998 | A |
| 5854625 | Frisch et al. | Dec 1998 | A |
| 5880411 | Gillespie et al. | Mar 1999 | A |
| 5880539 | Chen | Mar 1999 | A |
| 5898434 | Small et al. | Apr 1999 | A |
| 5920309 | Bisset et al. | Jul 1999 | A |
| 5923319 | Bishop et al. | Jul 1999 | A |
| 5933134 | Shieh | Aug 1999 | A |
| 5943044 | Martinelli et al. | Aug 1999 | A |
| 6002389 | Kasser | Dec 1999 | A |
| 6002808 | Freeman | Dec 1999 | A |
| 6020881 | Naughton et al. | Feb 2000 | A |
| 6031524 | Kunert | Feb 2000 | A |
| 6037882 | Levy | Mar 2000 | A |
| 6050825 | Nichol et al. | Apr 2000 | A |
| 6052339 | Frenkel et al. | Apr 2000 | A |
| 6072494 | Nguyen | Jun 2000 | A |
| 6084576 | Leu et al. | Jul 2000 | A |
| 6107997 | Ure | Aug 2000 | A |
| 6128003 | Smith et al. | Oct 2000 | A |
| 6131299 | Raab et al. | Oct 2000 | A |
| 6135958 | Mikula-Curtis et al. | Oct 2000 | A |
| 6144380 | Shwarts et al. | Nov 2000 | A |
| 6188391 | Seely et al. | Feb 2001 | B1 |
| 6198515 | Cole | Mar 2001 | B1 |
| 6208329 | Ballare | Mar 2001 | B1 |
| 6222465 | Kumar et al. | Apr 2001 | B1 |
| 6239790 | Martinelli et al. | May 2001 | B1 |
| 6243073 | Shwarts et al. | Jun 2001 | B1 |
| 6246862 | Grivas et al. | Jun 2001 | B1 |
| 6249606 | Kiraly et al. | Jun 2001 | B1 |
| 6288707 | Philipp | Sep 2001 | B1 |
| 6289326 | LaFleur | Sep 2001 | B1 |
| 6292178 | Bernstein et al. | Sep 2001 | B1 |
| 6323846 | Westerman et al. | Nov 2001 | B1 |
| 6323849 | He et al. | Nov 2001 | B1 |
| 6347290 | Bartlett | Feb 2002 | B1 |
| 6356259 | Maeda et al. | Mar 2002 | B1 |
| 6377009 | Philipp | Apr 2002 | B1 |
| 6380931 | Gillespie et al. | Apr 2002 | B1 |
| 6411287 | Scharff et al. | Jun 2002 | B1 |
| 6414671 | Gillespie et al. | Jul 2002 | B1 |
| 6421234 | Ricks et al. | Jul 2002 | B1 |
| 6429846 | Rosenberg et al. | Aug 2002 | B2 |
| 6452514 | Philipp | Sep 2002 | B1 |
| 6457355 | Philipp | Oct 2002 | B1 |
| 6466036 | Philipp | Oct 2002 | B1 |
| 6497521 | Lohr | Dec 2002 | B1 |
| 6515669 | Mohri | Feb 2003 | B1 |
| 6525749 | Moran et al. | Feb 2003 | B1 |
| 6535200 | Philipp | Mar 2003 | B2 |
| 6543684 | White et al. | Apr 2003 | B1 |
| 6543947 | Lee | Apr 2003 | B2 |
| 6570557 | Westerman et al. | May 2003 | B1 |
| 6593916 | Aroyan | Jul 2003 | B1 |
| 6610936 | Gillespie et al. | Aug 2003 | B2 |
| 6624833 | Kumar et al. | Sep 2003 | B1 |
| 6639577 | Eberhard | Oct 2003 | B2 |
| 6650319 | Hurst et al. | Nov 2003 | B1 |
| 6658994 | McMillan | Dec 2003 | B1 |
| 6670894 | Mehring | Dec 2003 | B2 |
| 6677932 | Westerman | Jan 2004 | B1 |
| 6677934 | Blanchard | Jan 2004 | B1 |
| 6706552 | Getz et al. | Mar 2004 | B2 |
| 6724366 | Crawford | Apr 2004 | B2 |
| 6757002 | Oross et al. | Jun 2004 | B1 |
| 6803906 | Morrison et al. | Oct 2004 | B1 |
| 6822635 | Shahoian et al. | Nov 2004 | B2 |
| 6842672 | Straub et al. | Jan 2005 | B1 |
| 6856259 | Sharp | Feb 2005 | B1 |
| 6888536 | Westerman et al. | May 2005 | B2 |
| 6900795 | Knight, III et al. | May 2005 | B1 |
| 6927761 | Badaye et al. | Aug 2005 | B2 |
| 6942571 | McAllister et al. | Sep 2005 | B1 |
| 6965375 | Gettemy et al. | Nov 2005 | B1 |
| 6972401 | Akitt et al. | Dec 2005 | B2 |
| 6977666 | Hedrick | Dec 2005 | B1 |
| 6985801 | Straub et al. | Jan 2006 | B1 |
| 6992659 | Gettemy | Jan 2006 | B2 |
| 7031228 | Born et al. | Apr 2006 | B2 |
| 7129416 | Steinfeld et al. | Oct 2006 | B1 |
| 7148875 | Rosenberg et al. | Dec 2006 | B2 |
| 7339572 | Schena | Mar 2008 | B2 |
| RE40993 | Westerman | Nov 2009 | E |
| 7630200 | Eldershaw | Dec 2009 | B1 |
| 7683888 | Kennedy | Mar 2010 | B1 |
| 7748634 | Zehr et al. | Jul 2010 | B1 |
| 7808479 | Hotelling et al. | Oct 2010 | B1 |
| 7978181 | Westerman | Jul 2011 | B2 |
| 8005518 | Birsel | Aug 2011 | B1 |
| 8115745 | Gray | Feb 2012 | B2 |
| 20020118175 | Liebenow | Aug 2002 | A1 |
| 20020118848 | Karpenstein | Aug 2002 | A1 |
| 20020149570 | Knowles et al. | Oct 2002 | A1 |
| 20030006974 | Clough et al. | Jan 2003 | A1 |
| 20030076301 | Tsuk et al. | Apr 2003 | A1 |
| 20030076303 | Huppi | Apr 2003 | A1 |
| 20030076306 | Zadesky et al. | Apr 2003 | A1 |
| 20030095095 | Pihlaja | May 2003 | A1 |
| 20030095096 | Robbin et al. | May 2003 | A1 |
| 20030098858 | Perski et al. | May 2003 | A1 |
| 20030206202 | Moriya | Nov 2003 | A1 |
| 20030234768 | Rekimoto et al. | Dec 2003 | A1 |
| 20040208681 | Dechene | Oct 2004 | A1 |
| 20040229663 | Tosey et al. | Nov 2004 | A1 |
| 20040263481 | Nishikawa et al. | Dec 2004 | A1 |
| 20040263484 | Mantysalo et al. | Dec 2004 | A1 |
| 20050012723 | Pallakoff | Jan 2005 | A1 |
| 20050017947 | Shahoian et al. | Jan 2005 | A1 |
| 20050052425 | Zadesky et al. | Mar 2005 | A1 |
| 20050052430 | Shahoian et al. | Mar 2005 | A1 |
| 20050104867 | Westerman et al. | May 2005 | A1 |
| 20050110768 | Marriott et al. | May 2005 | A1 |
| 20050190159 | Skarine | Sep 2005 | A1 |
| 20060022955 | Kennedy | Feb 2006 | A1 |
| 20060022956 | Lengeling et al. | Feb 2006 | A1 |
| 20060026521 | Hotelling et al. | Feb 2006 | A1 |
| 20060026535 | Hotelling et al. | Feb 2006 | A1 |
| 20060026536 | Hotelling et al. | Feb 2006 | A1 |
| 20060032680 | Elias et al. | Feb 2006 | A1 |
| 20060033724 | Chaudhri et al. | Feb 2006 | A1 |
| 20060053387 | Ording | Mar 2006 | A1 |
| 20060066582 | Lyon et al. | Mar 2006 | A1 |
| 20060084482 | Saila | Apr 2006 | A1 |
| 20060085757 | Andre et al. | Apr 2006 | A1 |
| 20060097991 | Hotelling et al. | May 2006 | A1 |
| 20060097993 | Hietala | May 2006 | A1 |
| 20060119589 | Rosenberg et al. | Jun 2006 | A1 |
| 20060187215 | Rosenberg et al. | Aug 2006 | A1 |
| 20060192771 | Rosenberg et al. | Aug 2006 | A1 |
| 20060197753 | Hotelling | Sep 2006 | A1 |
| 20060232567 | Westerman et al. | Oct 2006 | A1 |
| 20060238510 | Panotopoulos et al. | Oct 2006 | A1 |
| 20060238517 | King et al. | Oct 2006 | A1 |
| 20060238518 | Westerman et al. | Oct 2006 | A1 |
| 20060238519 | Westerman et al. | Oct 2006 | A1 |
| 20060238520 | Westerman et al. | Oct 2006 | A1 |
| 20060238521 | Westerman et al. | Oct 2006 | A1 |
| 20060238522 | Westerman et al. | Oct 2006 | A1 |
| 20070004451 | Anderson | Jan 2007 | A1 |
| 20070013665 | Vetelainen | Jan 2007 | A1 |
| 20070013677 | Rosenberg et al. | Jan 2007 | A1 |
| 20070040815 | Rosenberg et al. | Feb 2007 | A1 |
| 20070091070 | Larsen | Apr 2007 | A1 |
| 20070229464 | Hotelling et al. | Oct 2007 | A1 |
| 20070229478 | Rosenberg et al. | Oct 2007 | A1 |
| 20070236466 | Hotelling | Oct 2007 | A1 |
| 20070247429 | Westerman | Oct 2007 | A1 |
| 20070257890 | Hotelling et al. | Nov 2007 | A1 |
| 20080060856 | Shahoian et al. | Mar 2008 | A1 |
| 20080062134 | Duarte | Mar 2008 | A1 |
| 20080062143 | Shahoian et al. | Mar 2008 | A1 |
| 20080062144 | Shahoian et al. | Mar 2008 | A1 |
| 20080062145 | Shahoian et al. | Mar 2008 | A1 |
| 20080068348 | Rosenberg et al. | Mar 2008 | A1 |
| 20080068349 | Rosenberg et al. | Mar 2008 | A1 |
| 20080068350 | Rosenberg et al. | Mar 2008 | A1 |
| 20080068351 | Rosenberg et al. | Mar 2008 | A1 |
| 20080122796 | Jobs | May 2008 | A1 |
| 20080145127 | Liu | Jun 2008 | A1 |
| 20080303796 | Fyke | Dec 2008 | A1 |
| 20080316180 | Carmody | Dec 2008 | A1 |
| 20090046076 | Bychkov | Feb 2009 | A1 |
| 20090091537 | Huang | Apr 2009 | A1 |
| 20090128503 | Grant et al. | May 2009 | A1 |
| 20090167509 | Fadell | Jul 2009 | A1 |
| 20090183098 | Casparian | Jul 2009 | A1 |
| 20090213081 | Case, Jr. | Aug 2009 | A1 |
| 20090315830 | Westerman | Dec 2009 | A1 |
| 20100099463 | Kim | Apr 2010 | A1 |
| 20100134418 | Muresianu | Jun 2010 | A1 |
| 20100251152 | Cho | Sep 2010 | A1 |
| 20100277430 | Cruz-Hernandez | Nov 2010 | A1 |
| 20110021251 | Linden | Jan 2011 | A1 |
| 20110032211 | Christoffersen | Feb 2011 | A1 |
| 20110141679 | Dai et al. | Jun 2011 | A1 |
| 20110161809 | Gilmour | Jun 2011 | A1 |
| 20110221682 | Gao | Sep 2011 | A1 |
| 20110234498 | Gray | Sep 2011 | A1 |
| 20110261058 | Luo | Oct 2011 | A1 |
| Number | Date | Country |
|---|---|---|
| 1243096 | Oct 1988 | CA |
| 102 51 296 | May 2004 | DE |
| 10324580 | Dec 2004 | DE |
| 0 288 692 | Jul 1993 | EP |
| 0 664 504 | Jul 1995 | EP |
| 0 464 908 | Sep 1996 | EP |
| 0915490 | May 1999 | EP |
| 1 014 295 | Jan 2002 | EP |
| 2009047 | Jun 1979 | GB |
| 1997018547 | May 1997 | WO |
| 1997023738 | Jul 1997 | WO |
| 199814863 | Apr 1998 | WO |
| 99938149 | Jul 1999 | WO |
| 2003088176 | Oct 2003 | WO |
| 200409488 | Dec 2004 | WO |
| 2006023569 | Mar 2006 | WO |
| 2007127585 | Nov 2007 | WO |
| 2010029448 | Mar 2010 | WO |
| Entry |
|---|
| Lee, et al. “A Multi-Touch Three Dimensional Touch-Sensitive Table,” in CHI '85 Proceedings, pp. 121-128, 2000. |
| Lee et al. “A Multi-Touch Three Dimensional Touch-Sensitive Tablet,” in CHI '85 Proceedings, pp. 121-128, 2000. |
| Matsushita, et ai, “HoloWall: Designing a Finger, Hand, Body and Object Sensitive Wall,” In Proceedings of UIST '97, Oct. 1997. |
| Quantum Research Group “QT510IQWheel™ Touch Slider IC”copyright 2004-2005, 14-pgs. |
| Qraek, “Unencumbered Gestural Interaction,” IEEE Multimedia, 3:36-47 (Winter 1996). |
| Radwin, “Activation Force and Travel Effects on Overexertion in Repetitive Key Tapping,” Human Factors, 39(1):130-140 (Mar 1997). |
| Rekimoto, “SmartSkin: An infrastructure for Freehand Manipulation on Interactive Surfaces” CHI 2002, Apr. 20-25, 2002. |
| Rekimoto, et al., “ToolStone: Effective Use of the Physical Manipulation Vocabularies of Input Devices,” In Proc. Of UIST 2000, 2000. |
| Rubine et ai, “Programmable Finger-Tracking Instrument Controllers,” Computer Music Journal, vol. 14, No. 1 (Spring 1990). |
| Rutledge, et al., “Force-To-Motion Functions for Pointing,” Human-Computer Interaction—INTERACT (1990). |
| Subatai Ahmed, “A Usable Real-Time 3D Hand Tracker,” Proceedings of the 28th Asilomar Conference on Signals, Systems and Computers—Part 2 (of 2), vol. 2 (Oct. 1994). |
| Texas Instruments “TSC2003/12C Touch Screen Controller” Data Sheet SBAS 162, dated Octobe 2001,20 pgs. |
| Wellner, “The Digital Desk Calculators: Tangivle Manipulation on a Desk Top Display” In ACM UIST '91 Proceedings, pp. 27-34, Nov. 1991. |
| Williams, “Applications for a Sitched-Capacitor Instrumentation Building Block” Linear Technology Application Note 3, Jul. 1985, pp. 1-16. |
| Yamada, et al_, “A Switched-Capacitor Interface for Capacitive Pressure Sensors” IEEE Transactions on Instrumentation and Measurement, vol. 41, No. 1, Feb. 1992, pp. 81-86. |
| Yeh, et al., “Switched Capacitor Interface Circuit for Capacitive Transducers” 1985 IEEE. |
| Zhai, et al., “Dual Stream Input for Pointing and Scrolling,” Proceedings of CHI '97 Extended Abstracts (1997). |
| Zimmerman, et al., “Applying Electric Field Sensing to Human-Computer Interfaces,” In CHI '85 Proceedings, pp. 280-287, 1995. |
| International Search Report dated Jan. 2, 2008 for peT Application No. PCT/US2007/066022, filed Apr. 5, 2007, 3 pages. |
| Jenna Wortham and Malt Richtel, Makers Hope Touch Screens Will Help Celilphone Sales, NY Times, Dec. 1, 2008, at B1. |
| “4-Wire Resistive Touchscreens” obtained from http://www.touchscreen.com/intro-touchtypes-4resistive.html generated Aug. 5, 2005. |
| “5-Wire Resistive Touchscreens” obtained from http://www.touchscreens.com/intro-touchtypes-resisitive.html generated Aug. 5, 2005. |
| “A Brief Overview of Gesture Recognition” obtained from http://www.dai.ed.ac.ak/Cvonline/LOCA_COPIES/COHEN/gesture_overview.html, generated Apr. 20, 2004. |
| “Capacitive Touchscreens” obtained from http://www.touchscreens.com/intro-touchtypes-capacitive.html generated Aug. 5, 2005. |
| “Capacitive Position Sensing” obtained from http://www.synaptics.com/technology/eps.cfm generated Aug. 5, 2005. |
| “Comparing Touch Technologies” obtained from http://www.touchscreens.com/intro-touchtypes.html generated Oct. 10, 2004. |
| “Gesture Recognition” http://www.fingerworks.com/gesture_recognition.html. |
| GlidePoint® obtained from http://www.cirque.com/technology/technology_gp.hmtl generated Aug. 5, 2005. |
| “How do touchscreen monitors know where you're touching?” obtained from http://www.electronics.howstuffworks.com/question716.html generated Aug. 5, 2005. |
| “How does a touchscreen work?” obtained from http://www.touchscreens.com/intro-anatomy.html generated Aug. 5, 2005. |
| “iGesture Products for Everyone (learn in minutes) Product Overview” FingerWorks.com. |
| “Infrared Touchscreens” obtained from http://www.touchscreens.com/intro-touchtypes-infrared.html generated Aug. 5, 2005. |
| “Mouse Emulation” FingerWorks obtained from http://www.fingerworks.com/gesture_guide_mouse.html generated Aug. 30, 2005. |
| “Mouse Gestures in Opera” obtained from http://www.opera.com/products/desktop/mouse/index.dml generated Aug. 30, 2005. |
| “Mouse Gestures,” Optim oz, May 21, 2004. |
| “MultiTouch Overview” FingerWorks obtained from http://www.fingerworks.com/multoverview.html generated Aug. 30, 2005. |
| “Near Field Imaging Touchscreens” obtained from http://www.touchscreens.com/intro-touchtypes-nfi.html generated Aug. 5, 2005. |
| “PenTouch Capacitive Touchscreens” obtained from http://www.touchscreens.com/intro-touchtypes-pentouch.html generated Aug. 5, 2005. |
| “Surface Acoustic Wave Touchscreens” obtained from http://www.touchscreens.com/intro-touchtypes-saw.html generated Aug. 5, 2005. |
| “Symbol Commander” obtained from http://www.sensiva.com/symbolcomander/, generated Aug. 30, 2005. |
| “Tips for Typing” FingerWorks http://www.fingerworks.com/mini_typing.hml generated Aug. 30, 2005. |
| “Touch Technologies Overview” 2001, 3M Touch Systems, Massachusetts. |
| “Wacom Components—Technology” obtained from http://www.wacom-components.com/english/tech.asp generated on Oct. 10, 2004. |
| “Watershed Algorithm” http://rsb.info.nih.gov/ij/plugins/watershed.html generated Aug. 5, 2005. |
| “FingerWorks—Gesture Guide—Application Switching,” obtained from http://www.fingerworks.com/gesture_guide_apps.html, generated on Aug. 27, 2004, 1-pg. |
| “FingerWorks—Gesture Guide—Editing,” obtained from http://www.fingerworks.com/gesture_guide_editing.html, generated on Aug. 27, 2004, 1-pg. |
| “FingerWorks—Gesture Guide—File Operations,” obtained from http://www.fingerworks.com/gesture_guide_files.html, generated on Aug. 27, 2004, 1-pg. |
| “FingerWorks—Gesture Guide—Text Manipulation,” obtained from http://www.fingerworks.com/gesture_guide_text_manip.html, generated on Aug. 27, 2004, 2-pg. |
| “FingerWorks—Gesture Guide—Tips and Tricks,” obtained form http://www.fingerworks.com/gesture_guide_tips.html, generated Aug. 27, 2004, 2-pgs. |
| “FingerWorks—Gesture Guide—Web,” obtained from http://www.fingerworks.com/gesture_guide_web.html, generated on Aug. 27, 2004, 1-pg. |
| “FingerWorks—Guide to Hand Gestures for USB Touchpads,” obtained from http://www.fingerworks.com/igesture_userguide.html, generated Aug. 27, 2004, 1-pg. |
| “FingerWorks—iGesture—Technical Details,” obtained from http://www.fingerworks.com/igesture_tech.html, generated Aug. 27, 2004, 1-pg. |
| “FingerWorks—The Only Touchpads with Ergonomic Full-Hand Resting and Relaxation!” obtained from http://www.fingerworks.com/resting.hml, Copyright 2001, 1-pg. |
| “FingerWorks—Tips for Typing on the Mini,” obtained from http://www.fingerworks.com/mini_typing.html, generated on Aug. 27, 2004, 2-pgs. |
| “iGesture Pad—the MultiFinger USB TouchPad with Whole-Hand Gestures,” obtained from http://www.fingerworks.com/igesture.html, generated Aug. 27, 2004, 2-pgs. |
| Bier, et al., “Toolglass and Magic Lenses: The see-through interface” In James Kijiya, editor, Computer Graphics (SIGGRAPH '93 Proceedings), vol. 27, pp. 73-80, Aug. 1993. |
| Douglas et al., The Ergonomics of Computer Pointing Devices (1997). |
| European Search Report received in EP 1 621 989 (@Beyer Weaver & Thomas, LLP) dated Mar. 27, 2006. |
| EVB Elektronic “TSOP6238 IR Receiver Modules for Infrared Remote Control Systems” dated Jan. 2004, 1-pg. |
| Fisher et al., “Repetitive Motion Disorders: The Design of Optimal Rate-Rest Profiles,” Human Factors, 35(2):283-304 (Jun. 1993). |
| Fukumoto, et al., “ActiveClick: Tactile Feedback for Touch Panels,” In CHI 2001 Summary, pp. 121-122, 2001. |
| Fukumoto and Yoshinobu Tonomura, “Body Coupled Fingering: Wireless Wearable Keyboard,” Chi 97, pp. 147-154 (Mar. 1997). |
| Hardy, “Fingerworks” Mar. 7, 2002; BBC World On Line. |
| Hillier and Gerald J. Lieberman, Introduction to Operations Research (1986). |
| International Search Report dated Mar. 3, 2006 (PCT/US 05/03325;119-0052WO). |
| Jacob et al., “Integrality and Separability of Input Devices,” ACM Transactions on Computer-Human Interaction, 1:3-26 (Mar. 1994). |
| Kinkely et al., “Touch-Sensing Input Devices,” in CHI '99 Proceedings, pp. 223-230, 1999. |
| Kjonx “KXP84 Series Summary Data Sheet” copyright 2005,dated Oct. 21, 2005, 4-pgs. |
| Number | Date | Country | |
|---|---|---|---|
| 20170083100 A1 | Mar 2017 | US |
| Number | Date | Country | |
|---|---|---|---|
| 61353509 | Jun 2010 | US |
| Number | Date | Country | |
|---|---|---|---|
| Parent | 12849530 | Aug 2010 | US |
| Child | 15369591 | US |
| Number | Date | Country | |
|---|---|---|---|
| Parent | 12759410 | Apr 2010 | US |
| Child | 12849530 | US | |
| Parent | 12339721 | Dec 2008 | US |
| Child | 12759410 | US | |
| Parent | 12142624 | Jun 2008 | US |
| Child | 12339721 | US |