Projected field haptic actuation

Information

  • Patent Grant
  • 8310444
  • Patent Number
    8,310,444
  • Date Filed
    Tuesday, January 27, 2009
    15 years ago
  • Date Issued
    Tuesday, November 13, 2012
    12 years ago
Abstract
An electronic device includes a touch surface that can be physically engaged by a user. The touch surface is operably connected to an actuator arm which, in turn, is connected to an actuator array. Drive electronics sense a user's movement relative to the touch surface and, responsively, drive the actuator array effective to move the actuator arm and, in turn, provide haptic feedback to the user through the touch surface.
Description
BACKGROUND

Some devices such as mobile phones, hand-held media players, personal digital assistants (PDAs), and the like, can be configured to provide physically-sensible feedback to a user, such as a vibration. Yet, the actuating mechanisms that are utilized to cause this feedback, such as electromagnetic actuators and others, can be prohibitively large in size, can consume a significant amount of power, and/or can be expensive to use.


SUMMARY

This Summary is provided to introduce a selection of concepts in a simplified form that are further described below in the Detailed Description. This Summary is not intended to identify key features or essential features of the claimed subject matter, nor is it intended to be used to limit the scope of the claimed subject matter.


In one or more embodiments, an electronic device includes a touch surface that can be physically engaged by a user. The touch surface is operably connected to an actuator arm which, in turn, is connected to an actuator array. Drive electronics sense a user's movement relative to the touch surface and, responsively, drive the actuator array effective to move the actuator arm and, in turn, provide haptic feedback to the user through the touch surface.





BRIEF DESCRIPTION OF THE DRAWINGS

The same numbers are used throughout the drawings to reference like features.



FIG. 1 illustrates an example electronic device in accordance with one or more embodiments.



FIG. 2 illustrates an exploded view of the electronic device of FIG. 1 in accordance with one or more embodiments.



FIG. 3 illustrates an actuator array in accordance with one or more embodiments.



FIG. 4 illustrates an electronic device with its housing removed in accordance with one or more embodiments.



FIG. 5 illustrates an electronic device in a cut-away view as seen from the front of the device in accordance with one or more embodiments.



FIG. 6 illustrates a high-level block diagram of example system in accordance with one or more embodiments.



FIG. 7 illustrates an example voltage regulator in accordance with one or more embodiments.



FIG. 8 illustrates a USB interface that can allow real-time changes of haptic profiles in accordance with one or more embodiments.



FIG. 9 illustrates an electronic circuit in accordance with one or more embodiments.



FIG. 10 illustrates an electronic circuit in accordance with one or more embodiments.



FIG. 11 illustrates an electronic circuit in accordance with one or more embodiments.



FIG. 12
a illustrates an example touch surface layout in accordance with one or more embodiments.



FIG. 12
b illustrates an example touch surface layout in accordance with one or more embodiments.



FIG. 13 is a flow diagram that describes steps in a method in accordance with one or more embodiments.





DETAILED DESCRIPTION

Overview


In one or more embodiments, an electronic device includes a touch surface that can be physically engaged by a user. The touch surface can comprise any suitable type of touch surface and can be formed from any suitable type of material such as, by way of example and not limitation, glass, plastic, and the like. In at least some embodiments, the touch surface can take the form of a touch screen. Touch screens are typically employed to enable a user to provide input by touching portions of the screen. Touch screens can be employed in many different types of devices such as, for example, hand-held devices, printers, copiers, multifunction peripheral devices, and the like.


For example, a touch screen may display a numerical dialing pad to emulate a telephone. By touching individual displayed numbers, the user can enter a telephone number. A touch screen can also be used to provide a virtual keyboard on which a user can type. These types of touch screens are typically employed in hand-held devices such as cellular telephones, smartphones and can be found in computer monitors, tablet PC's, GPS devices, notebook PC's automotive dashboards, etc.


Alternately or additionally, the touch surface can take the form of a touch pad. Touch pads also enable users to provide input to an electronic device. For example, many laptop computers employ some type of touchpad to enable a user to move a cursor around an associated screen.


In one or more embodiments, the touch surface is operably connected to an actuator arm which, in turn, is connected to an actuator array. The actuator array includes an electrically-deformable material that can be electrically deformed responsive to being electrically driven by a voltage. Any suitable type of electrically-deformable material can be utilized such as piezoelectric-electric materials, electromagnetic materials, electro restrictive polymers, electrostatic materials and the like. In at least some embodiments, the electrically-deformable material comprises an electro-active polymer or “EAP.” EAP refers to a class of polymers which are formulated to exhibit different physical, electrical, and/or electro-optical behaviors and properties. In general, when EAP is driven by an applied voltage, the EAP undergoes a deformation in a particular direction. This deformation causes the EAP to move in the particular direction. In various embodiments, the electrically-deformable material is driven by one or more drive voltages to effect movement of the touch surface, as will become apparent below. EAP is available from a company named Artificial Muscle Inc. located in Sunnyvale Calif.


Drive electronics sense a user's movement relative to the touch surface and, responsively, drive the actuator array effective to move the actuator arm and, in turn, provide haptic feedback to the user through the touch surface. For example, responsive to a user moving within a predefined area relative to the touch surface, the touch surface can be moved under the influence of the drive electronics to provide haptic feedback to the user.


As an example, consider the case in which a user dials a telephone number using their cellular phone. Assume that their cellular phone includes a touchscreen such as one described above and below. As the user engages the touchscreen to dial a telephone number, haptic feedback is provided to the user via movement of the touchscreen. This haptic feedback can be designed to simulate a button press. Specifically, by moving the touchscreen in a particular manner relative to the cellular phone's housing, the user feels like they are actually dialing on push buttons.


The particular manner in which the touchscreen is moved can vary and can be adjustable, in at least some embodiments, in accordance with predefined or definable profiles. For example, movement of the touch screen can occur in different directions, e.g. toward the user a defined distance, away from the user a define distance, and then back toward the user for a movement cycle.


In the discussion that follows, a section entitled “Example Electronic Device” describes an example electronic device that is configured to provide haptic feedback in accordance with one or more embodiments. Following this, a section entitled “Example Circuitry” describes example circuitry that can be used to implement one or more embodiments. Next, a section entitled “Example Touch Surfaces” is provided and describes some example touch surface structures that can be employed in accordance with one or more embodiments. Following this, a section entitled “Example Method” describes an example method in accordance with one or more embodiments.


Example Electronic Device



FIG. 1 illustrates an example electronic device in accordance with one or more embodiments generally at 100. In this example electronic device 100 includes a housing 102 and a touch surface 104 mounted within housing 102. As mounted, touch surface 104 can move in one or more directions, relative to the housing, under the influence electronic circuitry that is described below. Examples of movement directions are shown by the arrows. In one or more embodiments, directional movement occurs along a vector or vectors that is (are) generally parallel with a plane defined by the top surface of touch surface 104.


In this particular example, touch surface 104 resides in the form of a touch screen that can be engaged by a user. FIG. 1 is shown in a cut-away view to expose internal components which are described in more detail in relation to FIG. 2.



FIG. 2 illustrates an exploded view of electronic device 100. In this particular example, the electronic device 100 includes an actuator arm 200, an actuator array 202, a subassembly 204 that includes an LCD, various electronics, and a battery assembly, and a front housing 206 that supports touch surface 104. In this particular example, front housing 206 includes a pair of touch surface connectors 208, 210. Touch surface 104 is fixedly mounted to each of touch surface connectors 208, 210. In turn, when the electronic device 100 is assembled, the touch surface connectors 208, 210 are connected to actuator arm 200 by way of a pair of screws that are received in associated apertures in the actuator arm 200. The actuator arm 200 is fixedly mounted to the actuator array 202 so that when drive electronics (described below) sense a user's movement relative to the touch surface 104, the actuator array is driven with a voltage or voltages effective to move the actuator arm 200 and hence the touch surface 104 by way of the touch surface connectors 208, 210. Doing so provides haptic feedback to the user through the touch surface. By using the illustrated actuator array 202 and by effecting movement of the touch surface 104 along a vector or vectors that are generally parallel to the plane of the top surface of touch surface 104, the elevational thickness of electronic device can, in at least some embodiments, be reduced. Specifically, the illustrated actuator array 202 constitutes a departure from previously employed collapsible dome technologies which required elevationally thicker devices.


It is to be appreciated and understood that the particular arrangement and order of the components shown in FIG. 2 is to provide but one example of how various components can be arranged within a suitably-configured housing. Accordingly, other arrangements and component orders can be utilized without departing from the spirit and scope of the claimed subject matter. Other arrangements can include, by way of example and not limitation, mounting the actuator or actuators around the perimeter edges of the touch surface, anywhere below the touch surface, and/or mounting the actuator or actuators to a mass within the device to create movement, to name just a few possibilities.


Further, in at least some embodiments, actuator array 202 can be connected to other different components within the electronic device in order to impart a different haptic feedback experience. For example, in at least some embodiments a so-called floating battery arrangement can be used in which the actuator array 202 is fixedly connected to the device's battery which is movably mounted within the device's housing. Haptic feedback can be provided to the user by moving the battery under the influence of the actuator array 202.



FIG. 3 illustrates the actuator array 202 in more detail in accordance with one or more embodiments. In the view shown in FIG. 3, each side of the actuator array 202 is shown. The left-most view illustrates the side of the actuator array shown in FIG. 2; the right-most view illustrates the reverse side of the actuator array 202. In this example, the actuator array 202 includes an actuator frame 300, an electrically-deformable region 302, such as EAP, that is driven by the drive electronics described below, and an actuator disk 304 that is fixedly mounted to corresponding regions on actuator arm 200 (FIG. 2). Any suitable number of electrically-deformable regions can be employed. In the illustrated example, six electrically-deformable regions are employed and mounted on the actuator frame 300. In addition, multiple different actuator frames can be employed in a stacked arrangement to increase the force with which movement occurs.



FIGS. 4 and 5 illustrate different views of electronic device 100 with various portions of the structure removed to show detail.



FIG. 4 illustrates the electronic device with its housing 102 (FIG. 2) removed. Actuator arm 200 is shown mounted on actuator array 202. FIG. 5 illustrates the electronic device in a cut-away view as seen from the front of the device where a portion of touch surface 104 is shown. In this example, actuator array 202 is shown in its disposition relative to touch surface connectors 208, 210. Recall that the actuator arm 200 (FIG. 4) is fixedly mounted to the actuator array 202 and to the touch surface connectors 208, 210 to permit movement of the touch surface 104 under the influence of the drive electronics.



FIG. 6 illustrates a high-level block diagram of an example system, generally at 600, that can be incorporated in the electronic device 100 (FIG. 1) and utilized to implement the functionality described above and below. In the illustrated and described example, system 600 includes a microcontroller 602 which, in turn, includes a haptics customizing engine 604, a computer-readable storage media in the form of an EEPROM 606, an HID keyboard component 608, a key scanning component 610, and a haptics engine 612. In addition, system 600 includes an adjustable DC/DC converter 614, high side switches 616, 618, low side switches 620, 622, and an actuator 624. An example of an actuator 624 is described above in the form of actuator array 202 (FIG. 2).


In addition, a switch is illustrated generally at 630 and represents aspects of a touch surface that is configured to detect a user's engagement. Detection of a user's engagement can occur using any suitable type of sensor or detection apparatus. For example, in at least some embodiments, a capacitive-type sensor or a projected field-type sensor, surface acoustic wave, infrared display, optical/imaging resolution, and/or image sensing can be employed to sense a user's engagement. The operating principles of these types of sensors are generally known and, for the sake of brevity, are not described in detail here other than the explanation that appears just below.


In at least some embodiments, the detection apparatus establishes a sensory field that overlays a portion or all of touch surface 104 effective to define a sensor layer. The sensor layer can be considered as a region in which the presence and/or movement of a user, such as a user's finger, can be detected by the sensor layer. When the user's presence and/or movement is sensed by the sensor layer, an electrical signal can be sent to the drive electronics to effectively drive the electric-deformable material to cause the touch surface 104 to move in a desired fashion.


As shown, haptics customizing engine 604 is connected to the adjustable DC/DC converter 614 which, in turn, is connected to high side and low side switches 616, 618 and 620, 622 respectively. Actuator 624 is operably connected to the high side and low side switches as shown. The switches, both high side and low side are connected to haptics engine 612.


In operation, in one or more embodiments, haptics customizing engine 604 is configured to load predefined haptic profiles from EEPROM 606 or modify parameters of existing profiles, either upon user/host request or by request from a self-adapting haptic hardware/software system. In addition, in one or more embodiments, haptics customizing engine 604 is configured to load new profiles to the haptics engine 612 or save new profiles to the EEPROM 606 as defined by a user or hardware/software developer. EEPROM 606 is configured to store haptic profile information for use in the haptic engine 612. This information can be predefined at production time, as well as updated or supplemented at runtime by users, host system, developers, or an adaptive module of the haptic system.


HID keyboard components 608 is configured to provide Human Interface Device functionality to the host system (if necessary) in order to allow the haptic system to act in the same manner as a keypad, keyboard, touchpad, mouse, and also to provide haptic information to the host for display, modification, or other use.


Key scanning component 610 is configured to provide a mechanism for the haptic system to know when it should trigger playback of a haptic profile. The haptic system does not need to directly scan keys itself. Rather, the haptic system can alternatively take key/switch/input state information from another device, such as a keyboard controller, touch screen controller, or other user input device.


Haptics engine 612 is configured to control the input signals to the haptic actuator based on profile data supplied by the EEPROM 606, haptics customization engine 604, and/or whatever other sources of data exist.


The adjustable DC/DC converter is configured to supply the actuator operating voltage. The output voltage may or may not be regulated, may or may not be adjustable on the fly, or may or may not be adjustable at all. The DC/DC converter may or may not have any common or uncommon features of typical power supplies, such as over current protection, under voltage protection, sleep mode, off mode, voltage feedback, etc. On the fly adjustment allows the output voltage to be adjustable such that the host or haptics customization engine 604 can modify the output voltage.


In operation, in one or more embodiments, the high side and low side switches are configured to drive the voltage of an actuator phase to the actuator's maximum positive operating voltage, maximum negative operating voltage, or any voltage in between, including ground or a high impedance (floating) potential.


Having described an example electronic device, consider now a discussion of example circuitry that can be utilized to implement the embodiments described above.


Example Circuitry



FIG. 7 illustrates an example voltage regulator in accordance with one or more embodiments. In this example, the adjustable, low voltage regulator feeds a high voltage DC/DC converter, such as converter 614 in FIG. 6, to allow a real-time adjustable high voltage level. In this example, a linear regulator with resistor-adjusted output voltage is used to drive a DC/DC converter whose output voltage is proportional to its input voltage. Additionally, the resistor path that controls the output voltage of the linear regulator contains an electrically-controlled potentiometer with a serial interface. This allows a microcontroller to serially set the resistance of the feedback branch and control the output of the linear regulator which in turn drives the DC/DC converter and controls the actuator drive voltage. It is to be appreciated there are many other ways to use regulated and unregulated supplies to provide the necessary operating voltage, and also that an adjustable high voltage rail is not necessary for every implementation, although if adjustability is required there are additionally many ways of providing adjustability.



FIG. 8 along with FIG. 10 illustrate a USB device that can allow real-time changes of haptic profiles and can act as an HID compliant keyboard. This circuit is an example implementation of one way to provide the system user with a means to interact with the haptic device. A USB device is provided which defines two interfaces. One is a standard HID keyboard, the other is a generic HID device functioning as a haptic customization engine. The standard keyboard interface allows the key presses on the haptic device to register on the host as keypresses of a keyboard. The haptic customization engine interface allows host software to send a variety of commands to define, redefine, modify, select or read haptic profile information that is stored/used in the haptic device.



FIG. 9 illustrates an example schematic, combined with FIG. 11, of the high-side and low-side switches used to drive the actuator. The components, including the optoisolators, constitute but one implementation. Accordingly, other implementations can be utilized without departing from the spirit and scope of the claimed subject matter.



FIG. 10 illustrates an example schematic of a microcontroller and supporting hardware used to implement the haptic customization engine, the haptic engine, the USB interface, the key scan circuitry, and the EEPROM. Other circuitry can be used without departing from the spirit and scope of the claimed subject matter.



FIG. 11 illustrates the details of FIG. 9. This schematic is an example implementation of a solid state switch stacking scheme that allows inexpensive, low voltage parts to be used together in order to switch high voltage. This particular stacking scheme utilizes capacitor coupled MOSFET gates and is uniquely designed for this switching application to be very power efficient during idle and active state due to the elimination of resistors while providing reliable switching function to capacitive loads which include many electrically-deformable devices such as, by way of example and not limitation, electroactive polymers, piezo materials, and electrostatic actuators. It is to be appreciated and understood that capacitive coupling is not the only way to stack switches for increased voltage handling, nor are stacked switches the only way to handle switching of high voltage.


Example Touch Surfaces


In the discussion above, an example touch surface was illustrated in the form of a single touch surface. It is to be appreciated and understood that multiple different touch surfaces can be provided on a single device.


As an example, consider FIG. 12a which illustrates an example touch surface layout for a device 1200 which may reside in the form of a hand-held device. In this example, multiple different touch surfaces are provided including a display surface 1202 and a typing surface 1204. The display surface 1202 can be utilized to display images for the user such as, for example, webpages to which a user may browse. Typing surface 1204 can be used to render a virtual keyboard or other type of virtual input mechanism for a user to provide input to the device. Both the display surface 1202 and the typing surface 1204 can be configured to provide haptic feedback as described above. For example, each individual surface may have its own actuator array to provide individualized haptic feedback to the user.



FIG. 12
b illustrates an example touch surface layout for a device that includes a control panel 1250. The device can be any suitable type of device such as, by way of example and not limitation, a printer, a copier, a multifunction peripheral device, a vending machine, an ATM machine, appliance white goods, GPS devices, portable gaming consoles, touch pads, mouse buttons, portable media players (MP3), Medical equipment, personal computing devices, automotive dash boards, and the like.


In this particular example, control panel 1250 includes display surface 1252, typing surface 1254, display surface 1256, and typing surface 1258. Any suitable number of surfaces can be provided and can operate as described above.


Example Method



FIG. 13 is a flow diagram that describes steps a method in accordance with one embodiment. The method can be implemented in connection with any suitable hardware, software, firmware or combination thereof. In at least some embodiments, the method can be implemented by a system, such as those systems shown and described above. It is to be appreciated and understood that the described method can be implemented by systems other than those described above without departing from the spirit and scope of the claimed subject matter.


Step 1300 detects user movement or physical engagement of a touch surface. An example of how user movement can be detected is provided above. In addition, various examples of touch surfaces are provided above as well. Step 1302 activates electrically-deformable material responsive to detecting the user movement or physical engagement. Examples of electrically-deformable material are provided above. Step 1304 moves the touch surface responsive to activation of the electrically-deformable material. Examples of how a touch surface can be moved are provided above. It is to be appreciated and understood that any suitable movement of the touch surface can occur. For example, the touch surface can be moved in a single direction. Alternately or additionally, the touch surface can be moved in multiple different directions along different movement vectors. For example, one movement vector can be away from the user and another movement vector can be toward the user.


By moving the touch surface in accordance with the embodiments described above, haptic feedback can be provided to the user to provide the user with a realistic-feeling that enhances the user's experience.


Conclusion


In one or more embodiments, an electronic device includes a touch surface that can be physically engaged by a user. The touch surface is operably connected to an actuator arm which, in turn, is connected to an actuator array. Drive electronics sense a user's movement relative to the touch surface and, responsively, drive the actuator array effective to move the actuator arm and, in turn, provide haptic feedback to the user through the touch surface.


Although the subject matter has been described in language specific to structural features and/or methodological acts, it is to be understood that the subject matter defined in the appended claims is not necessarily limited to the specific features or acts described above. Rather, the specific features and acts described above are disclosed as example forms of implementing the claims.

Claims
  • 1. An electronic device comprising: a device housing;a touch surface mounted within the device housing, the touch surface comprising a generally planar top surface;a floating battery arrangement moveably mounted within the device housing;an actuator array fixedly mounted to the floating battery arrangement wherein the floating battery arrangement is moved under the influence of the actuator array; anddrive electronics operably coupled to the actuator array and configured to electrically drive the electrically-deformable material, responsive to sensing a user's movement relative to the touch surface, effective to move the touch surface along one or more vectors that are generally parallel to a plane defined by the top surface.
  • 2. The device of claim 1, wherein the electrically-deformable material comprises an electro-active polymer.
  • 3. The device of claim 1, wherein the actuator array comprises an electrically-deformable material.
  • 4. The device of claim 1, wherein the actuator array comprises an electrically-deformable material comprising an electro-active polymer.
  • 5. The device of claim 1, wherein the floating battery arrangement can be moved in multiple different directions along different movement vectors.
  • 6. A method comprising: detecting a user's movement relative to, or physical engagement of, a touch surface on an electronic device; andresponsive to detecting, activating an electrically-deformable material within the electronic device effective to move a floating battery arrangement of the electronic device.
  • 7. The method of claim 6, wherein the touch surface comprises a generally planar top surface and said activating is effective to move the floating battery arrangement in one or more directions generally parallel to a plane defined by the generally planar top surface.
RELATED APPLICATION

This application claims priority to U.S. Provisional Application No. 61/024,411, filed on Jan. 29, 2008, the disclosure of which is incorporated by reference herein.

US Referenced Citations (98)
Number Name Date Kind
4200778 Bovio et al. Apr 1980 A
4529849 Kamei et al. Jul 1985 A
5057657 Skulic Oct 1991 A
5239152 Caldwell et al. Aug 1993 A
5612692 Dugas et al. Mar 1997 A
5676476 Uke Oct 1997 A
5729222 Iggulden et al. Mar 1998 A
5943233 Ebina et al. Aug 1999 A
6003390 Cousy Dec 1999 A
6218966 Goodwin et al. Apr 2001 B1
6684166 Bellwood et al. Jan 2004 B2
6791480 Uke Sep 2004 B1
7166795 Lengeling Jan 2007 B2
7182691 Schena Feb 2007 B1
7196688 Schena Mar 2007 B2
7339572 Schena Mar 2008 B2
7342573 Ryynanen Mar 2008 B2
7741979 Schlosser et al. Jun 2010 B2
7791588 Tierling et al. Sep 2010 B2
7834857 Prados Nov 2010 B2
8199033 Peterson et al. Jun 2012 B2
8203531 Peterson et al. Jun 2012 B2
8248277 Peterson et al. Aug 2012 B2
8248278 Schlosser et al. Aug 2012 B2
20020054060 Schena May 2002 A1
20020149561 Fukumoto et al. Oct 2002 A1
20030208324 Bellwood et al. Nov 2003 A1
20030209131 Asashi Nov 2003 A1
20040031673 Levy Feb 2004 A1
20040085716 Uke May 2004 A1
20040130526 Rosenberg Jul 2004 A1
20040252104 Nakamura et al. Dec 2004 A1
20050017947 Shahoian et al. Jan 2005 A1
20050134561 Tierling et al. Jun 2005 A1
20050157893 Pelrine et al. Jul 2005 A1
20050204906 Lengeling Sep 2005 A1
20050237309 Sharma Oct 2005 A1
20060187201 Rosenberg et al. Aug 2006 A1
20060256075 Anastas et al. Nov 2006 A1
20060261983 Griffin et al. Nov 2006 A1
20060267949 Rosenberg Nov 2006 A1
20060279538 Chang et al. Dec 2006 A1
20060290662 Houston et al. Dec 2006 A1
20070091070 Larsen et al. Apr 2007 A1
20070146317 Schena Jun 2007 A1
20070152974 Kim Jul 2007 A1
20070193436 Chu Aug 2007 A1
20070203011 Gudgel et al. Aug 2007 A1
20070234887 Sawada et al. Oct 2007 A1
20070234890 Yamashita Oct 2007 A1
20070236449 Lacroix Oct 2007 A1
20070236450 Colgate et al. Oct 2007 A1
20070251810 Corcoran et al. Nov 2007 A1
20070285284 Matteo et al. Dec 2007 A1
20080010593 Uusitalo et al. Jan 2008 A1
20080042978 Perez-Noguera Feb 2008 A1
20080060856 Shahoian et al. Mar 2008 A1
20080062144 Shahoian et al. Mar 2008 A1
20080062145 Shahoian et al. Mar 2008 A1
20080083314 Hayashi et al. Apr 2008 A1
20080084384 Gregorio et al. Apr 2008 A1
20080092720 Yamashita et al. Apr 2008 A1
20080197901 Cruz-Hernandez et al. Aug 2008 A1
20080198139 Lacroix et al. Aug 2008 A1
20080223706 Hagiwara et al. Sep 2008 A1
20080251364 Takala et al. Oct 2008 A1
20080289952 Pelrine et al. Nov 2008 A1
20080303782 Grant et al. Dec 2008 A1
20090002199 Lainonen et al. Jan 2009 A1
20090002205 Klinghult et al. Jan 2009 A1
20090008234 Tolbert et al. Jan 2009 A1
20090072662 Sadler et al. Mar 2009 A1
20090085878 Heubel et al. Apr 2009 A1
20090085882 Grant et al. Apr 2009 A1
20090106655 Grant et al. Apr 2009 A1
20090135142 Fu et al. May 2009 A1
20090160763 Cauwels et al. Jun 2009 A1
20090167704 Terlizzi et al. Jul 2009 A1
20090174672 Schmidt Jul 2009 A1
20090178913 Peterson et al. Jul 2009 A1
20090188374 Folkesson Jul 2009 A1
20090189790 Peterson Jul 2009 A1
20090189873 Peterson Jul 2009 A1
20090210568 Peterson et al. Aug 2009 A1
20090231113 Olien et al. Sep 2009 A1
20090231277 Peterson Sep 2009 A1
20090267921 Pryor Oct 2009 A1
20090303187 Pallakoff Dec 2009 A1
20100045612 Molne Feb 2010 A1
20100108408 Colgate et al. May 2010 A1
20100130280 Arezina et al. May 2010 A1
20100160016 Shimabukuro et al. Jun 2010 A1
20100171715 Peterson et al. Jul 2010 A1
20100177050 Heubel et al. Jul 2010 A1
20110073454 Chen et al. Mar 2011 A1
20110107958 Pance et al. May 2011 A1
20110148607 Zeleny Jun 2011 A1
20110227763 Schlosser et al. Sep 2011 A1
Foreign Referenced Citations (14)
Number Date Country
19704253 Aug 1998 DE
10126670 Dec 2002 DE
2005002417 Apr 2005 DE
2004005501 Aug 2005 DE
0654727 May 1995 EP
1310860 May 2003 EP
1548776 Jun 2005 EP
61000825 Jan 1986 JP
WO-0191100 Nov 2001 WO
WO-2009043605 Apr 2009 WO
WO 2009097358 Aug 2009 WO
WO-2009097359 Aug 2009 WO
WO 2009097361 Aug 2009 WO
WO-2009114827 Sep 2009 WO
Related Publications (1)
Number Date Country
20090189873 A1 Jul 2009 US
Provisional Applications (1)
Number Date Country
61024411 Jan 2008 US