1. Field of the Invention
The present invention relates generally to touch pads that provide visual feedback. More particularly, the present invention relates to illuminated touch pads that use light to provide feedback.
2. Description of the Related Art
There exist today many styles of input devices for performing operations in a computer system. The operations generally correspond to moving a cursor and/or making selections on a display screen. By way of example, the input devices may include buttons or keys, mice, trackballs, touch pads, joy sticks, touch screens and the like.
Touch pads, in particular, are becoming increasingly popular because of their ease and versatility of operation as well as to their declining price. Touch pads allow a user to make selections and move a cursor by simply touching an input surface via a finger or stylus. In general, the touch pad recognizes the touch and position of the touch on the input surface and the computer system interprets the touch and thereafter performs an action based on the touch event.
Touch pads typically include an opaque touch panel, a controller and a software driver. The touch panel registers touch events and sends these signals to the controller. The controller processes these signals and sends the data to the computer system. The software driver translates the touch events into computer events.
Although touch pads work well, improvements to their form feel and functionality are desired. By way of example, it may be desirable to provide visual stimuli at the touch pad so that a user can better operate the touch pad. For example, the visual stimuli may be used (among others) to alert a user when the touch pad is registering a touch, alert a user where the touch is occurring on the touch pad, provide feedback related to the touch event, indicate the state of the touch pad, and/or the like.
The invention relates, in one embodiment, to an illuminated input device. The illuminated input device includes an object sensing mechanism capable of sensing a user input over an input surface. The illuminated input device also includes a visual feedback system configured to illuminate the input surface in association with a user input.
The invention relates, in another embodiment, to a method of operating an input device. The method includes sensing an object over an input surface. The method also includes and illuminating at least a portion of the input surface when an object is sensed.
The invention relates, in another embodiment, to a method of operating an input device. The method includes illuminating at least a portion of an input surface when an object is detected over the input surface. The method also includes adjusting the illumination when the object is moved over the input surface.
The invention relates, in another embodiment, to a method of operating an input device. The method includes detecting a user input over the input surface. The method also includes determining an input state of the input device based on the user input. The method additionally includes illuminating the input surface based on the input state of the input device. Each input state having a different illumination profile.
The invention may best be understood by reference to the following description taken in conjunction with the accompanying drawings in which:
The present invention pertains to improved feedback mechanisms for touch pads. One aspect relates to devices capable of illuminating the touch sensitive surface of the touch pad. Not just in backlighting so that the user knows where the touch pad is located in low light conditions, but also to give other feedback related to how the touch pad is being used. Another aspect relates to methods for providing feedback at the touch pad. For example, changing intensity or color based on motion characteristics- and/or pressure, providing an illumination point that follows a finger as it is moved about the touch sensitive surface, showing different states with varying levels of brightness or color, etc.
Embodiments of the invention are discussed below with reference to
In order to generate inputs as for example initiating commands, making selections or tracking, the input device 10 includes an object sensing mechanism 12 configured to detect one or more objects in close proximity to and/or in contact with an input surface 13. The object sensing mechanism 12 may be based on proximity sensing and/or touch sensing.
In the case of proximity sensing, the input surface 13 may be the surface directly underneath a proximity sensing field. The object sensing mechanism 12 generates input signals when an object such as a finger (or stylus) is moved above the input surface and within the sensing field (e.g., x and y plane), from an object holding a particular position above the surface and within the sensing field and/or by an object moving through or in and out of the sensing field (e.g., z direction). Proximity detection may be based on technologies including but not limited to capacitive, electric field, inductive, hall effect, reed, eddy current, magneto resistive, optical shadow, optical visual light, optical IR, optical color recognition, ultrasonic, acoustic emission, radar, heat, sonar, conductive or resistive and the like.
In the case of touch sensing, the input surface 13 may be a touch surface that is sensitive to direct physical touch. The object sensing mechanism 12 generates input signals when an object in contact with the input surface 13 is moved across the input surface (e.g., x and y plane), from an object holding a particular position on the input surface and/or by an object tapping on the input surface. Touch sensing may be based on technologies including but not limited to resistive, capacitive, infrared and surface acoustic wave. Examples of touch sensing devices that utilize these technologies include touch pads, touch screens, and the like.
To elaborate, the sensing region, i.e., input surface or the sensing field above the input surface, is typically divided into several independent and spatially distinct sensing points, nodes or regions. The sensing points, which are typically hidden from view, are dispersed about the sensing region with each sensing point representing a different position in the sensing region. The sensing points may be positioned in a grid or a pixel array where each pixilated sensing point is capable of generating a signal. In the simplest case, a signal is produced each time an object is positioned over a sensing point. When an object is placed over multiple sensing points or when the object is moved between or over multiple sensing points, multiple signals can be generated. The sensing points generally map the sensing region into a coordinate system such as a Cartesian coordinate system, a Polar coordinate system or some other coordinate system. Furthermore, the touch sensing means may be based on single point sensing or multipoint sensing. Single point sensing is capable of only distinguishing a single object at any given time, while multipoint sensing is capable of distinguishing multiple objects at the same time.
The input device 10 also includes a visual feedback system 14 configured to output visual effects at the input surface 13 in association with the object sensing system 12. The visual feedback system 14 is dedicated to enhancing the operation of the input device 10 by providing visual feedback to the user when making touch or proximity inputs via the object sensing system 12. For example, the visual effects may be used to indicate the location of the input surface 13 thereby making inputting easier for the user (e.g., backlighting).
Alternatively or additionally, the visual effects may be used during and after the input event to dynamically indicate characteristics associated with the input events. The characteristics may for example include when and where and the number of inputs being made relative to the input surface 13. This type of feedback also improves inputting by providing visual cues to the user about the input device 10 as the input device 10 is used.
Alternatively or additionally, the visual effects may be used before an input event to invoke the user to perform a particular input event at the input surface 13. This type of feedback also improves inputting by helping the user make appropriate inputs or helping them learn input gestures (e.g., timing, location and movements).
In most cases, the visual effects (outputs) are linked or tied to and associated with the input events being performed. Although capable of, they typically do not provide external outputs associated with events occurring outside the input device 10. That is, the visual effects typically do not operate separately as an output for any device other than the input device 10 (e.g., the visual feedback system should not be considered a separate display).
The visual feedback system 14 includes one or more visual changing elements 15. The visual changing elements 15 may be separate of integral with the sensing elements of the object sensing system 12. In some cases, the one or more of the visual changing elements 15 may be mapped, associated with or tied to one or more of the sensing nodes of the object sensing system. The number of visual changing elements tied to a particular sensing node may be less than, equal to, or more than the number of sensing nodes.
The resolution of the visual changing elements 15 can be widely varied. In one embodiment, the resolution of the visual changing elements 15 is greater than the sensing nodes (e.g., the number of visual changing elements is greater than the number of sensing nodes). In another embodiment, the resolution of the visual changing elements 15 is substantially equal to the resolution of the sensing nodes (e.g., the number of visual changing elements is substantially equal to the number of sensing nodes). In yet another embodiment, the resolution of the visual changing elements 15 is less than the resolution of the sensing nodes (e.g., the number of visual changing elements is less than the number of sensing nodes). The resolution generally depends on the needs of the particular input device. In some cases, high resolution is need to create dynamic visual effects such a graphical effects. In other cases, only low resolution is required, as for example to visually change a region as large as a finger.
The visual feedback system 14 may be widely varied. In one embodiment, the visual feedback system 14 is a light based system that illuminates the input surface 13. In this embodiment, the visual changing elements 15 are embodied as light emitting devices. The light emitting devices may include one or more light sources, and a light distribution system for distributing the light at the input surface 13. In some cases, the light from the light sources may be diffused so that the input surface 13 emits a characteristic glow (not a precise point of light, but rather a blurred glowing effect or phosphorous glow). That is, the input surface can generate glowing special effects that may for example provide backlighting to the input surface 13 and/or provide an outline, trace or shadow of the sensed object on the input surface 13. The glowing special effects may even indicate a state of the input device as for example when the input device is in a tracking state or gesture state.
Alternatively, the visual feedback system may be a graphically based system that generates graphics at the input surface. LCDs, OLEDs and electronic inks are examples of graphically based systems. These devices however can be cost prohibitive and more complex to implement when compared to light based systems.
The input device 10 also includes a controller 16 that is operatively coupled to the object sensing device 12 and visual feedback system 14. The controller 16 monitors signals generated at the input surface 13 and sends corresponding control signals associated therewith to the host device, which interprets the signals in accordance with its programming (e.g., input events). The controller 16 also generates visual effect commands for controlling the visual effects outputted by the visual feedback system 14. Single or multiple commands can be generated to change one, some or all of the visual changing elements at the same time. Further, the commands may be based on the signals generated via the object sensing device 12.
In one embodiment, the controller 16 may instruct the visual changing elements to change in a non-trivial manner in the region of the detected object in order to indicate a location of the object relative to the input surface 13. In another embodiment, the commands may be based on instructions from the host device. For example, the host device may instruct the input device 10 to backlight the input surface 13 or alternatively to alter the input surface in such a way as to prompt the user to perform a particular event relative to the input surface (e.g., the host device may use the input surface to teach the user how to perform a particular gesture).
In one embodiment, the visual effects are performed in the region of the detected object. For example, in the case of a light based system, the area under and/or around the detected object may be illuminated. As should be appreciated, the visual effects can be made to follow the object as it is moved around the sensing region. In fact in some cases, the visual effects may include a leading edge, a body, and/or a trailing edge. The leading edge indicates where the object is directed, the body indicates the current location of the object, and the trailing edge indicates where the object has been.
In another embodiment, the visual effects are performed to indicate the state of the object sensing event. For example, if one object is detected, a first visual effect may be performed, and if multiple objects are detected, a second visual effect may be performed. This may be beneficial in cases where single objects are used for tracking and multiple objects are used for gesturing. See for example U.S. patent application Ser. No. 10/903,964, which is herein incorporated by reference.
In block 56, control signals are generated based on the detected objects. This may for example be accomplished with the controller described above. The signals are reported to a host device as an input event, and the host device interprets the signals in accordance with its programming.
In one embodiment, the visual effects are based on a control signal from the host device. For example, the host device may instruct the visual feedback system to output visual effects to encourage a user to place an object at a particular location at the input surface or to perform a gesture during a particular operation in the host device (e.g., training sequence).
Following block 72, the method proceeds to blocks 74 and 76. In block 74, the sensing region is monitored. This may for example be accomplished with the object sensing device described above. In block 76, control signals are generated when objects are detected in the sensing region. This may for example be accomplished with the controller described above. The signals may be used by the host device as an input event.
The touch sensing device may be widely varied. The touch sensing device may for example be selected from any of those used for touch screens. An example of a touch screen that may be used can be found in U.S. patent application Ser. No. 10/840,862, which is herein incorporated by reference.
The light panel may also be widely varied. In one embodiment, the light panel is a pixilated light device that includes a plurality of light sources that are distributed over an extended area such as the touch sensitive surface of the touch sensing device. The light panel may include a plurality of light emitting diodes (LEDs) that are laid out in a matrix such as rows and columns. Any number of LEDs may be used. The number generally depends on the desired resolution of the light panel. In the simplest case, LEDs are placed next to or adjacent one another in rows and columns on a PCB that is sized similarly to the touch sensing device (e.g., covers the same area).
The touch sensing device may be widely varied. The touch sensing device may for example be selected from any of those used for touch pads or touch screens. An example of a touch pad that may be used can be found in U.S. patent application Ser. Nos. 10/188,182, 10/722,948 and 10/643,256, all of which are herein incorporated by reference.
The light panel may also be widely varied. Unlike the light panel discussed in
In one embodiment, the touch screen 152 includes a plurality of capacitance sensing nodes. The capacitive sensing nodes may be widely varied. For example, the capacitive sensing nodes may be based on self-capacitance or mutual capacitance. In self-capacitance, the “self” capacitance of a single electrode is measured as for example relative to ground. In mutual capacitance, the mutual capacitance between at least first and second electrodes is measured. In either cases, each of the nodes works independent of the other nodes so as to produce simultaneously occurring signals representative of different points on the touch screen 152.
In order to produce a transparent touch screen 152, the capacitance sensing nodes may be formed with a transparent conductive medium such as indium tin oxide (ITO).
In self-capacitance sensing arrangements, the transparent conductive medium is patterned into spatially separated electrodes and traces. Each of the electrodes represents a different coordinate and the traces connect the electrodes to a capacitive sensing circuit. The coordinates may be associated with Cartesian coordinate system (x and y), Polar coordinate system (r, θ) or some other coordinate system. During operation, the capacitive sensing circuit monitors changes in capacitance that occur at each of the electrodes. The positions where changes occur and the magnitude of those changes are used to help recognize the touch events. A change in capacitance typically occurs at an electrode when a user places an object such as a finger in close proximity to the electrode, i.e., the object steals charge thereby affecting the capacitance.
In mutual capacitance, the transparent conductive medium is patterned into a group of spatially separated lines formed on two different layers. Driving lines are formed on a first layer and sensing lines are formed on a second layer. Although separated by being on different layers, the sensing lines traverse, intersect or cut across the driving lines thereby forming a capacitive coupling node. The manner in which the sensing lines cut across the driving lines generally depends on the coordinate system used. For example, in a Cartesian coordinate system, the sensing lines are perpendicular to the driving lines thereby forming nodes with distinct x and y coordinates. Alternatively, in a polar coordinate system, the sensing lines may be concentric circles and the driving lines may be radially extending lines (or vice versa). The driving lines are connected to a voltage source and the sensing lines are connected to capacitive sensing circuit. During operation, a current is driven through one driving line at a time, and because of capacitive coupling, the current is carried through to the sensing lines at each of the nodes (e.g., intersection points). Furthermore, the sensing circuit monitors changes in capacitance that occurs at each of the nodes. The positions where changes occur and the magnitude of those changes are used to help recognize the multiple touch events. A change in capacitance typically occurs at a capacitive coupling node when a user places an object such as a finger in close proximity to the capacitive coupling node, i.e., the object steals charge thereby affecting the capacitance.
Referring now to the light panel 154, the light panel 154 includes a light emitting surface that is typically divided into several independent and spatially distinct illumination points, nodes or regions 156. The illumination points 156 are dispersed about the light emitting surface with each illumination point 156 representing a different position in the light emitting surface. The illumination points 156 may be positioned in a grid or a pixel array where each pixilated illumination point is capable of emitting light. The illumination points 156 generally map the illumination region into a coordinate system such as a Cartesian coordinate system, a Polar coordinate system or some other coordinate system. In some cases, the illuminations points may be laid out in a pattern similar to the sensing points of the touch panel 152 (e.g., same coordinate system, same number of points). In other cases, the illumination points 156 may be laid out in a pattern that is different than the sensing points of the touch panel 152 (e.g., different coordinate system, different number of points).
The light panel 154 may be widely varied. In the illustrated embodiment, the illumination points 156 are embodied as individual light emitting diodes that are placed in a grid like manner thereby forming a pixilated illumination area, i.e., each of the light emitting diodes forms an illumination node. The grid may be oriented rows and columns (x and y) or angular/radial segments (as shown). Furthermore, the LEDs are attached to the printed circuit board 160 and operatively coupled to the controller 158 located on the backside of the printed circuit board 160.
The touch screen 152 is also operatively coupled to the controller 158 as for example using a flex circuit that attached to the printed circuit board 160. During operation, the controller 158 monitors the changes in capacitance and generates control signals based on these changes. The controller 158 also separately adjusts the intensity of each of the LEDs to illuminate portions or all of the touch screen 152 in a controlled manner. That is, the light panel 154 can produce any number of various light effects by selectively controlling the intensities of the LED's via the controller 158. Because the touch screen 154 is translucent, the light can be seen through the touch screen 154.
In some cases, the touch pad 150 may further include a light diffuser 162. The light diffuser 162 is configured to diffuse the light being emitted by the light panel 154. This may be done to normalize the light intensity of the LEDs, to produce a characteristic glow at the input surface, and/or to hide the physical parts of the touch pad located underneath the light diffuser.
Although the light diffuser 162 can include color components, in most cases, the light diffuser appears as a white or semi transparent white material. When embodied with white elements, the light diffuser 162 takes on the color of light emitted by the LEDs. Generally speaking, the light diffuser 162 is positioned somewhere between the LEDs and the input surface. More particularly, the light diffuser 162 can be placed above, within or underneath the touch screen. For example, a light diffuser 162 can be placed on the upper surface, lower surface, or in the layers of the touch screen. Alternatively or additionally, the light diffuser 162 may be integrated with or attached to the light panel or even be a separate component disposed between the light panel 154 and touch screen 152 (as shown).
The light diffuser 162 may be embodied in many different forms including for example surface treatments on one or more layers of the touch screen, additives in one or more layers of the touch screen, an additional layer in the touch screen, rigid plastic inserts disposed above or below the touch screen, flexible labels disposed above or below the touch screen, and the like. The light diffuser 162 may even be the ITO coating used to form the sensing components of the touch screen (e.g., the greater the density of the ITO coating, the greater the amount of light that is diffused).
In the illustrated embodiment, the light diffuser 162 is a plastic insert that includes light scattering additives. Furthermore, the light diffuser 162 is disposed between the light panel 154 and the touch screen 152.
It should be pointed out that LED's offer many advantages over other light sources. For example, LED's are relatively small devices that are energy efficient and long lasting. LED's also run relatively cool and are low in cost. Furthermore, LED's come in various colors such as white, blue, green, red and the like. The pixilated LEDs may be configured to emit that same color of light or a different color of light.
Furthermore, although shown as single LEDs, it should be noted that the LEDs may be embodied as an integrated array of LEDs that are grouped together as for example an array of red, blue, green and/or white LEDs that cooperate to produce a resultant color (via color mixing). The resultant color may be a wide range of colors, as for example, a majority of the colors from the color spectrum. During operation, the controller can produced almost any color by adjusting the intensity of each of the colored LED's. By way of example, in order to produce the highest shade of red, the intensities of the green and blue are reduced to zero intensity and the intensity of the red is increased to its peak intensity. The highest shades of green and blue can be implemented in a similar manner. In addition, in order to produce a shade of red and green, the intensities of the green and red are increased to levels above zero intensity while the intensity of blue is reduced to zero intensity. Shades of green and blue and blue and red can be implemented in a similar manner. Furthermore, in order to produce shades of white, the intensities of the red, green and blue are increased to the same levels above zero intensity, or alternatively the red, green and blue LED's are turned off and a white LED is turned on.
Although the integrated LED array is described as using the three primary colors, it should be noted that this is not a limitation and that other combinations may be used. For example, the integrated LED array may be configured to include only two of the primary colors or it may only include LED's with a single color.
When the LEDs are capable of generating any color, unique input surfaces can be produced. By way of example, the touch pad can produce an input surface with rainbow stripes, different colored spots, different colored quadrants or sections and the like. The touch pad can also produce an input surface that has a dynamically changing pattern. This is typically accomplished by activating distinct LED's at different times or by adjusting the intensities of distinct LED's at different times.
The electrode layer 204 includes a plurality of spatially separated electrodes 205 configured to detect changes in capacitance at an upper surface 208 of the light panel 202. Each of the electrodes 205 is operatively coupled to a controller 210 located on the backside of the printed circuit board 206. During operation, the controller 210 monitors the changes in capacitance and generates control signals based on these changes.
The light panel 202 includes a light distribution panel 212 disposed over the electrode layer 204 and one or more side mounted light emitting diodes 214 disposed around the periphery of the light distribution panel 212. The side mounted light emitting diodes 214 are configured to direct light into a different portion of the light distribution panel 212. Alternatively, a light pipe may be used to direct light from an LED located away from the light distribution panel 212. The light distribution panel 212 is configured to redirect the light made incident thereon via the light emitting diodes 214 to an upper surface of the light distribution panel 212 thereby illuminating the touch pad surface 201. The light distribution panel 212 is also configured to serve as a dielectric layer that covers the electrode layer 204 in order to help form the capacitance sensing circuit of the touch pad 200.
As shown, the LEDs 214 are attached to the printed circuit board 206 and operatively coupled to the controller 210 located on the backside of the printed circuit board 206. During operation, the controller 210 selectively adjusts the intensity of each of the LEDs 214 to illuminate portions of or all of the light distribution panel 212 in a controlled manner.
Although shown as single LEDs, the LEDs may be embodied as an array of LEDs as for example an array of red, blue and green LEDs. Arrayed LEDs such as this may be capable of generating most colors in the color spectrum.
The light distribution panel 212 can be widely varied. In one embodiment, the light distribution panel 212 is a separate component disposed within the housing 211 of the touch pad 200. For example, the light distribution panel 212 is inserted within an opening in the housing 211 (as shown). In this arrangement, it may be preferable to place the upper surface of the light distribution panel 212 flush with or recessed below the outer surface of the housing 211. Furthermore, in order to provide a tight fit that limits dust and particles from entering the touch pad 200, the light distribution panel 212 may include edges that extend over the outer surface of the housing 211.
In another embodiment, the light distribution panel 212 is an integral part of the housing 211. For example, the housing 211 is formed from a transparent or semi-transparent material. This particular embodiment provides a continuous surface without gaps or breaks, which can be aesthetically pleasing to the user.
In either embodiment, the light distribution panel 212 typically includes a portion 213 that extends below the inner surface of the housing 211. This portion 213 provides a light receiving area at the sides of the light distribution panel 212 for receiving light emitted by the side mounted LED's 214.
The light distribution panel 212, which can be formed from a single or multiple layers, is typically formed from translucent or semi-translucent dielectric materials including for example plastic materials such as polycarbonate, acrylic or ABS plastic. It should be appreciated, however, that these materials are not a limitation and that any optically transmittable dielectric material may be used.
In most cases, the light distribution panel 212 or some other component of the touch pad 200 includes light diffusing elements to diffuse the light made incident thereon in order to normalize the light intensity of the LEDs, to produce a characteristic glow at the input surface, and/or to hide the physical parts of the touch pad located underneath the input surface. The light diffusing elements may be provided on an inner surface, outer surface or they may be embedded inside the light distribution panel 212. Additionally or alternatively, the light diffusing elements can also be applied to a separate optical component disposed above the light distribution panel 212.
In one embodiment, the light diffusing element is an additive disposed inside the light distribution panel 212. For example, the light distribution panel 212 may include a plurality of light scattering particles dispersed between the top and bottom surfaces of the light distribution panel. When the light is made incident on the inner surface, it is transmitted through the light distribution panel 212 until is intersects a light scattering particle disposed inside the panel. After intersecting the light scattering particle, the light is scattered outwards in a plurality of directions, i.e., the light is reflected off the surface and/or refracted through the light scattering particle thereby creating the characteristic glow. By way of example, the light scattering particles may be formed from small glass particles or white pigments. Furthermore, by changing the amount of light scattering particles disposed in the panel, the characteristics of the glow can be altered, i.e., the greater the particles the greater the light scattering.
In another embodiment, the light diffusing element is a layer, coating and/or texture that is applied to the inner, side or outer surfaces of the panel 212. For example, the panel 212 may include a light scattering coating or a light scattering texture disposed on the side or outer surface of the panel. By way of example, the light scattering coating may be a paint, film or spray coating. In addition, the light scattering texture may be a molded surface of the wall or a sandblasted surface of the panel. When light is made incident on the inner or outer surface, it intersects the light scattering coating or texture applied on the surface. After intersecting the light scattering coating or the light scattering texture, the light is scattered outwards in a plurality of directions, i.e., the light is reflected off the surface and/or refracted through the light scattering particle thereby creating a characteristic glow.
In the illustrated embodiment, the light diffusing element is embodied as a light diffusing label 216. The light diffusing label 216 is at least adhered to the top surface of the light distribution panel 212. In some cases, the label 216 may even extend over and be adhered to a top edge of the housing wall 211. In cases such as this, the light diffusing label 216 may even be placed in a pocket formed by recesses 217 at the top edge of the housing wall 211 in order to make the top surface of the light diffusing label 216 flush with the external surface of the housing wall 211. The label 216 can have a graphic printed thereon, can have multiple colors and can have varying thickness to assist in controlling the intensity and color of the illumination. The label 216 may be formed from transparent or semitransparent dielectric materials such as Mylar or Polycarbonate or any other dielectric material that is thin, optically transmittable and includes some sort of light diffusing means.
Further, the light distribution panel 212 may be configured as a single node, or it may be broken up into plurality of distinct nodes 218, each of which includes its own dedicated light emitting diode for individual illumination thereof. During operation, when light is released by a light emitting diode 214, the light is made incident on the side of the light distribution panel 212 at the node 218. The node 218 redirects and transmits the light from its side to an upper surface of the node 218. In order to prevent light bleeding between adjacent nodes 218, each node 218 may be optically separated by a reflecting or masking region disposed therebetween.
Each of the nodes 218 may be formed from a solid piece of material or it may be formed from a combination of elements. In one embodiment, each of the nodes 218 is formed from a translucent or semi-translucent plastic insert that when combined with the other inserts forms the light distribution panel 212. In another embodiment, each of the nodes 218 is formed from a bundle of fiber optic strands.
The configuration of the nodes 218 including layout, shape and size may be widely varied. With regards to layout, the nodes 218 may be based on a Polar or Cartesian coordinate system (or some other coordinate system). With regards to shape, any shape including for example standard shapes such as circles, squares, rectangles, triangles, may be used. With regards to size, the nodes 218 may be larger than a finger or stylus, about the same size as a finger or stylus, or smaller than a finger or stylus. In one embodiment, the nodes 218 are set up similarly to the electrodes 205 of the electrode layer 204, i.e., the nodes 218 have generally the same layout, number, size and shape as the electrodes 205. In another embodiment, the nodes are set up differently. For example, the nodes 218 may have a different layout, different number, different shape and/or different size when compared to the electrodes 205.
In the illustrated embodiment, the touch pad 200 is circular and the nodes 218 are embodied as distinct angular segments (e.g., pie shaped). Any number of angular segments may be used. The number generally depends on the desired resolution of the illuminating surface. In this particular embodiment, the resolution of the light panel 202 is low and therefore each of the angular segments cover a plurality of sensing electrodes 205.
In one embodiment, all the LEDs 214 are powered at the same time to produce a fully illuminated touch pad 200. This may be analogous to backlighting. In another embodiment, the LEDs 214 are powered in accordance with the capacitance changes measured by each of the electrodes 205. For example, the segments above the detected area may be illuminated while the segments above the undetected areas may be turned off. This provides indication to the user as to their exact location of the touch surface. In yet another embodiment, selected segments may be illuminated to encourage a user to place their finger in a particular area of the touch pad.
Although only a single light panel 202 is shown, it should be appreciated that this is not a limitation and that additional light panels may be used. For example, one or more light panels may be further positioned underneath the first light panel described above. In one embodiment, each light panel in a group of light panels is configured to distribute a different color. For example, three light panels—including a red, green and blue light panel may be used. Using this arrangement, different colored segments may be produced. By controlling their intensity, almost any color can be produced (mixed) at the touch surface. In another embodiment, each light panel in the group of light panels may have a different orientation. For example, the angularly segmented nodes of the light distribution panel may be rotated relative to the other light panels so that they are placed at different positions about an axis (e.g., partially overlapping and angularly offset). Using this arrangement, leading and trailing illumination can be produced.
In this embodiment, both the light distribution panel 262 and the electrode layer 254 have an annular shape that creates a void at the center of the touch pad 250. The void provides a space for placement of the extra light emitting diodes 258 and 270 as well as the mechanical button 268. As shown, the inner LEDs 258 are disposed along the inner periphery of the light distribution panel 262 next to distinct inner nodes 264 of the light distribution pane 262. Furthermore, the outer LEDs 260 are disposed along the outer periphery of the light distribution panel 262 next to distinct outer nodes 266 of the light distribution panel 262.
Furthermore, the center LED 270 is disposed near the center of the touch pad 250 underneath a translucent button cap 272 of the mechanical button 270. The button cap 272 is movable trapped between a diffusing label layer 274 and a spring loaded switch 276 that is also located near the center of the touch pad 250. When the button cap is pressed, it moves against the actuator of the spring loaded switch thereby generating a button event.
In the illustrated embodiment, the electrode layer 254, LEDs 258, 260 and 270 and mechanical switch 276 are all attached to the printed circuit board 256, and operatively coupled to a controller 280 located on the backside of the PCB 256. During operation, the controller 280 monitors the signals generated at the electrode layer 254 and switch 276, and provides commands for controlling the LEDs 258, 260 and 270.
It also differs from the touch pad of
In order to transmit light from the second set of inner LEDs 310 to the inner/middle nodes 320, the inner/middle nodes 320 may include a light transmitting portion 326 that extends underneath the inner nodes 318. In most cases, the light transmitting portions 326 are optically separated from the inner nodes 318 so that the light does not bleed into the inner nodes 318 when the light is passing through the light transmitting portions 326.
In order to transmit light from the second set of outer LEDs 314 to the outer/middle nodes 322, the outer/middle nodes 322 may include a light transmitting portion 328 that extends underneath the outer nodes 324. In most cases, the light transmitting portions 328 are optically separated from the outer nodes 324 so that the light does not bleed into the outer nodes 324 when the light is passing through the light transmitting portions 328. Alternatively, a light pipe may be used.
The light distribution panel 316 may be embodied in a variety of ways. In one embodiment, the light distribution panel 316 includes an outer translucent ring disposed over and around a stepped outer/middle translucent ring, and an inner translucent ring disposed over and around a stepped inner/middle translucent ring that is adjacent the outer/middle ring. A masking layer may be placed between the various rings to prevent bleed through.
Although the touch has been described with only four radial segments, it should be noted that any number of radial segments may be used as well as any number of angular segments to obtain the desired resolution.
To elaborate, the light diffuser layer 352 is disposed over the electrode layer 354, the opaque electrode layer 354 is disposed on the PCB 356, and the PCB 356 is disposed over the light panel 358. In order to illuminate the light diffuser layer 352, the electrode layer 354 includes one or more openings 360 through which light may pass when emitted from the light panel 358. The openings 360 may be the gaps that would normally be formed between the spatially separated electrodes or they may be predetermined gaps around which the spatially separated electrodes are positioned when printed on the PCB 356. Furthermore, in order to allow light to pass through the PCB 356, the PCB 356 either is formed from a translucent material or it also includes one or more openings 362 that correspond with the openings 360 of the electrode layer 354. Moreover, the light panel 358 is typically laid out similarly to the openings 360/362 so that each of the openings 360/362 includes an individual light source 364 of the light panel 358. During operation, the light emitted from each of the light sources 364 travels through the PCB 356 and through the electrode layer 354 where it illuminates either the diffuser layer 352 or a button cap 366.
The light multiplexer 404 includes a single light emitting device 408 and a plurality of light pipes 410 with corresponding light switches 412. The light pipe 410 may for example include one or more optical fibers, and the light emitting device 408 may be embodied in many different forms including for example one or more individual LEDs or one or more LED arrays.
The first end of each light pipe 410 is optically connected to a different point, node or region of the diffuser panel 402. The light pipes 410 can therefore form a pixilated pattern of illumination points, nodes or regions across the light diffuser panel 402. By way of example, the position of the light pipes 410 may be based on Cartesian coordinates, Polar coordinates, or some other coordinate system. The second and opposite end of each of the light pipes 410 is optically connected to a distinct light switch 412. The light switches 412 are therefore dedicated to a particular illumination point, node or region of the diffuser panel 402. Further, the light switches 412 are all in optical communication with the light emitting device 408. In some cases, the light emitting device 408 extends across the light switches 412. In other cases, the light emitted by the light emitting device 408 is focused onto the various light switches 412 via a lens or light guide 414.
Furthermore, the light switches 412 and light emitting device 408 are operatively coupled to the controller 406. During operation, the controller 406 selectively controls the light emitted by the light emitting device 408 (e.g., color and intensity), and at the same time selectively controls the opening and closing of the light switches 412. As such, the illumination provided at the diffuser panel 402 can be controlled in a pixilated manner using a single light emitting device 408. Any number of switches can be opened or closed at any particular point in time to provide the desired illumination pattern (by opening and closing different light switches, various patterns can be created). When the light is turned on and a light switch is opened, light is allowed to pass through the light switch into the associated light pipe, which carries the light from the light switch to a distinct illumination point node or region of the diffuser panel. When the light is turned on and a light switch is closed, light is blocked from entering the light pipe and therefore no illumination is provided at the corresponding point, node or region of the diffuser panel.
It should be noted that the multiplexer can include any number of switches, and the light panel can include any number of multiplexers to arrive at the desired resolution of the light panel.
In one embodiment, the input surface is broken up into illumination regions, and whichever region is closest to the detected object is illuminated. By way of example, and referring to
In another embodiment, the input surface is broken up into illumination nodes or points (pixilated), and those points contained within and/or surrounding the detected object area are illuminated. In one implementation, at least the area adjacent the object is illuminated. By way of example, and referring to
The method may additionally include blocks 456 and 458. In block 456, a second object is detected over the input surface at the same time as the first object. This may for example be accomplished with a multipoint capacitance sensing device. In block 458, at least a portion of the input surface proximate the location of the second detected object is illuminated. As a result, the user will be informed where distinct multiple objects are located within the sensing plane at all times.
In one embodiment, the input surface is broken up into illumination regions, and the regions closest to the detected objects are illuminated. By way of example, and referring to
In another embodiment, the input surface is broken up into illumination nodes or points (pixilated), and those points contained within and/or surrounding the detected objects are illuminated. By way of example, and referring to
In block 506, a determination is made as to whether or not the object is moving. If the object is not moving, the method proceeds to block 508 where a determination is made as to whether or not the object is still detected. If the object is still detected, the method proceeds back to block 504 where the same portion of the input surface is illuminated. If the object is no longer detected, the method proceeds to block 510 where the illumination is stopped. This may occur immediately after determining that an object is no longer detected, or it may occur after a period of time (e.g., time out). Furthermore, the illumination may be stopped using an illumination effect such as fading out. Thereafter, the method proceeds back to block 502.
Referring back to block 506, if the object is moving across the input surface, the method proceeds to block 512 where motion characteristics of the object are determined. The motion characteristics may for example include acceleration, direction, and the like. Thereafter, in block 514, the characteristics of the illumination are adjusted based on one or more motion characteristics. Following blocks 514, the method proceeds back to block 506.
In one embodiment, block 514 includes moving the illumination area in accordance with the location of the moving object. That is, the illuminated portion follows the finger as the finger is moved about the input surface (i.e., the illumination tracks object movement). As a result, the user always knows where the object is located relative to the input surface. In some cases, block 514 may further include providing directional indicators around the illuminated portion in order to indicate previous and/or future locations of the object based on the motion characteristics of the moving object (e.g., acceleration, direction, etc.).
The leading edge, body and trailing edge may have different illumination profiles. For example, the leading edge may have a high intensity level, the body may have a medium intensity level and the trailing edge may have a low intensity level. Alternatively, the leading edge may have a low intensity level, the body may have a high intensity level, and the trailing edge may have a low intensity level. Alternatively or additionally, the colors of these components may differ. For example, the leading edge may be red, the body may be orange and the trailing edge may be yellow. Furthermore, the trailing edge may include an illumination tail. For example, the trailing edge may be segmented into regions that go from higher intensity to lower intensity levels (e.g., fades outwardly from body).
In block 556, the input surface is illuminated based on the state of the touch pad. As a result, the user is alerted to the current state of the touch pad, and therefore the type of inputs that can be made. By way of example, each state may include a different illumination profile. An illumination profile defines the illumination characteristics of the illumination to be provided. The illumination characteristics include for example intensity and/or color and/or illumination effects (e.g., fading, blinking, rastering, etc). In one example, a first state includes a first illumination profile (e.g., first color), a second state includes a second illumination profile (e.g., second color), and the third state includes a third illumination profile (e.g., third color).
Referring to 19A, one example of determining state will be described. In blocks 570 and 574, one or more objects are detected. In block 572, if a single object is detected, the touch pad is placed in a tracking state such that object motion is used to perform tracking operations. In block 576, if multiple objects are detected, the touch pad is placed in a gesturing state such that object motion is used to perform gesturing operations. By way of example, and referring to
Referring to 19B, one example of illuminating based on states will be described. In blocks 580 and 582, a determination is made as to whether the touch pad is in a tracking state or a gesturing state. In block 584, if the touch pad is in a tracking state, the touch pad is illuminated with a first illumination profile. In block 586, if the touch pad is in a gesturing state, the touch pad is illuminated with a second illumination profile that is different than the first illumination profile. The illumination profiles contain illumination information such as color, intensity and effects (e.g., blinking, fading, etc). By way of example, and referring to
In one embodiment, the method of changing illumination based on states of the touch pad may be further developed. For example, the method may include capturing a first touch image; determining the touch mode based on the first touch image; and illuminating the touch surface based on the first touch mode. The method may also include capturing a second touch image; determining the touch mode based on the second touch image; determining if the touch mode changed between the first and second touch images; if the touch mode stayed the same, comparing the first and second touch images and performing a control function based on the comparison; and if the touch mode changed, illuminating the touch surface based on the second touch mode. The method additionally includes capturing a third touch image; determining the touch mode based on the third touch image; determining if the touch mode changed between the second and third touch images; if the touch mode stayed the same, comparing the second and third touch images and performing a control function based on the comparison; and if the touch mode changed, illuminating the touch surface based on the third touch mode.
By way of example, and referring to
Rotational user inputs are further described in U.S. patent application Ser. Nos. 10/256,716 and 10/259,159, which are herein incorporated by reference.
By way of example and referring to
While this invention has been described in terms of several preferred embodiments, there are alterations, permutations, and equivalents, which fall within the scope of this invention.
For example, although the invention was primarily directed at touch pads, it should be pointed out that this is not a limitation and that invention may be applied to other touch sensing devices as for example touch sensitive housings and touch sensing palm rests. An example of a touch sensitive housing may be found in U.S. patent application Ser. No. 11/115,539, which is herein incorporated by reference.
It should also be noted that there are many alternative ways of implementing the methods and apparatuses of the present invention.
For example, different areas of the touch pad may be illuminated with different illumination profiles at the same time. By way of example, the touch pad may be segmented into illuminated quadrants that correspond to button functionality of the touch pad such as menu, play/pause, forward and reverse. See for example U.S. patent application Ser. No. 10/643,256, which is herein incorporated by reference.
Furthermore, the touch pad may be used as an indicator for a handheld computing device such as a media player. For example, the touch pad may be configured to ebb in and out when the device is in sleep mode or vary from high to low intensity based on the battery level. Other examples of controlling light may be found in U.S. patent application Ser. Nos. 10/889,933, 10/075,964 and 10/075,520, all of which are herein incorporated by reference.
Moreover, the touch pad may be used as a timer or clock. In the case of a clock, the touch pad may include segments corresponding to the position of a clock, and the segments can be illuminated in a controlled manner to indicate the current time. For example, to indicate 12:30, a 12 o'clock segment may be illuminated with a first illumination profile and 6 o'clock segment may be illuminated with a second illumination profile. In the case of a timer, the touch pad may be used to show how much time is left in a playing media item such as a song. For example, the entire touch pad may be illuminated when the song starts and consecutive segments may be turned off as the song plays. When the song is over, the touch pad is no longer illuminated. Alternatively, consecutive segments may be turned on as the song plays until the song is over and the touch pad is fully illuminated. The may be useful in a media player such as a music player.
In addition, the illumination of the touch pad may be further controlled by a sensor such as a light sensor. The light sensor measures the ambient light level, and the intensity of the illumination is adjusted based on the ambient light level. Examples of light arrangements that utilize ambient light sensors may be found in U.S. patent application Ser. No. 10/402,311, which is herein incorporated by reference.
It is therefore intended that the following appended claims be interpreted as including all such alterations, permutations, and equivalents as fall within the true spirit and scope of the present invention.
This application is a continuation of U.S. patent application Ser. No. 11/394,493, filed Mar. 31, 2006, which claims priority benefit of U.S. Provisional Patent Application No. 60/755,656, filed Dec. 30, 2005, entitled “TOUCH PAD WITH FEEDBACK,” which is hereby incorporated herein by reference. This application is related to the following applications, all of which are herein incorporated by reference: U.S. patent application Ser. No. 10/188,182, titled “TOUCH PAD FOR HANDHELD DEVICE”, filed Feb. 25, 2002; U.S. patent application Ser. No. 10/722,948, titled “TOUCH PAD FOR HANDHELD DEVICE”, filed Nov. 25, 2003; U.S. patent application Ser. No. 10/643,256, titled “MOVABLE TOUCH PAD WITH ADDED FUNCTIONALITY”, filed Aug. 18, 2003; U.S. patent application Ser. No. 10/840,862, titled “MULTIPOINT TOUCHSCREEN”, filed May 6, 2004; and U.S. patent application Ser. No. 11/115,539, titled “HAND HELD ELECTRONIC DEVICE WITH MULTIPLE TOUCH SENSING DEVICES”, filed Apr. 26, 2005.
Number | Name | Date | Kind |
---|---|---|---|
1061578 | Wischhusen et al. | May 1913 | A |
2063276 | Thomas | Dec 1936 | A |
2798907 | Schneider | Jul 1957 | A |
2903229 | Landge | Sep 1959 | A |
2945111 | McCormick | Jul 1960 | A |
3005055 | Mattke | Oct 1961 | A |
3965399 | Walker et al. | Jun 1976 | A |
3996441 | Ohashi | Dec 1976 | A |
4029915 | Ojima | Jun 1977 | A |
4103252 | Bobick | Jul 1978 | A |
4110749 | Janko et al. | Aug 1978 | A |
4115670 | Chandler | Sep 1978 | A |
4121204 | Welch et al. | Oct 1978 | A |
4129747 | Pepper | Dec 1978 | A |
4158216 | Bigelow | Jun 1979 | A |
4242676 | Piguet et al. | Dec 1980 | A |
4246452 | Chandler | Jan 1981 | A |
4264903 | Bigelow | Apr 1981 | A |
4266144 | Bristol | May 1981 | A |
4293734 | Pepper, Jr. | Oct 1981 | A |
D264969 | McGourty | Jun 1982 | S |
4338502 | Hashimoto et al. | Jul 1982 | A |
4380007 | Steinegger | Apr 1983 | A |
4380040 | Posset | Apr 1983 | A |
4394649 | Suchoff et al. | Jul 1983 | A |
4475008 | Doi et al. | Oct 1984 | A |
4570149 | Thornburg et al. | Feb 1986 | A |
4583161 | Gunderson et al. | Apr 1986 | A |
4587378 | Moore | May 1986 | A |
4604786 | Howie, Jr. | Aug 1986 | A |
4613736 | Shichijo et al. | Sep 1986 | A |
4644100 | Brenner et al. | Feb 1987 | A |
4719524 | Morishima et al. | Jan 1988 | A |
4734034 | Maness et al. | Mar 1988 | A |
4736191 | Matzke et al. | Apr 1988 | A |
4739191 | Puar | Apr 1988 | A |
4739299 | Eventoff et al. | Apr 1988 | A |
4752655 | Tajiri et al. | Jun 1988 | A |
4755765 | Ferland | Jul 1988 | A |
4764717 | Tucker et al. | Aug 1988 | A |
4771139 | DeSmet | Sep 1988 | A |
4798919 | Miessler et al. | Jan 1989 | A |
4810992 | Eventoff | Mar 1989 | A |
4822957 | Talmage, Jr. et al. | Apr 1989 | A |
4831359 | Newell | May 1989 | A |
4849852 | Mullins | Jul 1989 | A |
4856993 | Maness et al. | Aug 1989 | A |
4860768 | Hon et al. | Aug 1989 | A |
4866602 | Hall | Sep 1989 | A |
4876524 | Jenkins | Oct 1989 | A |
4897511 | Itaya et al. | Jan 1990 | A |
4914624 | Dunthorn | Apr 1990 | A |
4917516 | Retter | Apr 1990 | A |
4943889 | Ohmatoi | Jul 1990 | A |
4951036 | Grueter et al. | Aug 1990 | A |
4954823 | Binstead | Sep 1990 | A |
4976435 | Shatford et al. | Dec 1990 | A |
4990900 | Kikuchi | Feb 1991 | A |
5008497 | Asher | Apr 1991 | A |
5036321 | Leach et al. | Jul 1991 | A |
5053757 | Meadows | Oct 1991 | A |
5086870 | Bolduc | Feb 1992 | A |
5125077 | Hall | Jun 1992 | A |
5159159 | Asher | Oct 1992 | A |
5179648 | Hauck | Jan 1993 | A |
5186646 | Pederson | Feb 1993 | A |
5192082 | Inoue et al. | Mar 1993 | A |
5193669 | Demeo et al. | Mar 1993 | A |
5231326 | Echols | Jul 1993 | A |
5237311 | Mailey et al. | Aug 1993 | A |
5278362 | Ohashi | Jan 1994 | A |
5305017 | Gerpheide | Apr 1994 | A |
5313027 | Inoue et al. | May 1994 | A |
D349280 | Kaneko | Aug 1994 | S |
5339213 | O'Callaghan | Aug 1994 | A |
5367199 | Lefkowitz et al. | Nov 1994 | A |
5374787 | Miller et al. | Dec 1994 | A |
5379057 | Clough et al. | Jan 1995 | A |
5404152 | Nagai | Apr 1995 | A |
5408621 | Ben-Arie | Apr 1995 | A |
5414445 | Kaneko et al. | May 1995 | A |
5416498 | Grant | May 1995 | A |
5424756 | Ho et al. | Jun 1995 | A |
5432531 | Calder et al. | Jul 1995 | A |
5438331 | Gilligan et al. | Aug 1995 | A |
D362431 | Kaneko et al. | Sep 1995 | S |
5450075 | Waddington | Sep 1995 | A |
5453761 | Tanaka | Sep 1995 | A |
5473343 | Kimmich et al. | Dec 1995 | A |
5473344 | Bacon et al. | Dec 1995 | A |
5479192 | Carroll, Jr. et al. | Dec 1995 | A |
5494157 | Golenz et al. | Feb 1996 | A |
5495566 | Kwatinetz | Feb 1996 | A |
5508703 | Okamura et al. | Apr 1996 | A |
5508717 | Miller | Apr 1996 | A |
5543588 | Bisset et al. | Aug 1996 | A |
5543591 | Gillespie et al. | Aug 1996 | A |
5555004 | Ono et al. | Sep 1996 | A |
5559301 | Bryan, Jr. et al. | Sep 1996 | A |
5559943 | Cyr et al. | Sep 1996 | A |
5561445 | Miwa et al. | Oct 1996 | A |
5564112 | Hayes et al. | Oct 1996 | A |
5565887 | McCambridge et al. | Oct 1996 | A |
5578817 | Bidiville et al. | Nov 1996 | A |
5581670 | Bier et al. | Dec 1996 | A |
5585823 | Duchon et al. | Dec 1996 | A |
5589856 | Stein et al. | Dec 1996 | A |
5589893 | Gaughan et al. | Dec 1996 | A |
5596347 | Robertson et al. | Jan 1997 | A |
5596697 | Foster et al. | Jan 1997 | A |
5598183 | Robertson et al. | Jan 1997 | A |
5611040 | Brewer et al. | Mar 1997 | A |
5611060 | Belfiore et al. | Mar 1997 | A |
5613137 | Bertram et al. | Mar 1997 | A |
5617114 | Bier et al. | Apr 1997 | A |
5627531 | Posso et al. | May 1997 | A |
5632679 | Tremmel | May 1997 | A |
5640258 | Kurashima et al. | Jun 1997 | A |
5648642 | Miller et al. | Jul 1997 | A |
D382550 | Kaneko et al. | Aug 1997 | S |
5657012 | Tait | Aug 1997 | A |
5661632 | Register | Aug 1997 | A |
D385542 | Kaneko et al. | Oct 1997 | S |
5675362 | Clough et al. | Oct 1997 | A |
5689285 | Asher | Nov 1997 | A |
5721849 | Amro | Feb 1998 | A |
5726687 | Belfiore et al. | Mar 1998 | A |
5729219 | Armstrong et al. | Mar 1998 | A |
5730165 | Philipp | Mar 1998 | A |
5748185 | Stephan et al. | May 1998 | A |
5751274 | Davis | May 1998 | A |
5754890 | Holmdahl et al. | May 1998 | A |
5764066 | Novak et al. | Jun 1998 | A |
5777605 | Yoshinobu et al. | Jul 1998 | A |
5786818 | Brewer et al. | Jul 1998 | A |
5790769 | Buxton et al. | Aug 1998 | A |
5798752 | Buxton et al. | Aug 1998 | A |
5805144 | Scholder et al. | Sep 1998 | A |
5808602 | Sellers | Sep 1998 | A |
5812239 | Eger | Sep 1998 | A |
5812498 | Teres | Sep 1998 | A |
5815141 | Phares | Sep 1998 | A |
5825351 | Tam | Oct 1998 | A |
5825352 | Bisset et al. | Oct 1998 | A |
5825353 | Will | Oct 1998 | A |
5828364 | Siddiqui | Oct 1998 | A |
5838304 | Hall | Nov 1998 | A |
5841078 | Miller et al. | Nov 1998 | A |
5841423 | Carroll, Jr. et al. | Nov 1998 | A |
D402281 | Ledbetter et al. | Dec 1998 | S |
5850213 | Imai et al. | Dec 1998 | A |
5856645 | Norton | Jan 1999 | A |
5856822 | Du et al. | Jan 1999 | A |
5859629 | Tognazzini | Jan 1999 | A |
5861875 | Gerpheide | Jan 1999 | A |
5869791 | Young | Feb 1999 | A |
5875311 | Bertram et al. | Feb 1999 | A |
5883619 | Ho et al. | Mar 1999 | A |
5889236 | Gillespie et al. | Mar 1999 | A |
5889511 | Ong et al. | Mar 1999 | A |
5894117 | Kamishima | Apr 1999 | A |
5903229 | Kishi | May 1999 | A |
5907152 | Dandliker et al. | May 1999 | A |
5907318 | Medina | May 1999 | A |
5909211 | Combs et al. | Jun 1999 | A |
5910802 | Shields et al. | Jun 1999 | A |
5914706 | Kono | Jun 1999 | A |
5923388 | Kurashima et al. | Jul 1999 | A |
D412940 | Kato et al. | Aug 1999 | S |
5933102 | Miller et al. | Aug 1999 | A |
5933141 | Smith | Aug 1999 | A |
5936619 | Nagasaki et al. | Aug 1999 | A |
5943044 | Martinelli et al. | Aug 1999 | A |
5953000 | Weirich | Sep 1999 | A |
5956019 | Bang et al. | Sep 1999 | A |
5959610 | Silfvast | Sep 1999 | A |
5959611 | Smailagic et al. | Sep 1999 | A |
5964661 | Dodge | Oct 1999 | A |
5973668 | Watanabe | Oct 1999 | A |
6000000 | Hawkins et al. | Dec 1999 | A |
6002093 | Hrehor et al. | Dec 1999 | A |
6002389 | Kasser et al. | Dec 1999 | A |
6005299 | Hengst | Dec 1999 | A |
6025832 | Sudo et al. | Feb 2000 | A |
6031518 | Adams et al. | Feb 2000 | A |
6034672 | Gaultier et al. | Mar 2000 | A |
6057829 | Silfvast | May 2000 | A |
6075533 | Chang | Jun 2000 | A |
6084574 | Bidiville | Jul 2000 | A |
D430169 | Scibora | Aug 2000 | S |
6097372 | Suzuki | Aug 2000 | A |
6104790 | Narayanaswami | Aug 2000 | A |
6122526 | Parulski et al. | Sep 2000 | A |
6124587 | Bidiville et al. | Sep 2000 | A |
6128006 | Rosenberg et al. | Oct 2000 | A |
6131048 | Sudo et al. | Oct 2000 | A |
6141068 | Iijima | Oct 2000 | A |
6147856 | Karidis | Nov 2000 | A |
6163312 | Furuya | Dec 2000 | A |
6166721 | Kuroiwa et al. | Dec 2000 | A |
6179496 | Chou | Jan 2001 | B1 |
6181322 | Nanavati | Jan 2001 | B1 |
D437860 | Suzuki et al. | Feb 2001 | S |
6188391 | Seely et al. | Feb 2001 | B1 |
6188393 | Shu | Feb 2001 | B1 |
6191774 | Schena et al. | Feb 2001 | B1 |
6198054 | Janniere | Mar 2001 | B1 |
6198473 | Armstrong | Mar 2001 | B1 |
6211861 | Rosenberg et al. | Apr 2001 | B1 |
6219038 | Cho | Apr 2001 | B1 |
6222528 | Gerpheide et al. | Apr 2001 | B1 |
D442592 | Ledbetter et al. | May 2001 | S |
6225976 | Yates et al. | May 2001 | B1 |
6225980 | Weiss et al. | May 2001 | B1 |
6226534 | Aizawa | May 2001 | B1 |
6227966 | Yokoi | May 2001 | B1 |
D443616 | Fisher et al. | Jun 2001 | S |
6243078 | Rosenberg | Jun 2001 | B1 |
6243080 | Molne | Jun 2001 | B1 |
6243646 | Ozaki et al. | Jun 2001 | B1 |
6248017 | Roach | Jun 2001 | B1 |
6254477 | Sasaki et al. | Jul 2001 | B1 |
6256011 | Culver | Jul 2001 | B1 |
6259491 | Ekedahl et al. | Jul 2001 | B1 |
6262717 | Donohue et al. | Jul 2001 | B1 |
6262785 | Kim | Jul 2001 | B1 |
6266050 | Oh et al. | Jul 2001 | B1 |
6285211 | Sample et al. | Sep 2001 | B1 |
D448810 | Goto | Oct 2001 | S |
6297795 | Kato et al. | Oct 2001 | B1 |
6297811 | Kent et al. | Oct 2001 | B1 |
6300946 | Lincke et al. | Oct 2001 | B1 |
6307539 | Suzuki | Oct 2001 | B2 |
D450713 | Masamitsu et al. | Nov 2001 | S |
6314483 | Goto et al. | Nov 2001 | B1 |
6321441 | Davidson et al. | Nov 2001 | B1 |
6323845 | Robbins | Nov 2001 | B1 |
D452250 | Chan | Dec 2001 | S |
6340800 | Zhai et al. | Jan 2002 | B1 |
D454568 | Andre et al. | Mar 2002 | S |
6357887 | Novak | Mar 2002 | B1 |
D455793 | Lin | Apr 2002 | S |
6373265 | Morimoto et al. | Apr 2002 | B1 |
6373470 | Andre et al. | Apr 2002 | B1 |
6377530 | Burrows | Apr 2002 | B1 |
6396523 | Segal et al. | May 2002 | B1 |
6424338 | Anderson | Jul 2002 | B1 |
6429846 | Rosenberg et al. | Aug 2002 | B2 |
6429852 | Adams et al. | Aug 2002 | B1 |
6452514 | Philipp | Sep 2002 | B1 |
6465271 | Ko et al. | Oct 2002 | B1 |
6473069 | Gerpheide | Oct 2002 | B1 |
6476831 | Wirth et al. | Nov 2002 | B1 |
6492602 | Asai et al. | Dec 2002 | B2 |
6492979 | Kent et al. | Dec 2002 | B1 |
6496181 | Bomer et al. | Dec 2002 | B1 |
6497412 | Bramm | Dec 2002 | B1 |
D468365 | Bransky et al. | Jan 2003 | S |
D469109 | Andre et al. | Jan 2003 | S |
D472245 | Andre et al. | Mar 2003 | S |
6546231 | Someya et al. | Apr 2003 | B1 |
6563487 | Martin et al. | May 2003 | B2 |
6587091 | Serpa | Jul 2003 | B2 |
6606244 | Liu et al. | Aug 2003 | B1 |
6618909 | Yang | Sep 2003 | B1 |
6636197 | Goldenberg et al. | Oct 2003 | B1 |
6639584 | Li | Oct 2003 | B1 |
6640250 | Chang et al. | Oct 2003 | B1 |
6650975 | Ruffner | Nov 2003 | B2 |
D483809 | Lim | Dec 2003 | S |
6658773 | Rohne et al. | Dec 2003 | B2 |
6664951 | Fujii et al. | Dec 2003 | B1 |
6677927 | Bruck et al. | Jan 2004 | B1 |
6678891 | Wilcox et al. | Jan 2004 | B1 |
6686904 | Sherman et al. | Feb 2004 | B1 |
6686906 | Salminen et al. | Feb 2004 | B2 |
6703550 | Chu | Mar 2004 | B2 |
6724817 | Simpson et al. | Apr 2004 | B1 |
6727889 | Shaw | Apr 2004 | B2 |
D489731 | Huang | May 2004 | S |
6738045 | Hinckley et al. | May 2004 | B2 |
6750803 | Yates et al. | Jun 2004 | B2 |
6781576 | Tamura | Aug 2004 | B2 |
6784384 | Park et al. | Aug 2004 | B2 |
6788288 | Ano | Sep 2004 | B2 |
6791533 | Su | Sep 2004 | B2 |
6795057 | Gordon | Sep 2004 | B2 |
D497618 | Andre et al. | Oct 2004 | S |
6810271 | Wood et al. | Oct 2004 | B1 |
6822640 | Derocher | Nov 2004 | B2 |
6834975 | Chu-Chia et al. | Dec 2004 | B2 |
6844872 | Farag et al. | Jan 2005 | B1 |
6850225 | Whitcroft | Feb 2005 | B1 |
6855899 | Sotome | Feb 2005 | B2 |
6865718 | Levi Montalcini | Mar 2005 | B2 |
6886842 | Vey et al. | May 2005 | B2 |
6894916 | Reohr et al. | May 2005 | B2 |
D506476 | Andre et al. | Jun 2005 | S |
6922189 | Fujiyoshi | Jul 2005 | B2 |
6930494 | Tesdahl et al. | Aug 2005 | B2 |
6958614 | Morimoto | Oct 2005 | B2 |
6977808 | Lam et al. | Dec 2005 | B2 |
6978127 | Bulthuis et al. | Dec 2005 | B1 |
6985137 | Kaikuranta | Jan 2006 | B2 |
7006077 | Uusimaki | Feb 2006 | B1 |
7019225 | Matsumoto et al. | Mar 2006 | B2 |
7046230 | Zadesky et al. | May 2006 | B2 |
7050292 | Shimura et al. | May 2006 | B2 |
7069044 | Okada et al. | Jun 2006 | B2 |
7078633 | Ihalainen | Jul 2006 | B2 |
7084856 | Huppi | Aug 2006 | B2 |
7113196 | Kerr | Sep 2006 | B2 |
7117136 | Rosedale | Oct 2006 | B1 |
7119792 | Andre et al. | Oct 2006 | B1 |
7215319 | Kamijo et al. | May 2007 | B2 |
7233318 | Farag et al. | Jun 2007 | B1 |
7236154 | Kerr et al. | Jun 2007 | B1 |
7236159 | Siversson | Jun 2007 | B1 |
7253643 | Seguine | Aug 2007 | B1 |
7279647 | Philipp | Oct 2007 | B2 |
7288732 | Hashida | Oct 2007 | B2 |
7297883 | Rochon et al. | Nov 2007 | B2 |
7310089 | Baker et al. | Dec 2007 | B2 |
7312785 | Tsuk et al. | Dec 2007 | B2 |
7321103 | Nakanishi et al. | Jan 2008 | B2 |
7333092 | Zadesky et al. | Feb 2008 | B2 |
7348898 | Ono | Mar 2008 | B2 |
7382139 | Mackey | Jun 2008 | B2 |
7394038 | Chang | Jul 2008 | B2 |
7395081 | Bonnelykke Kristensen et al. | Jul 2008 | B2 |
7397467 | Park et al. | Jul 2008 | B2 |
7439963 | Geaghan et al. | Oct 2008 | B2 |
7466307 | Trent et al. | Dec 2008 | B2 |
7479949 | Jobs et al. | Jan 2009 | B2 |
7486323 | Lee et al. | Feb 2009 | B2 |
7502016 | Trent, Jr. et al. | Mar 2009 | B2 |
7503193 | Schoene et al. | Mar 2009 | B2 |
7593782 | Jobs et al. | Sep 2009 | B2 |
7616097 | Whang | Nov 2009 | B1 |
7645955 | Huang et al. | Jan 2010 | B2 |
7671837 | Forsblad et al. | Mar 2010 | B2 |
7708051 | Katsumi et al. | May 2010 | B2 |
7772507 | Orr et al. | Aug 2010 | B2 |
20010011991 | Wang et al. | Aug 2001 | A1 |
20010011993 | Saarinen | Aug 2001 | A1 |
20010033270 | Osawa et al. | Oct 2001 | A1 |
20010043545 | Aratani | Nov 2001 | A1 |
20010050673 | Davenport | Dec 2001 | A1 |
20010051046 | Watanabe et al. | Dec 2001 | A1 |
20020000978 | Gerpheide | Jan 2002 | A1 |
20020011993 | Lui et al. | Jan 2002 | A1 |
20020027547 | Kamijo et al. | Mar 2002 | A1 |
20020030665 | Ano | Mar 2002 | A1 |
20020033848 | Sciammarella et al. | Mar 2002 | A1 |
20020039493 | Tanaka | Apr 2002 | A1 |
20020045960 | Phillips et al. | Apr 2002 | A1 |
20020071550 | Pletikosa | Jun 2002 | A1 |
20020089545 | Levi Montalcini | Jul 2002 | A1 |
20020103796 | Hartley | Aug 2002 | A1 |
20020118131 | Yates et al. | Aug 2002 | A1 |
20020118169 | Hinckley et al. | Aug 2002 | A1 |
20020145594 | Derocher | Oct 2002 | A1 |
20020154090 | Lin | Oct 2002 | A1 |
20020158844 | McLoone et al. | Oct 2002 | A1 |
20020164156 | Bilbrey | Nov 2002 | A1 |
20020168947 | Lemley | Nov 2002 | A1 |
20020180701 | Hayama et al. | Dec 2002 | A1 |
20020196239 | Lee | Dec 2002 | A1 |
20030002246 | Kerr | Jan 2003 | A1 |
20030016211 | Woolley | Jan 2003 | A1 |
20030025679 | Taylor et al. | Feb 2003 | A1 |
20030028346 | Sinclair et al. | Feb 2003 | A1 |
20030043121 | Chen | Mar 2003 | A1 |
20030043174 | Hinckley et al. | Mar 2003 | A1 |
20030050092 | Yun | Mar 2003 | A1 |
20030076301 | Tsuk et al. | Apr 2003 | A1 |
20030076303 | Huppi | Apr 2003 | A1 |
20030091377 | Hsu et al. | May 2003 | A1 |
20030095095 | Pihlaja | May 2003 | A1 |
20030095096 | Robbin et al. | May 2003 | A1 |
20030098851 | Brink | May 2003 | A1 |
20030103043 | Mulligan et al. | Jun 2003 | A1 |
20030117367 | Yan | Jun 2003 | A1 |
20030122792 | Yamamoto et al. | Jul 2003 | A1 |
20030135292 | Husgafvel et al. | Jul 2003 | A1 |
20030142081 | Iizuka et al. | Jul 2003 | A1 |
20030184517 | Senzui et al. | Oct 2003 | A1 |
20030197740 | Reponen | Oct 2003 | A1 |
20030206202 | Moriya | Nov 2003 | A1 |
20030210537 | Engelmann | Nov 2003 | A1 |
20030224831 | Engstrom et al. | Dec 2003 | A1 |
20040027341 | Derocher | Feb 2004 | A1 |
20040074756 | Kawakami et al. | Apr 2004 | A1 |
20040080682 | Dalton | Apr 2004 | A1 |
20040109357 | Cernea et al. | Jun 2004 | A1 |
20040150619 | Baudisch et al. | Aug 2004 | A1 |
20040156192 | Kerr et al. | Aug 2004 | A1 |
20040178997 | Gillespie et al. | Sep 2004 | A1 |
20040200699 | Matsumoto et al. | Oct 2004 | A1 |
20040215986 | Shakkarwar | Oct 2004 | A1 |
20040224638 | Fadell et al. | Nov 2004 | A1 |
20040239622 | Proctor et al. | Dec 2004 | A1 |
20040252109 | Trent, Jr. et al. | Dec 2004 | A1 |
20040252867 | Lan et al. | Dec 2004 | A1 |
20040253989 | Tupler et al. | Dec 2004 | A1 |
20040263388 | Krumm et al. | Dec 2004 | A1 |
20040267874 | Westberg et al. | Dec 2004 | A1 |
20050012644 | Hurst et al. | Jan 2005 | A1 |
20050017957 | Yi | Jan 2005 | A1 |
20050024341 | Gillespie et al. | Feb 2005 | A1 |
20050030048 | Bolender | Feb 2005 | A1 |
20050052425 | Zadesky et al. | Mar 2005 | A1 |
20050052426 | Hagermoser et al. | Mar 2005 | A1 |
20050052429 | Philipp | Mar 2005 | A1 |
20050068304 | Lewis et al. | Mar 2005 | A1 |
20050083299 | Nagasaka | Apr 2005 | A1 |
20050083307 | Aufderheide | Apr 2005 | A1 |
20050090288 | Stohr et al. | Apr 2005 | A1 |
20050104867 | Westerman et al. | May 2005 | A1 |
20050110768 | Marriott et al. | May 2005 | A1 |
20050129199 | Abe | Jun 2005 | A1 |
20050139460 | Hosaka | Jun 2005 | A1 |
20050140657 | Park et al. | Jun 2005 | A1 |
20050143124 | Kennedy et al. | Jun 2005 | A1 |
20050156881 | Trent et al. | Jul 2005 | A1 |
20050162402 | Watanachote | Jul 2005 | A1 |
20050183035 | Ringel et al. | Aug 2005 | A1 |
20050204309 | Szeto | Sep 2005 | A1 |
20050237308 | Autio et al. | Oct 2005 | A1 |
20060026521 | Hotelling et al. | Feb 2006 | A1 |
20060032680 | Elias et al. | Feb 2006 | A1 |
20060034042 | Hisano et al. | Feb 2006 | A1 |
20060038791 | Mackey | Feb 2006 | A1 |
20060095848 | Naik | May 2006 | A1 |
20060097991 | Hotelling et al. | May 2006 | A1 |
20060131156 | Voelckers | Jun 2006 | A1 |
20060143574 | Ito et al. | Jun 2006 | A1 |
20060174568 | Kinoshita et al. | Aug 2006 | A1 |
20060181517 | Zadesky et al. | Aug 2006 | A1 |
20060197750 | Kerr et al. | Sep 2006 | A1 |
20060232557 | Fallot-Burghardt | Oct 2006 | A1 |
20060236262 | Bathiche et al. | Oct 2006 | A1 |
20060250377 | Zadesky et al. | Nov 2006 | A1 |
20060274042 | Krah et al. | Dec 2006 | A1 |
20060274905 | Lindahl et al. | Dec 2006 | A1 |
20060279896 | Bruwer | Dec 2006 | A1 |
20060284836 | Philipp | Dec 2006 | A1 |
20070013671 | Zadesky et al. | Jan 2007 | A1 |
20070018970 | Tabasso et al. | Jan 2007 | A1 |
20070052044 | Forsblad et al. | Mar 2007 | A1 |
20070052691 | Zadesky et al. | Mar 2007 | A1 |
20070080936 | Tsuk et al. | Apr 2007 | A1 |
20070080938 | Robbin et al. | Apr 2007 | A1 |
20070080952 | Lynch et al. | Apr 2007 | A1 |
20070083822 | Robbin et al. | Apr 2007 | A1 |
20070085841 | Tsuk et al. | Apr 2007 | A1 |
20070097086 | Battles et al. | May 2007 | A1 |
20070120834 | Boillot | May 2007 | A1 |
20070126696 | Boillot | Jun 2007 | A1 |
20070152975 | Ogihara | Jul 2007 | A1 |
20070152977 | Ng et al. | Jul 2007 | A1 |
20070152983 | McKillop et al. | Jul 2007 | A1 |
20070155434 | Jobs et al. | Jul 2007 | A1 |
20070157089 | Van Os et al. | Jul 2007 | A1 |
20070242057 | Zadesky et al. | Oct 2007 | A1 |
20070247421 | Orsley et al. | Oct 2007 | A1 |
20070247443 | Philipp | Oct 2007 | A1 |
20070271516 | Carmichael | Nov 2007 | A1 |
20070273671 | Zadesky et al. | Nov 2007 | A1 |
20070276525 | Zadesky et al. | Nov 2007 | A1 |
20070279394 | Lampell | Dec 2007 | A1 |
20070285404 | Rimon et al. | Dec 2007 | A1 |
20070290990 | Robbin et al. | Dec 2007 | A1 |
20070291016 | Philipp | Dec 2007 | A1 |
20070296709 | GuangHai | Dec 2007 | A1 |
20080006453 | Hotelling et al. | Jan 2008 | A1 |
20080006454 | Hotelling | Jan 2008 | A1 |
20080007533 | Hotelling et al. | Jan 2008 | A1 |
20080007539 | Hotelling et al. | Jan 2008 | A1 |
20080012837 | Marriott et al. | Jan 2008 | A1 |
20080018615 | Zadesky et al. | Jan 2008 | A1 |
20080018616 | Lampell et al. | Jan 2008 | A1 |
20080018617 | Ng et al. | Jan 2008 | A1 |
20080036473 | Jansson | Feb 2008 | A1 |
20080036734 | Forsblad et al. | Feb 2008 | A1 |
20080060925 | Weber et al. | Mar 2008 | A1 |
20080069412 | Champagne et al. | Mar 2008 | A1 |
20080079699 | Mackey | Apr 2008 | A1 |
20080087476 | Prest | Apr 2008 | A1 |
20080088582 | Prest | Apr 2008 | A1 |
20080088596 | Prest | Apr 2008 | A1 |
20080088597 | Prest | Apr 2008 | A1 |
20080088600 | Prest | Apr 2008 | A1 |
20080094352 | Tsuk et al. | Apr 2008 | A1 |
20080098330 | Tsuk et al. | Apr 2008 | A1 |
20080110739 | Peng et al. | May 2008 | A1 |
20080111795 | Bollinger | May 2008 | A1 |
20080143681 | XiaoPing | Jun 2008 | A1 |
20080165158 | Hotelling et al. | Jul 2008 | A1 |
20080196945 | Konstas | Aug 2008 | A1 |
20080202824 | Philipp et al. | Aug 2008 | A1 |
20080209442 | Setlur et al. | Aug 2008 | A1 |
20080264767 | Chen et al. | Oct 2008 | A1 |
20080280651 | Duarte | Nov 2008 | A1 |
20080284742 | Prest | Nov 2008 | A1 |
20080293274 | Milan | Nov 2008 | A1 |
20090021267 | Golovchenko et al. | Jan 2009 | A1 |
20090026558 | Bauer et al. | Jan 2009 | A1 |
20090033635 | Wai | Feb 2009 | A1 |
20090036176 | Ure | Feb 2009 | A1 |
20090058687 | Rothkopf et al. | Mar 2009 | A1 |
20090058801 | Bull | Mar 2009 | A1 |
20090058802 | Orsley et al. | Mar 2009 | A1 |
20090073130 | Weber et al. | Mar 2009 | A1 |
20090078551 | Kang | Mar 2009 | A1 |
20090109181 | Hui et al. | Apr 2009 | A1 |
20090141046 | Rathnam et al. | Jun 2009 | A1 |
20090160771 | Hinckley et al. | Jun 2009 | A1 |
20090179854 | Weber et al. | Jul 2009 | A1 |
20090197059 | Weber et al. | Aug 2009 | A1 |
20090229892 | Fisher et al. | Sep 2009 | A1 |
20090273573 | Hotelling | Nov 2009 | A1 |
20100058251 | Rottler et al. | Mar 2010 | A1 |
20100060568 | Fisher et al. | Mar 2010 | A1 |
20100073319 | Lyon et al. | Mar 2010 | A1 |
20100149127 | Fisher et al. | Jun 2010 | A1 |
20100289759 | Fisher et al. | Nov 2010 | A1 |
20100313409 | Weber et al. | Dec 2010 | A1 |
20110005845 | Hotelling et al. | Jan 2011 | A1 |
Number | Date | Country |
---|---|---|
1139235 | Jan 1997 | CN |
1455615 | Nov 2003 | CN |
1499356 | May 2004 | CN |
1659506 | Aug 2005 | CN |
3615742 | Nov 1987 | DE |
19722636 | Dec 1998 | DE |
10022537 | Nov 2000 | DE |
20019074 | Feb 2001 | DE |
10 2004 043 | Apr 2006 | DE |
0178157 | Apr 1986 | EP |
0419145 | Mar 1991 | EP |
0 498 540 | Aug 1992 | EP |
0 521 683 | Jan 1993 | EP |
0 674 288 | Sep 1995 | EP |
0 731 407 | Sep 1996 | EP |
0 551 778 | Jan 1997 | EP |
0 880 091 | Nov 1998 | EP |
1 026 713 | Aug 2000 | EP |
1 081 922 | Mar 2001 | EP |
1 098 241 | May 2001 | EP |
1 133 057 | Sep 2001 | EP |
1 162 826 | Dec 2001 | EP |
1 168 396 | Jan 2002 | EP |
1 205 836 | May 2002 | EP |
1 244 053 | Sep 2002 | EP |
1 251 455 | Oct 2002 | EP |
1 263 193 | Dec 2002 | EP |
1 347 481 | Sep 2003 | EP |
1 376 326 | Jan 2004 | EP |
1 467 392 | Oct 2004 | EP |
1 482 401 | Dec 2004 | EP |
1 496 467 | Jan 2005 | EP |
1 517 228 | Mar 2005 | EP |
1 542 437 | Jun 2005 | EP |
1 589 407 | Oct 2005 | EP |
1 610 210 | Dec 2005 | EP |
1 784 058 | May 2007 | EP |
1 841 188 | Oct 2007 | EP |
1 850 218 | Oct 2007 | EP |
1 876 711 | Jan 2008 | EP |
2 686 440 | Jul 1993 | FR |
2015167 | Sep 1979 | GB |
2072389 | Sep 1981 | GB |
2315186 | Jan 1998 | GB |
2333215 | Jul 1999 | GB |
2391060 | Jan 2004 | GB |
2402105 | Dec 2004 | GB |
57-95722 | Jun 1982 | JP |
57-97626 | Jun 1982 | JP |
61-117619 | Jun 1986 | JP |
61-124009 | Jun 1986 | JP |
63-20411 | Jan 1988 | JP |
63-106826 | May 1988 | JP |
63-181022 | Jul 1988 | JP |
63-298518 | Dec 1988 | JP |
03-57617 | Jun 1991 | JP |
3-192418 | Aug 1991 | JP |
4-32920 | Feb 1992 | JP |
4-205408 | Jul 1992 | JP |
5-041135 | Feb 1993 | JP |
5-080938 | Apr 1993 | JP |
5-101741 | Apr 1993 | JP |
5-36623 | May 1993 | JP |
5-189110 | Jul 1993 | JP |
5-205565 | Aug 1993 | JP |
5-211021 | Aug 1993 | JP |
5-217464 | Aug 1993 | JP |
5-233141 | Sep 1993 | JP |
5-262276 | Oct 1993 | JP |
5-265656 | Oct 1993 | JP |
5-274956 | Oct 1993 | JP |
5-289811 | Nov 1993 | JP |
5-298955 | Nov 1993 | JP |
5-325723 | Dec 1993 | JP |
6-20570 | Jan 1994 | JP |
6-084428 | Mar 1994 | JP |
6-089636 | Mar 1994 | JP |
6-96639 | Apr 1994 | JP |
6-111695 | Apr 1994 | JP |
6-139879 | May 1994 | JP |
6-187078 | Jul 1994 | JP |
6-208433 | Jul 1994 | JP |
6-267382 | Sep 1994 | JP |
6-283993 | Oct 1994 | JP |
6-333459 | Dec 1994 | JP |
7-107574 | Apr 1995 | JP |
7-41882 | Jul 1995 | JP |
7-201249 | Aug 1995 | JP |
7-201256 | Aug 1995 | JP |
7-253838 | Oct 1995 | JP |
7-261899 | Oct 1995 | JP |
7-261922 | Oct 1995 | JP |
7-296670 | Nov 1995 | JP |
7-319001 | Dec 1995 | JP |
8-016292 | Jan 1996 | JP |
8-115158 | May 1996 | JP |
8-203387 | Aug 1996 | JP |
8-293226 | Nov 1996 | JP |
8-298045 | Nov 1996 | JP |
8-299541 | Nov 1996 | JP |
8-316664 | Nov 1996 | JP |
9-044289 | Feb 1997 | JP |
9-069023 | Mar 1997 | JP |
9-128148 | May 1997 | JP |
9-134248 | May 1997 | JP |
9-218747 | Aug 1997 | JP |
9-230993 | Sep 1997 | JP |
9-231858 | Sep 1997 | JP |
9-233161 | Sep 1997 | JP |
9-251347 | Sep 1997 | JP |
9-258895 | Oct 1997 | JP |
9-288926 | Nov 1997 | JP |
9-512979 | Dec 1997 | JP |
10-63467 | Mar 1998 | JP |
10-74127 | Mar 1998 | JP |
10-074429 | Mar 1998 | JP |
10-198507 | Jul 1998 | JP |
10-227878 | Aug 1998 | JP |
10-240693 | Sep 1998 | JP |
10-320322 | Dec 1998 | JP |
10-326149 | Dec 1998 | JP |
11-24834 | Jan 1999 | JP |
11-184607 | Jul 1999 | JP |
11-194863 | Jul 1999 | JP |
11-194872 | Jul 1999 | JP |
11-194882 | Jul 1999 | JP |
11-194883 | Jul 1999 | JP |
11-194891 | Jul 1999 | JP |
11-195353 | Jul 1999 | JP |
11-203045 | Jul 1999 | JP |
11-212725 | Aug 1999 | JP |
11-272378 | Oct 1999 | JP |
11-338628 | Dec 1999 | JP |
2000-200147 | Jul 2000 | JP |
2000-215549 | Aug 2000 | JP |
2000-267777 | Sep 2000 | JP |
2000-267786 | Sep 2000 | JP |
2000-267797 | Sep 2000 | JP |
2000-353045 | Dec 2000 | JP |
2001-11769 | Jan 2001 | JP |
2001-22508 | Jan 2001 | JP |
2001-184158 | Jul 2001 | JP |
3085481 | Feb 2002 | JP |
2002-215311 | Aug 2002 | JP |
2003-015796 | Jan 2003 | JP |
2003-060754 | Feb 2003 | JP |
2003-099198 | Apr 2003 | JP |
2003-150303 | May 2003 | JP |
2003-517674 | May 2003 | JP |
2003-280799 | Oct 2003 | JP |
2003-280807 | Oct 2003 | JP |
2004-362097 | Dec 2004 | JP |
2005-251218 | Sep 2005 | JP |
2005-285140 | Oct 2005 | JP |
2005-293606 | Oct 2005 | JP |
2006-004453 | Jan 2006 | JP |
2006-178962 | Jul 2006 | JP |
3852854 | Dec 2006 | JP |
2007-123473 | May 2007 | JP |
1998-71394 | Oct 1998 | KR |
1999-50198 | Jul 1999 | KR |
2000-0008579 | Feb 2000 | KR |
2001-0052016 | Jun 2001 | KR |
2001-108361 | Dec 2001 | KR |
2002-65059 | Aug 2002 | KR |
10-2006-0021678 | Mar 2006 | KR |
431607 | Apr 2001 | TW |
00470193 | Dec 2001 | TW |
547716 | Aug 2003 | TW |
1220491 | Aug 2004 | TW |
WO-9417494 | Aug 1994 | WO |
WO-9500897 | Jan 1995 | WO |
WO-9627968 | Sep 1996 | WO |
WO-9814863 | Apr 1998 | WO |
WO-9949443 | Sep 1999 | WO |
WO-0079772 | Dec 2000 | WO |
WO-0102949 | Jan 2001 | WO |
WO-0144912 | Jun 2001 | WO |
WO-0208881 | Jan 2002 | WO |
WO-03025960 | Mar 2003 | WO |
WO-03044645 | May 2003 | WO |
WO-03044956 | May 2003 | WO |
WO-03088176 | Oct 2003 | WO |
WO-03090008 | Oct 2003 | WO |
WO-2004001573 | Dec 2003 | WO |
WO-2004040606 | May 2004 | WO |
WO-2004091956 | Oct 2004 | WO |
WO-2005055620 | Jun 2005 | WO |
WO-2005076117 | Aug 2005 | WO |
WO-2005114369 | Dec 2005 | WO |
WO-2005124526 | Dec 2005 | WO |
WO-2006020305 | Feb 2006 | WO |
WO-2006021211 | Mar 2006 | WO |
WO-2006037545 | Apr 2006 | WO |
WO 2006104745 | Oct 2006 | WO |
WO-2006135127 | Dec 2006 | WO |
WO-2007025858 | Mar 2007 | WO |
WO-2007078477 | Jul 2007 | WO |
WO-2007084467 | Jul 2007 | WO |
WO-2007089766 | Aug 2007 | WO |
WO-2008007372 | Jan 2008 | WO |
WO-2008045414 | Apr 2008 | WO |
WO-2008045833 | Apr 2008 | WO |
Entry |
---|
“About Quicktip®” www.loqicad3d.com/docs/qt.html, downloaded Apr. 8, 2002. |
Ahl, David, “Controller Update,” Creative Computing vo. 9, No. 12, Dec. 1983. |
Ahmad, “A Usable Real-Time 3D Hand Tracker,” Proceedings of the 28th Asilomar Conference on Signals, Systems and Computers—Part 2 (of 2) vol. 2 (Oct 1994). |
“Alps Electric introduces the GlidePoint Wave Keyboard; combines a gentily curved design with Alps' advanced GlidePoint Technology,” Business Wire (Oct. 21, 1996). |
“Alps Electric Ships GlidePoint Keyboard for the Macintosh; Includes a GlidePoint Touchpad, Erase-Eaze Backspace Key and Contoured Wrist Rest,” Business Wire (Jul. 1, 1996). |
Apple Presents iPod: Ultra-Portable MP3 Music Player Puts 1,000 Songs in Your Pocket, retreived from htto://www.apple.com/pr/library/2001/oct/23ipod.html on Oct. 23, 2001. |
“Apple Unveils Optical Mouse and New Pro Keyboard,” Press Release, Jul. 19, 2000. |
“APS show guide to exhibitors,” Physics Today 49(3) (Mar. 1996). |
“Atari VCS/2600 Peripherals,” www.classicgaming.com/gamingmuseum/2006p.html, downloaded Feb. 28, 2007, pp. 1-15. |
Baig, E.C., “Your PC Just Might Need a Mouse,” U.S. News & World Report 108(22) (Jun. 4, 1990). |
Bang & Olufsen Telecom a/s. (2000). BeoCom 6000 User Guide; 53 pages. |
Bartimo, Jim, “The Portables: Travelling Quickly,” Computerworld (Nov. 14, 1983). |
Boling, Douglas (1993) “Programming Microsoft Windows CE.NET,” p. 109. |
Bray, “Phosphors help switch on xenon,” Physics in Action, pp. 1-3, Apr. 1999. |
Brink et al., “Pumped-up portables,” U.S. News & World Report 116(21) (May 30, 1994). |
Brown et al., “Windows on Tablets as a Means of Achieving Virtual Input Devices,” Human-Computer Interaction—INTERACT '90 (1990). |
Buxton et al., “Issues and Techniques in Touch-Sensitive Tablet Input,” Computer Graphics 19(3), Proceedings of SIGGRAPH '85 (1985). |
Chapweske, Adam “PS/2 Mouse/Keyboard Protocol,” 1999, http://panda.cs.ndsu.nodak.edu/˜achapwes/PICmicro/PS2/ps2.htm. |
Chen et al., “A Study in Interactive 3-D Rotation Using 2-D Control Devices,” Computer Graphics 22(4) (Aug. 1988). |
De Meyer, Kevin, “Crystal Optical Mouse,” Feb. 14, 2002, Heatseekerz, Web Article 19. |
“Der Klangmeister,” Connect Magazine, Aug. 1998. |
“Design News literature plus,” Design News 51(24) (Dec. 18, 1995). |
“Diamond Multimedia Announces Rio PMP300 Portable MP3 Music Player,” located at http://news.harmony-central.com/Newp/1998/Rio-PMP300.html visited on May 5, 2008. (4 pages). |
Evans et al., “Tablet-based Valuators that Provide One, Two, or Three Degrees of Freedom,” Computer Grahics 15(3) (Aug. 1981). |
Evb Elektronik “TSOP6238 IR Receiver Modules for Infrared Remote Control Systems” dated Jan. 2004 1 page. |
Fiore, “Zen Touchpad,” Cornell University, May 2000. |
Gadgetboy, “Point and click with the latest mice,” CNET Asia Product Review, www.asia.cnet.com/reviews . . .are/qadgetboy/0,39001770,380235900,00.htm, downloaded Dec. 5, 2001. |
Gfroerer, “Photoluminescence in Analysis of Surfaces and Interfaces,” Encyclopedia of Analytical Chemistry, pp. 1-23, Copyright John Wiley & Sons Ltd, Chichester, 2000. |
Interlink Electronics, VersaPad: Integration Guide, ©1998 (VersaPad), pp. 1-35. |
Jesitus, John, “Broken promies?”, Industry Week/IW 246(20) (Nov. 3, 1997). |
Kobayashi (1996) “Design of Dynamic Soundscape: Mapping Time to Space for Audio Browsing with Simultaneous Listening,” Thesis submitted to Program in Media Arts and Sciences at the Massachusetts Institute of Technology, (58 pages). |
Kobayashi et al. (1994) “Development of the Touch Switches with the Click Response,” Koukuu Denshi Gihou No. 17, pp. 44-48 (published by the Japan Aviation Electronics Industry, Ltd.). |
Kobayashi et al. (1997) “Dynamic Soundscape: Mapping Time to Space for Audio Browsing,” Computer Human Interaction: 16 pages. |
Letter re: Bang & Olufsen a/s by David Safran, Nixon Peabody, LLP, May 21, 2004, with BeoCom 6000 Sales Training Brochure, 7 pages. |
Luna Technologies International, Inc., LUNA Photoluminescent Safety Products, “Photoluminescence-What is Photoluminescence?” from website at http://www.lunaplast.com/photoluminescence.com on Dec. 27, 2005. |
“Manufactures,” Laser Focus World, Buyers Guide '96, 31(12) (Dec. 1995). |
Mims, Forrest M., III, “A Few Quick Pointers; Mouses, Touch Screens, Touch Pads, Light Pads, and the Like Can Make System Easier to Use,” Computers & Electronics (22) (May 1984). |
Nass, Richard, “Touchpad input device goes digital to give portable systems a desktop 'mouse-like' feel,” Electronic Design 44(18) (Sep. 3, 1996). |
“National Design Engineering Show,” Design News 52(5) (Mar. 4, 1996). |
“Neuros MP3 Digital Audio Computer,” www.neurosaudio.com, downloaded Apr. 9, 2003. |
“OEM Touchpad Modules” website www.glidepoint.com/sales/modules.index.shtml, downloaded Feb. 13, 2002. |
Perenson, Melissa, “New & Improved: Touchpad Redux,” PC Magazine (Septmber 10, 1996). |
Petersen, Marty, “Koala Pad Touch Tablet & Micro Illustrator Software,” InfoWorld (Oct. 10, 1983). |
Petruzzellis, “Force-Sensing Resistors,” Electronics Now 64(3) (Mar. 1993). |
Photographs of Innovation 2000 Best of Show Award Presented at the 2000 Intl CES Innovations Design & Engineering Showcase, Jan. 6, 2000, 1 page. |
“Preview of exhibitor booths at the Philadelphia show,” Air Conditioning, Heating & News 200(2) (Jan. 13, 1997). |
“Preview of exhibitor booths at the Philadelphia show, ” Air Conditioning, Heating & News 200(2) (Jan. 13,1997). |
“Product news,” Design News 53(11) (Jun. 9, 1997). |
“Product news,” Design News 53(9) (May 5, 1997). |
“Product Overview—ErgoCommander®,” www.logicad3d.com/products/ErgoCommander.htm, downloaded Apr. 8, 2002. |
“Product Overview—SpaceMouse® Classic,” www.logicad3d.com/products/Classic.htm, downloaded Apr. 8, 2002. |
SanDisk Sansa Connect User Guide, 2007; 29 pages. |
Soderholm, Lars G., Sensing Systems for 'Touch and Feel,' Design News (May 8, 1989): pp. 72-76. |
Sony presents “Choice Without Compromise” at IBC '97 M2 Presswire (Jul. 24, 1997). |
Spiwak, Marc, “A Great New Wireless Keyboard,” Popular Electronics 14(12) (Dec. 1997). |
Spiwak, Marc, “A Pair of Unusual Controllers,” Popular Electronics 14(4) (Apr. 1997). |
Sylvania, “Intellvision™ Intelligent Television Master Component Service Manual,” pp. 1, 2 and 8, 1979. |
“Synaptics Tough Pad Interfacing Guide,” Second Edition, Mar. 25, 1998, Synaptics, Inc., San Jose, CA, pp. 1-90. |
“System Service and Troubleshooting Manual,” www.dsplib.com/intv/Master,. |
Tessler, Franklin, “Point Pad,” Macworld 12(10) (Oct. 1995). |
Tessler, Franklin, “Smart Input: How to Chose from the New Generation of Innovative Input Devices,” Macworld 13(5) (May 1996). |
Tessler, Franklin, “Touchpads,” Macworld 13(2) (Feb. 1996). |
Translation of Trekstor's Defense Statement to the District Court Mannheim of May 23, 2008; 37 pages. |
“Triax Custom Controllers due; Video Game Controllers,” HFD-The Weekly Home Furnishing Newspaper 67(1) (Jan. 4, 1993). |
Tsuk et al., U.S. Office Action mailed Sep. 30, 2004, directed to U.S. Appl. No. 10/256,716; 11 pages. |
“Touchpad,” Notebook PC Manual, Acer Information Co. Ltd., Feb. 16, 2005, pp. 11-12. |
Tsuk et al., U.S. Office Action mailed Jun. 24, 2005, directed to U.S. Appl. No. 10/256,716; 12 pages. |
Tsuk et al., U.S. Office Action mailed Jan. 10, 2006, directed to U.S. Appl. No. 10/256,716; 12 pages. |
Tsuk et al., U.S. Office Action mailed Aug. 3, 2006, directed to U.S. Appl. No. 10/256,716; 15 pages. |
Tsuk et al., U.S. Office Action mailed Oct. 13, 2006, directed to U.S. Appl. No. 10/256,716; 16 pages. |
Tsuk et al., U.S. Office Action mailed Aug. 7, 2009, directed to U.S. Appl. No. 11/610,181; 20 pages. |
Tsuk et al., U.S. Office Action mailed Jul. 7, 2009, directed to U.S. Appl. No. 11/610,190; 24 pages. |
Tsuk et al., U.S. Office Action mailed Dec. 31, 2009, directed to U.S. Appl. No. 11/610,190; 25 pages. |
Tsuk et al., U.S. Office Action mailed Apr. 28, 2010, directed to U.S. Appl. No. 11/610,190; 29 pages. |
Tsuk et al., U.S. Office Action mailed Aug. 6, 2010, directed to U.S. Appl. No. 11/610,190; 30 pages. |
Tsuk et al., U.S. Office Action mailed Apr. 19, 2011, directed to U.S. Appl. No. 11/610,190; 25 pages. |
Tsuk et al., U.S. Office Action mailed Nov. 1, 2010, directed to U.S. Appl. No. 11/959,918; 8 pages. |
Tsuk et al., U.S. Office Action mailed Mar. 31, 2011, directed to U.S. Appl. No. 11/959,918; 9 pages. |
Tsuk et al., U.S. Office Action mailed Oct. 26, 2010, directed to U.S. Appl. No. 11/959,942; 27 pages. |
Robbin et al., U.S. Office Action mailed Sep. 30, 2004, directed to U.S. Appl. No. 10/259,159; 14 pages. |
Robbin et al., U.S. Office Action mailed Jun. 16, 2005, directed to U.S. Appl. No. 10/259,159; 16 pages. |
Robbin et al., U.S. Office Action mailed Jan. 11, 2006, directed to U.S. Appl. No. 10/259,159; 15 pages. |
Robbin et al., U.S. Office Action mailed Aug. 3, 2006, directed to U.S. Appl. No. 10/259,159; 15 pages. |
Robbin et al., U.S. Office Action mailed Oct. 13, 2006, directed to U.S. Appl. No. 10/259,159; 18 pages. |
Robbin et al., U.S. Office Action mailed Jan. 18, 2007, directed to U.S. Appl. No. 10/259,159; 18 pages. |
Robbin et al., U.S. Office Action mailed Aug. 10, 2009, directed to U.S. Appl. No. 11/610,376; 11 pages. |
Robbin et al., U.S. Office Action mailed Aug. 12, 2009, directed to U.S. Appl. No. 11/610,384; 20 pages. |
Robbin et al., U.S. Office Action mailed Oct. 29, 2010, directed to U.S. Appl. No. 11/838,845; 8 pages. |
Robbin et al., U.S. Office Action mailed Apr. 26, 2011, directed to U.S. Appl. No. 11/838,845; 9 pages. |
Zadesky et al., U.S. Office Action mailed Oct. 27, 2006, directed to U.S. Appl. No. 10/643,256; 14 pages. |
Zadesky et al., U.S. Office Action mailed Mar. 23, 2007, directed to U.S. Appl. No. 10/643,256; 11 pages. |
Zadesky et al., U.S. Office Action mailed Jul. 13, 2007, directed to U.S. Appl. No. 10/643,256; 13 pages. |
Zadesky et al., U.S. Office Action mailed Dec. 12, 2007, directed to U.S. Appl. No. 10/643,256; 12 pages. |
Zadesky et al., U.S. Office Action mailed Jul. 9, 2008, directed to U.S. Appl. No. 10/643,256; 12 pages. |
Zadesky et al., U.S. Office Action mailed Mar. 30, 2010, directed to U.S. Appl. No. 11/592,679; 13 pages. |
Zadesky et al., U.S. Office Action mailed Mar. 31, 2011, directed to U.S. Appl. No. 11/882,005; 7 pages. |
Ng et al., U.S. Office Action mailed Jan. 14, 2010, directed to U.S. Appl. No. 11/394,493; 20 pages. |
Ng et al., U.S. Office Action mailed Jun. 22, 2010, directed to U.S. Appl. No. 11/394,493; 14 pages. |
Ng et al., U.S. Office Action mailed Jan. 15, 2010, directed to U.S. Application No. 11/882,423; 22 pages. |
Ng et al., U.S. Office Action mailed Dec. 9, 2010, directed to U.S. Appl. No. 11/394,493; 13 pages. |
Ng et al., U.S. Office Action mailed Jul. 8, 2010, directed to U.S. Appl. No. 11/882,423; 19 pages. |
Ng et al., U.S. Office Action mailed Oct. 26, 2010, directed to U.S. Appl. No. 11/882,423;18 pages. |
Ng et al., U.S. Office Action mailed May 11, 2011, directed to U.S. Appl. No. 11/882,423; 20 pages. |
Ng et al., U.S. Office Action mailed Jun. 21, 2011, directed to U.S. Appl. No. 11/394,493; 18 pages. |
Ng et al., U.S. Office Action mailed Oct. 27, 2011, directed to U.S. Appl. No. 11/882,423; 24 pages. |
Forsblad et al., U.S. Office Action mailed Jan. 26, 2009, directed to U.S. Appl. No. 11/355,022; 15 pages. |
Forsblad et al., U.S. Office Action mailed Jun. 25, 2009, directed to U.S. Appl. No. 11/355,022; 18 pages. |
Forsblad et al., U.S. Office Action mailed Jan. 27, 2009, directed to U.S. Appl. No. 11/882,421; 15 pages. |
Lampell, U.S. Office Action mailed Sep. 15, 2009, directed to U.S. Appl. No. 11/530,807; 15 pages. |
Lampell, U.S. Office Action mailed Jun. 4, 2010, directed to U.S. Appl. No. 11/530,807; 15 pages. |
Lampell, U.S. Office Action mailed Dec. 3, 2010, directed to U.S. Appl. No. 11/530,807; 17 pages. |
Lampell et al., U.S. Office Action mailed Dec. 22, 2010, directed to U.S. Appl. No. 11/882,427; 16 pages. |
Zadesky et al., U.S. Office Action mailed Nov. 20, 2007, directed to U.S. Appl. No. 11/057,050; 33 pages. |
Zadesky et al, U.S. Office Action mailed Aug. 19, 2008, directed to U.S. Appl. No. 11/057,050; 23 pages. |
Zadesky et al., U.S. Office Action mailed Nov. 26, 2008, directed to U.S. Appl. No. 11/057,050; 25 pages. |
Zadesky et al., U.S. Office Action mailed Dec. 24, 2008, directed to U.S. Appl. No. 11/057,050; 25 pages. |
Zadesky et al., U.S. Office Action mailed Feb. 20, 2009, directed to U.S. Appl. No. 11/057,050; 25 pages. |
Zadesky et al., U.S. Office Action mailed Aug. 6, 2009, directed to U.S. Appl. No. 11/057,050; 30 pages. |
Zadesky et al., U.S. Office Action mailed Oct. 4, 2010, directed to U.S. Appl. No. 11/057,050; 31 pages. |
Zadesky et al., U.S. Office Action mailed Mar. 5, 2009, directed to U.S. Appl. No. 11/477,469; 12 pages. |
Zadesky et al., U.S. Office Action mailed Feb. 4, 2010, directed to U.S. Appl. No. 11/477,469; 14 pages. |
Zadesky et al., U.S. Office Action mailed Nov. 16, 2010, directed to U.S. Appl. No. 11/477,469; 13 pages. |
Prest et al., U.S. Office Action mailed Jun. 22, 2010, directed to U.S. Appl. No. 11/878,132; 32 pages. |
Prest et al., U.S. Office Action mailed Jun. 22, 2010, directed to U.S. Appl. No. 11/882,882; 32 pages. |
Prest et al., U.S. Office Action mailed Jun. 23, 2010, directed to U.S. Appl. No. 11/812,384; 29 pages. |
Prest et al., U.S. Office Action mailed Jun. 22, 2010, directed to U.S. Appl. No. 11/882,890; 15 pages. |
Prest et al., U.S. Office Action mailed Jun. 22, 2010, directed to U.S. Appl. No. 11/812,383; 21 pages. |
Prest et al., U.S. Office Action mailed Jun. 23, 2010, directed to U.S. Appl. No. 11/882,889; 13 pages. |
Bollinger et al., U.S. Office Action mailed Jun. 25, 2010, directed to U.S. Appl. No. 11/842,724; 22 pages. |
Bollinger et al., U.S. Office Action mailed Mar. 21, 2011, directed to U.S. Appl. No. 11/842,724; 22 pages. |
McKillop et al., U.S. Office Action mailed Sep. 16, 2010, directed to U.S. Appl. No. 11/591,752; 14 pages. |
McKillop et al., U.S. Office Action mailed Mar. 24, 2011, directed to U.S. Appl. No. 11/591,752; 11 pages. |
Zadesky et al., U.S. Office Action mailed Mar. 4, 2004, directed to U.S. Appl. No. 10/188,182; 8 pages. |
Zadesky et al., U.S. Office Action mailed Jul. 30, 2004, directed to U.S. Appl. No. 10/188,182; 7 pages. |
Zadesky et al., U.S. Office Action mailed Sep. 21, 2005, directed to U.S. Appl. No. 10/188,182; 10 pages. |
Zadesky et al., U.S. Office Action mailed Oct. 4, 2007, directed to U.S. Appl. No. 11/386,238; 12 pages. |
Zadesky et al.., U.S. Office Action mailed Oct. 4, 2007, directed to U.S. Appl. No. 11/806,957; 14 pages. |
Zadesky et al., U.S. Office Action mailed Aug. 2, 2010, directed to U.S. Appl. No. 11/882,004; 9 pages. |
Zadesky et al, U.S. Office Action mailed Feb. 1, 2011, directed to U.S. Appl. No. 11/882,004; 16 pages. |
Zadesky et al., U.S. Office Action mailed Sep. 29, 2010, directed to U.S. Appl. No. 11/882,003; 13 pages. |
Zadesky et al., U.S. Office Action mailed Mar. 16, 2011, directed to U.S. Appl. No. 11/882,003; 12 pages. |
Marriott et al., U.S. Office Action mailed Jun. 2, 2006, directed to U.S. Appl. No. 10/722,948; 12 pages. |
Marriott et al., U.S. Office Action mailed Dec. 12, 2006, directed to U.S. Appl. No. 10/722,948; 14 pages. |
Marriott et al., U.S. Office Action mailed Jul. 13, 2007, directed to U.S. Appl. No. 10/722,948; 15 pages. |
Marriott et al., U.S. Office Action mailed Jan. 30, 2008, directed to U.S. Appl. No. 10/722,948; 17 pages. |
Marriott et al., U.S. Office Action mailed Aug. 19, 2010, directed to U.S. Appl. No. 11/882,422; 13 pages. |
Hotelling, U.S. Office Action mailed Jan. 25, 2010, directed to U.S. Appl. No. 11/482,286; 17 pages. |
Hotelling, U.S. Office Action mailed Sep. 1, 2009, directed to U.S. Appl. No. 11/482,286; 14 pages. |
Hotelling, U.S. Office mailed Jun. 9, 2010, directed to U.S. Appl. No. 11/482,286; 21 pages. |
Hotelling, U.S. Office Action mailed Oct. 1, 2010, directed to U.S. Appl. No. 11/482,286; 28 pages. |
Hotelling, U.S. Office Action mailed Dec. 8, 2010, directed to U.S. Appl. No. 11/482,286; 33 pages. |
Hotelling, U.S. Office Action mailed Aug. 18, 2010, directed to U.S. Appl. No. 11/882,424; 16 pages. |
Hotelling et al., U.S. Office Action mailed Jul. 24, 2009, directed to U.S. Appl. No. 11/483,008; 17 pages. |
Hotelling et al., U.S. Office Action mailed Mar. 30, 2010, directed to U.S. Appl. No. 11/483,008; 20 pages. |
Hotelling et al., U.S. Office Action mailed Oct. 27, 2010, directed to U.S. Appl. No. 11/483,008; 23 pages. |
Hotelling et al., U.S. Office Action mailed Jul. 27, 2009, directed to U.S. Appl. No. 11/882,420; 17 pages. |
Elias et al., U.S. Office Action mailed Sep. 17, 2008, directed to U.S. Appl. No. 11/203,692; 8 pages. |
Elias et al., U.S. Office Action mailed Feb. 23, 2009, directed to U.S. Appl. No. 11/203,692; 13 pages. |
Elias et al., U.S. Office Action mailed Aug. 4, 2009, directed to U.S. Appl. No. 11/203,692; 12 pages. |
Elias et al., U.S. Office Action mailed Mar. 30, 2010, directed to U.S. Appl. No. 11/203,692; 15 pages. |
Elias et al., U.S. Office Action mailed Jun. 11, 2010, directed to U.S. Appl. No. 11/203,692; 17 pages. |
Elias et al., U.S. Office Action mailed Nov. 22, 2010, directed to U.S. Appl. No. 11/203,692; 6 pages. |
Bull, U.S. Office Action mailed Jul. 9, 2010, directed to U.S. Appl. No. 11/849,801; 13 pages. |
Bull, U.S. Office Action mailed Feb. 4, 2011, directed to U.S. Appl. No. 11/849,801; 22 pages. |
Weber et al., U.S. Office Action mailed Jun. 7, 2010, directed to U.S. Appl. No. 11/856,530; 15 pages. |
Weber et al, U.S. Office Action mailed Jan. 7, 2011, directed to U.S. Appl. No. 11/856,530; 13 pages. |
Rathnam et al., U.S. Office Action mailed Mar. 24, 2011, directed to U.S. Appl. No. 12/205,757; 14 pages. |
Weber et al., U.S. Office Action mailed Oct. 13, 2010, directed to U.S. Appl. No. 12/205,795; 15 pages. |
Weber et al., U.S. Office Action mailed Jan. 7, 2011, directed to U.S. Appl. No. 12/205,795; 21 pages. |
Weber et al., U.S. Office Action mailed Feb. 17, 2011, directed to U.S. Appl. No. 12/844,502; 11 pages. |
Lynch et al., U.S. Office Action mailed Jan. 27, 2010, directed to U.S. Appl. No. 11/499,360; 8 pages. |
Lynch et al., U.S. Office Action mailed Oct. 5, 2009, directed to U.S. Appl. No. 11/499,360; 7 pages. |
Bollinger, U.S. Appl. No. 60/858,404 filed Nov. 13, 2006, entitled “Method of Capacitively Sensing Finger Position”; 13 pages. |
Elias et al., U.S. App. No. 60/522,107 filed Aug. 16, 2004, entitled, “A Method for Increasing the Spatial Resolution of Touch Sensitive Devices”; 15 pages. |
Fisher et al., U.S. Appl. No. 61/036,804 filed Mar. 14, 2008 entitled “Switchable Sensor Configurations”; 46 pages. |
Grignon et al., U.S. Appl. No. 60/755,656 filed Dec. 30, 2005, entitled “Touch Pad with Feedback”; 109 pages. |
Hotelling, U.S. Appl. No. 60/658,777 titled Multi-Functional Hand-held Device, filed Mar. 4, 2005; 68 pages. |
Lampell et al., U.S. Appl. No. 60/810,423 filed Jun. 2, 2006, entitled “Techniques for Interactive Input to Portable Electronic Devices”; 53 pages. |
Prest et al., U.S. Appl. No. 60/850,662 filed Oct. 11, 2006, entitled,“Capacitive Scroll Wheel”; 21 pages. |
Rathnam et al., U.S. Appl. No. 60/992,056 filed Dec. 3, 2007, entitled, “Scroll Wheel Circuit Arrangements and Methods of Use Thereof”; 42 pages. |
Rathnam et al., U.S. Appl. No. 61/017,436 filed Dec. 28, 2007, entitled, “Multi-Touch Scroll Wheel Circuit Arrangements And Processing Methods”; 58 pages. |
Robbin et al., U.S. Appl. No. 60/387,692 entitled “Method and Apparatus for Use of Rotational User Inputs”, filed Jun. 10, 2002. |
Robbin et al., U.S. Appl. No. 60/399,806 entitled “Graphical User Interface and Methods of Use Thereof in a Multimedia Player”, filed Jul. 30, 2002. |
Robbin, U.S. Appl. No. 60/346,237 entitled, “Method and System for List Scrolling”, filed Oct. 22, 2001; 12 pages. |
Rothkopf, U.S. Appl. No. 60/935,854 titled “Compact Media Players”, filed Sep. 4, 2007; 36 pages. |
Weber et al., U.S. Appl. No. 61/020,531 filed Jan. 11, 2008 entitled “Modifiable Clickwheel Text ”, 11 pages. |
Weber et al., U.S. App. No. 61/025,531 filed Feb. 1, 2008 entitled “Co-Extruded Materials and Methods”; 11 pages. |
Zadesky et al., U.S. Appl. No. 60/359,551 entitled “Touchpad for Handheld Device,” filed Feb. 25, 2002; 34 pages. |
Zadesky et al., U.S. Appl. No. 60/714,609 entitled “Scrolling Input Arrangements Using Capacitive Sensors on a Flexible Membrane,” filed Sep. 6, 2005; 17 pages. |
Australian Office Action issued Aug. 19, 2009, directed to AU 20006333374; 2 pages. |
Australian Office Action issued Nov. 12, 2009, directed to AU 20006333374; 2 pages. |
Australian Office Action issued Apr. 15, 2010, directed to AU Application No. 2006333374; 2 pages. |
Canadian Office Action mailed Nov. 15, 2011, directed to CA Application No. 2,635,517; 5 pages. |
Chinese Office Action issue Dec. 29, 2006, directed to CN Application No. 200510103886.3, 25 pages. |
Chinese Office Action dated Aug. 10, 2010, directed to corresponding application No. 200680053630.9; 16 pages. |
Second Office Action dated Jan. 5, 2012, directed to Chinese Application No. 200680053630.9; 7 pages. |
German Office Action dated Oct. 19, 2009, directed to corresponding application No. 11 2006 003 531.2-53, 9 pages. ✓. |
German Office Action mailed Jul. 26, 2010 directed to counterpart application No. 112006003531.2; 6 pages. |
EP Communication Pursuant to Article 94(3) EPC dated Aug. 13, 2009, directed to Application No. 06 838 571.5; 5 pages. |
EP Communication under Rule 71(3) EPC dated Feb. 21, 2011, directed to application No. 06838571.5; 66 pages. |
GB Search Report mailed Feb. 9, 2009, directed to counterpart GB Application No. 0811605.5; 4 pages. |
KIPO's Notice of Preliminary Rejection dated Feb. 19, 2010, directed to counterpart application No. 10-2008-7018716; 9 pages. |
Third Office Action dated Aug. 2, 2012, directed to Chinese Patent Application No. 200680053630.9; 14 pages. |
Office Action mailed Jan. 22, 2013, directed to Canadian Application No. 2,635,517; 4 pages. |
First Examination Report dated May 10, 2013, directed to Indian Patent Application No. 2583/KOLNP/2008; 3 pages. |
Number | Date | Country | |
---|---|---|---|
20120206392 A1 | Aug 2012 | US |
Number | Date | Country | |
---|---|---|---|
60755656 | Dec 2005 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 11394493 | Mar 2006 | US |
Child | 13453760 | US |