Products and processes for providing force sensations in a user interface

Information

  • Patent Grant
  • 8917234
  • Patent Number
    8,917,234
  • Date Filed
    Wednesday, October 15, 2003
    21 years ago
  • Date Issued
    Tuesday, December 23, 2014
    10 years ago
Abstract
Products and processes for providing haptic feedback in a user interface device are disclosed. In one exemplary embodiment, a process comprises defining a first cell, mapping a first location of a matrix with the defined first cell, and mapping a second location of the matrix with the defined first cell. The first cell comprises a first parameter representing a first haptic effect.
Description
FIELD OF THE INVENTION

The invention generally relates to products and processes for providing haptic feedback.


BACKGROUND

Tactile cues and feedback enhance the human-machine interface. Providing tactile feedback increases the amount of information available to a user in the operation of a device. Some devices utilize structural tactile methods. One such example is to provide a raised surface on an input surface, e.g., keypad, of the device. Such methods, however, are inherently static, and thus, cannot offer a wide array of, or effective, tactile feedback. Whereas before, one had to rely solely on aural, visual cues, and/or structural tactile cues, active methods of tactile feedback, i.e., haptics, increasingly provide greater and more effective physical cues to users of various devices.


Enhancing the functionality of a device with haptics, however, generally requires additional processor memory and speed. Moreover, enhancing a device with haptics may further complicate design considerations by placing demands on space constraints, especially where manufacturers of such devices, to satisfy consumer demands, strive to reduce the physical size of devices while increasing device functionality.


In a one-dimensional environment, one may select an item from a set of items, e.g., a list of names, by using a “detent” effect, where each item in the item set can be rendered or mapped to a single detent in a one-dimensional device. In a two-dimensional environment, one may select an item in horizontal or vertical directions. Therefore, one may speak of two-dimensional detents, also referred to as matrix-detents.


To create a haptic effect with a single effect located at different locations on a workspace, or interface area (such as for example, a keypad), it is known to create all those haptic effects in the device, and play them each cycle. A shortcoming of this known approach is that to create a two-dimensional detent, more than one effect has to be created. This, in turn, demands more on the communications bandwidth, as well as more memory and computational time on an embedded system, i.e., a system disposed on or within a device or system. Some applications are not well-suited for communicating an amount of traffic demanded by the known approach.


SUMMARY OF THE INVENTION

The invention provides products and processes for providing haptic feedback in a user interface device. In one exemplary embodiment, a single effect can be repeated at several locations in a matrix. One process according to the present invention comprises defining a first cell comprising a first parameter representing a first haptic effect, mapping a first location of a matrix with the defined first cell, and mapping a second location of the matrix with the defined first cell.


In another embodiment, a process comprises providing a cell comprising an arc and first and second edges. The cell forms a wedge of a switch. The process also comprises providing a plurality of force vectors within the cell and delimiting a corner of the wedge. The force vectors are directed radially toward the first and second edges. The corner is formed by the first and second edges.


In another exemplary embodiment, an apparatus comprises a first primary channel disposed about a first axis, a second primary channel disposed about a second axis, a first secondary channel disposed proximate to the first primary channel, and a second secondary channel disposed proximate to the second primary channel.


These exemplary embodiments are mentioned not to limit the invention, but to provide an example of an embodiment of the invention to aid understanding. Exemplary embodiments are discussed in the Detailed Description, and further description of the invention is provided there. Advantages offered by the various embodiments of the present invention may be understood by examining this specification.





BRIEF DESCRIPTION OF THE DRAWINGS

The accompanying drawings, which constitute part of this specification, help to illustrate embodiments of the invention. In the drawings, like numerals are used to indicate like elements throughout.



FIG. 1 is a schematic drawing of a cell according to an embodiment of the present invention.



FIG. 2 is a schematic drawing of a matrix of the cell of FIG. 1.



FIG. 3 is a schematic drawing of a device employing the principles of the invention.



FIG. 4 is a schematic drawing of a cell according to another embodiment of the invention.



FIG. 5 is a schematic drawing of a matrix of the cell of FIG. 4.



FIG. 6 is a schematic drawing of a system according to an embodiment of the invention.



FIG. 7 is a block diagram of a method according to an embodiment of the invention.



FIG. 8 is a schematic drawing of a prior art switch.



FIG. 9 is another schematic drawing of the prior art switch of FIG. 8.



FIG. 10 is a schematic drawing of a switch according to an embodiment of the invention.



FIG. 11 is a block diagram of a method according to an embodiment of the invention.



FIG. 12 is a schematic drawing of a prior art switch.



FIG. 13 is a schematic drawing of a switch according to an embodiment of the invention.



FIG. 14 is a schematic drawing of a switch according another embodiment of the invention.



FIG. 15 is a schematic drawing of another switch according to another embodiment of the invention.





DETAILED DESCRIPTION

Embodiments of the present invention include products and processes for providing haptic feedback in a user interface device. In some interface devices, cutaneous feedback (such as, without limitation, vibration, texture, and heat), is also provided to the user, in addition to kinesthetic feedback (such as, without limitation, forces or resistances sensed by muscles, tendons, and/or joints) both subsumed under the phrase, and more generally known collectively as, “haptic feedback.” The present invention may be embodied in hand-held devices, such as mobile phones, personal digital assistants (“PDAs”), camcorders, and other devices, such as control knobs and computer mice and joysticks.


Referring now to FIG. 1, a two-dimensional cell 10 is shown. As used herein, a cell refers to a memory construct in which parameters represent or define haptic effects. In one embodiment, the cell 10 comprises a plurality of parameters, including a wall 12, a detent 14, a location 16, and a dead-band 18. Preferably, the wall 12 forms a perimeter around the cell 10. Typically, the wall 12 is a high-force constraining or defining a physical boundary of the cell 10.


In one embodiment, the detent 14 is a haptic effect, including, but not limited to, a kinesthetic force profile directing a user toward the location 16. Alternatively, the force profile of the detent 14 directs a user away from the location 16. Various haptic effects can be used for the force profiles of the detent 14. Examples of various haptic effects are described in U.S. Pat. Nos. 6,169,540 and 6,285,351, assigned to the assignee of the present invention and incorporated in their entirety herein by reference.


In general, the location 16 represents a physical area of the cell 10 where a haptic effect is disposed. In one embodiment, the dead-band 18 comprises a physical area of the cell 10 in which no forces are present. The dead-band 18 represents a stable position in the cell 10. As shown in FIG. 1, the dead-band 18 includes a horizontal component as well as a vertical component. Alternatively, other suitable parameters for and arrangements of the cell 10 can be used.


Referring now to FIG. 2, a matrix 20 of cells 21 is shown. Preferably, the matrix 20 is a two-dimensional configuration of the cell 10 described above. The size of the matrix 20 is defined by the number of rows and columns. Although a three-by-three (“3×3”) matrix is shown, other suitable arrangements can be used. Preferably, a repeated single haptic effect is repeated at multiple locations arranged in the matrix 20. Alternatively, different cells having different haptic effects can be arranged in the matrix 20, including an arrangement of active and inactive cells. Other suitable cells and arrangements of cells can be used.


Preferably, the matrix 20 comprises cells similar to that described above with reference to the cell 10. That is, the cells 21 each comprise parameters including a detent 24, a location 26, and a dead-band 28. However, unlike the cell 10, each of the cells 21 do not have a wall. Rather, a wall 22 defines a perimeter of the matrix 20, rather than defining each cell 21. Alternatively, other suitable cells, arrangements of cells, and cell parameters can be used.


Referring now to FIG. 3, a keypad 30 of an interface device (not shown) employing the principles of the invention is shown. In one embodiment, the keypad 30 can be disposed in a mobile telephone. In another embodiment, the keypad 30 can be disposed in a PDA. Alternatively, the keypad 30 can be used in other suitable embodiments, including in a graphical user interface (“GUI”).


The numerals of the keypad 30 correspond substantially to location 26 of the matrix 20 shown in FIG. 2. Thus, a cursor in a GUI or a finger (or other digit) of or a stylus used by a user is guided by the different parameters, i.e., the wall 22, detent 24, and dead-band 28, to the center, i.e., location 26, of each cell of the device. Alternatively, the parameters can be used to guide one away from a specific point or area (not shown) on a device.


Referring now to FIG. 4, a cell 40 according to another embodiment of the invention is shown. The cell 40 is preferably defined by a radius and an arc, depending on the configuration of a particular device. A width 41 (measured in degrees) of the cell 40 is measured as an angle. A detent 48 is defined by a height and the width 41. An angular dead-band 44 is disposed within the area defining detent 48. A matrix dead-band 43 is disposed between a location 46 and the height 42.


The parameters of the cell 40 are similar to that described above with reference to the cell 10. Thus, the location 46 comprises a haptic effect, the detent 48 comprises a force profile, and the matrix and angular dead-bands 43,44 lack any discernable force profile.


Referring now to FIG. 5, a matrix 50 of cells 51 is shown. The matrix 50 is a two-dimensional configuration of the cell 40 described above. The matrix 50 is a 3×4 circular matrix, which is defined as including three rings along the radius of the circle and four quadrants of the circle. Alternatively, other suitable numbers of rings can be used. Also alternatively, rather than dividing the circle into four equally-sized quadrants, the circle can be divided into any other suitable number of wedges (i.e., portions of the circle defined by the radius and an arc).


Each cell 51 comprises a matrix dead-band 53, a first radial force 52, a second radial force 54, a first tangential force 55, a second tangential force 57, and a stable location 56. A wall (not shown) can be disposed surrounding the cell 51. Alternatively, other suitable cells, arrangements of cells, and cell parameters can be used. The matrix 50 can be used in suitable interface devices (not shown), such as a mobile phone, PDA, or GUI.


The first and second radial forces 52,54 are disposed radially within each cell 51. The first and second radial forces 52,54 are equal and opposite to one another. The intersection of the first and second radial forces 52,54 create a stable position along a radius.


The first and second tangential forces 55,57 are disposed tangentially within each cell 51. The first and second tangential forces 55,57 are equal and opposite to one another. The intersection of the first and second tangential forces 55,57 create a stable position along an arc. The intersection of the first and second radial forces 52,54 and the first and second tangential forces 55,57 creates a stable location 56. There is no force profile at stable location 56. The combination of first and second radial forces 52,54, first and second tangential forces 55,57, and matrix dead-band 53 can be used to guide a user toward a particular location of the matrix 50, and thus, the device.


Referring now to FIG. 6, a schematic drawing of a system 60 implementing an embodiment of the invention is shown. The system 60 comprises a host 61, a communications channel 64, and a device 65. The host 61 comprises a host (or first) processor 62 and an effect library 63. The host 61 is located in a suitable location according to the needs of the device 65. Preferably, the host 61 is located remotely from the device 65. The host 61 is in communication with the device 65 through the communication channel 64. The communication channel 64 connects the host 61 and the device 65 through direct, indirect, wireless, or other suitable means.


The host processor 62 can be a computer or any other suitable processor, such as for example, digital logic processors capable of processing input, excuting algorithms, and generating output as needed. Such processors can include a microprocessor, an Application Specific Integrated Circuit (ASIC), and state machines. Such processors include, or can be in communication with media, for example computer readable media, which stores instructions that, when executed by the processor, cause the processor to perform the steps described herein as carried out, or assisted, by a processor.


One embodiment of a suitable computer-readable medium includes an electronic optical, magnetic, or other storage or transmission device capable of providing a processor, such as the processor in a web server, with computer-readable instructions. Other examples of suitable media include, but are not limited to, a floppy disk, CD-ROM, magnetic disk, memory chip, ROM, RAM, ASIC, configured processor, all optical media, all magnetic tape or other magnetic media, or any other medium from which a computer processor can read.


The host controller 62 is in communication with the effect library 63. The effect library 63 stores instructions for a variety of haptic effects. Preferably, the host controller 62 controls the effect library 63. In one embodiment, the host controller 62 communicates instructions to the effect library 63 to communicate a particular haptic effect to the device 65. As described above, the instructions for a particular haptic effect are communicated from the host 61 to the device 65 through the communications channel 64.


The device 65 comprises a local (or second) processor 66, a sensor 67, and an actuator 68. Preferably, the local processor 66 is embedded in (i.e., disposed within) the device 65. Alternatively, the local processor 66 is located in any other suitable location. The local processor 66 is operable to receive and executions received from the host processor 62. Generally, the local processor 66 is also operable to execute instructions autonomously of the host processor 62. Where the local processor 66 is embedded in the device 65, the local processor 66 generally is more limited in size, and thus computational power, than the host processor 62. Notwithstanding the limitations in size and computational power, the local processor 66 is preferably similar to that described above with reference to the host processor 62.


The local processor 66 is in communication with a sensor 67. The sensor 67 can be a single sensor or a plurality of sensors. The sensor 67 is operable to detect a wide variety of conditions, such as, but not limited to, position, pressure, motion, direction, displacement, and changes or deviations in such conditions. The information detected by the sensor 67 is communicated to the local processor 66, which then processes this information and/or communicates the information received from the sensor 67 to the host controller 62 via the communication channel 64.


The local processor 66 is also in communication with an actuator 68. The actuator 68 can be a single actuator or a plurality of actuators. The actuator 68 is operable to receive instructions (i.e., an actuation signal) from the local processor 66 and to output haptic feedback to the device 65. The local processor 66 can vary at least one of frequency, waveform, and magnitude of the actuation signal.


Suitable structures that can provide haptic feedback and that can produce a plurality of distinct haptic sensations, include, but are not limited to, a voice coil and a permanent magnet, rotating masses, a piezo material, such as quartz, Rochelle Salt, and synthetic polycrystalline ceramics, piezoelectric ceramics, piezoelectric films, and electroactive polymers. Alternatively, other suitable actuators can be used.


While system 60 is shown with first and second processors 62,66, an alternate embodiment comprises a single processor (not shown). For example, a stand-alone device can perform the tasks of both first and second processors 62,66. Thus, in this alternate embodiment, the communication channel 64 would be implemented in software rather than hardware as described in the embodiment above with reference to FIG. 6.


Referring now to FIG. 7, a block diagram of a method 70 according to an embodiment of the invention is shown. FIG. 7 shows an embodiment of a method that may be used to generate a matrix of cells, as described above. However, the method 70 may be used to form alternate matrices or cells. Items shown above in FIGS. 1-6, as well as the accompanying description above, are referred to in describing FIG. 7 to aid understanding of the embodiment of the method 70 shown. Thus, the method 70 is not limited to the embodiments described above and with reference to FIGS. 1-6.


As indicated by block 72, the method 70 comprises defining a first cell. In one embodiment, the first cell is defined by the first processor. The first cell comprises first parameter representing a first haptic effect. As described above, various haptic effects can be used. In one embodiment, the first cell comprises a first detent.


In one embodiment, the method 70 comprises communicating the defined first cell from a first processor to a second processor. In another embodiment, the defined first cell is communicated from the first processor to the second processor via a communication channel. In one embodiment, the communication channel is a wireless interface.


As described above, the first processor can be a host processor and the second processor can be a local processor embedded in a device. In one embodiment, the second processor is disposed remotely from the first processor.


As indicated by block 74, the method 70 comprises mapping a first location of a matrix with the defined first cell. In one embodiment, the second processor maps the first location with the defined first cell. In one embodiment, the matrix comprises a square shape, such as, for example, the 3×3 matrix described above and with reference to FIG. 2. In another embodiment, the matrix comprises a circular shape, such as, for example, the 3×4 circular matrix described above and with reference to FIG. 5. Alternatively, other suitable configurations, shapes, and sizes of matrix can be used.


As indicated by block 76, the method 70 comprises mapping a second location of the matrix with the defined first cell. In one embodiment, the second processor maps the second location with the defined first cell. In another embodiment, the method 70 further comprises defining a second cell. Preferably, the second cell is defined by the first processor. The second cell comprises a second haptic effect. Preferably, the second haptic effect is different than the first haptic effect. Alternatively, the first and second haptic effects can be the same. In one embodiment, the second cell comprises a second detent.


In another embodiment, the method 70 further comprises communicating the defined second cell from the first processor to the second processor. In one embodiment, the defined second cell is communicated from the first processor to the second processor via the communication channel. In one embodiment, the method 70 further comprises mapping a third location of the matrix with the defined second cell by the second processor. In another embodiment, the second cell is an inactive cell, whereas the first cell is an active cell.


In one embodiment, the third location is disposed between the first and second locations. Alternatively, the first and second cells can be disposed in other desired arrangements or configurations. The second processor is operable to arrange the first and second cells in suitable arrangements without instructions from the first processor, thus reducing the amount of communication between the first and second processors, and therefore bandwidth of the communication channel. Thus, multiple effects can be displayed in a two-dimensional device by specifying a limited number of parameters and a desired number of cells in the matrix.


In another embodiment, the method 70 further comprises providing an actuator in communication with the first, second and third locations. The actuator can be similar to that described above. In one embodiment, the actuator comprises a plurality of actuators. Preferably, the actuator is operated to provide a computer-modulated force to the first, second, and third locations. The actuator preferably receives an actuating signal from the second processor. Alternatively, the actuator receives instructions from the first processor.


Referring now to FIGS. 8 and 9, a prior art switch 80 is shown. The prior art switch 80 includes a wall 82, which forms a perimeter around the switch 80. The switch 80 includes a plurality of radial detents 84, a location 86, and a dead-band 88. The switch 80 includes a force profile (see FIG. 9) in separate and equally-sized wedges. The force profile includes first 87 and second 89 cartesian forces. The first and second Cartesian forces 87,89 are equal and opposite forces. The intersection of the first and second Cartesian forces 87,89 form the dead-band 88.


In the force profile of the prior art switch shown in FIG. 9, as one approaches the location 86 disposed in the center of the switch 80 the force profiles proximate to the corners approaching the location 86 exert antagonistic forces on the user, thus, providing incoherent and confusing haptic feedback to the user.


Referring now to FIG. 10, a switch 100 according to an embodiment of the invention is shown. The switch 100 comprises a wall 101 forming a perimeter around the switch 100. Preferably, the switch 100 comprises a plurality of cells 104 disposed adjacent to one another. In one embodiment, the switch 100 comprises a plurality of radial detents (not shown). The switch 100 comprises a location 109 and a plurality of dead-bands 108 disposed between the cells 104. In one embodiment, a spring (not shown) is disposed proximate to the location 109. In another embodiment, a centering spring is disposed on top of the switch 100.


Each cell 104 comprises a centerline 103, a corner 105, a first edge 106, and a second edge 107. As shown in FIG. 10, the corner 105 does not extend to the location 109. Further, the corner 105 does not form a sharp edge, but rather a gradual arc. The corner 105 joins first and second edges 106,107. The centerline 103 bisects the cell 104.


Each cell 104 also comprises a force profile (indicated by the vectors). The force profile of each cell 104 is directed outwardly from the centerline 103 toward the first and second edges 106,107 and the corner 105. Thus, the force profile of the switch guides a user toward the dead-bands 108 and the location 109.


Referring now to FIG. 11, a method 110 according to an embodiment of the invention is shown. FIG. 11 shows an embodiment of a method that may be used to make a switch, as described above. However, the method 110 may be used to form alternate switches. Items shown above in FIG. 10, as well as the accompanying description above, are referred to in describing FIG. 11 to aid understanding of the embodiment of the method 110 shown. Thus, the method 110 is not limited to the embodiments described above and with reference to FIG. 10.


As indicated by block 112, the method 110 comprises providing a cell comprising an arc and first and second edges. A plurality of cells form the switch. In one embodiment, the switch comprises a circular shape. In another embodiment, the switch comprises an eight-way switch. The eight-way switch is operable to select a channel about a first axis.


As indicated by block 114, the method 110 comprises providing a plurality of force vectors within the cell. The force vectors are directed radially toward the first and second edges. The force vectors direct a user toward the outside of the cell, i.e., toward a dead-band. As indicated by block 116, the method 110 comprises delimiting a corner of the cell. The corner is delimited by forming an arc joining the first and second edges. Thus, the force profile does not extend completely to a center of the switch. Preferably, the force vectors within the corner are directed toward the center of the switch.


In one embodiment, the method 110 comprises providing a biasing element proximate to the center of the switch. In another embodiment, the method 110 comprises providing a detent proximate to a radius of the switch. Other suitable configurations of switches can be used.


Referring now to FIG. 12, a prior art switch 120 is shown. The switch 120 includes a wall 121, a center 122, a first axis 124, a second axis 126, and a plurality of channels 128. The wall 121 forms a border about the switch 120. The first and second axes 124,126 are orthogonal to one another. The plurality of channels 128 intersect at the center 122 of the switch 120.


The switch 120 shows 12 different channels 128. Each channel is labeled with a letter beginning with the letter “A” and ending with the letter “L.” Thus, a user can select up to 12 options (i.e., any letter from A-L) with the switch 120.


Such a switch is referred to as an N-way switch, where “N” refers to the number of channels that can be selected or activated. As the number of choices increases, the more difficult N-way switches become for one to navigate. It is generally easier to select the four cardinal directions than to select channels that are located at angles, especially when the angles are close together. Of course, reducing the size of the switch to accommodate smaller-sized devices adds to the difficulty in navigating the switch.


Referring now to FIGS. 13 and 14, a switch 130 according to an embodiment of the invention is shown. The switch 130 can be referred to as an M×N switch to improve the navigability of a switch. The “M” dimension represents the number of available channels from a center of the workspace. Once one selects an initial or primary channel (generally along a cardinal direction), the “N” dimension offers additional channels to be explored and selected.


For example, each channel of the switch 130 is labeled with the letters “A” through “L.” The letters A, D, G, and J are labeled along the “M” dimension, while the “N” dimensions are labeled with the letters B, C, E, F, H, I, K, and L. Thus, a 4×3 switch will offer the same number of options as a 12-way switch. However, the 4×3 switch only requires one to select along the cardinal directions making the task of selecting a channel easier for the user.


For example, to spell the word “ace” using the prior art switch of FIG. 12 would require one to select the channels representing the letters “c” and “e,” which are disposed at predetermined angles off of a horizontal or vertical line, something which adds complexity to the task of navigating about the switch. Spelling the word “ace” using the switch of FIG. 13 only requires one to maneuver in a cardinal direction, something which is much simpler to do than in the prior art switch.


Referring again to FIG. 13, the switch 130 comprises a first primary channel 133 disposed about a first axis 131, a second primary channel 134 disposed about a second axis 132. In one embodiment, the first axis 131 is disposed substantially orthogonal to the second axis 132. Alternatively, the first and second axes 131,132 can be disposed in other suitable arrangements.


The switch 130 also comprises a first secondary channel 133a,133b disposed proximate to the first primary channel 133 and a second secondary channel 134a,134b disposed proximate to the second primary channel 134. In one embodiment, the first secondary channel 133a,133b is in communication with the first primary channel 133 and the second secondary channel 134a,134b is in communication with the second primary channel 134.


In one embodiment, the first and second primary channels 133,134 and the first and second secondary channels 133a,133b,134a,134b are activated by one's touch, i.e., by a digit or device, such as a stylus. Alternatively, the first and second primary channels 133,134 and the first and second secondary channels 133a,133b,134a,134b are activated by depressing the switch 130 causing the switch to pivot about the first and/or second axes 131,132.


In the embodiment shown in FIG. 13, the first secondary channel 133a,133b is disposed substantially orthogonal to the first primary channel 133 and the second secondary channel 134a,134b is disposed substantially orthogonal to the second primary channel 134. In the embodiment shown in FIG. 14, the first secondary channel 143a,143b is disposed obliquely to the first primary channel 143 and the second secondary channel 144a,144b is disposed obliquely to the second primary channel 144.


In one embodiment, the switch 130 further comprises a third primary channel 135 disposed substantially co-axial with the first primary channel 133 and a fourth primary channel 136 disposed substantially co-axial with the second primary channel 134. In another embodiment, the switch 130 further comprises a third secondary channel 135a,135b disposed proximate to the third primary channel 135 and a fourth secondary channel 136a,136b disposed proximate to the fourth primary channel 136. In one embodiment, the third secondary channel 135a,135b is in communication with the third primary channel 135 and the fourth secondary channel 136a,136b is in communication with the fourth primary channel 136.


In one embodiment, the third and fourth primary channels 135,136 and the third and fourth secondary channels 135a,135b,136a,136b are activated by one's touch. Alternatively, the third and fourth primary channels 135,136 and the third and fourth secondary channels 135a,135b,136a,136b are activated by depressing the switch 130 causing the switch 130 to pivot about the first and/or second axes 131,132.


In the embodiment shown in FIG. 13, the third secondary channel 135a,135b is disposed substantially orthogonal to the first primary channel 135 and the fourth secondary channel 136a,136b is disposed substantially orthogonal to the second primary channel 136. In the embodiment shown in FIG. 14, the third secondary channel 145a,145b is disposed obliquely to the third primary channel 145 and the second secondary channel 146a,146b is disposed obliquely to the second primary channel 146.


Referring now to FIG. 15, a switch 150 is shown. The switch 150 uses a 4-way switch 152 inside a circular switch 154 having 12 cells or channels. The 4-way switch 152 is similar to the embodiments described above and shows a force profile 153. The 4-way switch 152 allows one to access directly only four cells, after which one can select any of the 12 channels of the circular switch 154. The circular switch 154 shows a force profile 155, as previously described above. Thus, such an embodiment would be referred to as a 4×12 switch, and would provide one with a choice of 48 options. Other suitable variations of the arrangements and configurations of the embodiments described are possible.


While the present invention has been disclosed with reference to certain embodiments, numerous modifications, alterations, and changes to the described embodiments are possible without departing from the sphere and scope of the present invention, as defined by the appended claims. Accordingly, it is intended that the present invention not be limited to the described embodiments, but that it has the full scope defined by the language of the following claims, and equivalents thereof.

Claims
  • 1. A method comprising: defining a graphical user interface having a plurality of graphical input elements arranged in a matrix configuration;defining a first cell, the first cell comprising a first parameter representing a first haptic effect;assigning the first cell to a first graphical input element in the matrix configuration;assigning the first cell to a second graphical input element in the matrix configuration;receiving a sensor signal from a sensor, the sensor configured to detect a movement of a user manipulatable object of an interface device and the sensor signal associated with the movement;determining a position of a graphical object based at least in part on the sensor signal;determining an interaction between the position of the graphical object and at least one of the plurality of graphical input elements; andoutputting the first haptic effect based at least in part on the first parameter and the interaction, the haptic effect configured to resist or assist the movement of the user manipulatable object.
  • 2. The method of claim 1, further comprising communicating the first cell from a first processor to a second processor.
  • 3. The method of claim 2, further comprising: defining a second cell, the second cell comprising a second parameter representing a second haptic effect;communicating the second cell from the first processor to the second processor; andassigning the second cell to a third input element in the matrix configuration.
  • 4. The method of claim 3, wherein the first and second cells are defined by the first processor and the first, second, and third input elements are assigned by the second processor.
  • 5. The method of claim 3, wherein the third input element is disposed between the first and second input elements.
  • 6. The method of claim 1, wherein the first cell comprises a first detent and the second cell comprises a second detent.
  • 7. The method of claim 3, further comprising providing an actuator in communication with the first, second, and third input elements, the actuator operable to provide a computer-modulated force to the first, second, and third input elements.
  • 8. The method of claim 2, wherein the second processor is disposed remotely from the first processor.
  • 9. The method of claim 1, wherein the matrix configuration comprises a square shape.
  • 10. The method of claim 1, wherein the matrix configuration comprises a circular shape.
  • 11. The method of claim 1, wherein the first cell comprises an arc and first and second edges; and wherein the haptic effect comprises a plurality of force vectors within the first cell, the force vectors directed outward from a centerline of the first cell toward the first and second edges.
  • 12. A switch comprising: a sensor;an actuator configured to output a haptic effect; anda processor in communication with the sensor and the actuator, the processor configured to receive a sensor signal from the sensor, and to cause the actuator to generate a haptic effect based at least in part on the sensor signal, wherein the haptic effect is based on a plurality of detents defining: a first primary channel defined along a first axis,a second primary channel defined along a second axis,a first secondary channel proximate to the first primary channel, anda second secondary channel proximate to the second primary channel,the plurality of detents configured to substantially constrain movement of an interface device to one of the first primary channel, the second primary channel, the first secondary channel, or the second secondary channel, wherein: each channel is a substantially one-dimensional channel,the first primary channel intersects the second primary channel,the first secondary channel intersects one of the first or second primary channel, andthe second secondary channel intersects one of the first or second primary channels or the first secondary channel.
  • 13. The switch of claim 12, wherein the switch comprises a circular shape.
  • 14. The switch of claim 12, wherein the switch comprises an eight-way switch, the eight-way switch operable to select a channel about a first axis.
  • 15. The switch of claim 12, further comprising providing a biasing element proximate to a center of the switch.
  • 16. The switch of claim 12, further comprising providing a detent proximate to a radius of the switch.
  • 17. The switch of claim 12, further comprising: a third primary channel defined substantially co-axial with the first primary channel;a fourth primary channel defined substantially co-axial with the second primary channel;a third secondary channel defined proximate to the third primary channel; anda fourth secondary channel defined proximate to the fourth primary channel.
  • 18. The switch of claim 17, wherein the first axis is substantially orthogonal to the second axis.
  • 19. The switch of claim 17, wherein the third secondary channel is oblique to the third primary channel; and the fourth secondary channel is oblique to the fourth primary channel.
  • 20. The switch of claim 17, wherein the third secondary channel is substantially orthogonal to the third primary channel; and the fourth secondary channel is substantially orthogonal to the fourth primary channel.
  • 21. The switch of claim 12, wherein the first secondary channel is oblique to the first primary channel; and the second secondary channel is oblique to the second primary channel.
  • 22. The switch of claim 12, wherein the first secondary channel is substantially orthogonal to the first primary channel; and the second secondary channel is substantially orthogonal to the second primary channel.
  • 23. A non-transitory computer-readable medium comprising program code, the program code comprising: program code for defining a graphical user interface having a plurality of graphical input elements arranged in a matrix configuration;program code for defining a first cell, the first cell comprising a first parameter representing a first haptic effect;program code for assigning the first cell to a first graphical input element in the matrix configuration;program code for assigning the first cell to a second graphical input element in the matrix configuration;program code for receiving a sensor signal from a sensor, the sensor configured to detect a movement of a user manipulatable object of an interface device and the sensor signal associated with the movement;program code for determining an interaction between the position of the graphical object and at least one of the plurality of graphical input elements; andprogram code for outputting the first haptic effect based at least in part on the first parameter and the interaction, the haptic effect configured to resist or assist the movement of the user manipulatable object.
  • 24. The non-transitory computer-readable medium of claim 23, further comprising program code for communicating the first cell from a first processor to a second processor.
  • 25. The non-transitory computer-readable medium of claim 24, further comprising: program code for defining a second cell, the second cell comprising a second parameter representing a second haptic effect;program code for communicating the second cell from the first processor to the second processor; and
RELATED APPLICATIONS AND CLAIM PRIORITY

This application claims priority to U.S. Provisional Application No. 60/419,024, filed Oct. 15, 2002, the priority benefit of which is claimed by this application and which is incorporated in its entirety herein by reference.

US Referenced Citations (356)
Number Name Date Kind
2524782 Ferrar et al. Oct 1950 A
2972140 Hirsch Feb 1961 A
3157853 Hirsch Nov 1964 A
3220121 Cutler Nov 1965 A
3490059 Paulsen et al. Jan 1970 A
3497668 Hirsch Feb 1970 A
3517446 Corlyon et al. Jun 1970 A
3623046 Scourtes Nov 1971 A
3623064 Kagan Nov 1971 A
3875488 Crocker et al. Apr 1975 A
3902687 Hightower Sep 1975 A
3903614 Diamond et al. Sep 1975 A
3911416 Feder Oct 1975 A
4050265 Drennen et al. Sep 1977 A
4103155 Clark Jul 1978 A
4125800 Jones Nov 1978 A
4127752 Lowthorp Nov 1978 A
4148014 Burson Apr 1979 A
4160508 Salsbury Jul 1979 A
4236325 Hall et al. Dec 1980 A
4262549 Schwellenbach Apr 1981 A
4311980 Prusenziati Jan 1982 A
4333070 Barnes Jun 1982 A
4385836 Schmitt May 1983 A
4391282 Ando et al. Jul 1983 A
4400790 Chambers et al. Aug 1983 A
4443952 Schulien et al. Apr 1984 A
4464117 Forest Aug 1984 A
4484191 Vavra Nov 1984 A
4513235 Acklam et al. Apr 1985 A
4546347 Kirsch Oct 1985 A
4581491 Boothroyd Apr 1986 A
4599070 Hladky et al. Jul 1986 A
4637264 Takahashi et al. Jan 1987 A
4639884 Sagues Jan 1987 A
4676115 Morscheck et al. Jun 1987 A
4678908 LaPlante Jul 1987 A
4680466 Kuwahara et al. Jul 1987 A
4692726 Green et al. Sep 1987 A
4695266 Hui Sep 1987 A
4699043 Violante De Dionigi Oct 1987 A
4708656 De Vries et al. Nov 1987 A
4712101 Culver Dec 1987 A
4713007 Alban Dec 1987 A
4724715 Culver Feb 1988 A
4728954 Phelan et al. Mar 1988 A
4734685 Watanabe Mar 1988 A
4758165 Tieman et al. Jul 1988 A
4772205 Chlumsky et al. Sep 1988 A
4776701 Pettigrew Oct 1988 A
4794384 Jackson Dec 1988 A
4794392 Selinko Dec 1988 A
4795907 Maekawa Jan 1989 A
4799055 Nestler et al. Jan 1989 A
4803413 Kendig et al. Feb 1989 A
4811608 Hilton Mar 1989 A
4815006 Andersson et al. Mar 1989 A
4819195 Bell et al. Apr 1989 A
4823106 Lovell Apr 1989 A
4825157 Mikan Apr 1989 A
4840634 Muller et al. Jun 1989 A
4851771 Ikeda et al. Jul 1989 A
4860051 Taniguchi et al. Aug 1989 A
4885565 Embach Dec 1989 A
4891764 McIntosh Jan 1990 A
4891889 Tomelleri Jan 1990 A
4906843 Jones et al. Mar 1990 A
4914976 Wyllie Apr 1990 A
4926879 Sevrain et al. May 1990 A
4930770 Baker Jun 1990 A
4934694 McIntosh Jun 1990 A
4935725 Turnau Jun 1990 A
4935728 Kley Jun 1990 A
4937685 Barker et al. Jun 1990 A
4940234 Ishida et al. Jul 1990 A
4949119 Moncrief et al. Aug 1990 A
4962448 DeMaio et al. Oct 1990 A
4964837 Collier Oct 1990 A
4965446 Vyse Oct 1990 A
4982504 Soderberg et al. Jan 1991 A
5006703 Shikunami et al. Apr 1991 A
5019761 Kraft May 1991 A
5022384 Freels Jun 1991 A
5022407 Horch et al. Jun 1991 A
5024626 Robbins et al. Jun 1991 A
5035242 Franklin Jul 1991 A
5038089 Szakaly Aug 1991 A
5053975 Tsuchihashi et al. Oct 1991 A
5062830 Dunlap Nov 1991 A
5065145 Purcell Nov 1991 A
5068529 Ohno et al. Nov 1991 A
5078152 Bond Jan 1992 A
5079845 Childers Jan 1992 A
5086197 Liou Feb 1992 A
5095303 Clark et al. Mar 1992 A
5107080 Rosen Apr 1992 A
5113179 Scott-Jackson et al. May 1992 A
5116051 Moncrief et al. May 1992 A
5121091 Fujiyama Jun 1992 A
5125281 Mottate Jun 1992 A
5132927 Lenoski et al. Jul 1992 A
5138154 Hotelling Aug 1992 A
5139261 Openiano Aug 1992 A
5148377 McDonald Sep 1992 A
5155423 Karlen et al. Oct 1992 A
5165897 Johnson Nov 1992 A
5168268 Levy Dec 1992 A
5175459 Danial et al. Dec 1992 A
5182557 Lang Jan 1993 A
5186695 Mangseth et al. Feb 1993 A
5195179 Tokunaga Mar 1993 A
5195920 Collier Mar 1993 A
5202961 Mills et al. Apr 1993 A
5204600 Kahkoska Apr 1993 A
5209131 Baxter May 1993 A
5212473 Louis May 1993 A
5216337 Orton et al. Jun 1993 A
5223658 Suzuki Jun 1993 A
5229836 Nagano Jul 1993 A
5230623 Guthrie et al. Jul 1993 A
5235868 Culver Aug 1993 A
5239249 Ono Aug 1993 A
5240417 Smithson et al. Aug 1993 A
5246316 Smith Sep 1993 A
5247648 Watkins et al. Sep 1993 A
5254919 Bridges et al. Oct 1993 A
5271290 Fischer Dec 1993 A
5275174 Cook Jan 1994 A
5275565 Moncrief Jan 1994 A
5280276 Kwok Jan 1994 A
5283970 Aigner Feb 1994 A
5284330 Carlson et al. Feb 1994 A
5289273 Lang Feb 1994 A
5296846 Ledley Mar 1994 A
5299810 Pierce Apr 1994 A
5302132 Corder Apr 1994 A
5309140 Everett, Jr. et al. May 1994 A
5313229 Gilligan et al. May 1994 A
5313230 Venolia et al. May 1994 A
5317336 Hall May 1994 A
5329289 Sakamoto et al. Jul 1994 A
5334027 Wherlock Aug 1994 A
5341459 Backes Aug 1994 A
5351692 Dow et al. Oct 1994 A
5359193 Nyui et al. Oct 1994 A
5374942 Gilligan et al. Dec 1994 A
5379663 Hara Jan 1995 A
5384460 Tseng Jan 1995 A
5389849 Asano et al. Feb 1995 A
5390128 Ryan et al. Feb 1995 A
5390296 Crandall et al. Feb 1995 A
5396267 Bouton Mar 1995 A
5397323 Taylor et al. Mar 1995 A
5398044 Hill Mar 1995 A
5402499 Robison et al. Mar 1995 A
5402582 Raab Apr 1995 A
5402680 Korenaga Apr 1995 A
5414337 Schuler May 1995 A
5417696 Kashuba et al. May 1995 A
5428746 Dalrymple Jun 1995 A
5436542 Petelin et al. Jul 1995 A
5436622 Gutman et al. Jul 1995 A
5436640 Reeves Jul 1995 A
5437607 Taylor Aug 1995 A
5452615 Hilton Sep 1995 A
5457479 Cheng Oct 1995 A
5457793 Elko et al. Oct 1995 A
5461711 Wang et al. Oct 1995 A
5466213 Hogan Nov 1995 A
5467763 McMahon et al. Nov 1995 A
5473344 Bacon et al. Dec 1995 A
5474082 Junker Dec 1995 A
5481914 Ward Jan 1996 A
5491477 Clark et al. Feb 1996 A
5512919 Araki Apr 1996 A
5514150 Rostoker May 1996 A
5524195 Clanton, III et al. Jun 1996 A
5526022 Donahue et al. Jun 1996 A
5530455 Gillick et al. Jun 1996 A
5543821 Marchis et al. Aug 1996 A
5547382 Yamasaki Aug 1996 A
5547383 Yamaguchi Aug 1996 A
5550562 Aoki et al. Aug 1996 A
5550583 Matheny et al. Aug 1996 A
5565840 Thorner et al. Oct 1996 A
5570111 Barrett et al. Oct 1996 A
5575761 Hajianpour Nov 1996 A
5576727 Rosenberg et al. Nov 1996 A
5583407 Yamaguchi Dec 1996 A
5591924 Hilton Jan 1997 A
5592401 Kramer Jan 1997 A
5600777 Wang et al. Feb 1997 A
5604345 Matsuura Feb 1997 A
5611731 Bouton et al. Mar 1997 A
5623582 Rosenberg Apr 1997 A
5623642 Katz et al. Apr 1997 A
5627531 Posso et al. May 1997 A
5628686 Svancarek et al. May 1997 A
5635897 Kuo Jun 1997 A
5638060 Kataoka et al. Jun 1997 A
5638421 Serrano et al. Jun 1997 A
5652603 Abrams Jul 1997 A
5666138 Culver Sep 1997 A
5680141 Didomenico et al. Oct 1997 A
5690582 Ulrich et al. Nov 1997 A
5691747 Amano Nov 1997 A
5691898 Rosenberg et al. Nov 1997 A
5694153 Aoyagi et al. Dec 1997 A
5719561 Gonzales Feb 1998 A
5722071 Berg et al. Feb 1998 A
5724106 Autry et al. Mar 1998 A
5724264 Rosenberg et al. Mar 1998 A
5734108 Walker et al. Mar 1998 A
5736978 Hasser et al. Apr 1998 A
5740083 Anderson et al. Apr 1998 A
5745057 Sasaki et al. Apr 1998 A
5749577 Couch et al. May 1998 A
5755620 Yamamoto et al. May 1998 A
5763874 Luciano et al. Jun 1998 A
5766016 Sinclair Jun 1998 A
5767836 Scheffer et al. Jun 1998 A
5771037 Jackson Jun 1998 A
5785630 Bobick et al. Jul 1998 A
5795228 Trumbull et al. Aug 1998 A
5808568 Wu Sep 1998 A
5808603 Chen Sep 1998 A
5818426 Tierney et al. Oct 1998 A
5825305 Biferno Oct 1998 A
5828295 Mittel et al. Oct 1998 A
5831593 Rutledge Nov 1998 A
5841133 Omi Nov 1998 A
5841423 Carroll, Jr. et al. Nov 1998 A
5841428 Jaeger et al. Nov 1998 A
5844673 Ivers Dec 1998 A
5877748 Redlich Mar 1999 A
5879327 Moreau DeFarges et al. Mar 1999 A
5887995 Holehan Mar 1999 A
5889506 Lopresti et al. Mar 1999 A
5912661 Siddiqui Jun 1999 A
5917486 Rylander Jun 1999 A
5917906 Thronton Jun 1999 A
5919159 Lilley et al. Jul 1999 A
5929607 Rosenburg et al. Jul 1999 A
5936613 Jaeger et al. Aug 1999 A
5954689 Poulsen Sep 1999 A
5963196 Nishiumi et al. Oct 1999 A
5977867 Blouin Nov 1999 A
5986638 Cheng Nov 1999 A
6008800 Pryor Dec 1999 A
6017273 Pelkey Jan 2000 A
6031222 Carapelli Feb 2000 A
6067077 Martin et al. May 2000 A
6067081 Hahlganss et al. May 2000 A
6078311 Pelkey Jun 2000 A
6078876 Rosenberg et al. Jun 2000 A
6081536 Gorsuch et al. Jun 2000 A
6084587 Tarr et al. Jul 2000 A
6097499 Casey et al. Aug 2000 A
6097964 Nuovo et al. Aug 2000 A
6104379 Petrich et al. Aug 2000 A
6111577 Zilles et al. Aug 2000 A
6118435 Fujita et al. Sep 2000 A
6140987 Stein et al. Oct 2000 A
6151332 Gorsuch et al. Nov 2000 A
6154201 Levin et al. Nov 2000 A
6160489 Perry et al. Dec 2000 A
6169540 Rosenberg et al. Jan 2001 B1
6183364 Trovato Feb 2001 B1
6192432 Slivka et al. Feb 2001 B1
6198206 Saarmaa et al. Mar 2001 B1
6215778 Lomp et al. Apr 2001 B1
6218966 Goodwin et al. Apr 2001 B1
6219034 Elbing et al. Apr 2001 B1
6236647 Amalfitano May 2001 B1
6241574 Helbing Jun 2001 B1
6243080 Molne Jun 2001 B1
6256011 Culver Jul 2001 B1
6259433 Meyers Jul 2001 B1
6262717 Donohue et al. Jul 2001 B1
6280327 Leifer et al. Aug 2001 B1
6285351 Chang et al. Sep 2001 B1
6293798 Boyle et al. Sep 2001 B1
6295608 Parkes et al. Sep 2001 B1
6300038 Shimazu et al. Oct 2001 B1
6307465 Kayma et al. Oct 2001 B1
6324928 Hughes Dec 2001 B1
6326901 Gonzales Dec 2001 B1
6337678 Fish Jan 2002 B1
6349301 Mitchell et al. Feb 2002 B1
6373463 Beeks Apr 2002 B1
6388655 Leung May 2002 B1
6388999 Gorsuch et al. May 2002 B1
6414674 Kamper et al. Jul 2002 B1
6418329 Furuya Jul 2002 B1
6422941 Thorner et al. Jul 2002 B1
6429846 Rosenberg et al. Aug 2002 B2
6445284 Cruz-Hernandez et al. Sep 2002 B1
6469695 White Oct 2002 B1
6487421 Hess et al. Nov 2002 B2
6509892 Cooper et al. Jan 2003 B1
6518958 Miyajima et al. Feb 2003 B1
6535201 Cooper et al. Mar 2003 B1
6546390 Pollack et al. Apr 2003 B1
6563487 Martin et al. May 2003 B2
6590568 Astala et al. Jul 2003 B1
6613000 Reinkensmeyer et al. Sep 2003 B1
6628195 Coudon Sep 2003 B1
6633224 Hishida et al. Oct 2003 B1
6636202 Ishmael, Jr. et al. Oct 2003 B2
6637281 Yamamoto Oct 2003 B2
6639582 Shrader Oct 2003 B1
6647145 Gay Nov 2003 B1
6760751 Hachiya et al. Jul 2004 B1
6819312 Fish Nov 2004 B2
6904823 Levin et al. Jun 2005 B2
6954899 Anderson Oct 2005 B1
7038667 Vassallo et al. May 2006 B1
7081883 Chen Jul 2006 B2
7104152 Levin et al. Sep 2006 B2
7650810 Levin et al. Jan 2010 B2
20010018354 Pigni Aug 2001 A1
20010035854 Rosenberg et al. Nov 2001 A1
20010045978 McConnell et al. Nov 2001 A1
20020020236 Onodera Feb 2002 A1
20020021282 Masudaya Feb 2002 A1
20020033795 Shahoian et al. Mar 2002 A1
20020072674 Criton et al. Jun 2002 A1
20020142701 Rosenberg Oct 2002 A1
20020149561 Fukumoto et al. Oct 2002 A1
20020149570 Knowles et al. Oct 2002 A1
20020156807 Dieberger Oct 2002 A1
20030006892 Church Jan 2003 A1
20030016211 Woolley Jan 2003 A1
20030022701 Gupta Jan 2003 A1
20030025679 Taylor et al. Feb 2003 A1
20030030628 Sato et al. Feb 2003 A1
20030038776 Rosenberg et al. Feb 2003 A1
20030043206 Duarte Mar 2003 A1
20030048260 Matusis Mar 2003 A1
20030058265 Robinson et al. Mar 2003 A1
20030067449 Yoshikawa et al. Apr 2003 A1
20030071795 Baldauf et al. Apr 2003 A1
20030090460 Schena et al. May 2003 A1
20030095105 Vaananen May 2003 A1
20030112269 Lentz et al. Jun 2003 A1
20030128191 Strasser et al. Jul 2003 A1
20030128192 van Os Jul 2003 A1
20030151597 Roberts et al. Aug 2003 A1
20030172757 Yone Sep 2003 A1
20030174121 Poupyrev et al. Sep 2003 A1
20030179190 Franzen Sep 2003 A1
20030188594 Levin et al. Oct 2003 A1
20040076444 Badovinac et al. Apr 2004 A1
20040193393 Keane Sep 2004 A1
20050109145 Levin et al. May 2005 A1
20050187747 Paxson et al. Aug 2005 A1
Foreign Referenced Citations (41)
Number Date Country
41 40 780 Sep 1992 DE
196 36 779 Mar 1998 DE
0 085 518 Aug 1989 EP
0349086 Jan 1990 EP
0 265 086 Jul 1990 EP
0 470 257 Feb 1992 EP
0 358 989 Jul 1994 EP
0556999 May 1998 EP
1182851 Feb 2002 EP
0 875 819 Oct 2002 EP
569136 May 1945 GB
1 291 271 Oct 1972 GB
1 372 412 Oct 1974 GB
2 237 160 Apr 1991 GB
2 347 199 Aug 2000 GB
01-003664 Jul 1990 JP
02-109714 Jan 1992 JP
04-007371 Aug 1993 JP
5-231509 Sep 1993 JP
05-193862 Jan 1995 JP
11-078576 Mar 1999 JP
2001-350592 Dec 2001 JP
2002-02944 Feb 2002 JP
2002-09756 Feb 2002 JP
2002-259059 Sep 2002 JP
2003-260948 Sep 2003 JP
2004-511739 Apr 2004 JP
WO 9616397 May 1996 WO
WO 9624398 Aug 1996 WO
WO 9632679 Oct 1996 WO
WO 9718546 May 1997 WO
WO 0026891 May 2000 WO
WO 0077689 Dec 2000 WO
WO 0100630 Jan 2001 WO
WO 0167297 Sep 2001 WO
WO 0212991 Feb 2002 WO
WO 0227645 Apr 2002 WO
WO 0231807 Apr 2002 WO
WO 0233290 Apr 2002 WO
WO 03000319 Jan 2003 WO
WO 03085481 Oct 2003 WO
Non-Patent Literature Citations (117)
Entry
PCT International Search Report dated Jan. 12, 2004.
Lindemann, R.W. and Templemann, J.N., Vibrotactile Feedback for Handling Virtual Contact in Immersive Virtual Environments, in Usability Evaluation and Interlace Design: Cognitive Engineering, Intelligent Agents and Virtual Reality, Smith, M.J., Salvendy, G., Harris, D., and Koubek, R.J. (Eds.) 2001, pp. 21-25.
Lindemann, R.W., Templemann, J.N., Sibert, J.L. and Cutler, J.R., Handling of Virtual Contact in Immersive Virtual Environments: Beyond Visuals, Virtual Reality (2002) 6:130-139.
Partial PCT International Search Report dated Sep. 2, 2004.
Lake, “Cyberman From Logitech,” GameBytes, 1994.
Noll, .“Man-Machine Tactile,” SID Journal, Jul./Aug. 1972 Issue.
Rosenberg, Virtual Fixtures: Perceptual Overlays Enhance Operator Performance in Telepresence Tasks, Ph.D. Dissertation, Stanford University, Jun. 1994.
Yamakita et al., “Tele-Virtual Reality of Dynamic Mechanical Model,” Proceedings of the 1992 IEEE/RSJ International Conference on Intelligent Robots and Systems, Raleigh, NC, Jul. 7-10, 1992.
Patent Office of Peoples Republic of China, First Office Action, Chinese Patent Application No. 03807822, dated Feb. 2, 2007.
Shobayashi International Patent and Trademark Office, Office Action, Japanese Patent Application No. 2003-582601, mailed Dec. 10, 2008.
Japanese Patent Office, Decision of Dismissal of Amendment, Japanese Application No. 2003-582601, dated Jan. 15, 2008.
Shobayashi International Patent and Trademark Office, Notice of Reasons of Rejection, Japanese Patent Application No. 2003-582601, dated Apr. 21, 2009.
Patent Office of Peoples Republic of China, Second Office Action, Chinese Patent Application No. 03807822, dated Apr. 28, 2009.
Shobayashi International Patent and Trademark Office, Notice of Reasons of Rejection, Japanese Patent Application No. 2003-582601, dated Feb. 27, 2007.
Shobayashi International Patent and Trademark Office, Notice of Reasons of Rejection, Japanese Patent Application No. 2003-582601, dated May 30, 2006.
KIPOS Notice of Preliminary Rejection, Korean Patent Application No. 10-2004-7015606, dated Jun. 4, 2010.
KIPOS Notice of Preliminary Rejection, Korean Patent Application No. 10-2004-7015606, dated Jul. 10, 2009.
Adelstein, “A Virtual Environment System for the Study of Human Arm Tremor,” Ph.D. Dissertation, Dept. of Mechanical Engineering, MIT, Jun. 1989.
Adelstein, “Design and Implementation of a Force Reflecting Manipulandum for Manual Control research,” DSC-vol. 42, Advances in Robotics, Edited by H. Kazerooni, pp. 1-12, 1992.
Aukstakalnis et al., “Silicon Mirage: The Art and Science of Virtual Reality,” ISBN 0-938151-82-7, pp. 129-180, 1992.
Baigrie, “Electric Control Loading—A Low Cost, High Performance Alternative,” Proceedings, pp. 247-254, Nov. 6-8, 1990.
Bejczy et al., “Kinesthetic Coupling Between Operator and Remote Manipulator,” International Computer Technology Conference, the American Society of Mechanical Engineers, San Francisco, CA, Aug. 12-15, 1980.
Bejczy, “Sensors, Controls, and Man-Machine Interface for Advanced Teleoperation,” Science, vol. 208, No. 4450, pp. 1327-1335, 1980.
Bejczy, “Generalization of Bilateral Force-Reflecting Control of Manipulators,” Proceedings of Fourth CISM-IFToMM, Sep. 8-12, 1981.
Bejczy, et al., “Universal Computer Control System (UCCS) for Space Telerobots,” CH2413-3/87/0000/0318501.00 1987 IEEE, 1987.
Bejczy et al., “A Laboratory Breadboard System for Dual-Arm Teleoperation,” SOAR '89 Workshop, JSC, Houston, TX, Jul. 25-27, 1989.
Bliss, “Optical-to-Tactile Image Conversion for the Blind,” IEEE Transactions on Man-Machine Systems, vol. MMS-11, No. 1, Mar. 1970.
Brooks et al., “Hand Controllers for Teleoperation—A State-of-the-Art Technology Survey and Evaluation,” JPL Publication 85-11; NASA-CR-175890; N85-28559, pp. 1-84, Mar. 1, 1985.
Burdea et al., “Distributed Virtual Force Feedback, Lecture Notes for Workshop on Force Display in Virtual Environments and its Application to Robotic Teleoperation,” 1993 IEEE International Conference on Robotics and Automation, pp. 25-44, May 2, 1993.
Cadler, “Design of a Force-Feedback Touch-Introducing Actuator for Teleoperator Robot Control,” Bachelor of Science Thesis, MIT, Jun. 23, 1983.
Caldwell et al., “Enhanced Tactile Feedback (Tele-Taction) Using a Multi-Functional Sensory System,” 1050-4729/93, pp. 955-960, 1993.
“Cyberman Technical Specification,” Logitech Cyberman SWIFT Supplement, Apr. 5, 1994.
Eberhardt et al., “OMAR—A Haptic display for speech perception by deaf and deaf-blind individuals,” IEEE Virtual Reality Annual International Symposium, Seattle, WA, Sep. 18-22, 1993.
Eberhardt et al., “Including Dynamic Haptic Perception by the Hand: System Description and Some Results,” DSC-vol. 55-1, Dynamic Systems and Control: vol. 1, ASME 1994.
Fokumoto, “Active Click: Tactile Feedback for Touch Panels,” ACM CHI2001 Extended Abstracts, pp. 121-122, Apr. 2001.
Force Feedback Touch Panel, Represented by CSC Division, Sales Department., SIXIK Corporation, Tokyo, Japan, www.smk.co.jp.
Gobel et al., “Tactile Feedback Applied to Computer Mice,” International Journal of Human-Computer Interaction, vol. 7, No. 1, pp. 1-24, 1995.
Gotow et al., “Controlled Impedance Test Apparatus for Studying Human Interpretation of Kinesthetic Feedback,” WA11-11:00, pp. 332-337, 1989.
Howe, “A Force-Reflecting Teleoperated Hand System for the Study of Tactile Sensing in Precision Manipulation,” Proceedings of the 1992 IEEE International Conference on Robotics and Automation, Nice, France, May 1992.
IBM Technical Disclosure Bulletin, “Mouse Ball-Actuating Device With Force and Tactile Feedback,” vol. 32, No. 9B, Feb. 1990.
Iwata, “Pen-based Haptic Virtual Environment,” 0-7803-1363-1/93 IEEE, pp. 287-292, 1993.
Jacobsen et al., “High Performance, Dextrous Telerobotic Manipulator With Force Reflection,” Intervention/ROV '91 Conference & Exposition, Hollywood, Florida, May 21-23, 1991.
Johnson, “Shape-Memory Alloy Tactile Feedback Actuator,” Armstrong Aerospace Medical Research Laboratory, AAMRL-TR-90-039, Aug. 1990.
Jones et al., “A perceptual analysis of stiffness,” ISSN 0014-4819 Springer International (Springer-Verlag); Experimental Brain Research, vol. 79, No. 1, pp. 150-156, 1990.
Kaczmarek et al., “Tactile Displays,” Virtual Environment Technologies, 1995.
Kontarinis et al., “Display of High-Frequency Tactile Information to Teleoperators,” Telemanipulator Technology and Space Telerobotics, Won S. Kim, Editor, Proc. SPIE vol. 2057, pp. 40-50, Sep. 7-9, 1993.
Kontarinis et al., “Tactile Display of Vibratory Information in Teleoperation and Virtual Environments,” PRESENCE, 4(4):387-402, 1995.
Marcus, “Touch Feedback in Surgery,” Proceedings of Virtual Reality and Medicine the Cutting Edge, Sep. 8-11, 1994.
McAffee, “Teleoperator Subsystem/Telerobot Demonstrator: Force Reflecting Hand Controller Equipment Manual,” JPL D-5172, pp. 1-50, A1-A36, B1-B5, C1-C36, Jan. 1988.
Minsky, “Computational Haptics: The Sandpaper System for Synthesizing Texture for a Force-Feedback Display,” Ph.D. Dissertation, MIT, Jun. 1995.
Ouh-Young, “Force Display in Molecular Docking,” Order No. 9034744, p. 1-369, 1990.
Ouh-Young, “A Low-Cost Force Feedback Joystick and its Use in PC Video Games,” IEEE Transactions on Consumer Electronics, vol. 41, No. 3, Aug. 1995.
Ouhyoung et al., “The Development of a Low-Cost Force Feedback Joystick and its Use in the Virtual Reality Environment,” Proceedings of the Third Pacific Conference on Computer Graphics and Applications, Pacific Graphics '95, Seoul, Korea, Aug. 21-24, 1995.
Patrick et al., “Design and Testing of a Non-reactive, Fingertip, Tactile Display for Interaction with Remote Environments,” Cooperative Intelligent Robotics in Space, Rui J. deFigueiredo et al., Editor, Proc. SPIE vol. 1387, pp. 215-222, 1990.
Patrick, “Design, Construction, and Testing of a Fingertip Tactile Display for Interaction with Virtual and Remote Environments,” Master of Science Thesis, MIT, Nov. 8, 1990.
Pimentel et al., “Virtual Reality: through the new looking glass,” 2nd Edition; McGraw-Hill, ISBN 0-07-050167-X, pp. 41-202, 1994.
Rabinowitz et al., “Multidimensional tactile displays: Identification of vibratory intensity, frequency, and contactor area,” Journal of the Acoustical Society of America, vol. 82, No. 4, Oct. 1987.
Russo, “The Design and Implementation of a Three Degree of Freedom Force Output Joystick,” MIT Libraries Archives Aug. 14, 1990, pp. 1-131, May 1990.
Russo, “Controlling Dissipative Magnetic Particle Brakes in Force Reflective Devices,” DSC-vol. 42, Advances in Robotics, pp. 63-70, ASME 1992.
Safe Flight Instruments Corporation, “Coaxial Control Shaker,” Part No. C-25502, Jul. 1, 1967.
Scannell, “Taking a Joystick Ride,” Computer Currents, Boston Edition, vol. 9, No. 11, Nov. 1994.
Shimoga, “Finger Force and Touch Feedback Issues in Dexterous Telemanipulation,” Proceedings of Fourth Annual Conference on Intelligent Robotic Systems for Space Exploration, Rensselaer Polytechnic Institute, Sep. 30-Oct. 1, 1992.
SMK Corporation, “Multi-Functional Touch Panel, Force-Feedback Type, Developed: A Touch Panel Providing a Clicking Feeling,” http://www.smk.co.jp/whatsnew—e/628csc—e.html, Sep. 30, 2002.
SMK Corporation, “Force Feedback Type Optical Touch Panel Developed,” SMK Corporation Website, Oct. 30, 2002.
Snow et al., “Model-X Force-Reflecting-Hand-Controller,” NT Control No. MPO-17851; JPL Case No. 5348, pp. 1-4, Jun. 15, 1989.
Stanley et al., “Computer Simulation of Interacting Dynamic Mechanical Systems Using Distributed Memory Parallel Processors,” DSC-vol. 42, Advances in Robotics, pp. 55-61, ASME 1992.
Tadros, “Control System Design for a Three Degree of Freedom Virtual Environment Simulator Using Motor/Brake Pair Actuators”, MIT Archive © Massachusetts Institute of Technology, pp. 1-88, Feb. 1990.
Terry et al., “Tactile Feedback in a Computer Mouse,” Proceedings of Fourteenth Annual Northeast Bioengineering Conference, University of New Hampshire, Mar. 10-11, 1988.
Wiker, “Teletouch Display Development: Phase 1 Report,” Technical Report 1230, Naval Ocean Systems Center, San Diego, Apr. 17, 1989.
Adelstein, B., A Virtual Environment System for the Study of Human Arm Tremor, Submitted to the Dept. of Mechanical Engineering in partial fulfillment of the requirements for the degree of Doctor of Philosophy at the Massachusetts Institute of Technology, Jun. 1989, pp. 1-253.
Adelstein, B. et al., Design and Implementation of a Force Reflecting Manipulandum for Manual Control Research, DSC-vol. 42, Advances in Robotics, ASME 1992, pp. 1-12.
Akamatsu et al., Multimodal Mouse: A Mouse-Type Device with Tactile and Force Display, Presence, vol. 3, No. 1 pp. 73-80, 1994.
ATIP9B.059: Virtual Reality (VR) Development at SERI (Korea), Asian Technology Information Program (ATIP) Jul. 20, 1996, pp. 1-9.
Aukstakalnis, S. et al., The Art and Science of Virtual Reality Silicon Mirage, 1992, Peachpit Press, Inc., Berkeley, CA, pp. 129-180.
Baigrie, S. et al., Electric Control Loading-A Low Cost, High Performance Alternative, Proceedings, Nov. 6-8, 1990, pp. 247-254.
Bejczy, A., Sensors, Controls, and Man-Machine Interface for Advanced Teleoperation, Science, vol. 208, No. 4450, 1980, pp. 1327-1335.
Bejczy, A. et al., Kinesthetic Coupling Between Operator and Remote Manipulator, International Computer Technology Conference, the American Society of Mechanical Engineers, San Francisco, CA, Aug. 12-15, 1980, pp. 1-9.
Bejczy, A. et al., A Laboratory Breadboard System for Dual-Arm Teleoperation, SOAR '89 Workshop, JSC, Houston, Jul. 25-27, 1989.
Bejczy, A. et al., Universal Computer Control System (UCCS) for Space Telerobots, Jet Propulsion Laboratory, California Institute of Technology, Pasadena, CA, pp. 317-324, 1987.
Bjork, S. et al., An Alternative to Scroll Bars on Small Screens, Play: Applied Research on Art and Technology, Viktoria Institute, Box 620, SE-405 30 Gothenburg, Sweden, pp. 1-2, 1999.
Bouguila, L. et al., Effect of Coupling Haptics and Stereopsis on Depth Perception in Virtual Environment, Precision and Intelligence Laboratory, Tokyo Institute of Technology, 4259 Nagatsuta cho Midori ku Yokohama shi 226-8503-Japan, 2000.
Brooks. T. et al., Hand Controllers for Teleoperation: A State-of-the-Art Technology Survey and Evaluation, 1985, NASA Jet Propulsion Laboratory, California Institute of Technology, Pasadena, CA.
Burdea, G. et al., Distributed Virtual Force Feedback, IEEE Workshop on “Force Display in Virtual Environments and its Application to Robotic Teleoperation,” May 2, 1993, Atlanta, GA.
Calder, B. et al., Design of a Force-Feedback Touch-Inducing Actuator for Teleoperator Robot Control, Submitted to the Department of Mechanical Engineering and Electrical Engineering in partial Fulfillment of the requirements of the degree of Bachelors of Science in Mechanical Engineering and Bachelor of Science in Electrical Engineering at the Massachusetts Institute of Technology, May 1983.
Caldwell, D. et al., Enhanced Tactile Feedback (Tele-Taction) using a Multi-Functional Sensory System, Dept. of Electronic Eng., University of Salford, Salford, M5 4WT, UK, 1993.
Cyberman Technical Specification, Logitech Cyberman SWIFT Supplement, Revision 1.0, Apr. 5, 1994, pp. 1-29.
Eberhardt, S. et al., OMAR-A Haptic Display for Speech Perception by Deaf and Deaf-Blind Individuals, IEEE Virtual Reality Annual International Symposium, Sep. 18-22, 1993, Seattle Washington.
Eberhardt, S. et al., Inducing Dynamic Haptic Perception by the Hand: System Description and Some Results, Dynamic Systems and Control, 1994, vol. 1, presented at 1994 International Mechanical Engineering Congress and Exposition, Chicago Illinois, Nov. 6-11, 1994.
Fukumoto. M. et al., Active Click Tactile Feedback for Touch Panels, NTT DoCoMo Multimedia Labs, Japan, 2001.
Gobel, M. et al., Tactile Feedback Applied to Computer Mice, International Journal of Human-Computer Interaction, vol. 7, No, 1, pp. 1-24, 1995.
Gotow, J. et al., Controlled Impedance Test Apparatus for Studying Human Interpretation of Kinesthetic Feedback, the Robotics Institute and Deptartmetn of Mechanical Engineering, Carnegie Mellon University, Pittsburgh, PA 15213, pp. 332-337, 1989.
Hansen, W., Enhancing Docuemtns with Embedded Programs: How Ness extends Insets in the Andrew Toolkit, 1990. Information Technology Center, Carnegie Mellon University, Pittsburgh, PA 15213.
Hasser, C. et al., Tactile Feedback with Adaptive Controller for a Force-Reflecting Haptic Display Part 1: Design, 1996, Armstrong Laboratory, Human Systems Center, Air Force Materiel Command, Wright-Patterson AFB OH 45433.
Hasser. C. et al., Tactile Feedback for a Force-Reflecting Haptic Display, Thesis Submitted to the School of Engineering of the University of Daytona, Dayton OH, Dec. 1995.
Hasser, C., Force-Reflecting Anthropomorphic Hend Masters, Crew Systems Directorate Biodynamics and Biocommunications Division, Wright-Patterson AFB OH 45433-7901, Jul. 1995, Interim Report for the Period Jun. 1991-Jul. 1995.
Hinckley, K. et al., Haptic Issues for Virtual Manipulation, a Dissertation presented to the Faculty of the School of Engineering and Applied Science at the University of Virginia, in Partial Fulfillment of the Requirement for the Degree Doctor of Philosophy (Computer Science), Dec. 1996.
Howe, R., A Force-Reflecting Teleoperated Hand System for the Study of Tactile Sensing in Precision Manipulation, Proceedings of the 1992 IEEE Conference in Robotics and Automation, Nice, France-May 1992.
Iwata, H., Pen-Based Haptic Virtual Environment, Institute of Engineering Mechanics, University of Tsukuba, Japan, 1993.
Jacobsen, S. et al., High Performance, Dextrous Telerobotic Manipulator with Force Reflection. Intervention/ROV '91, Conference & Exposition, May 21-23, 1991, Hollywood, FL.
Johnson, A., Shape-Memory Alloy Tactical Feedback Actuator, Phase I-Final Report, Air Force SABIR Contract F33-88-C-0541, Armstrong Aerospace Medical Research Laboratory, Human Systems Division, Air Force Systems Command, Wright-Patterson Air Force Base, OH 45433, 1990.
Jones, L. et al., A Perceptual Analysis of Stiffness, Experimental Brain Research, 1990, vol. 79, pp. 150-156.
Kaczmarek, K. et al., Tactile Displays, Virtual Environment Technologies, pp. 349-414, 1995.
Kelley, A. et al., MagicMouse: Tactile and Kinesthetic Feedback in the Human-Computer Interface using an Electromagnetically Actuated Input/Output Device, Department of Electrical Engineering, University of British Columbia, Canada, Oct. 19, 1993.
Lake, S.L., Cyberman from Logitech, web site at http://www.ibiblio.org/GameBytes/issue21/greviews/cyberman/html, as available via the Internet and printed May 29, 2002.
MacLean, K., Designing with Haptic Feedback, Interval Research Corporation, 1801 Page Mill Road, Palo Alto, CA 94304, 2000.
Mine, M., Isaac, A Virtual Environment Tool for the Interactive Construction of Virtual Worlds, Department of Computer Science, University of North Carolina Chapel Hill, 1995.
Picinbono, G. et al., Extrapolation: A Solution for Force Feedback, Virtual Reality and Prototyping, Jun. 1999, Laval, France.
Wloka, M., Interacting with Virtual Reality, Science and Technology Center for Computer Graphics and Scientific Visualization, Brown University Site. Department of Computer Science, 1995.
eRENA, Pushing Mixed Reality Boundaries, Deliverable 7b.1, Final, Version 1.0, 1999.
Real Time Graphics, the Newsletter of Virtual Environment Technologies and Markets, Aug. 1998, vol. 7, No. 2.
1998 IEEE International Conference on Robotics and Automation, May 16-20, 1998, Lueven, Belgium.
Office Action, U.S. Appl. No. 10/116,237 mailed Jun. 30, 2003.
Office Action, U.S. Appl. No. 10/116,237. mailed Dec. 2, 2003.
Office Action, U.S. Appl. No. 11/445,522, mailed Feb. 2, 2007.
Office Action, U.S. Appl. No. 11/445,522, mailed May 2, 2007.
Office Action, U.S. Appl. No. 11/445,522, mailed Oct. 10, 2007.
Patent Cooperation Treaty, International Search Report, International Application No. PCT/US03/10173, mailed Nov. 17, 2003.
Related Publications (1)
Number Date Country
20040145600 A1 Jul 2004 US
Provisional Applications (1)
Number Date Country
60419024 Oct 2002 US