Methods and systems for providing a virtual touch haptic effect to handheld communication devices

Information

  • Patent Grant
  • 8830161
  • Patent Number
    8,830,161
  • Date Filed
    Monday, December 8, 2003
    20 years ago
  • Date Issued
    Tuesday, September 9, 2014
    10 years ago
Abstract
Embodiments of the invention relate to methods and systems for providing customized “haptic messaging” to users of handheld communication devices in a variety of applications. In one embodiment, a method of providing virtual touch to a handheld communication device includes: receiving an input signal associated with a virtual touch; outputting a request relating to a contact with a user-interface member coupled to a handheld communication device; and providing a control signal associated with the contact to an actuator coupled to the handheld communication device, the control signal being configured to cause the actuator to output a haptic effect associated with the virtual touch.
Description
FIELD OF THE INVENTION

This invention relates generally to haptic-feedback systems. More specifically, embodiments of the present invention relate to using customized haptic effects in a variety of applications to convey information to users of handheld communication devices.


BACKGROUND

As handheld communication devices become part of everyday life, device manufactures and service providers strive to enhance the versatility and performance of such devices.


Handheld communication devices in the art (e.g., mobile phones, pagers, personal digital assistants (PDAs), etc.) typically use auditory and visual cues to alert a user when incoming messages, such as voice calls and emails, are received. Such auditory and visual alerts, however, have the disadvantages of being distracting in some situations (e.g., during driving), or annoying in others (e.g., during a meeting or a concert). Although vibratory alerts are made available in some communication devices such as cellular phones, such vibratory effects cannot be customized or personalized according to applications, thus conveying little information to the user. A need, therefore, exists in the art for a new sensory modality that delivers information to users of handheld communication devices in a personalized fashion.


SUMMARY

Embodiments of the invention relate to methods and systems for providing customized “haptic messaging” to users of handheld communication devices in a variety of applications.


In one embodiment, a method of providing virtual touch to a handheld communication device includes: receiving an input signal associated with a virtual touch; outputting a request relating to a contact with a user-interface member coupled to a handheld communication device; and providing a control signal associated with the contact to an actuator coupled to the handheld communication device, the control signal being configured to cause the actuator to output a haptic effect associated with the virtual touch.


In another embodiment, a method of providing virtual touch to a handheld communication device includes: receiving a virtual touch indicator; performing an initialization responsive to the virtual touch indicator on a handheld communication device; receiving a virtual touch signal associated with the initialization; and outputting a control signal associated with the virtual touch signal to an actuator coupled to the handheld communication device.


Further details and advantages of embodiments of the invention are set forth below.





BRIEF DESCRIPTION OF THE FIGURES

These and other features, aspects, and advantages of the present invention are better understood when the following Detailed Description is read with reference to the accompanying drawings, wherein:



FIG. 1 depicts a block diagram of a haptic handheld communication device according to an embodiment of the present invention;



FIG. 2 shows a flowchart depicting a method of using customized haptic effects to convey information to users of handheld communication devices, according to an embodiment of the invention;



FIG. 3 shows a flowchart depicting a method of using haptic logos to relate information to users of handheld communication devices, according to an embodiment of the invention;



FIG. 4 shows a flowchart depicting a method of haptically encoding communication signals, according to an embodiment of the invention;



FIG. 5 shows a flowchart depicting a method of providing haptic messaging to users of handheld communication devices, according to a further embodiment of the invention;



FIG. 6 shows a flowchart illustrating a method of providing an interactive virtual touch in one embodiment of the present invention;



FIG. 7 depicts a flowchart illustrating a method of carrying out a chat session using handheld communication devices, according to an embodiment of the invention;



FIG. 8 shows a flowchart depicting a method of using haptic effects to relate navigation information, according to an embodiment of the invention; and



FIG. 9 shows a flowchart illustrating a method for providing haptic effects to a remote control in one embodiment of the present invention.





DETAILED DESCRIPTION

Embodiments described in the following description are provided by way of example to illustrate some general principles of the invention, and should not be construed as limiting the scope of the invention in any manner. One skilled in the art would also recognize that various changes and modifications can be made herein, without departing from the principles and scope of the invention.



FIG. 1 depicts a block diagram of a handheld communication device 100 according to an embodiment of the invention. It will be appreciated that various elements are shown in schematic form for illustrative purposes and are not drawn to scale. It will also be appreciated that many alternative ways of practicing the present invention exit. Accordingly, various changes and modifications may be made herein, without departing from the principles and scope of the invention.


Device 100 includes a device body including a housing 110 and a user-interface 112; a processor 120; at least one actuator 130 in communication with processor 120; and a memory 140 in communication with processor 120. Device 100 also includes an antenna 150 and a transceiver 160, in communication with processor 120. Device 100 additionally includes a display module 170 and an audio module 180, in communication with processor 120. Display module 170 may include, for example, a liquid crystal device. Audio means 180 may include, for example, a speaker, a microphone, and the like.


For purpose of illustration in the embodiment of FIG. 1, processor 120, actuator 130, and memory 140 are shown to be enclosed within and coupled to the device body. Such an illustration, however, should not be construed as limiting the scope of the invention in any manner. In alternative embodiments, actuator 130 may, for example, be coupled to the outside of housing 110, or embedded in housing 110 via a suitable mechanism. Further, user-interface 112 may include one or more user-interface members. As used herein, a user-interface member includes, without limitation, a key pad having one or more keys, one or more buttons, a touch screen or touch pad, a scroll wheel, a direction pad, a trackball, a knob, a miniature joystick, or other user-interface means known in the art.


Device 100 further includes an API (Application Program Interface) 190, working in conjunction with an operating system 195. A device driver (not shown) may optionally provide an interface between operating system 195 and processor 120.


Memory 140 of device 100 stores a program code that includes instructions to cause processor 120 to perform various tasks. The following description provides some examples.



FIG. 2 shows a flowchart 200 depicting a method of using customized haptic effects to convey information to users of handheld communication devices, according to an embodiment of the invention. At step 210, an input signal associated with an event is received. At step 220, a source of the event is determined and a control signal is selected based on the determination. At step 230, a control signal is output to an actuator coupled to a handheld communication device (see FIG. 1 for an embodiment of such device). The control signal is configured to cause the actuator to output a haptic effect associated with the event.


Furthermore at step 240, a collection of haptic effects is provided, each haptic effect being associated with a control signal. For example, memory 140 of FIG. 1 can store a program code that includes instructions to generate the control signals (e.g., each characterized by a distinct waveform) for rendering the corresponding haptic effects. Haptic effects (along with associated control signals) may also be downloaded or transmitted from a remote source, such as a service provider, a network resource, a Web server, a remote handheld communication device or computer. Such downloaded or transmitted haptic effects can be further edited or modified. At step 250, a mapping between an event of interest and one of the stored haptic effects is received. By way of example, memory 140 of FIG. 1 may also store a program code that enables a user to map an event of interest to one of the haptic effects as provided, e.g., via user-interface 112 through API 190, where the event may be identified by its source. At step 260, the one-to-one mappings made between various events of interest and the corresponding haptic effects are compiled into a haptic lookup table, which can, for example, be stored in memory 140 of FIG. 1.


In the embodiment of FIG. 2, the term “selecting” includes, without limitation, looking up a predetermined mapping between the event of interest and a corresponding haptic effect based on the source determination, and selecting/generating a control signal that is configured to render the desired haptic effect associated with the event (e.g., upon being applied to an actuator). Selection can be made based upon the aforementioned haptic lookup table, for example.


In one embodiment, the input signal may include a communication signal associated with a call event, such as a voice call, an e-mail, or a message in text or multimedia form, which may be received via antenna 150 and transceiver 160 of FIG. 1, for example. The “source” of a call event may be related to a characteristic that distinctly identifies or characterizes the call event, such as the caller's phone number, the sender's e-mail address, a graphical feature or an icon associated with the incoming message, etc.


In another embodiment, the input signal may be associated with a reminder event, which may be a self-generated message on the handheld communication device serving as a reminder for a pre-scheduled activity (e.g., an appointment or a meeting). The source in this scenario may be associated with the type of a pre-scheduled activity (e.g., a business meeting vs. a restaurant reservation), or the time at which the pre-scheduled activity takes place.


In yet another embodiment, the input signal may include a communication signal associated with a status event, for example, received via antenna 150 and transceiver 160 of FIG. 1. Examples of a status event include, but are not limited to: an advertisement (e.g., sale) event, a one-to-one marketing event, a business-transaction event, a stock-trading event, a weather-forecast event, a sports (or game) event, an entertainment event, and an emergency (e.g., 911) event. In this scenario, the source may be associated with a characteristic that distinctly identifies the sender and/or the nature of a status event, such as the phone number of the handheld user's stock broker, the e-mail address of the user's favorite store, the logo associated with the user's favorite TV or radio station, and so on.


In one embodiment, an event of interest can be accompanied by a distinct haptic effect, or overlapping haptic effects, conveying to the user customized information such as “who is calling,” “what is happening,” and so on. The user can also be allowed to update the haptic lookup table, e.g., to include new events, and/or to modify the mappings between the existing events of interest and the corresponding haptic effects.


Moreover, a specific haptic effect can be assigned to any incoming signal event whose source is unknown, so as to alert the user that the incoming message is from an un-identifiable or sender.


As used herein, the term “handheld communication device” includes, without limitation, a mobile phone such as a cellular phone or a satellite phone, a personal digital assistant (PDA), a cordless telephone, a pager, a two-way radio, a handheld or portable computer, a game console controller, a personal gaming device, an MP3 player, or other personal electronic devices known in the art that are equipped with communication or networking capabilities.


In one embodiment, the aforementioned haptic effects can be used as haptic ringers (e.g., counterparts to auditory ring tones) that are customized or personalized to convey information to the user about various events of interest. By way of example, a haptic ringer associated with a call from a loved one (e.g., the user's spouse) may comprise low-amplitude and high frequency vibrations that impart gentle sensations to the user. In contrast, a haptic ringer associated with an emergency event (such as a 911-call) may comprise jolt-like pulses that impart pounding sensations to the user.


In contrast with conventional auditory ring tones, the aforementioned haptic effects (e.g., haptic ringers) are more desirable in an environment where extraneous auditory signals are prohibited (e.g., during a meeting or a concert), and/or where it is difficult to distinguish auditory signals (e.g., in a loud environment such as an airport). The haptic ringers are also more suitable in distracting situations such as driving, so that the user of a handheld communication device can keep eyes on the road without having to look at the device. Moreover, such haptic ringers convey customized information to the user, so that the user is aware of “who is calling,” “what is happening,” and so on, as the following examples further illustrate.


A handheld communication device such as a mobile phone may be configured to allow a user to include haptic information or a haptic code in an outgoing communication signal, e.g., carrying a voice call, an e-mail, or a message. The encoding of a communication signal with haptic information may be based on an established scheme or protocol, and/or on a per-system basis. The haptic code is configured to cause a haptic effect to be output when the communication signal is delivered to another handheld communication device. In one embodiment, businesses and organizations may each be associated with a distinct haptic logo (e.g., a particular vibration pattern) and include their haptic logos in various messages sent to the handheld communication devices of their customers. Such haptic logos can serve as counterparts to conventional logos known in the art, for example. Various status events mentioned above may also be transmitted in this manner. By way of example, a merchant may include its haptic logo in various advertisement events and business transaction events to be transmitted to the handheld communication devices of its customers. Stock brokers (or brokerage firms), TV or radio stations, and marketing/advertising agencies may likewise include their haptic logos in various stock-trading events, weather-forecast events, sports events, entertainment events, and one-to-one marketing events to be transmitted to the handheld users.



FIG. 3 shows a flowchart 300 depicting a method of using haptic logos to relate information to users of handheld communication devices, according to an embodiment of the invention. A handheld communication device receives an input signal at step 310, the input signal being associated with a status event. The handheld communication device extracts a haptic code from the input signal at step 320, where the haptic code is associated with a haptic logo. At step 330, the handheld communication device provides a haptic effect associated with the haptic logo. Step 330 may include providing a control signal to an actuator coupled to the handheld communication device, where the control signal is based at least in part on the haptic code and configured to cause the actuator to output the haptic effect.


In one embodiment, the extracted haptic code may be directly applied to the actuator for rendering the desired haptic effect. In another embodiment, the haptic code may be configured according to a predetermined scheme or protocol that includes, for example, a table of haptic codes (some of which may be associated with one or more haptic logos) versus control signals for rendering the corresponding haptic effects. In this way, a processor in the handheld communication device can look up the corresponding control signal from the table based on the extracted haptic code, and output the selected control signal to the actuator for rendering the desired haptic effect.


In the embodiments of FIG. 2 or 3, the handheld communication device (or the haptic code) may be programmed such that the haptic effect is output immediately, or at a prescribed time after receiving the input signal, as desired in applications. The haptic effects can also be triggered by, or synchronized with, other occurrences.


A handheld communication device may be further configured such that some of its user-interface members (such as those described above) are each associated with a haptic code, e.g., according to a predetermined scheme or protocol. In one embodiment, some of these haptic codes may be associated with haptic effects that emulate expressions or behaviors, such as “laugh,” “giggle,” “hug,” “high-five,” “heartbeat,” “pet purring,” etc. This allows haptic effects to be transmitted and experienced, e.g., in an interactive conversation or a chat session, by pressing or manipulating such members.


By way of example, suppose that user A (termed “Alice” herein) is engaged in a chat session with user B (termed “Bob” herein) via their respective mobile phones. In one embodiment, when Bob tells Alice a joke, Alice can respond by sending a “laugh” sensation to Bob, e.g., by pressing a key on her mobile phone that is assigned with a haptic code corresponding to a laugh sensation. This causes a signal to be transmitted from Alice's phone to Bob's phone, and a corresponding haptic effect to be output to Bob's phone (and thereby experienced by Bob). In alternative embodiments, Alice can include a haptic code in an outgoing message (which may also contain a video image such as a picture taken by her mobile phone, and/or a graphical feature such as an emoticon emulating a smiley face) to be transmitted to Bob, e.g., by pressing the corresponding user-interface member. The haptic code causes a haptic effect to be output when the message is delivered to a remote device such as Bob's mobile phone. In one embodiment, the haptic effect may be correlated or synchronized with the displaying of a video image contained in the message. In another embodiment, the generation of the haptic effect based on the haptic code may be carried out in a manner similar to that described above with respect to the embodiment of FIG. 3.



FIG. 4 depicts a flowchart 400 illustrating a method of a method of haptically encoding communication signals, according to an embodiment of the invention. At step 410, an input signal associated with an actuation of a user-interface member is received. By way of example, the input signal may be associated with Alice's pressing or manipulating a particular user-interface member. At step 420, a haptic code associated with the actuation is determined. At step 430, the haptic code is included in an output signal, and the output signal is sent to a remote handheld communication device. As described above, the output signal may also include a message, a video image, and/or a graphical feature.


A handheld communication device may also be configured such that a haptic effect, along with a message, is output upon a contact with a user-interface member being made (e.g., by a user or an input device). FIG. 5 depicts a flowchart 500 illustrating a method of haptic message that can be associated with this situation, according to an embodiment of the invention. At step 510 of the flowchart 500, a handheld communication device receives an input signal. At step 520, the handheld communication device outputs a request for a contact with a user-interface member coupled to the handheld communication device. At step 530, the handheld communication device provides a control signal associated with the contact to an actuator coupled to the handheld communication device. The control signal is configured to cause the actuator to output a haptic effect associated with the input signal. Step 520 may include having a visual effect displayed, an auditory effect played, and/or a distinctive haptic ringer output, which requests a contact with the user-interface member being made.


In one embodiment, the input signal in FIG. 5 may include a haptic code, along with a message, a video image, and/or a graphical feature, etc. For example, the haptic code may be configured to cause a “hug” sensation to be output when the video image contained in the input signal is displayed. The input signal may also contain a provision or protocol that specifies that the incoming message along with the corresponding haptic effect is output upon a contact with a particular user-interface member (e.g., the #5 key) being made. Alternatively, the handheld communication device may determine the user-interface member to be contacted, before outputting incoming message along with the corresponding haptic effect.


In another embodiment, the input signal of FIG. 5 may be associated with a “virtual touch,” e.g., to mimic a handshake, a “high-five,” a pat on the back, a pulse or heartbeat sensation, a pet purring sensation, or other touch sensations associated with human (and/or human-animal) interactions. In one scenario, the input signal at step 510 may include a “virtual touch indicator,” based on which the request for a contact with a particular user-interface member is made. The virtual touch indicator may be in the form of a haptic code, a message, or other informative means. The control signal at step 530 may be generated, e.g., based on the virtual touch indicator, a haptic code associated with the user-interface member at play, or other predetermined scheme. The input signal at step 510 may also include a virtual touch indicator along with a virtual touch signal for rendering the desired haptic effect. In this case, the control signal at step 530 may be based on the virtual touch signal.


Referring back to the chat session between Alice and Bob, by way of example at the end of their chat session, Alice may wish to send Bob a “high-five.” She sends to Bob's mobile phone a signal including a virtual touch indicator, which in turn prompts a request that Bob be in contact with a user-interface member coupled to his phone, such as a direction pad (e.g., by putting his fingers on the individual keys of the direction pad), a key pad, a touch screen, a trackball, a joystick, or the like. The control signal for rendering a haptic effect that emulates a “high-five” may be based on the haptic code associated with the user-interface member, transmitted with the input signal from Alice, and/or other predetermined scheme.


Interactive virtual touch can also be engaged between users of handheld communication devices, where the manipulation of a user-interface member on one handheld communication device is transmitted possibly in substantially real-time to another handheld device and experienced by its user, and vice versa. FIG. 6 depicts a flowchart 600 illustrating a method of providing interactive virtual touch in one embodiment of the present invention. In the embodiment shown, a handheld communication device first receives an input signal including a virtual touch indicator at step 610. A distinctive haptic ringer may, for example, accompany the arrival of the virtual touch indicator, identifying the sender and the nature of the input signal. The handheld communication device may then perform any necessary initialization to enable the communication at step 620, which may also include requesting a contact with a particular user-interface member coupled to the handheld communication device at step 625. The handheld communication device subsequently receives a virtual touch signal in the communication associated with the desired haptic effect at step 630. The handheld communication device provides the haptic effect at step 640, e.g., by applying the virtual touch signal to an actuator coupled to the user-interface member.


In one embodiment, the virtual touch signal may be associated with the manipulation of a user-interface member on a remote handheld device and transmitted in substantially real-time. And the user on the receiving end may respond by acting in a similar fashion, so as to emulate an interactive touch. Any schemes for delivering virtual touch to users of handheld communication devices may be used.


Haptic effects can also be used to enhance and complement the information content communicated between handheld communication devices. In one embodiment, a plurality of handheld communication users may be engaged in a chat session via their handheld communication devices. The users may each have a graphical representation or avatar displayed on other handheld communication devices. Such avatars can also be haptically enabled, for example, whereby their expressions and/or behaviors are accompanied and enhanced by corresponding haptic effects. FIG. 7 is a flowchart 700 depicting a method of carrying out a chat session using handheld communication devices, according to an embodiment of the invention. In the embodiment shown, a handheld communication device receives an input signal associated with a chat message at step 710. The handheld communication device displays an avatar associated with the chat message at step 720. The avatar may be shown on display 170 of FIG. 1, in one embodiment. At step 730, the handheld communication device provides a haptic effect associated with the chat message. Step 730 may include outputting a control signal to an actuator coupled to the handheld communication device, where the control signal is configured to cause the actuator to output the haptic effect. In one embodiment, the haptic effect may be correlated with an expression or behavior of the avatar, such as a laugh or giggle, a cry, a pet purring, or the like.


Handheld communication devices are increasingly equipped with navigation capability, for example, in communication with the Global Position System (GPS) or other navigation systems. Haptic effects can also be used to convey navigation information, such as positional and/or directional information, to handheld users. By way of example, FIG. 8 shows a flowchart 800 depicting a method of haptic navigation, according to an embodiment of the present invention. The flowchart 800 discloses receiving an input signal associated with a position of a handheld communication device at step 810; determining the position of a handheld communication device relative to a predetermined location at step 820; and providing a haptic effect associated with the determination at step 830. Step 830 may include outputting a control signal associated with the determination to an actuator coupled to the handheld communication device, the control signal being configured to cause the actuator to output the haptic effect. Further, the input signal at step 810 may be received from GPS, a digital compass, or other navigation systems known in the art.


In one embodiment, the haptic effect may be associated with a distance between the position of the handheld communication device and a predetermined location (termed “destination” herein). For example, the haptic effect may include a vibration having a magnitude and a frequency, where at least one of the magnitude and the frequency decreases as the distance from the destination diminishes. Additionally, the haptic effect may be configured to convey a quantitative measure of the distance. By way of example, the haptic effect may include one or more pulse or jolt sensations, where the number of pulses is proportional to the number of miles between the position of the handheld device and the destination.


Processors described above (including processor 120 of FIG. 1) can include, for example, one or more digital logical processors capable of processing input, execute algorithms, and generate output as necessary to perform various tasks, such as those described above. Such processors/controllers may include a microprocessor, an Application Specific Integrated Circuit (ASIC), and state machines. Such processors include, or may be in communication with, media (including memory 140 of FIG. 1). Such media include, for example, computer readable media, which stores program code that, when executed by a processor, cause the processor to perform the steps described herein. Embodiments of computer-readable media include, but are not limited to, an electronic, optical, magnetic, or other storage or transmission device capable of providing a processor, such as the processor in a web server, with computer-readable instructions. Other examples of suitable media include, but are not limited to, a floppy disk, CD-ROM, magnetic disk, memory chip, ROM, RAM, ASIC, configured processor, all optical media, all magnetic tape or other magnetic media, or any other medium from which a computer processor can read. Also, various other forms of computer-readable media may transmit or carry instructions to a computer, including a router, private or public network, or other transmission device or channel.


Program code and associated application programs related to various applications may also reside on a remote source, such as a network resource, a Web server, a remote handheld communication device or computer, which can be transmitted or downloaded to a handheld communication device on a regular or predetermined basis. Haptic effects (along with associated control signals) can also be downloaded or transmitted from a remote source, as described above.


Actuators described above (including actuator 130 shown in FIG. 1) can include, for example, a pager motor, an eccentric rotating motor, a harmonic eccentric rotating motor, a voice coil, a solenoid, a resistive actuator, a piezoelectric actuator, an electro-active polymer actuator, or other types of active/passive actuators suitable for generating haptic effects. U.S. Pat. Nos. 6,429,846 and 6,424,333 disclose further details relating to some of these actuators, both of which are incorporated in full herein by reference. In some embodiments, one or more actuators may be implemented in a handheld communication device, configured to deliver appropriate haptic effects. It will be appreciated that various control schemes can be devised accordingly, for controlling the actuator(s) in a manner that best achieves the desired haptic effects.


Referring back to FIG. 1. In one embodiment, actuator 130 may be coupled to housing 110, thereby imparting haptic effects thus generated to the device body. Haptic ringers (or alerts) described above may be delivered in this manner, for instance. In another embodiment, actuator 130 may be coupled to user-interface 112 of the device body. For instance, an active and/or resistive actuator can be coupled to user-interface 112 to deliver a virtual touch described above. One or more actuators can also be coupled to user-interface 112, for example, to convey a virtual touch such to a user. In yet another embodiment, a plurality of actuators can be coupled to housing 110 as well as user-interface 112. In addition, one or more actuators may also be coupled to a headset, a wristband, or other accessory means associated with a handheld communication device.


Embodiments of the invention include the following.


In one embodiment, an individual (or “Bob”) can have a mobile phone according to the invention. The mobile phone also has an e-mail capability, for example, including both “receive” and “send”). The mobile phone is configured to provide a plurality of haptic effects, e.g., by including appropriate hardware (such as actuators described above) and program code. Bob can program the mobile phone, for example, via user-interface 112 through API 150 shown in FIG. 1, by inputting various events of interest and associating each with a distinct haptic effect. Thus, when an event of interest is subsequently received, the mobile phone provides the corresponding haptic effect.


In one embodiment, Bob's phone includes programming that provides a first haptic effect when an input signal is received from the mobile phone of Bob's wife (or “Alice”). Bob's phone also includes programming that provides a second haptic effect that is different and distinct from the first haptic effect, when an input signal is received from the mobile phone of Bob's supervisor at work (termed “Carol” herein). Bob's phone is further be configured to provide a third haptic effect that is different from the two mentioned above, e.g., when an e-mail is received from the e-mail address of Bob's stock broker (where the e-mail contains a “smiley-face” emoticon, for instance). The third haptic effect can be a vibration with high magnitude and short duration, e.g., to emulate a “high-five.”


In another embodiment, Bob can be watching a movie in a theater with his mobile phone in his pocket. It is set to make no noise, because Bob is in a theater. While Bob is watching the movie, Bob's mobile phone vibrates with the second haptic effect mentioned above. Bob chooses to ignore the call, because he does not wish to speak with his supervisor at a movie theater. Later, Bob's mobile phone vibrates with the first haptic effect. Bob wants to speak with Alice, for example, to make plans to meet later. So Bob answers the phone and quickly exits the theater to talk with Alice.


Bob's mobile phone can also include a personal schedule/calendar application. After speaking with Alice, Bob can enter an entry in the calendar at the 7:00 PM time mark—“Meet Alice”. Bob can also choose a fourth haptic effect to associate with the calendar entry. The mobile phone can be programmed to output the fourth haptic effect fifteen minutes before the time entry of the calendar (i.e., at 6:45 PM).


Bob's mobile phone can be equipped with GPS capability, along with an associated application program for location determination. Bob can also store addresses of various locations of interest in the application program. In one embodiment, Bob can be on the road. Bob's mobile phone vibrates with a distinct fifth haptic effect. Bob recognizes the fifth haptic effect being associated with the haptic logo of his favorite electronics store. He then checks with the application program, and receives a sixth haptic effect associated with the distance between his current position and the store location. Bob then decides to make a stop at the store.


A haptically-enabled handheld communication device of the invention may be further used as a two-way haptic remote control, for example, for controlling a remote system such as a Television set or a multimedia system. In one embodiment, the events as referred to above may be related to program channels shown on the remote system, each identified by a channel number (which may be used as the “source”), for instance. The corresponding haptic effects may be customized on a per-channel basis. Such haptic effects can serve to inform a user as to which channel is on, as a user is channel-surfing by way of this haptic remote control, so that the user need not to look up the display screen.



FIG. 9 depicts a flowchart illustrating a method for providing haptic effects to a remote control in one embodiment of the present invention. In the embodiment shown, the remote control sends a command signal to a remote system at step 910. As with a conventional remote control, the signal may or may not reach the television. The remote control then determines whether a feedback signal has been received at step 920. If the remote control receives a feedback signal, the remote control provides a first haptic effect at step 930. If not, the remote control provides a second haptic effect at step 940.


The first haptic effect can be further customized according to the received feedback signal. In one embodiment, the remote system provides information (e.g., via the feedback signal) to the remote control regarding the state of the display, e.g., based on a predetermined scheme. The remote control may use the information to determine a corresponding haptic effect to provide at step 930. In alternative embodiments, the remote system may determine the appropriate haptic effect to provide and include a corresponding haptic code in the feedback signal. The remote control provides the haptic effect at step 930 based on this haptic code.


The foregoing description of the preferred embodiments of the invention has been presented only for the purpose of illustration and description and is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Numerous modifications and adaptations thereof will be apparent to those skilled in the art without departing from the spirit and scope of the present invention.

Claims
  • 1. A method, comprising: receiving, by a first communication device at different times, a first input signal associated with a first virtual touch and a second input signal associated with a second virtual touch, the first communication device including a first user-interface member, a second user-interface member, and an actuator, wherein the first user-interface member and the second user-interface member are hardware, whereby the first and second virtual touches originate from a second communication device operated by a user to communicate the first and second virtual touches to the first communication device;outputting, at the first communication device, a request to initiate a contact with the first user-interface member when the first virtual touch is received and a second request to initiate a contact with the second user-interface member when the second virtual touch is received;receiving the contact; andproviding a control signal to the actuator in response to the contact, the control signal configured to cause the actuator to output a first haptic effect associated with the first virtual touch when the first virtual touch is received and a second haptic effect associated with the second virtual touch when the second virtual touch is received.
  • 2. The method of claim 1 further comprising extracting a haptic code from the first input signal, the control signal being based at least in part on the haptic code.
  • 3. The method of claim 1 wherein the first user-interface member includes one of a key, a button, a key pad, a direction pad, a touch screen, a scroll wheel, a mini-joystick, a trackball, and a knob.
  • 4. The method of claim 1 wherein the first virtual touch is associated with one of a handshake, a high-five, a pat on the back, a pulse sensation, a heartbeat sensation, and a pet purring sensation.
  • 5. A non-transitory computer-readable storage medium containing executable instructions which cause a data processing system to perform a method, the method comprising: receiving, by a first communication device at different times, a first input signal associated with a first virtual and a second input signal associated with a second virtual touch, the first communication device including a first user-interface member, a second user-interface member, and an actuator, wherein the first user-interface member and the second user-interface member are hardware, whereby the first and second virtual touches originate from a second communication device operated by a user to communicate the first and second virtual touches to the first communication device;outputting, at the first communication device, a request to initiate a contact with the first user-interface member when the first virtual touch is received and a second request to initiate a contact with the second user-interface member when the second virtual touch is received;receiving the contact; andproviding a control signal in response to the contact, the control signal configured to cause the actuator to output a first haptic effect associated with the first virtual touch when the first virtual touch is received and a second haptic effect associated with the second virtual touch when the second virtual touch is received.
  • 6. The non-transitory computer-readable storage medium of claim 5 further comprising extracting a haptic code from the first input signal, the control signal being based at least in part on the haptic code.
  • 7. The non-transitory computer-readable storage medium of claim 5 wherein the first virtual touch is associated with one of a handshake, a high-five, a pat on the back, a pulse sensation, a heartbeat sensation, and a pet purring sensation.
  • 8. An apparatus, comprising: a first user-interface member coupled to a body and a second user-interface member coupled to the body, wherein the first user-interface member and the second user-interface member are hardware;a processor;an actuator coupled to the body and in communication with the processor; anda memory in communication with the processor, the memory storing instructions configuring the processor to: receive, at different times, a first input signal associated with a first virtual touch and a second input signal associated with a second virtual touch at the apparatus, whereby the first and second virtual touches originate from a second apparatus operated by a user to communicate the first and second virtual touches to the apparatus;output a request to initiate a contact with the first user-interface member when the first virtual touch is received and a second request to initiate a contact with the second user-interface member when the second virtual touch is received;receive an indication that the contact was made; andprovide a control signal to the actuator in response to the contact, the control signal configured to cause the actuator to output a first haptic effect when the first virtual touch is received and a second haptic effect associated with the second virtual touch when the second virtual touch is received.
  • 9. The apparatus of claim 8 wherein the body is included in a handheld communication device.
  • 10. The apparatus of claim 9 wherein the handheld communication device includes one of a cellular phone, a satellite phone, a cordless phone, a personal digital assistant, a pager, a two-way radio, a portable computer, a game console controller, a personal gaming device, and an MP3 player.
  • 11. The apparatus of claim 9 wherein the first user-interface member includes at least one of a key, a button, a key pad, a direction pad, a touch screen, a scroll wheel, a mini-joystick, a trackball, and a knob.
  • 12. The apparatus of claim 8 wherein the first virtual touch is associated with one of a handshake, a high-five, a pat on the back, a pulse sensation, a heartbeat sensation, and a pet purring sensation.
CROSS-REFERENCE TO RELATED APPLICATIONS

This application claims priority based upon to U.S. Provisional Patent Application Ser. No. 60/431,662, filed on Dec. 8, 2002, the entire disclosure of which is incorporated herein by reference. This application also claims priority under the Patent Cooperation Treaty based upon co-pending PCT application no. PCT/US03/38862 filed Dec. 8, 2003.

PCT Information
Filing Document Filing Date Country Kind 371c Date
PCT/US03/38862 12/8/2003 WO 00 6/19/2006
Publishing Document Publishing Date Country Kind
WO2004/053829 6/24/2004 WO A
US Referenced Citations (464)
Number Name Date Kind
2972140 Hirsch Feb 1961 A
3157853 Hirsch Nov 1964 A
3220121 Cutler Nov 1965 A
3497668 Hirsch Feb 1970 A
3517446 Corlyon et al. Jun 1970 A
3623064 Kagan Nov 1971 A
3875488 Crocker et al. Apr 1975 A
3902687 Hightower Sep 1975 A
3903614 Diamond et al. Sep 1975 A
3911416 Feder Oct 1975 A
3919691 Noll Nov 1975 A
3923166 Fletcher et al. Dec 1975 A
4023290 Josephson May 1977 A
4101884 Benton, Jr. Jul 1978 A
4108146 Golden Aug 1978 A
4108164 Hall, Sr. Aug 1978 A
4127752 Lowthorp Nov 1978 A
4160508 Salisbury, Jr. et al. Jul 1979 A
4236325 Hall et al. Dec 1980 A
4242823 Bruno Jan 1981 A
4262549 Schwellenbach Apr 1981 A
4333070 Barnes Jun 1982 A
4334280 McDonald Jun 1982 A
4398889 Lam et al. Aug 1983 A
4414537 Grimes Nov 1983 A
4414984 Zarudiansky Nov 1983 A
4436188 Jones Mar 1984 A
4464117 Foerst Aug 1984 A
4477043 Repperger Oct 1984 A
4484179 Kasday Nov 1984 A
4484191 Vavra Nov 1984 A
4513235 Acklam et al. Apr 1985 A
4542375 Alles et al. Sep 1985 A
4545023 Mizzi Oct 1985 A
4550221 Mabusth Oct 1985 A
4557275 Dempsey, Jr. Dec 1985 A
4560983 Williams Dec 1985 A
4581491 Boothroyd Apr 1986 A
4584625 Kellogg Apr 1986 A
4599070 Hladky et al. Jul 1986 A
4603284 Perzley Jul 1986 A
4604016 Joyce Aug 1986 A
4689449 Rosen Aug 1987 A
4692756 Clark Sep 1987 A
4706294 Ouchida Nov 1987 A
4708656 De Vries et al. Nov 1987 A
4713007 Alban Dec 1987 A
4715235 Fukui et al. Dec 1987 A
4757453 Nasiff Jul 1988 A
4758165 Tieman et al. Jul 1988 A
4763356 Day, Jr. et al. Aug 1988 A
4771344 Fallacaro et al. Sep 1988 A
4772205 Chlumsky et al. Sep 1988 A
4782327 Kley et al. Nov 1988 A
4791416 Adler Dec 1988 A
4794384 Jackson Dec 1988 A
4794392 Selinko Dec 1988 A
4795296 Jau Jan 1989 A
4800721 Cemenska et al. Jan 1989 A
4821030 Batson et al. Apr 1989 A
4823634 Culver Apr 1989 A
4837734 Ichikawa et al. Jun 1989 A
4839838 Labiche et al. Jun 1989 A
4851820 Fernandez Jul 1989 A
4853874 Iwamoto et al. Aug 1989 A
4861269 Meenen, Jr. Aug 1989 A
4868549 Affinito et al. Sep 1989 A
4871992 Petersen Oct 1989 A
4885565 Embach Dec 1989 A
4891764 McIntosh Jan 1990 A
4896554 Culver Jan 1990 A
4906843 Jones et al. Mar 1990 A
4914624 Dunthorn Apr 1990 A
4926879 Sevrain et al. May 1990 A
4930770 Baker Jun 1990 A
4934694 McIntosh Jun 1990 A
4935728 Kley Jun 1990 A
4949119 Moncrief et al. Aug 1990 A
4961038 MacMinn Oct 1990 A
4977298 Fujiyama Dec 1990 A
4983901 Lehmer Jan 1991 A
5004391 Burdea Apr 1991 A
5007300 Siva Apr 1991 A
5019761 Kraft May 1991 A
5022384 Freels Jun 1991 A
5022407 Horch et al. Jun 1991 A
5035242 Franklin Jul 1991 A
5038089 Szakaly Aug 1991 A
5044956 Behensky et al. Sep 1991 A
5065145 Purcell Nov 1991 A
5076517 Ferranti et al. Dec 1991 A
5078152 Bond Jan 1992 A
5095303 Clark et al. Mar 1992 A
5103404 McIntosh Apr 1992 A
5107080 Rosen Apr 1992 A
5107262 Cadoz et al. Apr 1992 A
5116180 Fung et al. May 1992 A
5121091 Fujiyama Jun 1992 A
5133076 Hawkins et al. Jul 1992 A
5139261 Openiano Aug 1992 A
5143505 Burdea et al. Sep 1992 A
5146566 Hollis, Jr. et al. Sep 1992 A
5159159 Asher Oct 1992 A
5165897 Johnson Nov 1992 A
5172092 Nguyen et al. Dec 1992 A
5175459 Danial et al. Dec 1992 A
5184319 Kramer Feb 1993 A
5184868 Nishiyama Feb 1993 A
5185561 Good et al. Feb 1993 A
5186629 Rohen Feb 1993 A
5186695 Mangseth et al. Feb 1993 A
5189355 Larkins et al. Feb 1993 A
5193963 McAffee et al. Mar 1993 A
5197003 Moncrief et al. Mar 1993 A
5203563 Loper, III Apr 1993 A
5212473 Louis May 1993 A
5220260 Schuler Jun 1993 A
5223658 Suzuki Jun 1993 A
5223776 Radke et al. Jun 1993 A
5235868 Culver Aug 1993 A
5237327 Saitoh et al. Aug 1993 A
5240417 Smithson et al. Aug 1993 A
5262777 Low et al. Nov 1993 A
5264768 Gregory et al. Nov 1993 A
5270710 Gaultier et al. Dec 1993 A
5271290 Fischer Dec 1993 A
5275174 Cook Jan 1994 A
5275565 Moncrief Jan 1994 A
5283970 Aigner Feb 1994 A
5286203 Fuller et al. Feb 1994 A
5296871 Paley Mar 1994 A
5299810 Pierce et al. Apr 1994 A
5302132 Corder Apr 1994 A
5309140 Everett May 1994 A
5313230 Venolia et al. May 1994 A
5316017 Edwards et al. May 1994 A
5334027 Wherlock Aug 1994 A
5341459 Backes Aug 1994 A
5354162 Burdea et al. Oct 1994 A
5355148 Anderson Oct 1994 A
5376948 Roberts et al. Dec 1994 A
5381080 Schnell et al. Jan 1995 A
5389849 Asano et al. Feb 1995 A
5389865 Jacobus et al. Feb 1995 A
5396266 Brimhall Mar 1995 A
5398044 Hill Mar 1995 A
5399091 Mitsumoto Mar 1995 A
5405152 Katanics et al. Apr 1995 A
5414337 Schuler May 1995 A
5432531 Calder et al. Jul 1995 A
5436622 Gutman et al. Jul 1995 A
5437607 Taylor Aug 1995 A
5450613 Takahara et al. Sep 1995 A
5451924 Massimino et al. Sep 1995 A
5457479 Cheng Oct 1995 A
5459382 Jacobus et al. Oct 1995 A
5461711 Wang et al. Oct 1995 A
5466213 Hogan et al. Nov 1995 A
5471571 Smith et al. Nov 1995 A
5473235 Lance et al. Dec 1995 A
5473344 Bacon et al. Dec 1995 A
5491477 Clark et al. Feb 1996 A
5506605 Paley Apr 1996 A
5512919 Araki Apr 1996 A
5513100 Parker et al. Apr 1996 A
5521336 Buchanan et al. May 1996 A
5530455 Gillick et al. Jun 1996 A
5542672 Meredith Aug 1996 A
5547382 Yamasaki Aug 1996 A
5557365 Ohsawa Sep 1996 A
5562707 Prochazka et al. Oct 1996 A
5563632 Roberts Oct 1996 A
5565887 McCambridge et al. Oct 1996 A
5575761 Hajianpour Nov 1996 A
5576727 Rosenberg et al. Nov 1996 A
5577981 Jarvik Nov 1996 A
5580251 Gilkes et al. Dec 1996 A
5583407 Yamaguchi Dec 1996 A
5587937 Massie et al. Dec 1996 A
5589828 Armstrong Dec 1996 A
5589854 Tsai Dec 1996 A
5591082 Jensen et al. Jan 1997 A
5596347 Robertson et al. Jan 1997 A
5600777 Wang et al. Feb 1997 A
5608788 Demlow et al. Mar 1997 A
5619180 Massimino et al. Apr 1997 A
5625576 Massie et al. Apr 1997 A
5629594 Jacobus et al. May 1997 A
5638060 Kataoka et al. Jun 1997 A
5642469 Hannaford et al. Jun 1997 A
5643087 Marcus et al. Jul 1997 A
5656901 Kurita Aug 1997 A
5666138 Culver Sep 1997 A
5666473 Wallace Sep 1997 A
5670755 Kwon Sep 1997 A
5690582 Ulrich et al. Nov 1997 A
5691747 Amano Nov 1997 A
5691898 Rosenberg et al. Nov 1997 A
5694013 Stewart et al. Dec 1997 A
5699059 Hiller Dec 1997 A
5709219 Chen et al. Jan 1998 A
5714978 Yamanaka Feb 1998 A
5719561 Gonzales Feb 1998 A
5721566 Rosenberg et al. Feb 1998 A
5724106 Autry et al. Mar 1998 A
5724278 Ohgose et al. Mar 1998 A
5729249 Yasutake Mar 1998 A
5731804 Rosenberg Mar 1998 A
5732347 Bartle et al. Mar 1998 A
5734373 Rosenberg et al. Mar 1998 A
5736978 Hasser et al. Apr 1998 A
5739811 Rosenberg et al. Apr 1998 A
5742278 Chen et al. Apr 1998 A
5745715 Pickover et al. Apr 1998 A
5748185 Stephan et al. May 1998 A
5754023 Roston et al. May 1998 A
5755577 Gillio May 1998 A
5757358 Osga May 1998 A
5760764 Martinelli Jun 1998 A
5766016 Sinclair Jun 1998 A
5767457 Gerpheide et al. Jun 1998 A
5767839 Rosenberg Jun 1998 A
5769640 Jacobus et al. Jun 1998 A
5771037 Jackson Jun 1998 A
5781172 Engel et al. Jul 1998 A
5784052 Keyson Jul 1998 A
5785630 Bobick et al. Jul 1998 A
5790108 Salcudean Aug 1998 A
5791992 Crump et al. Aug 1998 A
5802353 Avila et al. Sep 1998 A
5804780 Bartha Sep 1998 A
5805140 Rosenberg et al. Sep 1998 A
5805165 Thorne, III et al. Sep 1998 A
5805416 Friend et al. Sep 1998 A
5805601 Takeda et al. Sep 1998 A
5808601 Leah et al. Sep 1998 A
5808603 Chen Sep 1998 A
5821921 Osborn et al. Oct 1998 A
5823876 Unbehand Oct 1998 A
5825308 Rosenberg Oct 1998 A
5828197 Martin Oct 1998 A
5828364 Siddiqui Oct 1998 A
5831408 Jacobus et al. Nov 1998 A
5832386 Nojima et al. Nov 1998 A
5835080 Beeteson et al. Nov 1998 A
5844392 Peurach et al. Dec 1998 A
5844506 Binstead Dec 1998 A
5877748 Redlich Mar 1999 A
5880714 Rosenberg et al. Mar 1999 A
5884029 Brush, II et al. Mar 1999 A
5887995 Holehan Mar 1999 A
5889236 Gillespie et al. Mar 1999 A
5889670 Schuler et al. Mar 1999 A
5896125 Niedzwiecki Apr 1999 A
5897437 Nishiumi et al. Apr 1999 A
5903257 Nishiumi et al. May 1999 A
5907615 Kaschke May 1999 A
5912661 Siddiqui Jun 1999 A
5914705 Johnson et al. Jun 1999 A
5914708 Legrange et al. Jun 1999 A
5917906 Thornton Jun 1999 A
5929846 Rosenberg et al. Jul 1999 A
5942733 Allen et al. Aug 1999 A
5943044 Martinelli et al. Aug 1999 A
5944151 Jakobs et al. Aug 1999 A
5953413 Peyer et al. Sep 1999 A
5956016 Kuenzner et al. Sep 1999 A
5956484 Rosenberg et al. Sep 1999 A
5959613 Rosenberg et al. Sep 1999 A
5973670 Barber et al. Oct 1999 A
5973689 Gallery Oct 1999 A
5977867 Blouin Nov 1999 A
5982304 Selker et al. Nov 1999 A
5982352 Pryor Nov 1999 A
5984785 Takeda et al. Nov 1999 A
5986643 Harvill et al. Nov 1999 A
5988902 Holehan Nov 1999 A
5990869 Kubica et al. Nov 1999 A
5999168 Rosenberg et al. Dec 1999 A
6001014 Ogata et al. Dec 1999 A
6004134 Marcus et al. Dec 1999 A
6005551 Osborne et al. Dec 1999 A
6008800 Pryor Dec 1999 A
6018711 French-St. George et al. Jan 2000 A
6020876 Rosenberg et al. Feb 2000 A
6024576 Bevirt et al. Feb 2000 A
6028531 Wanderlich Feb 2000 A
6028593 Rosenberg et al. Feb 2000 A
6037927 Rosenberg Mar 2000 A
6061004 Rosenberg May 2000 A
6067081 Hahlganss et al. May 2000 A
6071194 Sanderson et al. Jun 2000 A
6072475 Van Ketwich Jun 2000 A
6081536 Gorssuch et al. Jun 2000 A
6084587 Tarr et al. Jul 2000 A
6088017 Tremblay et al. Jul 2000 A
6088019 Rosenberg Jul 2000 A
6094565 Alberth et al. Jul 2000 A
6100874 Schena et al. Aug 2000 A
6102803 Takeda et al. Aug 2000 A
6110130 Kramer Aug 2000 A
6111577 Zilles et al. Aug 2000 A
6118435 Fujita et al. Sep 2000 A
6125264 Watanabe et al. Sep 2000 A
6125385 Wies et al. Sep 2000 A
6128006 Rosenberg Oct 2000 A
6131097 Peurach et al. Oct 2000 A
6140987 Stein et al. Oct 2000 A
6147422 Delson et al. Nov 2000 A
6147674 Rosenberg et al. Nov 2000 A
6151332 Gorsuch et al. Nov 2000 A
6154201 Levin et al. Nov 2000 A
6160489 Perry et al. Dec 2000 A
6161126 Wies et al. Dec 2000 A
6166723 Schena et al. Dec 2000 A
6169540 Rosenberg et al. Jan 2001 B1
6171191 Ogata et al. Jan 2001 B1
6178157 Berg et al. Jan 2001 B1
6184868 Shahoian et al. Feb 2001 B1
6198206 Saarmaa et al. Mar 2001 B1
6215778 Lomp et al. Apr 2001 B1
6218966 Goodwin et al. Apr 2001 B1
6219032 Rosenberg et al. Apr 2001 B1
6219034 Elbing et al. Apr 2001 B1
6236647 Amalfitano May 2001 B1
6239790 Martinelli et al. May 2001 B1
6243080 Molne Jun 2001 B1
6262717 Donohue et al. Jul 2001 B1
6262718 Findlay et al. Jul 2001 B1
6292174 Mallett et al. Sep 2001 B1
6300936 Braun et al. Oct 2001 B1
6307465 Kayama et al. Oct 2001 B1
6323846 Westerman et al. Nov 2001 B1
6326901 Gonzales Dec 2001 B1
6332075 Verdonk Dec 2001 B1
6337678 Fish Jan 2002 B1
6359550 Brisebois et al. Mar 2002 B1
6373463 Beeks Apr 2002 B1
6388655 Leung May 2002 B1
6388999 Gorsuch et al. May 2002 B1
6411198 Hirai et al. Jun 2002 B1
6411276 Braun et al. Jun 2002 B1
6414674 Kamper et al. Jul 2002 B1
6415138 Sirola et al. Jul 2002 B2
6418323 Bright et al. Jul 2002 B1
6422941 Thorner et al. Jul 2002 B1
6424356 Chang et al. Jul 2002 B2
6429846 Rosenberg et al. Aug 2002 B2
6435794 Springer Aug 2002 B1
6438390 Awan Aug 2002 B1
6438392 Toba Aug 2002 B1
6441599 Kropidlowski Aug 2002 B1
6445284 Cruz-Hernandez et al. Sep 2002 B1
6447069 Terris et al. Sep 2002 B1
6448977 Braun et al. Sep 2002 B1
6469695 White Oct 2002 B1
6473069 Gerpheide Oct 2002 B1
6487421 Hess et al. Nov 2002 B2
6509892 Cooper et al. Jan 2003 B1
6518958 Miyajima et al. Feb 2003 B1
6529122 Magnussen et al. Mar 2003 B1
6535201 Cooper et al. Mar 2003 B1
6563487 Martin May 2003 B2
6574489 Uriya Jun 2003 B1
6606508 Becker et al. Aug 2003 B2
6610936 Gillespie et al. Aug 2003 B2
6626358 Breimesser et al. Sep 2003 B1
6628195 Coudon Sep 2003 B1
6636197 Goldenbert et al. Oct 2003 B1
6636202 Ishmael, Jr. et al. Oct 2003 B2
6639581 Moore Oct 2003 B1
6639582 Shrader Oct 2003 B1
6647145 Gay Nov 2003 B1
6671618 Hoisko Dec 2003 B2
6680729 Shahoian et al. Jan 2004 B1
6690955 Komiyama Feb 2004 B1
6697043 Shahoian Feb 2004 B1
6710518 Morton et al. Mar 2004 B2
6723937 Englemann et al. Apr 2004 B2
6727916 Ballard Apr 2004 B1
6792294 Kushita Sep 2004 B1
6819922 Janz Nov 2004 B1
6822635 Shahoian et al. Nov 2004 B2
6834373 Dieberger Dec 2004 B2
6850150 Ronkainen Feb 2005 B1
6850781 Goto Feb 2005 B2
6859819 Rosenberg et al. Feb 2005 B1
6876847 Kudoh Apr 2005 B2
6882713 Sakai et al. Apr 2005 B2
6885876 Aaltonen et al. Apr 2005 B2
6940497 Vincent et al. Sep 2005 B2
6944482 Engstrom et al. Sep 2005 B2
6963762 Kaaresoja et al. Nov 2005 B2
6963839 Ostermann et al. Nov 2005 B1
6982696 Shahoian Jan 2006 B1
7009595 Roberts et al. Mar 2006 B2
7061466 Moore Jun 2006 B1
7088816 Donnelly Aug 2006 B2
7096045 Yoshinaga Aug 2006 B2
7103389 Shibata Sep 2006 B2
7113177 Franzen Sep 2006 B2
7127271 Fujisaki Oct 2006 B1
7148875 Rosenberg et al. Dec 2006 B2
7149208 Mattaway et al. Dec 2006 B2
7151528 Taylor et al. Dec 2006 B2
7162274 Hama et al. Jan 2007 B2
7171191 Olson Jan 2007 B2
7184765 Birnie et al. Feb 2007 B1
7215329 Yoshikawa et al. May 2007 B2
7218310 Tierling May 2007 B2
7225000 Katayanagi May 2007 B2
7289796 Kudoh Oct 2007 B2
7292227 Fukumoto et al. Nov 2007 B2
7312790 Sato et al. Dec 2007 B2
7328020 Masuda et al. Feb 2008 B2
7336260 Martin et al. Feb 2008 B2
7336977 Katayanagi Feb 2008 B2
7369115 Cruz-Hernandez et al. May 2008 B2
7373120 Messel et al. May 2008 B2
7424272 Horiguchi Sep 2008 B2
7450110 Shahoian et al. Nov 2008 B2
7463249 Knowles et al. Dec 2008 B2
7548232 Shahoian et al. Jun 2009 B2
7728820 Rosenberg et al. Jun 2010 B2
7835771 Jauk et al. Nov 2010 B2
20010035854 Rosenberg et al. Nov 2001 A1
20010036832 McKay Nov 2001 A1
20010044328 Tsukamoto Nov 2001 A1
20020012323 Petite et al. Jan 2002 A1
20020039914 Hama et al. Apr 2002 A1
20020044155 Becker Apr 2002 A1
20020107936 Amon Aug 2002 A1
20020111737 Hoisko Aug 2002 A1
20020130904 Becker et al. Sep 2002 A1
20020177471 Kaaresoja et al. Nov 2002 A1
20020194246 Moskowitz et al. Dec 2002 A1
20030002682 Smith et al. Jan 2003 A1
20030006892 Church Jan 2003 A1
20030016211 Woolley Jan 2003 A1
20030022701 Gupta Jan 2003 A1
20030038776 Rosenberg et al. Feb 2003 A1
20030045266 Staskal et al. Mar 2003 A1
20030048260 Matusis Mar 2003 A1
20030058265 Robinson et al. Mar 2003 A1
20030067440 Rank Apr 2003 A1
20030071795 Baldauf et al. Apr 2003 A1
20030076298 Rosenberg Apr 2003 A1
20030078071 Uchiyama Apr 2003 A1
20030095105 Vaananen May 2003 A1
20030128191 Strasser et al. Jul 2003 A1
20030128192 Van Os Jul 2003 A1
20030174121 Poupyrev et al. Sep 2003 A1
20030184574 Phillips et al. Oct 2003 A1
20030236729 Epstein et al. Dec 2003 A1
20040059790 Austin-Lane et al. Mar 2004 A1
20040067780 Eiden Apr 2004 A1
20040203656 Andrew et al. Oct 2004 A1
20040204049 Hsu et al. Oct 2004 A1
20050058268 Koch Mar 2005 A1
20050134561 Tierling et al. Jun 2005 A1
20060248183 Barton Nov 2006 A1
20070015507 Petronelli Jan 2007 A1
20070229478 Rosenberg et al. Oct 2007 A1
20080068350 Rosenberg et al. Mar 2008 A1
Foreign Referenced Citations (17)
Number Date Country
11-004282 Jan 1999 JP
11-299305 Feb 1999 JP
2000-174864 Jun 2000 JP
2002-176679 Jun 2002 JP
2004-186831 Jul 2004 JP
2004-289577 Oct 2004 JP
2005-72944 Mar 2005 JP
WO 9520788 Aug 1995 WO
0041788 Jul 2000 WO
0203172 Jan 2002 WO
02059869 Aug 2002 WO
02078810 Oct 2002 WO
2004052193 Jun 2004 WO
2004053644 Jun 2004 WO
2004053671 Jun 2004 WO
2004053829 Jun 2004 WO
2004053830 Jun 2004 WO
Non-Patent Literature Citations (16)
Entry
Erikson, Carl, “Polygonal Simplification: An Overview,” Dept. of Computer Science, TR96-016, 1996, pp. 1-32.
Hardman, Lynda et al., “Do You Have the Time? Composition and Linking in Time-based Hypermedia,” Proceedings of the 10th ACM Conference on Hypertext and Hypermedia, Feb. 1999, pp. 189-196.
MacLean, Karon E. et al., “An Architecture for Haptic Control of Media,” in: The Proceedings of the ASMA Dynamic Syatems and Control Division: 1999 International Mechanical Engineering Congress and Exposition, Eighth Annual Symposium on Haptic Interfaces for Virtual Environment and Teleoperator Systems, Nov. 14-19, 1999, pp. 1-10.
Snibbe, Scott S., “Haptic Techniques for Media Control,” in Proceeding of the 14th Annual ACM Symposium on User Interface Software and Technology, 2001, pp. 1-10.
Chang, A. et al., “Audio-Haptic Feedback in Mobile Phones”, Proceedings of ACM CHI 2005 Conference on Human Factors in Computing Systems, Apr. 2-7, 2005, Portland, Oregon, pp. 1264-1267, 2005.
Search Report, International Searching Authority, PCT/US01/41099, Jan. 2, 2002.
Search Report, International Searching Authority, PCT/US03/38961, Apr. 5, 2004.
Search Report, International Searching Authority, PCT/US03/38862, Apr. 12, 2004.
Search Report, International Searching Authority, PCT/US03/38900, Apr. 14, 2004.
Search Report, International Searching Authority, PCT/US03/38899, Apr. 19, 2004.
Search Report, International Searching Authority, PCT/US03/38868, Sep. 27, 2004.
Search Report, International Searching Authority, PCT/US2006/045644, Apr. 23, 2007.
Search Report, International Searching Authority, PCT/US06/35645, Jun. 23, 2008.
Notice of Reasons for Rejection as issued for Japanese Patent Application No. 2012-216677, dated Oct. 15, 2013.
Non-Final Office Action as issued for U.S. Appl. No. 11/226,893, dated Feb. 28, 2013.
Extended Search Report, including the Search Opinion, as issued for European Patent Application No. 06814576.2, dated Jun. 6, 2012.
Related Publications (1)
Number Date Country
20060284849 A1 Dec 2006 US
Provisional Applications (1)
Number Date Country
60431662 Dec 2002 US