The present disclosure relates generally to systems and methods for haptic remote control gaming.
A remotely controllable device is typically connected with a remote control through a wireless connection so that an operator can stand in one place and remotely control the operation of the remotely controllable device using the remote control. Some remotely controllable devices include a camera and video data from the camera can be transmitted to the remote control through the wireless connection so that the operator can view the video data on a display associated with the remote control.
Embodiments provide systems and methods for haptic remote control gaming systems. For example, one disclosed method comprises receiving, by a portable multifunction device, information from a remotely controllable device, the portable multifunction device being operable as a remote control for the remotely controllable device; determining, by the portable multifunction device, a haptic effect based at least in part on the received information; generating, by the portable multifunction device, a signal configured to cause an actuator to output the haptic effect; and outputting, by the first portable multifunction device, the signal. In another embodiment, a computer readable medium comprises program code for causing a processor to perform such a method.
These illustrative embodiments are mentioned not to limit or define the invention, but rather to provide examples to aid understanding thereof. Illustrative embodiments are discussed in the Detailed Description, which provides further description of the invention. Advantages offered by various embodiments of this invention may be further understood by examining this specification.
The accompanying drawings, which are incorporated into and constitute a part of this specification, illustrate one or more examples of embodiments and, together with the description of example embodiments, serve to explain the principles and implementations of the embodiments.
Example embodiments are described herein in the context of systems and methods for haptic remote control gaming. Those of ordinary skill in the art will realize that the following description is illustrative only and is not intended to be in any way limiting. Other embodiments will readily suggest themselves to such skilled persons having the benefit of this disclosure. Reference will now be made in detail to implementations of example embodiments as illustrated in the accompanying drawings. The same reference indicators will be used throughout the drawings and the following description to refer to the same or like items.
In the interest of clarity, not all of the routine features of the implementations described herein are shown and described. It will, of course, be appreciated that in the development of any such actual implementation, numerous implementation-specific decisions must be made in order to achieve the developer's specific goals, such as compliance with application- and business-related constraints, and that these specific goals will vary from one implementation to another and from one developer to another.
Referring now to
In addition to being able to control the RC car, the smartphone can also receive information from the RC car. For example, in this embodiment, the RC car has multiple collision sensors on it to sense when it collides with an obstacle. It also has a sensor on its roof to sense when the car has rolled over and is no longer drivable. The RC car has a processor that receives signals from the sensors and then is able to use a transmitter also in the RC car to transmit the sensor information back to the smartphone. The smartphone receives the information from the RC car and can output haptic effects to indicate collisions or roll-overs to the user. Thus, the user experiences a more immersive experience when driving the car: the user can feel impacts with obstacles via haptic effects, and can be notified of error conditions, such as when the car has rolled over and needs to be flipped back onto its wheels.
This illustrative example is given to introduce the reader to the general subject matter discussed herein. The disclosure is not limited to this example. The following sections describe various additional non-limiting embodiments and examples of devices, systems, and methods for haptic remote control gaming.
Referring now to
In the embodiment shown in
The remote control 200 can be any device that is capable of receiving user input and communicating with a remotely controllable device. For example, the remote control 200 in
In some embodiments, one or more touch-sensitive surfaces may be included on or disposed within one or more sides of the remote control 200. For example, in one embodiment, a touch-sensitive surface is disposed within or comprises a rear surface of the remote control 200. In another embodiment, a first touch-sensitive surface is disposed within or comprises a rear surface of the remote control 200 and a second touch-sensitive surface is disposed within or comprises a side surface of the remote control 200. In some embodiments, the remote control device may comprise two or more housing components, such as in a clamshell arrangement or in a slideable arrangement. For example, one embodiment comprises a remote control device having a clamshell configuration with a touch-sensitive display disposed in each of the portions of the clamshell. Furthermore, in embodiments where the remote control 200 comprises at least one touch-sensitive surface on one or more sides of the remote control 200 or in embodiments where the remote control 200 is in communication with an external touch-sensitive surface, the display 230 may or may not comprise a touch-sensitive surface. In some embodiments, one or more touch-sensitive surfaces may have a flexible touch-sensitive surface. In other embodiments, one or more touch-sensitive surfaces may be rigid. In various embodiments, the remote control 200 may comprise both flexible and rigid touch-sensitive surfaces.
In various embodiments, the remote control 200 may comprise or be in communication with fewer or additional components than the embodiment shown in
The housing 205 of the remote control 200 shown in
In the embodiment shown in
In the embodiment shown in
A haptic output device, such as haptic output devices 240 or 260, can be any component or collection of components that is capable of outputting one or more haptic effects. For example, a haptic output device can be one of various types including, but not limited to, an eccentric rotational mass (ERM) actuator, a linear resonant actuator (LRA), a piezoelectric actuator, a voice coil actuator, an electro-active polymer (EAP) actuator, a memory shape alloy, a pager, a DC motor, an AC motor, a moving magnet actuator, an E-core actuator, a smartgel, an electrostatic actuator, an electrotactile actuator, a deformable surface, an electrostatic friction (ESF) device, an ultrasonic friction (USF) device, or any other haptic output device or collection of components that perform the functions of a haptic output device or that are capable of outputting a haptic effect. Multiple haptic output devices or different-sized haptic output devices may be used to provide a range of vibrational frequencies, which may be actuated individually or simultaneously. Various embodiments may include a single or multiple haptic output devices and may have the same type or a combination of different types of haptic output devices.
In various embodiments, one or more haptic effects may be produced in any number of ways or in a combination of ways. For example, in one embodiment, one or more vibrations may be used to produce a haptic effect, such as by rotating an eccentric mass or by linearly oscillating a mass. In some such embodiments, the haptic effect may be configured to impart a vibration to the entire remote control or to only one surface or a limited part of the remote control. In another embodiment, friction between two or more components or friction between at least one component and at least one contact may be used to produce a haptic effect, such as by applying a brake to a moving component, such as to provide resistance to movement of a component or to provide a torque. In order to generate vibration effects, many devices utilize some type of actuator or haptic output device. Known haptic output devices used for this purpose include an electromagnetic actuator such as an Eccentric Rotating Mass (“ERM”) in which an eccentric mass is moved by a motor, a Linear Resonant Actuator (“LRA”) in which a mass attached to a spring is driven back and forth, or a “smart material” such as piezoelectric, electro-active polymers or shape memory alloys.
In other embodiments, deformation of one or more components can be used to produce a haptic effect. For example, one or more haptic effects may be output to change the shape of a surface or a coefficient of friction of a surface. In an embodiment, one or more haptic effects are produced by creating electrostatic forces and/or ultrasonic forces that are used to change friction on a surface. In other embodiments, an array of transparent deforming elements may be used to produce a haptic effect, such as one or more areas comprising a smartgel. Haptic output devices also broadly include non-mechanical or non-vibratory devices such as those that use electrostatic friction (ESF), ultrasonic surface friction (USF), or those that induce acoustic radiation pressure with an ultrasonic haptic transducer, or those that use a haptic substrate and a flexible or deformable surface, or those that provide projected haptic output such as a puff of air using an air jet, and so on. U.S. patent application Ser. No. 13/092,484 describes ways that one or more haptic effects can be produced and describes various haptic output devices. The entirety of U.S. patent application Ser. No. 13/092,484, filed Apr. 22, 2011, is hereby incorporated by reference.
In
Referring now to
In the embodiment shown in
In the embodiment shown in
In the embodiment shown in
The remotely controllable device 300 can be any device that is capable of communicating with a remote control such as remote control 200 shown in
In various embodiments, the remotely controllable device 300 may comprise or be in communication with fewer or additional components than the embodiment shown in
The housing 305 of the remotely controllable device 300 shown in
In the embodiment shown in
In the embodiment shown in
In one embodiment, haptic output device 340 and/or haptic output device 360 provides status information regarding the remotely controllable device 300. For example, in one embodiment, the remotely controllable device 300 is a passenger vehicle and a haptic effect is output by haptic output device 340 and/or haptic output device 360 when the battery 315 of the vehicle is below a threshold level. As another example, in one embodiment, the remotely controllable device 300 comprises a stuffed animal, doll, or similar toy. In one such embodiment, a user of remote control 200 could send a hug and/or a kiss to a user of the stuffed animal. For example, a parent may be able to use a mobile phone to send a hug and/or a kiss to a child holding the stuffed animal. The stuffed animal can receive the hug and/or the kiss and output one or more haptic effects indicating that a hug and/or kiss has been received. In another embodiment, haptic output device 340 and/or haptic output device 360 provides cues from one or more remote controls. For example, if one remote control cedes control of the remotely controllable device to another remote control, then haptic output device 340 and/or haptic output device 360 can output a haptic effect configured to indicate to a passenger of the remotely controllable device 300 that a new remote control is now controlling the remotely controllable device 300. As shown in
In
Referring now to
One or more remote controls may control or otherwise operate a remotely controllable device. For example, in one embodiment, remote control 410 may control remotely controllable device 420. In this embodiment, remote control 410 may cede control of the remotely controllable device 420 to remote control 415 in response to receiving a request from remote control 415 and/or remotely controllable device 420. In some embodiments, remote control 410 cedes control to remote control 415 in response to an event. For example, if a particular round in a game being played for a user of remote control 410 ends, then remote control 410 may cede control of the remotely controllable device 420 to remote control 415 such that a user of remote control 415 can play the game. As another example, if remotely controllable device 420 receives too much damage while being controlled by remote control 410, then control may be switched to remote control 415. In other embodiments, if remotely controllable device 420 is involved in a predetermined number of crashes within a predetermined period of time while being controlled by remote control 410, then remote control 410 cedes control of the remotely controllable device 420 to remote control 415. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
In embodiments, only one remote control controls the remotely controllable device 420 at a time, but at various times different remote controls can control the remotely controllable device 420. In other embodiments, two or more remote controls may control a remotely controllable device 420 at the same time. For example, remote control 410 may control the direction (e.g., forward, backward, left, right, etc.) of the remotely controllable device 420 while remote control 415 may control the speed of the remotely controllable device 420. As another example, remote control 410 may control the direction and speed of the remotely controllable device 420 while remote control 420 controls the use of a real or a simulated weapon associated with the remotely controllable device 420. In yet another example, remote control 410 may control the driving of a remotely controllable device 420 while remote control 420 controls taking pictures with a camera of the remotely controllable device 420. Thus, in embodiments, operation of a remotely controllable device 420 may be shared by two, three, four, or more remote controls.
In the embodiment shown in
In one embodiment, remotely controllable device 420 is a helicopter and comprises a camera. In this embodiment, remote control 410 controls the flying of the helicopter and remote control 415 controls the camera. For example, remote control 410 may be able to control the speed of a propeller of the helicopter and the direction that the helicopter is travelling. Remote control 415 can rotate the camera to view different locations from the helicopter and can take pictures with the camera. As a user of remote control 410 and/or a user of remote control 415 interacts with the remote control(s) to control the helicopter, remote control 410 and/or remote control 415 may output haptic effects to provide the user(s) cues regarding various events. For example, if the helicopter crashes, then remote control 410 and/or remote control 415 can output a haptic effect to indicate that a crash has occurred. As another example, if the speed of the propeller is below a predetermined threshold necessary to sustain lift, then remote control 410 and/or remote control 415 may output a haptic effect indicating that the speed of the propeller needs to be increased. In one embodiment, if the helicopter travels above a certain altitude, then remote control 410 and/or remote control 415 outputs a haptic effect indicating that the altitude of the helicopter needs to be lowered. In another embodiment, remote control 410 and/or remote control 415 outputs a haptic effect when the user of remote control 415 takes a picture with the camera.
In some embodiments, the helicopter is flown outside and various haptic effects, such as those discussed herein, are output in response to events within and/or interactions between the helicopter and the outside environment. For example, a haptic effect may be output when the helicopter is actually flown above a certain altitude outside. As another example, a haptic effect may be output if the helicopter physically crashes into the ground. In other embodiments, various haptic effects are output in response to events with and/or interactions between the helicopter and an augmented environment. For example, a haptic effect may be output when the helicopter is flown above a certain altitude within the augmented environment regardless of the actual physical altitude that the helicopter is actually flying. As another example, a haptic effect may be output when the helicopter crashes into a virtual mountain shown on a display of remote control 410 and/or remote control 415, regardless of whether the helicopter physically crashes into an object in reality. As described herein, remote control 410, remote control 415, and/or remotely controllable device 420 can be used in a real environment, in a virtual reality environment, and/or in an augmented reality. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
In one embodiment, remotely controllable device 420 is a vehicle and comprises multiple gun turrets. In this embodiment, remote control 410 controls a first gun turret and remote control 420 controls a second gun turret. As a user of remote control 410 and/or a user of remote control 415 interacts with the remote control(s) to control their respective gun turret, remote control 410, remote control 415, and/or remotely controllable device 420 may output haptic effects to provide the user(s) and/or passenger(s) cues regarding various events. For example, the vehicle may output a haptic effect indicating that a gun has been fired when the first gun turret and/or the second gun turret is fired. As another example, remote control 410 can output a haptic effect when the first gun turret is fired and remote control 415 can output a haptic effect when the second gun turret is fired. In one embodiment, remote control 410 outputs a haptic effect when the second gun turret is fired and remote control 415 outputs a haptic effect when the first gun turret is fired.
In some embodiments, the vehicle is driven on a road or through terrain outside and various haptic effects, such as those discussed herein, are output in response to events within and/or interactions between the vehicle and the outside environment. For example, a haptic effect may be output when the vehicle is driven into an object, such as a rock. As another example, a haptic effect may be output may be output when the first gun turret and/or the second gun turret in the vehicle is actually, physically fired. In other embodiments, various haptic effects are output in response to sensed events with and/or sensed interactions between the vehicle and an augmented environment. For example, a haptic effect may be output when the vehicle is driven into a virtual rock shown on a display of remote control 410 and/or remote control 415, regardless of whether the vehicle actually is driven into a physical rock. As another example, a haptic effect may be output when the first gun turret and/or the second gun turret is virtually fired in an augmented reality, regardless of whether the first gun turret and/or the second gun turret is actually fired in reality. As described herein, remote control 410, remote control 415, and/or remotely controllable device 420 can be used in a real environment, in a virtual reality environment, and/or in an augmented reality. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
In another embodiment, remotely controllable device 420 comprises a stuffed animal, doll, or similar toy. In one such embodiment, a user such as a parent can control a stuffed animal using remote control 415. For example, a user may be able to use a smartphone to control the stuffed animal. In this embodiment, the user of the smartphone can send messages to a user of the stuffed animal which are output as haptic effects. For example, a parent using a smartphone or other suitable remote control 415 can send a hug or a kiss to a user (such as a child) holding or otherwise contacting the remotely controllable device 420. Thus, if the remotely controllable device 420 is a stuffed animal, then the stuffed animal can receive the message from the smartphone and output a haptic effect configured to indicate that a message (such as a kiss or a hug) has been received. If the child is holding the stuffed animal, then the child may feel the haptic effects output by the stuffed animal and know that he or she has been sent a hug and/or a kiss. In some embodiments, the stuffed animal or other toy may also transmit haptic messages back to the remote control device, such as to indicate that the hug or kiss is being received. In other embodiments, the remotely controllable device may be a robotic toy or any other suitable remotely controllable device. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
As another example, remotely controllable device 420 may be a stuffed animal, such as a cat. In one embodiment, remote control 410 comprises an application having a graphical user interface including an image of an animal corresponding to the stuffed animal. For example, if the stuffed animal is a cat, then the graphical user interface on the remote control 410 may display an image of a cat. In one embodiment, as a user interacts with the image of the cat displayed on the remote control 410, one or more haptic effects are output by the remotely controllable device 420. For example, if the remote control 410 comprises a touchscreen display, and a user of the remote control 410 pets the image of the cat displayed on the touchscreen display, then the remotely controllable device 420 (e.g., cat) may output a haptic effect that feels like a purr. In some embodiments, the remote control 410 may output a haptic effect that feels like a purr when a user of the remote control 410 pets the image of the cat displayed on the touchscreen display. In another embodiment, one or more haptic effects may be output by remote control 410 when a user interacts with the remotely controllable device 420 (e.g., cat). For example, as a user pets the cat, remote control 410 may output a haptic effect that feels like a purr. In some embodiments, the remotely controllable device 420 may output a haptic effect that feels like a purr when a user of the remotely controllable device 420 pets the cat. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
Referring now to
A remote control, such as remote control 510 and/or remote control 515, can be any suitable device such as remote control 200 discussed herein with respect to
In the system 500 shown in
In embodiments, computer 540 maintains a virtual and/or augmented reality for one or more of the remote controls and/or remotely controllable devices. For example, the computer 540 may maintain an augmented reality by stitching together or otherwise combining information from various remote controls and/or remotely controllable devices into a single environment. In this embodiment, absolute movements of each device at their own location are translated into relative movements in the augmented reality and the remote controls are provided with information about its own remotely controllable device as well as other devices within the augmented reality. For example, at least a portion of the augmented reality can be displayed on a display of a remote control based at least in part on the information received from the computer 540. In embodiments, users of the remote controls can interact with each other through an augmented environment and/or a virtual environment. For example, users may be able to have car races or mock helicopter battles in various augmented environments. As another example, in one embodiment, projectile movement of a weapon is simulated by computer 540 when a user on a remote control presses a trigger button indicating that a weapon should be fired. In this embodiment, the projectile movement simulated by computer 540 may be displayed on a display of one or more of the remote controls. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
One or more remote controls may control or otherwise operate at least a portion of one or more remotely controllable devices. For example, in one embodiment, remote control 510 may control at least a portion of remotely controllable device 520 and/or remotely controllable device 525. In this embodiment, remote control 510 may cede control of at least a portion of one or more of the remotely controllable device 520, 525 to remote control 515. Thus, in embodiments, one remote control controls at least a portion of remotely controllable device 520 and another remote control controls another portion of remotely controllable device 525. In other embodiments, each remote control may control at least a portion of separate remotely controllable devices. For example, remote control 510 may control at least a portion of remotely controllable device 520 and remote control 515 may control at least a portion of remotely controllable device 525.
One or more remote controls may control or otherwise operate at least a portion of one or more remotely controllable devices. For example, in one embodiment, remote control 510 may control at least a portion of remotely controllable device 520 and/or remotely controllable device 525. In this embodiment, remote control 510 may cede control of at least a portion of one or more of the remotely controllable devices 520, 525 to remote control 515 in response to receiving a request from remote control 515, remotely controllable device 520, and/or remotely controllable device 525. In some embodiments, remote control 510 cedes control to remote control 515 in response to an event. For example, if a particular round in a game being played for a user of remote control 510 ends, then remote control 510 may cede control of remotely controllable device 520 and/or remotely controllable device 525 to remote control 515 such that a user of remote control 515 can play the game. As another example, if remotely controllable device 520 receives too much damage while being controlled by remote control 510, then control may be switched to remote control 515. In other embodiments, if remotely controllable device 520 is involved in a predetermined number of crashes within a predetermined period of time while being controlled by remote control 510, then remote control 510 cedes control of the remotely controllable device 520 to remote control 515. In another embodiment remote control 510 initially controls remotely controllable device 520 and remotely controllable device 525 and control is ceded to remote control 515 if communication between remote control 510 and remotely controllable device 525 is lost. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
In embodiments, two or more remote controls may control one or more remotely controllable devices at the same time. For example, remote control 510 may control the direction (e.g., forward, backward, left, right, etc.) of the remotely controllable device 520 while remote control 515 may control the speed of the remotely controllable device 520. As another example, remote control 510 may control the direction of the remotely controllable device 520 while remote control 515 controls the use of a real or a simulated weapon associated with the remotely controllable device 520. In some embodiments, additional remote controls may be used to control portions of a remotely controllable device. For example, a third remote control (not shown) may be used to control a camera associated with remotely controllable device 520. Thus, in embodiments, operation of one or more remotely controllable devices, such as remotely controllable device 520 and/or remotely controllable device 525, may be shared by two, three, four, or more remote controls.
Various types of information can be sent and/or received between one or more remote controls and one or more remotely controllable devices. For example, sensor information from remotely controllable device 520 and/or remotely controllable device 525 can be sent to remote control 510 and/or remote control 515. As another example, information such as a video and/or audio can be sent to remote control 510 and/or remote control 515. In some embodiments, information can be sent to and/from a remote control to one or more other remote controls and/or one or more remotely controllable devices. For example, information regarding the location of one remotely controllable device may be sent from the remotely controllable device and/or a remote control in communication with the remotely controllable device to other remote controls and/or other remotely controllable devices. In some embodiments, where one or more remote controls comprise or are in communication with a display, information received from one or more remote controls and/or one or more remotely controllable devices may be displayed on the display. For example, a location, speed, direction, temperature, video, other sensor information, etc. may be displayed on the display of a remote control.
In some embodiments, one or more graphical images are overlaid with information received from another device. For example, referring to
In the embodiment shown in
In one embodiment, remotely controllable device 520 and remotely controllable device 525 are each helicopters. In this embodiment, remote control 510 controls remotely controllable device 520 and remote control 515 controls remotely controllable device 525. For example, remote control 510 may be able to control the driving of remotely controllable device 520 and remote control 515 can control the driving of remotely controllable device 525. As a user of remote control 510 and/or a user of remote control 515 interacts with their respective remote control to control the helicopters, remote control 510 and/or remote control 515 may output haptic effects to provide the user(s) with cues regarding various events. For example, if remotely controllable device 520 collides with remotely controllable device 525, then remote control 510 and/or remote control 515 may output a haptic effect to indicate that a crash has occurred. As another example, if remotely controllable device 520 is approaching remotely controllable device 525, then remote control 510 and/or remote control 515 may output a haptic effect to indicate that remotely controllable device 520 is approaching remotely controllable device 525.
In some embodiments, remotely controllable device 520 is physically in the same location as remotely controllable device 525. For example, remotely controllable device 520 and remotely controllable device 525 may both be flown in the same park or in the same yard. In this embodiment, various haptic effects, such as those discussed herein, are output in response to events within and/or interactions between remotely controllable device 520, remotely controllable device 525, and/or the outside environment. For example, a haptic effect may be output when remotely controllable device 520 physically crashes into remotely controllable device 525. As another example, a haptic effect may be output if both remotely controllable device 520 and remotely controllable device 525 physically are flown within a predetermined distance of each other.
In other embodiments, various haptic effects are output in response to events with and/or interactions between remotely controllable device 520, remotely controllable device 525, and/or an augmented environment. For example, a haptic effect may be output when remotely controllable device 520 and/or remotely controllable device 525 is flown above a certain altitude within the augmented environment regardless of the actual physical altitude that the remotely controllable device 520 and/or remotely controllable device 525 is actually flying. As another example, a haptic effect may be output when remotely controllable device 520 and remotely controllable device 525 crash into each other in the augmented environment regardless of whether an actual, physical crash between the devices 520, 525 occurs. In one embodiment, a haptic effect is output when remotely controllable device 520 crashes into a virtual object, such as a virtual rock, shown on a display of remote control 510 and/or remote control 515, regardless of whether remotely controllable device 520 physically crashes into an object in reality. As described herein, remote control 510, remote control 515, remotely controllable device 520, and/or remotely controllable device 525 can be used in a real environment, in a virtual reality environment, and/or in an augmented reality environment. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
In one embodiment, remotely controllable device 520 is a helicopter and comprises a gun and remotely controllable device 525 is a vehicle and comprises a gun. In this embodiment, remote control 510 controls the flying of the helicopter and another remote control (not shown) controls the gun of the helicopter. In addition, in this embodiment, remote control 515 controls the driving of the vehicle and another remote control (not shown) controls the gun of the vehicle. As users of the various remote controls interact with their respective remote control to control the helicopter, vehicle, and/or guns, haptic effects can be output by one or more of the remote controls to provide the user(s) cues regarding various events. For example, if a user of the remote control controlling the gun of the helicopter interacts with the remote control to fire the gun at the car, then a haptic effect may be output by remote control 510, remote control 515, and/or other remote controls to indicate that a shot has been fired. As another example, if the vehicle is hit by a bullet fired by the gun of the helicopter, then remote control 510, remote control 515, and/or other remote controls may output one or more haptic effects to indicate that the vehicle has been damaged. In one embodiment, if the helicopter is approaching the vehicle within a predetermined distance, then remote control 510, remote control 515, and/or other remote controls may output one or more haptic effects to indicate that the helicopter is approaching.
In some embodiments, the helicopter and the vehicle are both physically in the same location. For example, the helicopter and the vehicle may both physically be in the same park or in the same yard. In this embodiment, various haptic effects, such as those discussed herein, are output in response to events within and/or interactions between the helicopter, vehicle, and/or physical environment. For example, a haptic effect may be output if the helicopter crashes into the ground or if the vehicle overturns. As another example, a haptic effect can be output if the helicopter crashes into the vehicle. In one embodiment, a haptic effect is output if the helicopter and the vehicle come within a predetermined distance of each other. As another example, a haptic effect may be output if a bullet fired by a gun of the helicopter physically hits the vehicle.
In other embodiments, various haptic effects are output in response to events with and/or interactions between the helicopter, the vehicle, and/or an augmented environment. For example, the helicopter and the vehicle may be physically operated in the same or in different locations. Information from these devices can be received by remote control 510, remote control 515 and/or other remote controls and combined or otherwise incorporated into a virtual environment to create an augmented environment. For example, movements of the helicopter and/or the vehicle can be received by the remote controls and incorporated into an augmented environment. Thus, if a remote control is operated to fly the helicopter to the right in a physical environment, then the helicopter can flown to the right in an augmented environment by updating a display of a remote control to reflect that the helicopter is moving to the right in the augmented environment.
Various haptic effects, such as those discussed herein, may be output in response to events within and/or interactions between the helicopter, vehicle, physical environment, and/or augmented environment. For example, a haptic effect may be output if the helicopter crashes into the ground in the augmented environment regardless of whether the helicopter physically crashes into the ground in the physical environment. In one embodiment, a haptic effect is output if the helicopter physically crashes into an object in the physical environment regardless of whether the helicopter crashes into an object in the augmented environment. As another example, a haptic effect can be output if the helicopter and the vehicle come within a predetermined distance of each other in the augmented environment regardless of the actual physical distance between the helicopter and the vehicle in reality. In one embodiment, a haptic effect is output if a virtual or real bullet fired by the helicopter hits the vehicle in an augmented reality environment. As described herein, remote control 510, remote control 515, remotely controllable device 520, remotely controllable device 525, other remote control(s), other remotely controllable device(s), other devices, or a combination thereof, can be used in a real environment, in a virtual environment, and/or in an augmented reality environment. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
In embodiments, remotely controllable device 520 and remotely controllable device 525 are in a same location, such as in a same building, in a same park, on the same road, etc. In other embodiments, remotely controllable device 520 and remotely controllable device 525 are in different physical locations. For example, remotely controllable device 520 may be located indoors and remotely controllable device 525 may be located outdoors. As another example, remotely controllable device 520 may be located in one city and remotely controllable device 520 may be located in another city. In some embodiments, remote control 510 and remote control 515 are in a same location, such as in a same building, in a same park, etc. In other embodiments, remote control 510 and remote control 515 are in different physical locations. For example, remote control 510 may be located indoors and remotely control 515 may be located outdoors. As another example, remote control 510 may be located in one city and remote control 515 may be located in another city. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
Referring now to
The method 600 begins in block 610 when a remote control and/or a remotely controllable device receives information from another remote control and/or another remotely controllable device. In one embodiment, referring to
The information received by a remote control 200 and/or a remotely controllable device 300 can include information associated with one or more accelerometers, gyroscopes, digital compasses, sensors usable to determine location, pressure, speed, wind speed, temperature, force and/or size, resistive panels, capacitive panels, infrared LEDs, photodetectors, image sensors, optical cameras, other cameras, microphones, speakers. The information can be received by an application, an applet, a plug-in, or a script being executed by a processor on remote control 200 and/or remotely controllable device 300. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
Referring back to method 600, once the remote control and/or remotely controllable device has received the information 610, the method 600 proceeds to block 620. In block 620, the remote control and/or the remotely controllable device determines one or more haptic effects 620. For example, in an embodiment, remote control 510 receives sensor information indicating a speed and direction of remotely controllable device 520 from remotely controllable device 520. In this embodiment, remote control 510 determines a haptic effect based at least in part on the received speed and/or the direction. For example, remote control 510 may determine based at least in part on the speed and direction that the remotely controllable device 520 is approaching an object. In this embodiment, the remote control 510 determines a haptic effect configured to warn a user of the remote control that the remotely controllable device 520 is approaching an object.
One or more haptic effects may be determined by a remote control 200 and/or a remotely controllable device 300 in any number of ways. In one embodiment, one or more haptic effects are determined by a remote control 200 based at least in part on information received from a remotely controllable device 300. For example, remote control 200 may determine a haptic effect when sensor information received from a remotely controllable device 300 indicates that the remotely controllable device 300 is approaching an object. As another example, remote control 200 may determine a haptic effect when sensor information received from a remotely controllable device 300 indicates that the remotely controllable device 300 has collided with another object. In some embodiments, remote control 200 determines a haptic effect based at least in part on state information associated with a remotely controllable device 300. For example, if the remote control 200 receives information from the remotely controllable device 300 that indicates that an engine associated with the remotely controllable device 300 has stalled, then the remote control 200 may determine a haptic effect configured to indicate to a user of the remote control 200 that the engine of the remotely controllable device 300 has stalled. In embodiments, a haptic effect may be determined by remote control 200 based at least in part on augmented reality. For example, if the remotely controllable device 300 is a remote control car, then the remote control 200 may display a virtual race track and the remote control car on a display associated with the remote control 200. In this embodiment, if the remote control car collides with a virtual car in the augmented reality environment, then the remote control 200 determines a haptic effect configured to indicate that a collision has occurred. As another example, the remote control 200 may determine a haptic effect as the remote control car approaches a virtual car on the race track. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
In some embodiments, one or more haptic effects are determined based at least in part on information from another remote control. For example, a first remote control 200 may control the driving a remotely controllable device 300 and a second remote control 200 may control a camera associated with the remotely controllable device 300. Thus, in embodiments, operational control of a remotely controllable device 300 may be shared between two, three, four, or more remote controls. In this embodiment, one or more haptic effects may be determined by the first remote control based at least in part on the second remote control. For example, the first remote control may receive a request from the second remote control and/or the remotely controllable device 300 to cede control of at least a portion of the operation to the second remote control. In this embodiment, the first remote control may determine a haptic effect indicating that at least partial control of the remotely controllable device 300 is being ceded to the second remote control. For example, in one embodiment, a first controller cedes control to a second controller based on the remotely controllable device being out of range of the first controller but still in range of the second controller. As another example, a first controller may cede control to a second controller in response to a user selecting a button to switch controllers. In one embodiment, the second remote control may determine a haptic effect configured to indicate a collision when a user of the first remote control drives the remotely controllable device 300 into a real or virtual object.
In some embodiments, one or more haptic effects are determined based at least in part on interactions between two or more remote controls and/or two or more remotely controllable devices. For example, referring to
In embodiments, one or more haptic effects may be based at least in part on information received from a remote control 200, such as the various information described above with respect to
Referring back to method 600, once the remote control and/or the remotely controllable device determines one or more haptic effects 620, the method 600 proceeds to block 630. In block 630, a signal is generated. For example, in an embodiment where an event involves a collision between a remotely controllable device and another object—such as the ground, a rock, a building, another remotely controllable device, etc.—then a signal may be generated when the collision between the remotely controllable device and the object occurs. In one embodiment, a collision physically occurs. For example, a remotely controllable device may physically contact a rock on the ground. As another example, a remotely controllable device may physically contact another remotely controllable device. In some embodiments, a collision occurs in augmented reality. For example, a collision may occur when a remotely controllable device contacts an augmented rock being displayed on a display of the remote control. As another example, a collision may occur when a remotely controllable device virtually contacts another remotely controllable device. In one embodiment, a remotely controllable device is struck by a gunshot from another remotely controlled device. The gunshot may be from a real, physical bullet or from a virtual bullet. In this embodiment, haptic effect may be determined and a signal generated to indicate that a remotely controllable device has been hit. Thus, in various embodiments, interactions between multiple remotely controllable devices can be detected by a remotely controllable device. The interactions can be reported to one or more remote controls by a remotely controllable device.
A remotely controllable device may be in communication with one, two, three, or more remote controls. In such embodiments, a remotely controllable device may be controlled by one, two, three, or more remote controls. In some embodiments, one or more haptic effects are determined based at least in part on environmental feedback from one or more of the remotely controllable toys, gameplay events from one or more remote controls, gameplay events from one or more interactions between one or more of the remotely controllable devices, or a combination thereof. Any determined haptic effect(s) may be output to any number of remote controls, remotely controllable devices, other devices, or a combination thereof.
Multiple remotely controllable devices and/or remote controls may be in communication with each other over one or more networks. In such embodiments, one or more haptic effects may be determined based at least in part on multiplayer interactions between the remotely controllable devices and/or the remote controls. For example, if sensor information received from one remotely controllable device indicates that a collision has occurred, then one or more of remote controls may output a haptic effect configured to indicate that a collision has occurred.
In one embodiment, a remote control comprises a display and a remotely controllable device comprises a camera. In such an embodiment, the remote control can display information from one or more of the remotely controllable devices. For example, at least a portion of one or more video feeds from one or more remotely controllable devices may be displayed on a display of a remote control. A video feed from one or more remotely controllable devices can be overlaid with other graphical images and displayed on the display of the remote control. Thus, in embodiments, an augmented reality using information received from one or more remotely controllable devices and/or one or more remote controls is displayed on a display of a remote control. In addition, haptic effects can be determined based at least in part on the augmented reality, environmental feedback, gameplay events, state information of one or more devices, other information, or a combination thereof. Any determined haptic effect(s) may be output to any number of remote controls, remotely controllable devices, other devices, or a combination thereof. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
In an embodiment, a signal is generated the first time an event occurs. For example, if the event comprises a collision between a remotely controllable device and an object, then the first time that a collision between the remotely controllable device and the object occurs, the processor 210 generates a signal. In one embodiment, if a subsequent collision occurs and/or if a subsequent collision occurs within a predetermined period of time between the remotely controllable device and the object, then another signal is not generated. In other embodiments, if a subsequent collision occurs between the remotely controllable device and the object, then the processor 210 generates a signal based on the subsequent collision.
In one embodiment, a signal is generated each time an event occurs. Thus, referring to the example above, each time a collision occurs between the remotely controllable device and the object, the processor 210 generates a signal. Therefore, if the remotely controllable device collides with a rock and then collides again with the rock for a second time, then the processor 210 would generate a signal twice. In another embodiment, a signal is generated only the first time the event occurs.
In embodiments, one or more signals are generated at any number of times based at least in part on information received from a remotely controllable device, a remote control, user input, other devices, or a combination thereof. The information can include environmental information such as a temperature, wind speed, terrain conditions (e.g., water, mud, dry land, mountains, hills, etc.), information regarding the state of a device, information regarding the proximity of a device to another device, information regarding events of one or more users of one or more remote controls, information regarding events initiated by a user's own interaction with a remote control, information regarding events received from another remote control being operated by a second user, other suitable information, or a combination thereof. Information can be information corresponding to real life conditions such as an actual temperature and/or information corresponding to virtual conditions such as a remotely controllable device colliding with a virtual rock. In one embodiment, one or more signals are generated when an event occurs. In some embodiments, one or more signals are generated prior to an event occurring. In other embodiments, one or more signals are generated after an event occurs. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
In some embodiments, the processor 210 generates a single signal when the event occurs. For example, in one embodiment, the processor 210 generates a signal configured to cause a haptic output device, such as haptic output device 240 or haptic output device 260, to output a haptic effect. The haptic effect may indicate that an object is currently displayed on the display 230, that an object is about to be displayed on the display 230, that an object is approaching, that an event has occurred, that an event is about to occur, or a combination thereof.
In other embodiments, the processor 210 generates two, three, or more signals. For example, in one embodiment, the processor 210 generates a first signal configured to cause a first haptic effect and a second signal configured to cause a second haptic effect. In some embodiments, the processor 210 generates a different signal for each event that occurs. In various embodiments, the processor 210 generates one or more signals configured to cause the touch-sensitive display 230, the communication interface 250, the haptic output device 240, the haptic output device 260, the speaker 270, other components of the device 200, other components of devices in communication with the device 200, or a combination thereof to output one or more of the generated signals, such as a video signal, audio signal, haptic output signal, and/or a communication signal. For example, in one embodiment, the processor 210 generates a signal when the event occurs where the signal is configured to cause a haptic output device in another device to cause a haptic effect. In one embodiment, the processor 210 sends the signal to the other device through the communication interface 250.
In one embodiment, a generated signal includes a command for a device or component to perform a specified function, such as to output a haptic effect or transmit a message to a remote device. In another embodiment, a generated signal includes parameters which are used by a device or component receiving the command to determine a response or some aspect of a response. Parameters may include various data related to, for example, magnitudes, frequencies, durations, or other parameters that a haptic output device can use to determine a haptic effect, output a haptic effect, or both. For example, in one embodiment, the processor 210 generates a signal configured to cause haptic output device 240 to output a haptic effect. In such an embodiment, the signal may include a pressure parameter that the haptic output device 240 uses to determine the intensity of the haptic effect to output. For example, according to one embodiment, the larger the pressure parameter the haptic output device 240 receives, the more intense the haptic effect that is output.
In one embodiment, an intensity parameter is used by a haptic output device to determine the intensity of a haptic effect. In this embodiment, the greater the intensity parameter, the more intense the haptic effect that is output. In one embodiment, the intensity parameter is based at least in part on sensor information, such as speed, direction, etc., of a remotely controllable device when an event occurs. Thus, according to one embodiment, a larger intensity parameter is sent to a haptic output device when an event occurs while the remotely controllable device is travelling at a faster speed than when an event occurs while the remotely controllable device is travelling at a slower speed. A signal may include data that is configured to be processed by a haptic output device, display, communication interface, speaker, or other components of a device or in communication with a device in order to determine an aspect of a particular response.
It will be recognized that any type of input synthesis method may be used to generate the interaction parameter for one or more haptic effect signals including, but not limited to, the method of synthesis examples listed in TABLE 1 below. A drive signal may be applied to a haptic actuator according to the interaction parameter. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
Referring again to
In various embodiments, the processor 210 may output one or more generated signals to any number of devices. For example, the processor 210 may output one signal to the communication interface 250. In one embodiment, the processor 210 may output one generated signal to the touch-sensitive display 230, another generated signal to the communication interface 250, and another generated signal to the haptic output device 260. In other embodiments, the processor 210 may output a single generated signal to multiple components or devices. For example, in one embodiment, the processor 210 outputs one generated signal to both haptic output device 240 and haptic output device 260. In another embodiment, the processor 210 outputs one generated signal to haptic output device 240, haptic output device 260, and communication interface 250. In still another embodiment, the processor 210 outputs one generated signal to both haptic output device 240 and haptic output device 260 and outputs a second generated signal to the touch-sensitive display 230.
As discussed above, the processor 210 may output one or more signals to the communication interface 250. For example, the processor 210 may output a signal to the communication interface 250 instructing the communication interface 250 to send data to another component or device in communication with the device 200. In such an embodiment, the communication interface 250 may send data to the other device and the other device may perform a function such as updating a display associated with the other device or the other device may output a haptic effect. Thus, in embodiments, a second device may output a haptic effect based at least in part upon an interaction with a first device in communication with the second device. In other embodiments, a second device may perform any number of functions such as, for example, updating a display associated with the second device or outputting a sound to a speaker associated with the second device based at least in part on an interaction with a first remote control 200.
In various embodiments, after the processor 210 outputs a signal to a component, the component may send the processor 210 a confirmation indicating that the component received the signal. For example, in one embodiment, haptic output device 260 may receive a command from the processor 210 to output a haptic effect. Once haptic output device 260 receives the command, the haptic output device 260 may send a confirmation response to the processor 210 that the command was received by the haptic output device 260. In another embodiment, the processor 210 may receive completion data indicating that a component not only received an instruction but that the component has performed a response. For example, in one embodiment, haptic output device 240 may receive various parameters from the processor 210. Based on these parameters haptic output device 240 may output a haptic effect and send the processor 210 completion data indicating that haptic output device 240 received the parameters and outputted a haptic effect.
It will be recognized that any type of input synthesis method may be used to generate the interaction parameter for one or more haptic effect signals including, but not limited to, the method of synthesis examples listed in TABLE 1 below. A drive signal may be applied to a haptic actuator according to the interaction parameter. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
Table 1—Methods of Synthesis
Additive synthesis—combining inputs, typically of varying amplitudes
Subtractive synthesis—filtering of complex signals or multiple signal inputs
Frequency modulation synthesis—modulating a carrier wave signal with one or more operators
Sampling—using recorded inputs as input sources subject to modification
Composite synthesis—using artificial and sampled inputs to establish a resultant “new” input
Phase distortion—altering the speed of waveforms stored in wavetables during playback
Waveshaping—intentional distortion of a signal to produce a modified result
Resynthesis—modification of digitally sampled inputs before playback
Granular synthesis—combining of several small input segments into a new input
Linear predictive coding similar technique as used for speech synthesis
Direct digital synthesis—computer modification of generated waveforms
Wave sequencing—linear combinations of several small segments to create a new input
Vector synthesis—technique for fading between any number of different input sources
Physical modeling—mathematical equations of the physical characteristics of virtual motion
Referring now to
The method 700 begins in block 710 when information is received from a first remote control. For example, referring to
In embodiments, when an action occurs in remotely controllable device 520, remote control 510 sends information to computer 540 directly and/or through network 530. For example, remote control 510 can send information regarding the action to computer 540. Thus, in one embodiment, as remotely controllable device 520 is driven or otherwise operated by a user using remote control 510, the remote control 510 sends information regarding the movements to computer 540. Remote control 510 can send sensor information received from remotely controllable device 520 to computer 540. For example, remote control 510 may receive sensor information from remotely controllable device 520 indicating that the remotely controllable device has collided with another object and at least a portion of the sensor information may be sent by remote control 510 to computer 540. In embodiments, computer 540 receives the sensor information and/or action information from remote control 510 through network 530. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
Referring back to method 700, once the information has been received from the first remote control 710, the method proceeds to block 720. In block 720, an environment is updated. For example, referring to
Referring back to method 700, once the environment has been updated 720, the method 700 proceeds to block 730. In block 730, information is sent to a second remote control. For example, referring to
As another example, if computer 540 receives information from remote control 510 indicating that a gun or other weapon is disabled, then computer 540 may update the augmented reality environment to reflect that the gun has been disabled and computer 540 may send status information to remote control 515 indicating that the gun has been disabled. In this embodiment, a display of remote control 515 may be updated to reflect that the gun has been disabled. For example, an “X” may be overlaid over the gun that has been disabled and displayed on the display. Numerous other embodiments are disclosed herein and variations are within the scope of this disclosure.
While the methods and systems herein are described in terms of software executing on various machines, the methods and systems may also be implemented as specifically-configured hardware, such as field-programmable gate array (FPGA) specifically to execute the various methods. For example, embodiments can be implemented in digital electronic circuitry, or in computer hardware, firmware, software, or in a combination thereof. In one embodiment, a device may comprise a processor or processors. The processor comprises a computer-readable medium, such as a random access memory (RAM) coupled to the processor. The processor executes computer-executable program instructions stored in memory, such as executing one or more computer programs for editing an image. Such processors may comprise a microprocessor, a digital signal processor (DSP), an application-specific integrated circuit (ASIC), field programmable gate arrays (FPGAs), and state machines. Such processors may further comprise programmable electronic devices such as PLCs, programmable interrupt controllers (PICs), programmable logic devices (PLDs), programmable read-only memories (PROMs), electronically programmable read-only memories (EPROMs or EEPROMs), or other similar devices.
Such processors may comprise, or may be in communication with, media, for example computer-readable media, that may store instructions that, when executed by the processor, can cause the processor to perform the steps described herein as carried out, or assisted, by a processor. Embodiments of computer-readable media may comprise, but are not limited to, an electronic, optical, magnetic, or other storage device capable of providing a processor, such as the processor in a web server, with computer-readable instructions. Other examples of media comprise, but are not limited to, a floppy disk, CD-ROM, magnetic disk, memory chip, ROM, RAM, ASIC, configured processor, all optical media, all magnetic tape or other magnetic media, or any other medium from which a computer processor can read. The processor, and the processing, described may be in one or more structures, and may be dispersed through one or more structures. The processor may comprise code for carrying out one or more of the methods (or parts of methods) described herein.
The foregoing description of some embodiments of the invention has been presented only for the purpose of illustration and description and is not intended to be exhaustive or to limit the invention to the precise forms disclosed. Numerous modifications and adaptations thereof will be apparent to those skilled in the art without departing from the spirit and scope of the invention.
Reference herein to “one embodiment” or “an embodiment” means that a particular feature, structure, operation, or other characteristic described in connection with the embodiment may be included in at least one implementation of the invention. The invention is not restricted to the particular embodiments described as such. The appearance of the phrase “in one embodiment” or “in an embodiment” in various places in the specification does not necessarily refer to the same embodiment. Any particular feature, structure, operation, or other characteristic described in this specification in relation to “one embodiment” may be combined with other features, structures, operations, or other characteristics described in respect of any other embodiment.
This application is a continuation of U.S. application Ser. No. 13/826,391, now U.S. Pat. No. 9,245,428, filed Mar. 14, 2013, entitled “Systems and Methods for Haptic Remote Control Gaming” which claims priority to U.S. Provisional Patent Application No. 61/678,908, filed Aug. 2, 2012, entitled “Method and Apparatus for a Haptic Cat,” and claims priority to U.S. Provisional Patent Application No. 61/679,382, filed Aug. 3, 2012, entitled “Method and Apparatus for a Haptic Cat,” the entirety of each of which is hereby incorporated by reference.
Number | Name | Date | Kind |
---|---|---|---|
3157853 | Hirsch | Nov 1964 | A |
3220121 | Cutler | Nov 1965 | A |
3497668 | Hirsch | Feb 1970 | A |
3517446 | Corlvon et al. | Jun 1970 | A |
3902687 | Hightower | Sep 1975 | A |
3903614 | Diamond et al. | Sep 1975 | A |
3919691 | Noll | Nov 1975 | A |
4018121 | Chowning | Apr 1977 | A |
4054749 | Suzuki | Oct 1977 | A |
4091302 | Yamashita | May 1978 | A |
4160508 | Salisbury, Jr. et al. | Jul 1979 | A |
4236325 | Hall et al. | Dec 1980 | A |
4334221 | Rosenhagen et al. | Jun 1982 | A |
4360345 | Hon | Nov 1982 | A |
4386346 | Levine | May 1983 | A |
4414984 | Zarudiansky | Nov 1983 | A |
4430595 | Nakasone | Feb 1984 | A |
4490810 | Hon | Dec 1984 | A |
4513235 | Acklam et al. | Apr 1985 | A |
4581491 | Boothroyd | Apr 1986 | A |
4599070 | Hladky et al. | Jul 1986 | A |
4691659 | Ito et al. | Sep 1987 | A |
4695266 | Hai | Sep 1987 | A |
4706294 | Ouchida | Nov 1987 | A |
4708656 | DeVries et al. | Nov 1987 | A |
4712101 | Culver | Dec 1987 | A |
4713007 | Alban | Dec 1987 | A |
4731603 | McRae et al. | Mar 1988 | A |
4795296 | Jau | Jan 1989 | A |
4823634 | Culver | Apr 1989 | A |
4824111 | Hoye et al. | Apr 1989 | A |
4868549 | Affmito et al. | Sep 1989 | A |
4885565 | Embach | Dec 1989 | A |
4891764 | Mcintosh | Jan 1990 | A |
4907973 | Hon | Mar 1990 | A |
4930770 | Baker | Jun 1990 | A |
4934694 | Mcintosh | Jun 1990 | A |
4938483 | Yavetz | Jul 1990 | A |
4940234 | Ishida et al. | Jul 1990 | A |
4964837 | Collier | Oct 1990 | A |
4983901 | Lehmer | Jan 1991 | A |
5019761 | Kraft | May 1991 | A |
5022407 | Horch et al. | Jun 1991 | A |
5024626 | Nagel | Jun 1991 | A |
5035242 | Franklin | Jul 1991 | A |
5038089 | Szakaly | Aug 1991 | A |
5047918 | Schwartz et al. | Sep 1991 | A |
5059958 | Jacobs et al. | Oct 1991 | A |
5078152 | Bond | Jan 1992 | A |
5103404 | Mcintosh | Apr 1992 | A |
5107262 | Cadoz et al. | Apr 1992 | A |
5146566 | Hollis, Jr. et al. | Sep 1992 | A |
5164530 | Iwase | Nov 1992 | A |
5184319 | Kramer | Feb 1993 | A |
5185561 | Good | Feb 1993 | A |
5186629 | Rohen | Feb 1993 | A |
5186695 | Mangseth et al. | Feb 1993 | A |
5189355 | Larkins | Feb 1993 | A |
5191641 | Yamamoto et al. | Mar 1993 | A |
5195920 | Collier | Mar 1993 | A |
5203563 | Loper, III | Apr 1993 | A |
5212473 | Louis | May 1993 | A |
5216337 | Orton et al. | Jun 1993 | A |
5220260 | Schuler | Jun 1993 | A |
5240417 | Smithson et al. | Aug 1993 | A |
5271290 | Fischer | Dec 1993 | A |
5273038 | Beavin | Dec 1993 | A |
5275174 | Cook | Jan 1994 | A |
5296846 | Ledle | Mar 1994 | A |
5296871 | Paley | Mar 1994 | A |
5299810 | Pierce | Apr 1994 | A |
5309140 | Everett | May 1994 | A |
5334027 | Wherlock | Aug 1994 | A |
5354162 | Burdea et al. | Oct 1994 | A |
5359527 | Takanabe | Oct 1994 | A |
5388992 | Franklin et al. | Feb 1995 | A |
5389865 | Jacobus et al. | Feb 1995 | A |
5399091 | Mitsumoto | Mar 1995 | A |
5405152 | Katanics et al. | Apr 1995 | A |
5436638 | Bolas et al. | Jul 1995 | A |
5440183 | Denne | Aug 1995 | A |
5442557 | Kaneko | Aug 1995 | A |
5459382 | Jacobus et al. | Oct 1995 | A |
5466213 | Hogan | Nov 1995 | A |
5499360 | Barbara et al. | Mar 1996 | A |
5547382 | Yamasaki | Aug 1996 | A |
5559412 | Schuler | Sep 1996 | A |
5565840 | Thorner et al. | Oct 1996 | A |
5572201 | Graham et al. | Nov 1996 | A |
5580251 | Gilkes et al. | Dec 1996 | A |
5583478 | Renzi | Dec 1996 | A |
5587937 | Massie et al. | Dec 1996 | A |
5589828 | Armstrong | Dec 1996 | A |
5614687 | Yamada et al. | Mar 1997 | A |
5619180 | Massimino et al. | Apr 1997 | A |
5629594 | Jacobus | May 1997 | A |
5631861 | Kramer | May 1997 | A |
5634051 | Thomson | May 1997 | A |
5643087 | Marcus et al. | Jul 1997 | A |
5661446 | Anderson et al. | Aug 1997 | A |
5666138 | Culver | Sep 1997 | A |
5669818 | Thorner et al. | Sep 1997 | A |
5684722 | Thorner et al. | Nov 1997 | A |
5692956 | Rifkin | Dec 1997 | A |
5701140 | Rosenberg | Dec 1997 | A |
5704791 | Gillio | Jan 1998 | A |
5709219 | Chen et al. | Jan 1998 | A |
5714978 | Yamanaka et al. | Feb 1998 | A |
5721566 | Rosenberg et al. | Feb 1998 | A |
5724264 | Rosenberg | Mar 1998 | A |
5728960 | Sitrick | Mar 1998 | A |
5731804 | Rosenberg | Mar 1998 | A |
5734373 | Rosenberg et al. | Mar 1998 | A |
5736978 | Hasser et al. | Apr 1998 | A |
5739811 | Rosenberg et al. | Apr 1998 | A |
5742278 | Chen et al. | Apr 1998 | A |
5747714 | Kniest et al. | May 1998 | A |
5754023 | Roston et al. | May 1998 | A |
5755577 | Gillio | May 1998 | A |
5766016 | Sinclair | Jun 1998 | A |
5767839 | Rosenberg | Jun 1998 | A |
5769640 | Jacobus | Jun 1998 | A |
5781172 | Engel et al. | Jul 1998 | A |
5784052 | Keyson | Jul 1998 | A |
5785630 | Bobick et al. | Jul 1998 | A |
5790108 | Salcudean et al. | Aug 1998 | A |
5791908 | Gillio | Aug 1998 | A |
5800177 | Gillio | Sep 1998 | A |
5805140 | Rosenberg et al. | Sep 1998 | A |
5821920 | Rosenberg | Oct 1998 | A |
5825308 | Rosenberg | Oct 1998 | A |
5828197 | Martin | Oct 1998 | A |
5857939 | Kaufman | Jan 1999 | A |
5857986 | Moriyasu | Jan 1999 | A |
5870740 | Rose et al. | Feb 1999 | A |
5874942 | Walker | Feb 1999 | A |
5882206 | Gillio | Mar 1999 | A |
5889670 | Schuler | Mar 1999 | A |
5889672 | Schuler et al. | Mar 1999 | A |
5894263 | Shimakawa et al. | Apr 1999 | A |
5897437 | Nishiumi et al. | Apr 1999 | A |
5914705 | Johnson et al. | Jun 1999 | A |
5928248 | Acker | Jul 1999 | A |
5945772 | Macnak et al. | Aug 1999 | A |
5952596 | Kondo | Sep 1999 | A |
5959613 | Rosenberg | Sep 1999 | A |
5973670 | Barber et al. | Oct 1999 | A |
5984880 | Lander et al. | Nov 1999 | A |
5986643 | Harvill et al. | Nov 1999 | A |
6001014 | Ogata et al. | Dec 1999 | A |
6004134 | Marcus et al. | Dec 1999 | A |
6024576 | Moore | Feb 2000 | A |
6025790 | Saneyoshi | Feb 2000 | A |
6036495 | Marus et al. | Mar 2000 | A |
6044646 | Silverbrook | Apr 2000 | A |
6078126 | Rollins et al. | Jun 2000 | A |
6088017 | Tremblay et al. | Jul 2000 | A |
6088019 | Rosenberg | Jul 2000 | A |
6096004 | Megland | Aug 2000 | A |
6104158 | Jacobus et al. | Aug 2000 | A |
6106301 | Merril | Aug 2000 | A |
6111577 | Zilles et al. | Aug 2000 | A |
6113459 | Nammoto | Sep 2000 | A |
6119114 | Smadja | Sep 2000 | A |
6121955 | Liu | Sep 2000 | A |
6122520 | Want | Sep 2000 | A |
6154201 | Levin et al. | Nov 2000 | A |
6160489 | Perry | Dec 2000 | A |
6160540 | Fishkin et al. | Dec 2000 | A |
6161126 | Weis et al. | Dec 2000 | A |
6163739 | Park et al. | Dec 2000 | A |
6184868 | Shahoian et al. | Feb 2001 | B1 |
6198206 | Saarmaa et al. | Mar 2001 | B1 |
6199067 | Geller | Mar 2001 | B1 |
6211861 | Rosenberg et al. | Apr 2001 | B1 |
6221861 | Seegmiller | Apr 2001 | B1 |
6241574 | Helbing | Jun 2001 | B1 |
6244742 | Yamada | Jun 2001 | B1 |
6256011 | Culver | Jul 2001 | B1 |
6275213 | Tremblay et al. | Aug 2001 | B1 |
6280327 | Leifer et al. | Aug 2001 | B1 |
RE37374 | Roston et al. | Sep 2001 | E |
6293798 | Boyle et al. | Sep 2001 | B1 |
6298323 | Kaemmerer | Oct 2001 | B1 |
6300938 | Culver | Oct 2001 | B1 |
6304520 | Watanabe | Oct 2001 | B1 |
6314094 | Boys et al. | Nov 2001 | B1 |
6317032 | Oishi | Nov 2001 | B1 |
6323412 | Loo | Nov 2001 | B1 |
6346025 | Tachau et al. | Feb 2002 | B1 |
6366272 | Rosenberg | Apr 2002 | B1 |
6369952 | Rallison et al. | Apr 2002 | B1 |
6376971 | Pelrine | Apr 2002 | B1 |
6401027 | Xu et al. | Jun 2002 | B1 |
6411896 | Shuman et al. | Jun 2002 | B1 |
6422941 | Thorner | Jul 2002 | B1 |
6424333 | Tremblay et al. | Jul 2002 | B1 |
6429846 | Rosenberg | Aug 2002 | B2 |
6470207 | Simon | Oct 2002 | B1 |
6470302 | Cunningham | Oct 2002 | B1 |
6498982 | Bellesfield et al. | Dec 2002 | B2 |
6501420 | Townsend | Dec 2002 | B2 |
6563487 | Martin | May 2003 | B2 |
6564210 | Korda et al. | May 2003 | B1 |
6585595 | Soma et al. | Jul 2003 | B1 |
6598707 | Nakagaki et al. | Jul 2003 | B2 |
6612925 | Forsberg | Sep 2003 | B1 |
6636835 | Ragsdale-Elliott et al. | Oct 2003 | B2 |
6640187 | Chenault | Oct 2003 | B1 |
6641480 | Murzanski et al. | Nov 2003 | B2 |
6646643 | Templeman | Nov 2003 | B2 |
6655817 | Devlin | Dec 2003 | B2 |
6665644 | Kanevsk et al. | Dec 2003 | B1 |
6686531 | Pennock et al. | Feb 2004 | B1 |
6686901 | Rosenberg | Feb 2004 | B2 |
6686911 | Levin | Feb 2004 | B1 |
6694228 | Rios | Feb 2004 | B2 |
6697043 | Shahoian et al. | Feb 2004 | B1 |
6697044 | Shahoian | Feb 2004 | B2 |
6697715 | Freeman | Feb 2004 | B1 |
6707443 | Bruneau et al. | Mar 2004 | B2 |
6717573 | Shahoian et al. | Apr 2004 | B1 |
6721706 | Strubbe et al. | Apr 2004 | B1 |
6732090 | Shanahan et al. | May 2004 | B2 |
6735568 | Buckwalter et al. | May 2004 | B1 |
6749537 | Hickman | Jun 2004 | B1 |
6768066 | Wehrenberg | Jul 2004 | B2 |
6768246 | Pelrine | Jul 2004 | B2 |
6772026 | Bradbury et al. | Aug 2004 | B2 |
6778226 | Eshelman | Aug 2004 | B1 |
6781289 | Heinz et al. | Aug 2004 | B2 |
6801837 | Carlstedt et al. | Oct 2004 | B2 |
6804643 | Kiss | Oct 2004 | B1 |
6808472 | Hickman | Oct 2004 | B1 |
6812394 | Weissflog | Nov 2004 | B2 |
6812624 | Pei et al. | Nov 2004 | B1 |
6816711 | Standke | Nov 2004 | B2 |
6819267 | Edmark | Nov 2004 | B1 |
6822635 | Shahoian | Nov 2004 | B2 |
6829599 | Chidlovskii | Dec 2004 | B2 |
6836982 | Augustine | Jan 2005 | B1 |
6857930 | Lawley, Jr. | Feb 2005 | B1 |
6858970 | Malkin | Feb 2005 | B2 |
6863220 | Selker | Mar 2005 | B2 |
6867733 | Sandhu | Mar 2005 | B2 |
6871142 | Kumada et al. | Mar 2005 | B2 |
6882086 | Kornbluh | Apr 2005 | B2 |
6885362 | Suomela | Apr 2005 | B2 |
6887158 | Goto et al. | May 2005 | B2 |
6915295 | Okamoto et al. | Jul 2005 | B2 |
6921351 | Hickman | Jul 2005 | B1 |
6929481 | Alexander | Aug 2005 | B1 |
6956538 | Moore | Oct 2005 | B2 |
6963762 | Kaaresoja et al. | Nov 2005 | B2 |
6970088 | Kovach | Nov 2005 | B2 |
6978320 | Nonaka | Dec 2005 | B2 |
6982697 | Wilson | Jan 2006 | B2 |
6983139 | Dowling et al. | Jan 2006 | B2 |
6985143 | Pharr | Jan 2006 | B2 |
7023423 | Rosenberg | Apr 2006 | B2 |
7062367 | Kim | Jun 2006 | B2 |
7094153 | Kunieda et al. | Aug 2006 | B2 |
7100835 | Selker | Sep 2006 | B2 |
7127333 | Arvidsson | Oct 2006 | B2 |
7137891 | Neveu et al. | Nov 2006 | B2 |
7176893 | Larkin | Feb 2007 | B1 |
7181438 | Szabo | Feb 2007 | B1 |
7235012 | DiDato | Jun 2007 | B2 |
7249126 | Ginsburg et al. | Jul 2007 | B1 |
7283997 | Howard et al. | Oct 2007 | B1 |
7312766 | Edwards | Dec 2007 | B1 |
7343232 | Duggan et al. | Mar 2008 | B2 |
7402964 | Calhoun | Jul 2008 | B1 |
7471216 | Chen et al. | Dec 2008 | B2 |
7779166 | Grant et al. | Aug 2010 | B2 |
20010003101 | Shinohara et al. | Jun 2001 | A1 |
20010045978 | McConnell et al. | Nov 2001 | A1 |
20020016786 | Pitkow et al. | Feb 2002 | A1 |
20020030663 | Tierling et al. | Mar 2002 | A1 |
20020054060 | Schena | May 2002 | A1 |
20020078045 | Dutta | Jun 2002 | A1 |
20020103025 | Murzanski et al. | Aug 2002 | A1 |
20020123988 | Dean | Sep 2002 | A1 |
20020133418 | Hammond et al. | Sep 2002 | A1 |
20020142701 | Rosenberg | Oct 2002 | A1 |
20020177471 | Kaaresoja et al. | Nov 2002 | A1 |
20030030619 | Martin et al. | Feb 2003 | A1 |
20030033287 | Shanahan et al. | Feb 2003 | A1 |
20030060906 | Kim | Mar 2003 | A1 |
20030067908 | Mattaway | Apr 2003 | A1 |
20030069077 | Korienek | Apr 2003 | A1 |
20030110038 | Sharma | Jun 2003 | A1 |
20030115193 | Okamoto et al. | Jun 2003 | A1 |
20030187837 | Cutliss | Oct 2003 | A1 |
20030201975 | Bailey et al. | Oct 2003 | A1 |
20030210228 | Ebersole et al. | Nov 2003 | A1 |
20040015714 | Abraham et al. | Jan 2004 | A1 |
20040017482 | Weitman | Jan 2004 | A1 |
20040059708 | Dean | Mar 2004 | A1 |
20040068486 | Chidlovskii | Apr 2004 | A1 |
20040097806 | Hunter | May 2004 | A1 |
20040103087 | Mukherjee et al. | May 2004 | A1 |
20040124248 | Selker | Jul 2004 | A1 |
20040125120 | Weiner | Jul 2004 | A1 |
20040145600 | Cruz-Hernandez et al. | Jul 2004 | A1 |
20040164971 | Hayward | Aug 2004 | A1 |
20040166937 | Rothschild | Aug 2004 | A1 |
20040186824 | Delic et al. | Sep 2004 | A1 |
20040224638 | Fadell | Nov 2004 | A1 |
20040225519 | Martin | Nov 2004 | A1 |
20040225635 | Toyama | Nov 2004 | A1 |
20040238732 | State et al. | Dec 2004 | A1 |
20050032528 | Dowling et al. | Feb 2005 | A1 |
20050059489 | Kim | Mar 2005 | A1 |
20050060299 | Filley | Mar 2005 | A1 |
20050071328 | Lawrence et al. | Mar 2005 | A1 |
20050080786 | Fish | Apr 2005 | A1 |
20050096047 | Haberman et al. | May 2005 | A1 |
20050107688 | Strommer | May 2005 | A1 |
20050114149 | Rodriguez | May 2005 | A1 |
20050139660 | Maxymych et al. | Jun 2005 | A1 |
20050149213 | Guzak et al. | Jul 2005 | A1 |
20050149499 | Franz et al. | Jul 2005 | A1 |
20050154636 | Hildinger et al. | Jul 2005 | A1 |
20050174975 | Mgrdechian et al. | Aug 2005 | A1 |
20050177614 | Bourne | Aug 2005 | A1 |
20050198376 | Kotzin | Sep 2005 | A1 |
20050212749 | Marvit et al. | Sep 2005 | A1 |
20050222981 | Lawrence et al. | Oct 2005 | A1 |
20050222989 | Haveliwala et al. | Oct 2005 | A1 |
20050227712 | Estevez | Oct 2005 | A1 |
20050277470 | Watanachote | Dec 2005 | A1 |
20050278317 | Gross et al. | Dec 2005 | A1 |
20050286546 | Basseli et al. | Dec 2005 | A1 |
20060015560 | MacAuley et al. | Jan 2006 | A1 |
20060017692 | Wehrenberg et al. | Jan 2006 | A1 |
20060022955 | Kennedy | Feb 2006 | A1 |
20060026521 | Hotelling et al. | Feb 2006 | A1 |
20060066569 | Eid et al. | Mar 2006 | A1 |
20060095412 | Zito et al. | May 2006 | A1 |
20060161621 | Rosenberg | Jul 2006 | A1 |
20060167576 | Rosenberg | Jul 2006 | A1 |
20060167943 | Rosenberg | Jul 2006 | A1 |
20060173828 | Rosenberg | Aug 2006 | A1 |
20060173837 | Berstis et al. | Aug 2006 | A1 |
20060189386 | Rosenberg | Aug 2006 | A1 |
20060195361 | Rosenberg | Aug 2006 | A1 |
20060223637 | Rosenberg | Oct 2006 | A1 |
20060253210 | Rosenberg | Nov 2006 | A1 |
20070067294 | Ward et al. | Mar 2007 | A1 |
20070125852 | Rosenberg | Jun 2007 | A1 |
20070135264 | Rosenberg | Jun 2007 | A1 |
20070143625 | Jung et al. | Jun 2007 | A1 |
20070156677 | Szabo | Jul 2007 | A1 |
20080300055 | Lutnick et al. | Dec 2008 | A1 |
20110061017 | Ullrich et al. | Mar 2011 | A1 |
Number | Date | Country |
---|---|---|
101711923 | May 2010 | CN |
0 265 011 | Apr 1988 | EP |
0 085 518 | Aug 1989 | EP |
0 349 086 | Jan 1990 | EP |
0 607 580 | Jul 1994 | EP |
0 626 634 | Nov 1994 | EP |
0 695 566 | Feb 1996 | EP |
0 940 162 | Sep 1999 | EP |
0 977 142 | Feb 2000 | EP |
2 237 160 | Apr 1991 | GB |
2 336 890 | Nov 1999 | GB |
H2-185278 | Jul 1990 | JP |
3033148 | Feb 1991 | JP |
H4-8381 | Jan 1992 | JP |
H5-192449 | Aug 1993 | JP |
05-300985 | Nov 1993 | JP |
H 05-300985 | Nov 1993 | JP |
06-039097 | May 1994 | JP |
H7-24147 | Jan 1995 | JP |
07-024708 | Mar 1995 | JP |
07-019512 | May 1995 | JP |
07-053189 | Jun 1995 | JP |
07-048297 | Feb 1996 | JP |
H 08-048297 | Feb 1996 | JP |
08-168545 | Jul 1996 | JP |
09-138767 | May 1997 | JP |
9-215870 | Aug 1997 | JP |
10-200882 | Jul 1998 | JP |
10-314463 | Dec 1998 | JP |
H 11-004966 | Jan 1999 | JP |
H 11-313001 | Nov 1999 | JP |
2000-020222 | Jan 2000 | JP |
2000-102677 | Apr 2000 | JP |
2000-317866 | Nov 2000 | JP |
2000-334163 | Dec 2000 | JP |
2001-502200 | Feb 2001 | JP |
2001-062160 | Mar 2001 | JP |
WO 9200559 | Jan 1992 | WO |
WO 9731333 | Aug 1997 | WO |
WO 9814253 | Apr 1998 | WO |
WO 9832112 | Jul 1998 | WO |
WO 9940504 | Aug 1999 | WO |
WO 0103105 | Jan 2001 | WO |
WO 0113354 | Feb 2001 | WO |
WO 0124158 | Apr 2001 | WO |
WO 0203172 | Jan 2002 | WO |
WO 0227705 | Apr 2002 | WO |
Entry |
---|
Adelstein, “A Virtual Environment System for the Study of Human Arm Tremor,” Ph.D. Dissertation, Dept. of Mechanical Engineering, MIT, Jun. 1989. |
Adelstein, “Design and Implementation of a Force Reflecting Manipulandum for Manual Control research,” DSC—vol. 42, Advances in Robotics, Edited by H. Kazerooni,pp. 1-12, 1992. |
Akamatsu, et al., “Multimodal Mouse: A Mouse-Type Device with Tactile and Force Display,” Presence, vol. 3, No. I, 1994,pp. 73-80. |
Aukstakalnis et al., “Silicon Mirage: The Art and Science of Virtual Reality,” ISBN 0-938151-82-7, pp. 129-180, 1992. |
Baigrie, “Electric Control Loading: A Low Cost. High Performance Alternative,” Proceedings, pp. 247-254, Nov. 6-8, 1990. |
Bejczy et al., “A Laboratory Breadboard System for Dual-Arm Teleoperation,” SOAR '89 Workshop, JSC, Houston, TX, Jul. 25-27, 1989. |
Bejczy et al., “Kinesthetic Coupling Between Operator and Remote Manipulator,” International Computer Technology Conference, The American Society of Mechanical Engineers. San Francisco, CA, Aug. 12-15, 1980. |
Bejczy, “Sensors, Controls, and Man-Machine Interface for Advanced Teleoperation,” Science, vol. 208, No. 4450, pp. 1327-1335, 1980. |
Bejczy, et al., “Universal Computer Control System (UCCS) for Space Telerobots.” CH 2413-3/87/0000/0316501.00 1987 IEEE, 1987. |
Bejczy, “Generalization of Bilateral Force-Reflecting Control of Manipulators,” Proceedings of Fourth CISM • IFToMM, Sep. 8-12, 1981. |
Brooks et al., “Hand Controllers for Teleoperation—A State-of-the-Art Technology Survey and Evaluation,” JPL Publication 85-11; NASA-CR-175890; N85-28559, pp. 1-84, Mar. 1, 1985. |
Burdea et al., “Distributed Virtual Force Feedback, Lecture Notes for Workshop on Force Display in Virtual Environments and its Application to Robotic Teleoperation,” 1993 IEEE International Conference on Robotics and Automation, pp. 25-44, May 2, 1993. |
Caldwell et al., “Enhanced Tactile Feedback (Tele-Taction) Using a Multi-Functional Sensory System,” 1050-4729/93, pp. 955-960, 1993. |
Dennerlein, et al., “Vibrotactile Feedback for Industrial Telemanipulators,” ASME IMECE, 6th Annual Symp. On Haptic Interfaces for Virtual Environment and Teleoperator Systems, Nov. 1997, pp. 1-7. |
Eberhardt et al., “Including Dynamic Haptic Perception by The Hand: System Description and Some Results,” DSC—vol. 55-1, Dynamic Systems and Control: vol. 1, ASME 1994. |
Eberhardt et al., “OMAR—A Haptic display for speech perception by deaf and deaf-blind individuals,” IEEE Virtual Reality Annual International Symposium, Seattle, WA, Sep. 18-22, 1993. |
Gobel et al., “Tactile Feedback Applied to Computer Mice,” International Journal of Human-Computer Interaction, vol. 7, No. 1, pp. 1-24, 1995. |
Gotow et al.,“Controlled Impedance Test Apparatus for Studying Human Interpretation of Kinesthetic Feedback,” WA11-11:00, pp. 332-337. |
Hasser, C. et al., “Tactile Feedback with Adaptive Controller for a Force-Reflecting Haptic Display,” Parts 1&2, IEEE 0-7803-3131-1, 1996, pp. 526-533. |
Hasser, C., “Force-Reflecting Anthropomorphic Hand Masters,” ALICF-TR-1995-0110, 1995, pp. 5-31. |
Howe. “A Force-Reflecting Teleoperated Hand System for the Study of Tactile Sensing in Precision Manipulation,” Proceedings of the 1992 IEEE International Conference on Robotics and Automation, Nice, France, May 1992. |
IBM Technical Disclosure Bulletin. “Mouse Ball-Actuating Device With Force and Tactile Feedback,” vol. 32, No. 98, Feb. 1990. |
Iwata, “Pen-based Haptic Virtual Environment,” 0-7803-1363-1/93 IEEE, pp. 287-292, 1993. |
Jacobsen et at., “High Performance, Dextrous Telerobotic Manipulator With Force Reflection,” Intervention/ROV'91 Conference & Exposition, Hollywood, Florida. May 21-23, 1991. |
Jones et al. “A perceptual analysis of stiffness,” ISSN 0014-4819 Springer International (Springer-Verlag); Experimental Brain Research, vol. 79, No. 1, pp. 150-156, 1990. |
Kaczmarek et al., “Tactile Displays,” Virtual Environment Technologies. |
Kelley, A. J. et al., “MagicMouse: Tactile and Kinesthetic Feedback in the Human—Computer Interface using an Electromagnetically Actuated Input/Output Device,” Dept. of Elec. Eng., Univ. of Brit. Columbia, 1993, pp. 1-27. |
Kim, W. “Telemanipulator Technology and Space Telerobotics,” SPIE Proceedings, 1993, vol. 2057, pp. 40-50. |
Kontarinis et al., “Display of High-Frequency Tactile Information to Teleoperators,” Telemanipulator Technology and Space Telerobotics, Won S. Kim, Editor, Proc. SPIE vol. 2057,pp. 40-50, Sep. 7-9, 1993. |
MaClean, et al., An Architecture for Haptic Control of Media, In the Proceedings of the ASME Dynamic Systems and Control Division: 1999 International Mechanical Engineering Congress and Exposition Eighth Annual Symposium on Haptic Interfaces for Virtual Environments and Teleoperator Systems Nov. 14-19, 1999, Nashville, TN. |
Marcus, “Touch Feedback in Surgery.” Proceedings of Virtual Reality and Medicine The Cutting Edge, Sep. 8-11, 1994. |
McAffee, “Teleoperator Subsystem/Telerobot Demonsdtrator: Force Reflecting Hand Controller Equipment Manual,” JPL D-5172, pp. 1-50, A1-A36, 81-B5, C1-C36, Jan. 1988. |
Minsky, Margaret et al., “Feeling and Seeing: Issues in Force Display,” ACM 089791-351-5, 1990, pp. 235-242. |
Minsky. “Computational Haptics: The Sandpaper System for Synthesizing Texture for a Force-Feedback Display,” Ph.D. Dissertation. MIT, Jun. 1995. |
Ouh-Young et al., “The Development of a Low-Cost Force Feedback Joystick and Its Use in the Virtual Reality Environment,” Proceedings of the Third Pacific Conference on Computer Graphics and Applications. Pacific Graphics '95, Seoul, Korea, Aug. 21-24, 1995. |
Ouh-Young, “A Low-Cost Force Feedback Joystick and Its Use in PC Video Games,” IEEE Transactions on Consumer Electronics, vol. 41, No. 3, Aug. 1995. |
Ouh-young, M. et al., “Creating an Illusion of Feel: Control Issues in Force Display,” Computer Science Dept., University of North Carolina, 1989, pp. 1-14. |
Ouh-Young,“Force Display in Molecular Docking,” Order No. 9034744,p. 1-369, 1990. |
Patrick et al., “Design and Testing of a Non-reactive. Fingertip, Tactile Display for Interaction with Remote Environments,” Cooperative Intelligent Robotics in Space, Rui J. deFigueiredo et al., Editor, Proc. SPIE vol. 1387, pp. 215-222, 1990. |
Pimentel et al., Virtual Reality: through the new looking glass, 2′• Edition; McGraw-Hill, ISBN 0-07-050167-X, pp. 41-202, 1994. |
Rabinowitz et al., “Multidimensional tactile displays: Identification of vibratory intensity, frequency, and contactor area,” Journal of the Acoustical Society of America, vol. 82, No. 4, Oct. 1987. |
Ramstein, C., “Combining Haptic and Braille Technologies: Design Issues and Pilot Study,” ASSETS '96, 2nd Annual ACM Conf. on Assistive Technologies, 1996, pp. 37-44. |
Russo,“Controlling Dissipative Magnetic Particle Brakes in Force Reflective Devices,” DSC—vol. 42, Advances in Robotics, pp. 63-70, ASME 1992. |
Russo, “The Design and Implementation of a Three Degree of Freedom Force Output Joystick,” MIT libraries Archives Aug. 14, 1990, pp. 1-131, May 1990. |
Scannell. ‘Taking a Joystick Ride,’ Computer Currents. Boston Edition. vol. 9, No. 11. Nov. 1994. |
Schmult et al., “Application Areas for a Force-Feedback Joystick,” ASME 1993, DSC—vol. 49,pp. 47-54. |
Shimoga, “Finger Force and Touch Feedback Issues in Dexterous Telemanipulation,” Proceedings of Fourth Annual Conference on Intelligent Robotic Systems for Space Exploration, Rensselaer Polytechnic Institute, Sep. 30-Oct. 1, 1992. |
Snow et al.,“Model-X Force-Reflecting-Hand-Controller,” NT Control No. MP0-17851; JPL Case No. 5348, pp. 1-4, Jun. 15, 1989. |
Stanley et al., “Computer Simulation of Interacting Dynamic Mechanical Systems Using Distributed Memory Parallel Processors,” DSC—vol. 42, Advances in Robotics, pp. 55-61. ASME 1992. |
Tadros, “Control System Design for a Three Degree of Freedom Virtual Environment Simulator Using Motor/Brake Pair Actuators”, MIT Archive © Massachusetts Institute of Technology, pp. 1-88, Feb. 1990. |
Terry et al., “Tactile Feedback in a Computer Mouse,” Proceedings of Fourteenth Annual Northeast Bioengineering Conference. University of New Hampshire, Mar. 10-11, 1988. |
Patent Cooperation Treaty, International Search Report, International Application No. PCT/US02/10394, date mailed Jul. 3, 2002. |
United States Patent and Trademark Office, Office Action, U.S. Appl. No. 09/823,943, mailed May 8, 2013. |
United States Patent and Trademark Office, Office Action, U.S. Appl. No. 09/823,943, mailed Aug. 1, 2012. |
United States Patent and Trademark Office, Office Action, U.S. Appl. No. 09/823,943, mailed Jan. 26, 2010. |
United States Patent and Trademark Office, Office Action, U.S. Appl. No. 09/823,943, mailed Jan. 7, 2009. |
United States Patent and Trademark Office, Office Action, U.S. Appl. No. 09/823,943, mailed Jul. 17, 2008. |
United States Patent and Trademark Office, Office Action, U.S. Appl. No. 09/823,943, mailed Jan. 11, 2008. |
United States Patent and Trademark Office, Office Action, U.S. Appl. No. 09/823,943, mailed Jul. 26, 2007. |
United States Patent and Trademark Office, Office Action, U.S. Appl. No. 09/823,943, mailed Oct. 20, 2006. |
United States Patent and Trademark Office, Office Action, U.S. Appl. No. 09/823,943, mailed Apr. 20, 2006. |
United States Patent and Trademark Office, Office Action, U.S. Appl. No. 09/823,943, mailed Jun. 29, 2005. |
United States Patent and Trademark Office, Office Action, U.S. Appl. No. 09/823,943, mailed Jul. 12, 2004. |
United States Patent and Trademark Office, Office Action, U.S. Appl. No. 09/823,943, mailed Sep. 13, 2002. |
United States Patent and Trademark Office, Office Action, U.S. Appl. No. 09/823,943, mailed Jan. 31, 2002. |
Japanese Patent Office, Notice of Reasons for Rejection, Application No. 2002-577071, issued Mar. 4, 2008. |
Japanese Patent Office, Notice of Reasons for Rejection, Application No. 2002-577071, issued Jul. 10, 2007. |
Japanese Patent Office, Notice of Reasons for Rejection, Application No. 2002-577071, issued Nov. 11, 2008. |
Japanese Patent Office, Decision for Rejection, Application No. 2002-577071, issued Jul. 14, 2009. |
State Intellectual Property Office of the People's Republic of China, Notification of First Office Action, Application No. 200910165759, issued Sep. 22, 2011. |
Patent Reexamination Board of the Chinese Patent Office, Notification of Reexamination, Application No. 02807519, issued Sep. 27, 2010. |
Patent Office of the People's Republic of China, Notification of First Office Action, Application No. 02807519, issued Aug. 18, 2006. |
Patent Office of the People's Republic of China, Notification of Second Office Action, Application No. 02807519, issued Mar. 14, 2008. |
Patent Office of the People's Republic of China, Notification of Third Office Action, Application No. 02807519, issued Jul. 4, 2008. |
Patent Office of the People's Republic of China, Decision of Rejection, Application No. 02807519, issued Apr. 24, 2009. |
Korean Intellectual Property Office, Notice of Preliminary Rejection, Application No. 10-2003-7012826, dated Jan. 4, 2008. |
Japanese Patent Office, Notice of Reasons for Rejection, Application No. 2009-260496, mailed Oct. 12, 2010. |
European Patent Office, Communication, Application No. 02726696, mailed Feb. 22, 2006. |
Webster's Third New International Dictionary of the English Language Unabridged, 1965, pp. 1375-1376. |
Haptics, web page at http://web.archive.org/web/19991 001194232/http://lslwww.epfl.ch/-penha/predoc/vr/haptic.html, as available via the Internet, 1999. |
BlueDrone: Android Controlled R/C Car by Stephen Mark—Kickstarter, web page at http://kickstarter.com/projects/1756214656/bluedrone-android-controlled-r-c-car, as available via the Internet and printed Aug. 1, 2013. |
Nxt-remote-control—NXT Remote Control for Android, web page at hppt://code.google.com/p/nxt-remote-control, as available via the Internet and printed Aug. 1, 2013. |
Lego's MINDroid Android app remotely controls Mindstorms NXT robots, web page at http://www.engadget.com/2010/10/12/legos-mindroid-android-app-remotely-controls-minds, as available via the Internet and printed Aug. 7, 2013. |
European Patent Office, Partial European Search Report, European Application No. EP13178283, dated Nov. 26, 2013. |
European Patent Office, Extended European Search Report, European Application No. EP13178283, dated Mar. 21, 2014. |
Chinese Patent Office Application No. 201310334554.0, First Office Action and translation dated Apr. 1, 2017, 26 pages. |
Number | Date | Country | |
---|---|---|---|
20160098085 A1 | Apr 2016 | US |
Number | Date | Country | |
---|---|---|---|
61678908 | Aug 2012 | US | |
61679382 | Aug 2012 | US |
Number | Date | Country | |
---|---|---|---|
Parent | 13826391 | Mar 2013 | US |
Child | 14967499 | US |